How to learn classifier chains using positive-unlabelled multi-label data? | ML in PL 22
How to learn classifier chains using positive-unlabelled multi-label data? by Paweł Teisseyre (Institute of Computer Science, Polish Academy of Sciences and Faculty of Mathematics and Information Sciences, Warsaw University of Technology)
Multi-label learning deals with data examples which are associated with multiple class labels simultaneously. The problem has attracted significant attention in recent years and dozens of algorithms have been proposed. However in traditional multi-label setting it is assumed that all relevant labels are assigned to the given instance. This assumption is not met in many real-world situations. In positive unlabelled multi-label setting, only some of relevant labels are assigned. The appearance of a label means that the instance is really associated with this label, while the absence of the label does not imply that this label is not proper for the instance. For example, when predicting multiple diseases in one patient, some diseases can be undiagnosed however it does not mean that the patient does not have these diseases. Among many existing multi-label methods, classifier chains gained the great popularity mainly due to their simplicity and high predictive power. However, it turns out that adaptation of classifier chains to positive unlabelled framework is not straightforward, due to the fact that the true target variables are observed only partially and therefore they cannot be used directly to train the models in the chain. The partial observability concerns not only the current target variable in the chain but also the feature space, which additionally increases the difficulty of the problem. We propose two modifications of classifier chains. In the first method we scale the output probabilities of the consecutive classifiers in the chain. In the second method we minimize weighted empirical risk, with weights depending on prior probabilities of the target variables. The predictive performance of the proposed methods is studied on real multi-label datasets for different positive unlabelled settings.
The talk was delivered during ML in PL Conference 2022 as a part of Contributed Talks. The conference was organized by a non-profit NGO called ML in PL Association.
ML in PL Association website: https://mlinpl.org/
ML in PL Conference 2022 website: https://conference2022.mlinpl.org/
ML In PL Conference 2023 website: https://conference2023.mlinpl.org/
---
ML in PL Association was founded based on the experiences in organizing of the ML in PL Conference (formerly PL in ML), the ML in PL Association is a non-profit organization devoted to fostering the machine learning community in Poland and Europe and promoting a deep understanding of ML methods. Even though ML in PL is based in Poland, it seeks to provide opportunities for international cooperation.
Видео How to learn classifier chains using positive-unlabelled multi-label data? | ML in PL 22 канала ML in PL
Multi-label learning deals with data examples which are associated with multiple class labels simultaneously. The problem has attracted significant attention in recent years and dozens of algorithms have been proposed. However in traditional multi-label setting it is assumed that all relevant labels are assigned to the given instance. This assumption is not met in many real-world situations. In positive unlabelled multi-label setting, only some of relevant labels are assigned. The appearance of a label means that the instance is really associated with this label, while the absence of the label does not imply that this label is not proper for the instance. For example, when predicting multiple diseases in one patient, some diseases can be undiagnosed however it does not mean that the patient does not have these diseases. Among many existing multi-label methods, classifier chains gained the great popularity mainly due to their simplicity and high predictive power. However, it turns out that adaptation of classifier chains to positive unlabelled framework is not straightforward, due to the fact that the true target variables are observed only partially and therefore they cannot be used directly to train the models in the chain. The partial observability concerns not only the current target variable in the chain but also the feature space, which additionally increases the difficulty of the problem. We propose two modifications of classifier chains. In the first method we scale the output probabilities of the consecutive classifiers in the chain. In the second method we minimize weighted empirical risk, with weights depending on prior probabilities of the target variables. The predictive performance of the proposed methods is studied on real multi-label datasets for different positive unlabelled settings.
The talk was delivered during ML in PL Conference 2022 as a part of Contributed Talks. The conference was organized by a non-profit NGO called ML in PL Association.
ML in PL Association website: https://mlinpl.org/
ML in PL Conference 2022 website: https://conference2022.mlinpl.org/
ML In PL Conference 2023 website: https://conference2023.mlinpl.org/
---
ML in PL Association was founded based on the experiences in organizing of the ML in PL Conference (formerly PL in ML), the ML in PL Association is a non-profit organization devoted to fostering the machine learning community in Poland and Europe and promoting a deep understanding of ML methods. Even though ML in PL is based in Poland, it seeks to provide opportunities for international cooperation.
Видео How to learn classifier chains using positive-unlabelled multi-label data? | ML in PL 22 канала ML in PL
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![Rafał Pilarczyk: Is Artificial Intelligence a threat to musicians? – Music generation techniques](https://i.ytimg.com/vi/J1fUVeGM9Wg/default.jpg)
![Philippe Preux - Bits of Reinforcement Learning | ML in PL 23](https://i.ytimg.com/vi/BZc2GWoEddc/default.jpg)
![B. Ludwiczuk, K. Jasinska-Kobus (Allegro) - Batch construction strategies in deep metric learning](https://i.ytimg.com/vi/ZU9faIBxXbc/default.jpg)
![Marcin Andrychowicz - Solving Rubik’s Cube with a Robot Hand](https://i.ytimg.com/vi/zORaE0v6QBA/default.jpg)
![Adam Paszke: PyTorch 1.0: now and in the future](https://i.ytimg.com/vi/cNeAPzlPl7A/default.jpg)
![Adam Podraza: Applied time series forecasting using machine learning](https://i.ytimg.com/vi/Qb9ekWYIxcU/default.jpg)
![Gül Varol - Learning human body representations from visual data](https://i.ytimg.com/vi/aUwqZbupkbE/default.jpg)
![David Haber - Opportunities and Challenges when Building AI for Autonomous Flight](https://i.ytimg.com/vi/pssmL5x2RFM/default.jpg)
![Adam Gonczarek (Alphamoon) – Intelligent Document Processing](https://i.ytimg.com/vi/QvhznidHYFM/default.jpg)
![Jonasz Pamuła (RTB House) – ML Challenges in cookieless world](https://i.ytimg.com/vi/seS3Hqn7bpY/default.jpg)
![João Henriques - Mapping environments with deep networks and spatial memories](https://i.ytimg.com/vi/vc0HwO_AJ40/default.jpg)
![Krzysztof Geras (NYU): "Towards Solving Breast Cancer Screening Diagnosis with Deep Learning"](https://i.ytimg.com/vi/IUELwCM1Efs/default.jpg)
![Stanisław Jastrzębski - Deep Learning in the Light of the Simplicity Bias | MLSS Kraków 2023](https://i.ytimg.com/vi/PV2EAkcgc7o/default.jpg)
![Yoshua Bengio – Cognitively-inspired inductive biases for higher-level cognition](https://i.ytimg.com/vi/02ABljCu5Zw/default.jpg)
![Tomasz Grel (Nvidia): Faster Deep Learning with mixed precision and multiple GPUs](https://i.ytimg.com/vi/zbBUExOG-To/default.jpg)
![Panel Discussion – Women in ML](https://i.ytimg.com/vi/GH1axtAEYhk/default.jpg)
![Michał Jamroż - Class fitting in residual convolutional networks | ML in PL 23](https://i.ytimg.com/vi/Zx5uG8VhDgA/default.jpg)
![Sebastian Cygert - Toward continually learning models | ML in PL 23](https://i.ytimg.com/vi/tf424AmS5ps/default.jpg)
![Barbara Rychalska - Neural Machine Translation: achievements, challenges and the way forward](https://i.ytimg.com/vi/0bZagsD7oPg/default.jpg)
![Stanisław Jastrzębski - Gradient Alignment: When Deep Networks Work, and When They Don't](https://i.ytimg.com/vi/9J0_jLyI5IE/default.jpg)