Addressing Multi-Label Learning with Partial Labels: From Sample Selection to Label Selection

Authors

  • Gengyu Lyu College of Computer Science, Beijing University of Technology School of Computer Science and Technology, Beijing Jiaotong University Idealism Beijing Technology Co., Ltd.
  • Bohang Sun College of Computer Science, Beijing University of Technology
  • Xiang Deng School of Computer Science and Technology, Beijing Jiaotong University Department of Automation, Tsinghua University
  • Songhe Feng School of Computer Science and Technology, Beijing Jiaotong University Key Laboratory of Big Data \& Artificial Intelligence in Transportation (Beijing Jiaotong University), Ministry of Education

DOI:

https://doi.org/10.1609/aaai.v39i18.34119

Abstract

Multi-label Learning with Partial Labels (ML-PL) learns from training data, where each sample is annotated with part of positive labels while leaving the rest of positive labels unannotated. Existing methods mainly focus on extending multi-label losses to estimate unannotated labels, further inducing a missing-robust network. However, training with single network could lead to confirmation bias (i.e., the model tends to confirm its mistakes). To tackle this issue, we propose a novel learning paradigm termed Co-Label Selection (CLS), where two networks feed forward all data and cooperate in a co-training manner for critical label selection. Different from traditional co-training based methods that networks select confident samples for each other, we start from a new perspective that two networks are encouraged to remove false-negative labels while keep training samples reserved. Meanwhile, considering the extreme positive-negative label imbalance in ML-PL that leads the model to focus on negative labels, we enforce the model to concentrate on positive labels by abandoning non-informative negative labels to alleviate such issue. By shifting the cooperation strategy from "Sample Selection'' to "Label Selection'', CLS avoids directly dropping samples and reserves training data in most extent, thus enhancing the utilization of supervised signals and the generalization of the learning model. Empirical results performed on various multi-label datasets demonstrate that our CLS is significantly superior to other state-of-the-art methods.

Downloads

Published

2025-04-11

How to Cite

Lyu, G., Sun, B., Deng, X., & Feng, S. (2025). Addressing Multi-Label Learning with Partial Labels: From Sample Selection to Label Selection. Proceedings of the AAAI Conference on Artificial Intelligence, 39(18), 19251-19259. https://doi.org/10.1609/aaai.v39i18.34119

Issue

Section

AAAI Technical Track on Machine Learning IV