Knowledge Amalgamation for Multi-Label Classification via Label Dependency Transfer

Authors

  • Jidapa Thadajarassiri Worcester Polytechnic Institute
  • Thomas Hartvigsen MIT
  • Walter Gerych Worcester Polytechnic Institute
  • Xiangnan Kong Worcester Polytechnic Institute
  • Elke Rundensteiner Worcester Polytechnic Institute

DOI:

https://doi.org/10.1609/aaai.v37i8.26190

Keywords:

ML: Multi-Class/Multi-Label Learning & Extreme Classification, KRR: Knowledge Acquisition, ML: Ensemble Methods

Abstract

Multi-label classification (MLC), which assigns multiple labels to each instance, is crucial to domains from computer vision to text mining. Conventional methods for MLC require huge amounts of labeled data to capture complex dependencies between labels. However, such labeled datasets are expensive, or even impossible, to acquire. Worse yet, these pre-trained MLC models can only be used for the particular label set covered in the training data. Despite this severe limitation, few methods exist for expanding the set of labels predicted by pre-trained models. Instead, we acquire vast amounts of new labeled data and retrain a new model from scratch. Here, we propose combining the knowledge from multiple pre-trained models (teachers) to train a new student model that covers the union of the labels predicted by this set of teachers. This student supports a broader label set than any one of its teachers without using labeled data. We call this new problem knowledge amalgamation for multi-label classification. Our new method, Adaptive KNowledge Transfer (ANT), trains a student by learning from each teacher’s partial knowledge of label dependencies to infer the global dependencies between all labels across the teachers. We show that ANT succeeds in unifying label dependencies among teachers, outperforming five state-of-the-art methods on eight real-world datasets.

Downloads

Published

2023-06-26

How to Cite

Thadajarassiri, J., Hartvigsen, T., Gerych, W., Kong, X., & Rundensteiner, E. (2023). Knowledge Amalgamation for Multi-Label Classification via Label Dependency Transfer. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9980-9988. https://doi.org/10.1609/aaai.v37i8.26190

Issue

Section

AAAI Technical Track on Machine Learning III