Fine-Grained Machine Teaching with Attention Modeling

Authors

  • Jiacheng Liu Shanghai Jiao Tong University
  • Xiaofeng Hou Shanghai Jiao Tong University
  • Feilong Tang Shanghai Jiao Tong University

DOI:

https://doi.org/10.1609/aaai.v34i03.5642

Abstract

The state-of-the-art machine teaching techniques overestimate the ability of learners in grasping a complex concept. On one side, since a complicated concept always contains multiple fine-grained concepts, students can only grasp parts of them during a practical teaching process. On the other side, because a single teaching sample contains unequal information in terms of various fine-grained concepts, learners accept them at different levels. Thus, with more and more complicated dataset, it is challenging for us to rethink the machine teaching frameworks. In this work, we propose a new machine teaching framework called Attentive Machine Teaching (AMT). Specifically, we argue that a complicated concept always consists of multiple features, which we call fine-grained concepts. We define attention to represent the learning level of a learner in studying a fine-grained concept. Afterwards, we propose AMT, an adaptive teaching framework to construct the personalized optimal teaching dataset for learners. During each iteration, we estimate the workers' ability with Graph Neural Network (GNN) and select the best sample using a pool-based searching approach. For corroborating our theoretical findings, we conduct extensive experiments with both synthetic datasets and real datasets. Our experimental results verify the effectiveness of AMT algorithms.

Downloads

Published

2020-04-03

How to Cite

Liu, J., Hou, X., & Tang, F. (2020). Fine-Grained Machine Teaching with Attention Modeling. Proceedings of the AAAI Conference on Artificial Intelligence, 34(03), 2585-2592. https://doi.org/10.1609/aaai.v34i03.5642

Issue

Section

AAAI Technical Track: Human-Computation and Crowd Sourcing