Relative Attributes for Enhanced Human-Machine Communication

Authors

  • Devi Parikh Toyota Technological Institute Chicago
  • Adriana Kovashka University of Texas at Austin
  • Amar Parkash IIIT-Delhi
  • Kristen Grauman University of Texas at Austin

DOI:

https://doi.org/10.1609/aaai.v26i1.8443

Keywords:

attributes, zero-shot learning, image description, feedback, image search, active learning

Abstract

We propose to model relative attributes that capture the relationships between images and objects in terms of human-nameable visual properties. For example, the models can capture that animal A is 'furrier' than animal B, or image X is 'brighter' than image B. Given training data stating how object/scene categories relate according to different attributes, we learn a ranking function per attribute. The learned ranking functions predict the relative strength of each property in novel images. We show how these relative attribute predictions enable a variety of novel applications, including zero-shot learning from relative comparisons, automatic image description, image search with interactive feedback, and active learning of discriminative classifiers. We overview results demonstrating these applications with images of faces and natural scenes. Overall, we find that relative attributes enhance the precision of communication between humans and computer vision algorithms, providing the richer language needed to fluidly "teach" a system about visual concepts.

Downloads

Published

2021-09-20

How to Cite

Parikh, D., Kovashka, A., Parkash, A., & Grauman, K. (2021). Relative Attributes for Enhanced Human-Machine Communication. Proceedings of the AAAI Conference on Artificial Intelligence, 26(1), 2153-2159. https://doi.org/10.1609/aaai.v26i1.8443