Consistency of Compositional Generalization Across Multiple Levels
DOI:
https://doi.org/10.1609/aaai.v39i5.32492Abstract
Compositional generalization is the capability of a model to understand novel compositions composed of seen concepts. There are multiple levels of novel compositions including phrase-phrase level, phrase-word level, and word-word level. Existing methods achieve promising compositional generalization, but the consistency of compositional generalization across multiple levels of novel compositions remains unexplored. The consistency refers to that a model should generalize to a phrase-phrase level novel composition, and phrase-word/word-word level novel compositions that can be derived from it simultaneously. In this paper, we propose a meta-learning based framework, for achieving consistent compositional generalization across multiple levels. The basic idea is to progressively learn compositions from simple to complex for consistency. Specifically, we divide the original training set into multiple validation sets based on compositional complexity, and introduce multiple meta-weight-nets to generate sample weights for samples in different validation sets. To fit the validation sets in order of increasing compositional complexity, we optimize the parameters of each meta-weight-net independently and sequentially in a multilevel optimization manner. We build a GQA-CCG dataset to quantitatively evaluate the consistency. Experimental results on visual question answering and temporal video grounding, demonstrate the effectiveness of the proposed framework.Downloads
Published
2025-04-11
How to Cite
Li, C., Li, Z., Jing, C., Fan, X., Ye, W., Wu, Y., & Jia, Y. (2025). Consistency of Compositional Generalization Across Multiple Levels. Proceedings of the AAAI Conference on Artificial Intelligence, 39(5), 4661-4669. https://doi.org/10.1609/aaai.v39i5.32492
Issue
Section
AAAI Technical Track on Computer Vision IV