Out-of-Distribution Generalization by Neural-Symbolic Joint Training
DOI:
https://doi.org/10.1609/aaai.v37i10.26444Keywords:
RU: Graphical Model, KRR: Automated Reasoning and Theorem Proving, KRR: Logic ProgrammingAbstract
This paper develops a novel methodology to simultaneously learn a neural network and extract generalized logic rules. Different from prior neural-symbolic methods that require background knowledge and candidate logical rules to be provided, we aim to induce task semantics with minimal priors. This is achieved by a two-step learning framework that iterates between optimizing neural predictions of task labels and searching for a more accurate representation of the hidden task semantics. Notably, supervision works in both directions: (partially) induced task semantics guide the learning of the neural network and induced neural predictions admit an improved semantic representation. We demonstrate that our proposed framework is capable of achieving superior out-of-distribution generalization performance on two tasks: (i) learning multi-digit addition, where it is trained on short sequences of digits and tested on long sequences of digits; (ii) predicting the optimal action in the Tower of Hanoi, where the model is challenged to discover a policy independent of the number of disks in the puzzle.Downloads
Published
2023-06-26
How to Cite
Liu, A., Xu, H., Van den Broeck, G., & Liang, Y. (2023). Out-of-Distribution Generalization by Neural-Symbolic Joint Training. Proceedings of the AAAI Conference on Artificial Intelligence, 37(10), 12252-12259. https://doi.org/10.1609/aaai.v37i10.26444
Issue
Section
AAAI Technical Track on Reasoning Under Uncertainty