Arithmetic Feature Interaction Is Necessary for Deep Tabular Learning

Authors

  • Yi Cheng Zhejiang University
  • Renjun Hu Alibaba Group
  • Haochao Ying Zhejiang University
  • Xing Shi Alibaba
  • Jian Wu Zhejiang University
  • Wei Lin Alibaba Group

DOI:

https://doi.org/10.1609/aaai.v38i10.29033

Keywords:

ML: Feature Construction/Reformulation, ML: Deep Learning Algorithms, ML: Deep Neural Architectures and Foundation Models, ML: Dimensionality Reduction/Feature Selection

Abstract

Until recently, the question of the effective inductive bias of deep models on tabular data has remained unanswered. This paper investigates the hypothesis that arithmetic feature interaction is necessary for deep tabular learning. To test this point, we create a synthetic tabular dataset with a mild feature interaction assumption and examine a modified transformer architecture enabling arithmetical feature interactions, referred to as AMFormer. Results show that AMFormer outperforms strong counterparts in fine-grained tabular data modeling, data efficiency in training, and generalization. This is attributed to its parallel additive and multiplicative attention operators and prompt-based optimization, which facilitate the separation of tabular samples in an extended space with arithmetically-engineered features. Our extensive experiments on real-world data also validate the consistent effectiveness, efficiency, and rationale of AMFormer, suggesting it has established a strong inductive bias for deep learning on tabular data. Code is available at https://github.com/aigc-apps/AMFormer.

Downloads

Published

2024-03-24

How to Cite

Cheng, Y., Hu, R., Ying, H., Shi, X., Wu, J., & Lin, W. (2024). Arithmetic Feature Interaction Is Necessary for Deep Tabular Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(10), 11516-11524. https://doi.org/10.1609/aaai.v38i10.29033

Issue

Section

AAAI Technical Track on Machine Learning I