Neural Tangent Kernels Under Stochastic Data Augmentation
DOI:
https://doi.org/10.1609/aaai.v40i25.39214Abstract
The learning dynamics of modern neural networks remain an open problem in deep learning. The Neural Tangent Kernel (NTK) offers an elegant description of training dynamics in the infinite‑width limit, yet its classical formulation assumes a static data set. Modern model training practice departs from this strong assumption through the use of on‑the‑fly data augmentations (e.g. additive noise). In this work, we conduct an NTK-driven analysis of how data transformations affect a neural net's evolution in the function space. Our theoretical contributions characterize how repeated Gaussian perturbations from NTK-derived covariances can steer neural-net optimizations toward user‑specified behavior. These theoretical insights are empirically validated by controlled experiments. Taken together, our results lay the foundation for a promising future research direction that transforms the NTK from a descriptive to a prescriptive tool, enabling control of neural net training trajectories and behavior of inference generalization with grounded interventions.Downloads
Published
2026-03-14
How to Cite
DeOliveira, J., Chakroborty, S., Gerych, W., & Rundensteiner, E. (2026). Neural Tangent Kernels Under Stochastic Data Augmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 20755–20762. https://doi.org/10.1609/aaai.v40i25.39214
Issue
Section
AAAI Technical Track on Machine Learning II