Neural Tangent Kernels Under Stochastic Data Augmentation

Authors

  • Joshua DeOliveira Worcester Polytechnic Institute, Worcester, MA
  • Sajal Chakroborty Worcester Polytechnic Institute, Worcester, MA
  • Walter Gerych Worcester Polytechnic Institute, Worcester, MA
  • Elke Rundensteiner Worcester Polytechnic Institute, Worcester, MA

DOI:

https://doi.org/10.1609/aaai.v40i25.39214

Abstract

The learning dynamics of modern neural networks remain an open problem in deep learning. The Neural Tangent Kernel (NTK) offers an elegant description of training dynamics in the infinite‑width limit, yet its classical formulation assumes a static data set. Modern model training practice departs from this strong assumption through the use of on‑the‑fly data augmentations (e.g. additive noise). In this work, we conduct an NTK-driven analysis of how data transformations affect a neural net's evolution in the function space. Our theoretical contributions characterize how repeated Gaussian perturbations from NTK-derived covariances can steer neural-net optimizations toward user‑specified behavior. These theoretical insights are empirically validated by controlled experiments. Taken together, our results lay the foundation for a promising future research direction that transforms the NTK from a descriptive to a prescriptive tool, enabling control of neural net training trajectories and behavior of inference generalization with grounded interventions.

Downloads

Published

2026-03-14

How to Cite

DeOliveira, J., Chakroborty, S., Gerych, W., & Rundensteiner, E. (2026). Neural Tangent Kernels Under Stochastic Data Augmentation. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 20755–20762. https://doi.org/10.1609/aaai.v40i25.39214

Issue

Section

AAAI Technical Track on Machine Learning II