Statistical Learning Theory for Distributional Classification
DOI:
https://doi.org/10.1609/aaai.v40i25.39255Abstract
In supervised learning with distributional inputs in the two-stage sampling setup, relevant to applications like learning-based medical screening or causal learning, the inputs (which are probability distributions) are not accessible in the learning phase, but only samples thereof. This problem is particularly amenable to kernel-based learning methods, where the distributions or samples are first embedded into a Hilbert space, often using kernel mean embeddings (KMEs), and then a standard kernel method like Support Vector Machines (SVMs) is applied, using a kernel defined on the embedding Hilbert space. In this work, we contribute to the theoretical analysis of this latter approach, with a particular focus on classification with distributional inputs using SVMs. We establish a new oracle inequality and derive consistency and learning rate results. Furthermore, for SVMs using the hinge loss and Gaussian kernels, we formulate a novel variant of an established noise assumption from the binary classification literature, under which we can establish learning rates. Finally, some of our technical tools like a new feature space for Gaussian kernels on Hilbert spaces are of independent interest.Published
2026-03-14
How to Cite
Fiedler, C. (2026). Statistical Learning Theory for Distributional Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 21120–21127. https://doi.org/10.1609/aaai.v40i25.39255
Issue
Section
AAAI Technical Track on Machine Learning II