On the Decreasing Power of Kernel and Distance Based Nonparametric Hypothesis Tests in High Dimensions

Authors

  • Aaditya Ramdas Carnegie Mellon University
  • Sashank Jakkam Reddi Carnegie Mellon University
  • Barnabas Poczos Carnegie Mellon University
  • Aarti Singh Carnegie Mellon University
  • Larry Wasserman Carnegie Mellon University

DOI:

https://doi.org/10.1609/aaai.v29i1.9692

Abstract

This paper is about two related decision theoretic problems, nonparametric two-sample testing and independence testing. There is a belief that two recently proposed solutions, based on kernels and distances between pairs of points, behave well in high-dimensional settings. We identify different sources of misconception that give rise to the above belief. Specifically, we differentiate the hardness of estimation of test statistics from the hardness of testing whether these statistics are zero or not, and explicitly discuss a notion of "fair" alternative hypotheses for these problems as dimension increases. We then demonstrate that the power of these tests actually drops polynomially with increasing dimension against fair alternatives. We end with some theoretical insights and shed light on the median heuristic for kernel bandwidth selection. Our work advances the current understanding of the power of modern nonparametric hypothesis tests in high dimensions.

Downloads

Published

2015-03-04

How to Cite

Ramdas, A., Jakkam Reddi, S., Poczos, B., Singh, A., & Wasserman, L. (2015). On the Decreasing Power of Kernel and Distance Based Nonparametric Hypothesis Tests in High Dimensions. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1). https://doi.org/10.1609/aaai.v29i1.9692

Issue

Section

AAAI Technical Track: Reasoning under Uncertainty