Learning Small Decision Trees with Few Outliers: A Parameterized Perspective

Authors

  • Harmender Gahlawat Ben-Gurion University, Beersheba
  • Meirav Zehavi Ben-Gurion University, Beersheba

DOI:

https://doi.org/10.1609/aaai.v38i11.29098

Keywords:

ML: Classification and Regression, ML: Representation Learning

Abstract

Decision trees is a fundamental tool in machine learning for representing, classifying, and generalizing data. It is desirable to construct ``small'' decision trees, by minimizing either the size (s) or the depth (d) of the decision tree (DT). Recently, the parameterized complexity of Decision Tree Learning has attracted a lot of attention. We consider a generalization of Decision Tree Learning where given a classification instance E and an integer t, the task is to find a ``small'' DT that disagrees with E in at most t examples. We consider two problems: DTSO and DTDO, where the goal is to construct a DT minimizing s and d, respectively. We first establish that both DTSO and DTDO are W[1]-hard when parameterized by s+y and d+y, respectively, where y is the maximum number of features in which two differently labeled examples can differ. We complement this result by showing that these problems become FPT if we include the parameter t. We also consider the kernelization complexity of these problems and establish several positive and negative results for both DTSO and DTDO.

Downloads

Published

2024-03-24

How to Cite

Gahlawat, H., & Zehavi, M. (2024). Learning Small Decision Trees with Few Outliers: A Parameterized Perspective. Proceedings of the AAAI Conference on Artificial Intelligence, 38(11), 12100–12108. https://doi.org/10.1609/aaai.v38i11.29098

Issue

Section

AAAI Technical Track on Machine Learning II