Per-Domain Generalizing Policies: On Validation Instances and Scaling Behavior

Authors

  • Timo P. Gros German Research Center for Artificial Intelligence (DFKI) Saarland University
  • Nicola J. Müller German Research Center for Artificial Intelligence (DFKI) Saarland University
  • Daniel Fišer Aalborg University
  • Isabel Valera Saarland University
  • Verena Wolf German Research Center for Artificial Intelligence (DFKI) Saarland University
  • Jörg Hoffmann German Research Center for Artificial Intelligence (DFKI) Saarland University

DOI:

https://doi.org/10.1609/icaps.v35i1.36118

Abstract

Recent work has shown that successful per-domain generalizing action policies can be learned. Scaling behavior, from small training instances to large test instances, is the key objective; and the use of validation instances larger than training instances is one key to achieve it. Prior work has used fixed validation sets. Here, we introduce a method generating the validation set dynamically, on the fly, increasing instance size so long as informative and feasible. We also introduce refined methodology for evaluating scaling behavior, generating test instances systematically to guarantee a given confidence in coverage performance for each instance size. In experiments, dynamic validation improves scaling behavior of GNN policies in all 9 domains used.

Downloads

Published

2025-09-16

How to Cite

Gros, T. P., Müller, N. J., Fišer, D., Valera, I., Wolf, V., & Hoffmann, J. (2025). Per-Domain Generalizing Policies: On Validation Instances and Scaling Behavior. Proceedings of the International Conference on Automated Planning and Scheduling, 35(1), 198-203. https://doi.org/10.1609/icaps.v35i1.36118