AIO-P: Expanding Neural Performance Predictors beyond Image Classification

Authors

  • Keith G. Mills Department of Electrical and Computer Engineering, University of Alberta Huawei Technologies, Edmonton, Alberta, Canada
  • Di Niu Department of Electrical and Computer Engineering, University of Alberta
  • Mohammad Salameh Huawei Technologies, Edmonton, Alberta, Canada
  • Weichen Qiu Department of Electrical and Computer Engineering, University of Alberta
  • Fred X. Han Huawei Technologies, Edmonton, Alberta, Canada
  • Puyuan Liu Huawei Technologies, Edmonton, Alberta, Canada
  • Jialin Zhang Huawei Kirin Solution, Shanghai, China
  • Wei Lu Huawei Technologies, Edmonton, Alberta, Canada
  • Shangling Jui Huawei Kirin Solution, Shanghai, China

DOI:

https://doi.org/10.1609/aaai.v37i8.26101

Keywords:

ML: Auto ML and Hyperparameter Tuning, CV: Applications, DMKM: Applications, ML: Graph-based Machine Learning, ML: Deep Neural Architectures, ML: Classification and Regression, ML: Optimization, CV: Segmentation, CV: Object Detection & Categorization

Abstract

Evaluating neural network performance is critical to deep neural network design but a costly procedure. Neural predictors provide an efficient solution by treating architectures as samples and learning to estimate their performance on a given task. However, existing predictors are task-dependent, predominantly estimating neural network performance on image classification benchmarks. They are also search-space dependent; each predictor is designed to make predictions for a specific architecture search space with predefined topologies and set of operations. In this paper, we propose a novel All-in-One Predictor (AIO-P), which aims to pretrain neural predictors on architecture examples from multiple, separate computer vision (CV) task domains and multiple architecture spaces, and then transfer to unseen downstream CV tasks or neural architectures. We describe our proposed techniques for general graph representation, efficient predictor pretraining and knowledge infusion techniques, as well as methods to transfer to downstream tasks/spaces. Extensive experimental results show that AIO-P can achieve Mean Absolute Error (MAE) and Spearman’s Rank Correlation (SRCC) below 1p% and above 0.5, respectively, on a breadth of target downstream CV tasks with or without fine-tuning, outperforming a number of baselines. Moreover, AIO-P can directly transfer to new architectures not seen during training, accurately rank them and serve as an effective performance estimator when paired with an algorithm designed to preserve performance while reducing FLOPs.

Downloads

Published

2023-06-26

How to Cite

Mills, K. G., Niu, D., Salameh, M., Qiu, W., Han, F. X., Liu, P., Zhang, J., Lu, W., & Jui, S. (2023). AIO-P: Expanding Neural Performance Predictors beyond Image Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9180-9189. https://doi.org/10.1609/aaai.v37i8.26101

Issue

Section

AAAI Technical Track on Machine Learning III