GENNAPE: Towards Generalized Neural Architecture Performance Estimators

Authors

  • Keith G. Mills Department of Electrical and Computer Engineering, University of Alberta Huawei Technologies, Edmonton, Alberta, Canada
  • Fred X. Han Huawei Technologies, Edmonton, Alberta, Canada
  • Jialin Zhang Huawei Kirin Solution, Shanghai, China
  • Fabian Chudak Huawei Technologies, Edmonton, Alberta, Canada
  • Ali Safari Mamaghani Department of Electrical and Computer Engineering, University of Alberta
  • Mohammad Salameh Huawei Technologies, Edmonton, Alberta, Canada
  • Wei Lu Huawei Technologies, Edmonton, Alberta, Canada
  • Shangling Jui Huawei Kirin Solution, Shanghai, China
  • Di Niu Department of Electrical and Computer Engineering, University of Alberta

DOI:

https://doi.org/10.1609/aaai.v37i8.26102

Keywords:

ML: Auto ML and Hyperparameter Tuning, CV: Applications, ML: Applications, ML: Classification and Regression, ML: Graph-based Machine Learning, ML: Deep Neural Architectures

Abstract

Predicting neural architecture performance is a challenging task and is crucial to neural architecture design and search. Existing approaches either rely on neural performance predictors which are limited to modeling architectures in a predefined design space involving specific sets of operators and connection rules, and cannot generalize to unseen architectures, or resort to Zero-Cost Proxies which are not always accurate. In this paper, we propose GENNAPE, a Generalized Neural Architecture Performance Estimator, which is pretrained on open neural architecture benchmarks, and aims to generalize to completely unseen architectures through combined innovations in network representation, contrastive pretraining, and a fuzzy clustering-based predictor ensemble. Specifically, GENNAPE represents a given neural network as a Computation Graph (CG) of atomic operations which can model an arbitrary architecture. It first learns a graph encoder via Contrastive Learning to encourage network separation by topological features, and then trains multiple predictor heads, which are soft-aggregated according to the fuzzy membership of a neural network. Experiments show that GENNAPE pretrained on NAS-Bench-101 can achieve superior transferability to 5 different public neural network benchmarks, including NAS-Bench-201, NAS-Bench-301, MobileNet and ResNet families under no or minimum fine-tuning. We further introduce 3 challenging newly labelled neural network benchmarks: HiAML, Inception and Two-Path, which can concentrate in narrow accuracy ranges. Extensive experiments show that GENNAPE can correctly discern high-performance architectures in these families. Finally, when paired with a search algorithm, GENNAPE can find architectures that improve accuracy while reducing FLOPs on three families.

Downloads

Published

2023-06-26

How to Cite

Mills, K. G., Han, F. X., Zhang, J., Chudak, F., Safari Mamaghani, A., Salameh, M., Lu, W., Jui, S., & Niu, D. (2023). GENNAPE: Towards Generalized Neural Architecture Performance Estimators. Proceedings of the AAAI Conference on Artificial Intelligence, 37(8), 9190-9199. https://doi.org/10.1609/aaai.v37i8.26102

Issue

Section

AAAI Technical Track on Machine Learning III