Bespoke: A Block-Level Neural Network Optimization Framework for Low-Cost Deployment

Authors

  • Jong-Ryul Lee Electronics and Telecommunications Research Institute (ETRI)
  • Yong-Hyuk Moon Electronics and Telecommunications Research Institute (ETRI) University of Science and Technology (UST)

DOI:

https://doi.org/10.1609/aaai.v37i7.26020

Keywords:

ML: Auto ML and Hyperparameter Tuning, ML: Deep Neural Architectures, ML: Deep Neural Network Algorithms, ML: Dimensionality Reduction/Feature Selection

Abstract

As deep learning models become popular, there is a lot of need for deploying them to diverse device environments. Because it is costly to develop and optimize a neural network for every single environment, there is a line of research to search neural networks for multiple target environments efficiently. However, existing works for such a situation still suffer from requiring many GPUs and expensive costs. Motivated by this, we propose a novel neural network optimization framework named Bespoke for low-cost deployment. Our framework searches for a lightweight model by replacing parts of an original model with randomly selected alternatives, each of which comes from a pretrained neural network or the original model. In the practical sense, Bespoke has two significant merits. One is that it requires near zero cost for designing the search space of neural networks. The other merit is that it exploits the sub-networks of public pretrained neural networks, so the total cost is minimal compared to the existing works. We conduct experiments exploring Bespoke's the merits, and the results show that it finds efficient models for multiple targets with meager cost.

Downloads

Published

2023-06-26

How to Cite

Lee, J.-R., & Moon, Y.-H. (2023). Bespoke: A Block-Level Neural Network Optimization Framework for Low-Cost Deployment. Proceedings of the AAAI Conference on Artificial Intelligence, 37(7), 8465-8472. https://doi.org/10.1609/aaai.v37i7.26020

Issue

Section

AAAI Technical Track on Machine Learning II