Skip to content

Latest commit

 

History

History
11 lines (8 loc) · 3.37 KB

Istrate2019TAPAS.md

File metadata and controls

11 lines (8 loc) · 3.37 KB

Title

TAPAS: Train-Less Accuracy Predictor for Architecture Search

Author

R. Istrate, F. Scheidegger, G. Mariani, D. Nikolopoulos, C. Bekas, A. C. I. Malossi

Abstract

In recent years an increasing number of researchers and practitioners have been suggesting algorithms for large-scale neural network architecture search: genetic algorithms, reinforcement learning, learning curve extrapolation, and accuracy predictors. None of them, however, demonstrated highperformance without training new experiments in the presence of unseen datasets. We propose a new deep neural network accuracy predictor, that estimates in fractions of a second classification performance for unseen input datasets, without training. In contrast to previously proposed approaches, our prediction is not only calibrated on the topological network information, but also on the characterization of the dataset-difficulty which allows us to re-tune the prediction without any training. Our predictor achieves a performance which exceeds 100 networks per second on a single GPU, thus creating the opportunity to perform large-scale architecture search within a few minutes. We present results of two searches performed in 400 seconds on a single GPU. Our best discovered networks reach 93.67% accuracy for CIFAR-10 and 81.01% for CIFAR-100, verified by training. These networks are performance competitive with other automatically discovered state-of-the-art networks however we only needed a small fraction of the time to solution and computational resources.

Bib

@article{Istrate_Scheidegger_Mariani_Nikolopoulos_Bekas_Malossi_2019, title={TAPAS: Train-Less Accuracy Predictor for Architecture Search}, volume={33}, url={https://ojs.aaai.org/index.php/AAAI/article/view/4282}, DOI={10.1609/aaai.v33i01.33013927}, abstractNote={<p>In recent years an increasing number of researchers and practitioners have been suggesting algorithms for large-scale neural network architecture search: genetic algorithms, reinforcement learning, learning curve extrapolation, and accuracy predictors. None of them, however, demonstrated highperformance without training new experiments in the presence of unseen datasets. We propose a new deep neural network accuracy predictor, that estimates in fractions of a second classification performance for unseen input datasets, without training. In contrast to previously proposed approaches, our prediction is not only calibrated on the topological network information, but also on the characterization of the dataset-difficulty which allows us to re-tune the prediction without any training. Our predictor achieves a performance which exceeds 100 networks per second on a single GPU, thus creating the opportunity to perform large-scale architecture search within a few minutes. We present results of two searches performed in 400 seconds on a single GPU. Our best discovered networks reach 93.67% accuracy for CIFAR-10 and 81.01% for CIFAR-100, verified by training. These networks are performance competitive with other automatically discovered state-of-the-art networks however we only needed a small fraction of the time to solution and computational resources.</p>}, number={01}, journal={Proceedings of the AAAI Conference on Artificial Intelligence}, author={Istrate, R. and Scheidegger, F. and Mariani, G. and Nikolopoulos, D. and Bekas, C. and Malossi, A. C. I.}, year={2019}, month={Jul.}, pages={3927-3934} }