Performance prediction based on neural architecture features

Neural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors...

Full description

Bibliographic Details
Main Authors: Duo Long, Shizhou Zhang, Yanning Zhang
Format: Article
Language:English
Published: Wiley 2020-04-01
Series:Cognitive Computation and Systems
Subjects:
Online Access:https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024
id doaj-857d6e71d0414bf785cc77ee7079a958
record_format Article
spelling doaj-857d6e71d0414bf785cc77ee7079a9582021-04-02T15:42:50ZengWileyCognitive Computation and Systems2517-75672020-04-0110.1049/ccs.2019.0024CCS.2019.0024Performance prediction based on neural architecture featuresDuo Long0Shizhou Zhang1Shizhou Zhang2Yanning Zhang3College of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityNeural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors try to use a performance prediction method to predict the model performance with little or even no training steps. They assume that the performance is determined once the architecture and other hyperparameters are chosen. So they first extract the sequence features of the chain-structured neural architecture by introducing the N-grams model to process architecture textual description. Subsequently, based on the extracted neural architecture features, they use the appropriate regression model to predict validation accuracies for a modelling learning curve. Through a series of experimental comparisons, they verify the effectiveness of the authors’ proposed performance prediction method and the effective acceleration of the NAS process.https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024neural netsregression analysislearning (artificial intelligence)n-grams modelarchitecture textual descriptionextracted neural architecture featuresappropriate regression modelmodelling learning curveperformance prediction methodneural architecture searchcandidate neural networkshigh-performance network architectureoptimising hyperparameterstraining stepssequence featureschain-structured neural architecture
collection DOAJ
language English
format Article
sources DOAJ
author Duo Long
Shizhou Zhang
Shizhou Zhang
Yanning Zhang
spellingShingle Duo Long
Shizhou Zhang
Shizhou Zhang
Yanning Zhang
Performance prediction based on neural architecture features
Cognitive Computation and Systems
neural nets
regression analysis
learning (artificial intelligence)
n-grams model
architecture textual description
extracted neural architecture features
appropriate regression model
modelling learning curve
performance prediction method
neural architecture search
candidate neural networks
high-performance network architecture
optimising hyperparameters
training steps
sequence features
chain-structured neural architecture
author_facet Duo Long
Shizhou Zhang
Shizhou Zhang
Yanning Zhang
author_sort Duo Long
title Performance prediction based on neural architecture features
title_short Performance prediction based on neural architecture features
title_full Performance prediction based on neural architecture features
title_fullStr Performance prediction based on neural architecture features
title_full_unstemmed Performance prediction based on neural architecture features
title_sort performance prediction based on neural architecture features
publisher Wiley
series Cognitive Computation and Systems
issn 2517-7567
publishDate 2020-04-01
description Neural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors try to use a performance prediction method to predict the model performance with little or even no training steps. They assume that the performance is determined once the architecture and other hyperparameters are chosen. So they first extract the sequence features of the chain-structured neural architecture by introducing the N-grams model to process architecture textual description. Subsequently, based on the extracted neural architecture features, they use the appropriate regression model to predict validation accuracies for a modelling learning curve. Through a series of experimental comparisons, they verify the effectiveness of the authors’ proposed performance prediction method and the effective acceleration of the NAS process.
topic neural nets
regression analysis
learning (artificial intelligence)
n-grams model
architecture textual description
extracted neural architecture features
appropriate regression model
modelling learning curve
performance prediction method
neural architecture search
candidate neural networks
high-performance network architecture
optimising hyperparameters
training steps
sequence features
chain-structured neural architecture
url https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024
work_keys_str_mv AT duolong performancepredictionbasedonneuralarchitecturefeatures
AT shizhouzhang performancepredictionbasedonneuralarchitecturefeatures
AT shizhouzhang performancepredictionbasedonneuralarchitecturefeatures
AT yanningzhang performancepredictionbasedonneuralarchitecturefeatures
_version_ 1721559343368765440