Performance prediction based on neural architecture features
Neural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors...
Main Authors: | , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Wiley
2020-04-01
|
Series: | Cognitive Computation and Systems |
Subjects: | |
Online Access: | https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024 |
id |
doaj-857d6e71d0414bf785cc77ee7079a958 |
---|---|
record_format |
Article |
spelling |
doaj-857d6e71d0414bf785cc77ee7079a9582021-04-02T15:42:50ZengWileyCognitive Computation and Systems2517-75672020-04-0110.1049/ccs.2019.0024CCS.2019.0024Performance prediction based on neural architecture featuresDuo Long0Shizhou Zhang1Shizhou Zhang2Yanning Zhang3College of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityCollege of Computer Science and Engineering, Northwestern Polytechnical UniversityNeural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors try to use a performance prediction method to predict the model performance with little or even no training steps. They assume that the performance is determined once the architecture and other hyperparameters are chosen. So they first extract the sequence features of the chain-structured neural architecture by introducing the N-grams model to process architecture textual description. Subsequently, based on the extracted neural architecture features, they use the appropriate regression model to predict validation accuracies for a modelling learning curve. Through a series of experimental comparisons, they verify the effectiveness of the authors’ proposed performance prediction method and the effective acceleration of the NAS process.https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024neural netsregression analysislearning (artificial intelligence)n-grams modelarchitecture textual descriptionextracted neural architecture featuresappropriate regression modelmodelling learning curveperformance prediction methodneural architecture searchcandidate neural networkshigh-performance network architectureoptimising hyperparameterstraining stepssequence featureschain-structured neural architecture |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Duo Long Shizhou Zhang Shizhou Zhang Yanning Zhang |
spellingShingle |
Duo Long Shizhou Zhang Shizhou Zhang Yanning Zhang Performance prediction based on neural architecture features Cognitive Computation and Systems neural nets regression analysis learning (artificial intelligence) n-grams model architecture textual description extracted neural architecture features appropriate regression model modelling learning curve performance prediction method neural architecture search candidate neural networks high-performance network architecture optimising hyperparameters training steps sequence features chain-structured neural architecture |
author_facet |
Duo Long Shizhou Zhang Shizhou Zhang Yanning Zhang |
author_sort |
Duo Long |
title |
Performance prediction based on neural architecture features |
title_short |
Performance prediction based on neural architecture features |
title_full |
Performance prediction based on neural architecture features |
title_fullStr |
Performance prediction based on neural architecture features |
title_full_unstemmed |
Performance prediction based on neural architecture features |
title_sort |
performance prediction based on neural architecture features |
publisher |
Wiley |
series |
Cognitive Computation and Systems |
issn |
2517-7567 |
publishDate |
2020-04-01 |
description |
Neural Architecture Search (NAS) usually requires to train quantities of candidate neural networks on a dataset for choosing a high-performance network architecture and optimising hyperparameters, which is very time consuming and computationally expensive. In order to resolve the issue, the authors try to use a performance prediction method to predict the model performance with little or even no training steps. They assume that the performance is determined once the architecture and other hyperparameters are chosen. So they first extract the sequence features of the chain-structured neural architecture by introducing the N-grams model to process architecture textual description. Subsequently, based on the extracted neural architecture features, they use the appropriate regression model to predict validation accuracies for a modelling learning curve. Through a series of experimental comparisons, they verify the effectiveness of the authors’ proposed performance prediction method and the effective acceleration of the NAS process. |
topic |
neural nets regression analysis learning (artificial intelligence) n-grams model architecture textual description extracted neural architecture features appropriate regression model modelling learning curve performance prediction method neural architecture search candidate neural networks high-performance network architecture optimising hyperparameters training steps sequence features chain-structured neural architecture |
url |
https://digital-library.theiet.org/content/journals/10.1049/ccs.2019.0024 |
work_keys_str_mv |
AT duolong performancepredictionbasedonneuralarchitecturefeatures AT shizhouzhang performancepredictionbasedonneuralarchitecturefeatures AT shizhouzhang performancepredictionbasedonneuralarchitecturefeatures AT yanningzhang performancepredictionbasedonneuralarchitecturefeatures |
_version_ |
1721559343368765440 |