A Feature Fusion Based Indicator for Training-Free Neural Architecture Search

Neural Architecture Search Without Training (NASWOT) has been proposed recently to replace the conventional Neural Architecture Search (NAS). Pioneer works only deploy one or two indicator(s) to search. Nevertheless, the quantitative assessment for indicators is not fully studied and evaluated. In t...

Full description

Bibliographic Details
Main Authors: Linh-Tam Tran, Muhammad Salman Ali, Sung-Ho Bae
Format: Article
Language:English
Published: IEEE 2021-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9548935/
id doaj-4f327ee6bdb545e9aa3ef25fa49f7338
record_format Article
spelling doaj-4f327ee6bdb545e9aa3ef25fa49f73382021-10-06T23:00:25ZengIEEEIEEE Access2169-35362021-01-01913391413392310.1109/ACCESS.2021.31159119548935A Feature Fusion Based Indicator for Training-Free Neural Architecture SearchLinh-Tam Tran0https://orcid.org/0000-0002-9699-1747Muhammad Salman Ali1https://orcid.org/0000-0002-8548-3827Sung-Ho Bae2https://orcid.org/0000-0003-2677-3186Department of Computer Science and Engineering, Kyung Hee University, Yongin, Republic of KoreaDepartment of Computer Science and Engineering, Kyung Hee University, Yongin, Republic of KoreaDepartment of Computer Science and Engineering, Kyung Hee University, Yongin, Republic of KoreaNeural Architecture Search Without Training (NASWOT) has been proposed recently to replace the conventional Neural Architecture Search (NAS). Pioneer works only deploy one or two indicator(s) to search. Nevertheless, the quantitative assessment for indicators is not fully studied and evaluated. In this paper, we first review several indicators, which are used to evaluate the network in a training-free manner, including the correlation of Jacobian, the output sensitivity, the number of linear regions, and the condition number of the neural tangent kernel. Our observation is that each indicator is responsible for characterizing a network in a specific aspect and there is no single indicator that achieves good performance in all cases, e.g. highly correlated with the test accuracy. This motivated us to develop a novel indicator where all properties of a network are taken into account. To obtain better indicator that can consider various characteristics of networks in a harmonized form, we propose a Fusion Indicator (FI). Specifically, the proposed FI is formed by combining multiple indicators in a weighted sum manner. We minimize the mean squared error loss between the predicted and actual accuracy of networks to acquire the weights. Moreover, as the conventional training-free NAS researches used limited metrics to evaluate the quality of indicators, we introduce more desirable metrics that can evaluate the quality of training-free NAS indicator in terms of fidelity, correlation and rank-order similarity between the predicted quality value and actual accuracy of networks. That is, we introduce the Pearson Linear Coefficient Correlation (PLCC), the Root Mean Square Error (RMSE), the Spearman Rank-Order Correlation Coefficient (SROCC), and Kendall Rank-Order Correlation Coefficient (KROCC). Extensive experiments on NAS-Bench-101 and NAS-Bench-201 demonstrate the effectiveness of our FI, outperforming existing methods by a large margin.https://ieeexplore.ieee.org/document/9548935/Neural architecture searchtraining-free neural architecture searchfusion indicatorevaluation metrics
collection DOAJ
language English
format Article
sources DOAJ
author Linh-Tam Tran
Muhammad Salman Ali
Sung-Ho Bae
spellingShingle Linh-Tam Tran
Muhammad Salman Ali
Sung-Ho Bae
A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
IEEE Access
Neural architecture search
training-free neural architecture search
fusion indicator
evaluation metrics
author_facet Linh-Tam Tran
Muhammad Salman Ali
Sung-Ho Bae
author_sort Linh-Tam Tran
title A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
title_short A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
title_full A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
title_fullStr A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
title_full_unstemmed A Feature Fusion Based Indicator for Training-Free Neural Architecture Search
title_sort feature fusion based indicator for training-free neural architecture search
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2021-01-01
description Neural Architecture Search Without Training (NASWOT) has been proposed recently to replace the conventional Neural Architecture Search (NAS). Pioneer works only deploy one or two indicator(s) to search. Nevertheless, the quantitative assessment for indicators is not fully studied and evaluated. In this paper, we first review several indicators, which are used to evaluate the network in a training-free manner, including the correlation of Jacobian, the output sensitivity, the number of linear regions, and the condition number of the neural tangent kernel. Our observation is that each indicator is responsible for characterizing a network in a specific aspect and there is no single indicator that achieves good performance in all cases, e.g. highly correlated with the test accuracy. This motivated us to develop a novel indicator where all properties of a network are taken into account. To obtain better indicator that can consider various characteristics of networks in a harmonized form, we propose a Fusion Indicator (FI). Specifically, the proposed FI is formed by combining multiple indicators in a weighted sum manner. We minimize the mean squared error loss between the predicted and actual accuracy of networks to acquire the weights. Moreover, as the conventional training-free NAS researches used limited metrics to evaluate the quality of indicators, we introduce more desirable metrics that can evaluate the quality of training-free NAS indicator in terms of fidelity, correlation and rank-order similarity between the predicted quality value and actual accuracy of networks. That is, we introduce the Pearson Linear Coefficient Correlation (PLCC), the Root Mean Square Error (RMSE), the Spearman Rank-Order Correlation Coefficient (SROCC), and Kendall Rank-Order Correlation Coefficient (KROCC). Extensive experiments on NAS-Bench-101 and NAS-Bench-201 demonstrate the effectiveness of our FI, outperforming existing methods by a large margin.
topic Neural architecture search
training-free neural architecture search
fusion indicator
evaluation metrics
url https://ieeexplore.ieee.org/document/9548935/
work_keys_str_mv AT linhtamtran afeaturefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
AT muhammadsalmanali afeaturefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
AT sunghobae afeaturefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
AT linhtamtran featurefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
AT muhammadsalmanali featurefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
AT sunghobae featurefusionbasedindicatorfortrainingfreeneuralarchitecturesearch
_version_ 1716840374710829056