A Low Effort Approach to Structured CNN Design Using PCA

Deep learning models hold state of the art performance in many fields, yet their design is still based on heuristics or grid search methods that often result in overparametrized networks. This work proposes a method to analyze a trained network and deduce an optimized, compressed architecture that p...

Full description

Bibliographic Details
Main Authors: Isha Garg, Priyadarshini Panda, Kaushik Roy
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
PCA
Online Access:https://ieeexplore.ieee.org/document/8941144/
id doaj-6a3ff45440a5453e86ca06b3b1e65a72
record_format Article
spelling doaj-6a3ff45440a5453e86ca06b3b1e65a722021-03-30T02:48:46ZengIEEEIEEE Access2169-35362020-01-0181347136010.1109/ACCESS.2019.29619608941144A Low Effort Approach to Structured CNN Design Using PCAIsha Garg0https://orcid.org/0000-0003-4702-9444Priyadarshini Panda1https://orcid.org/0000-0002-4167-6782Kaushik Roy2https://orcid.org/0000-0002-0735-9695School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN, USAElectrical Engineering Department, Yale University, New Haven, U.K.School of Electrical and Computer Engineering, Purdue University, West Lafayette, IN, USADeep learning models hold state of the art performance in many fields, yet their design is still based on heuristics or grid search methods that often result in overparametrized networks. This work proposes a method to analyze a trained network and deduce an optimized, compressed architecture that preserves accuracy while keeping computational costs tractable. Model compression is an active field of research that targets the problem of realizing deep learning models in hardware. However, most pruning methodologies tend to be experimental, requiring large compute and time intensive iterations of retraining the entire network. We introduce structure into model design by proposing a single shot analysis of a trained network that serves as a first order, low effort approach to dimensionality reduction, by using PCA (Principal Component Analysis). The proposed method simultaneously analyzes the activations of each layer and considers the dimensionality of the space described by the filters generating these activations. It optimizes the architecture in terms of number of layers, and number of filters per layer without any iterative retraining procedures, making it a viable, low effort technique to design efficient networks. We demonstrate the proposed methodology on AlexNet and VGG style networks on the CIFAR-10, CIFAR-100 and ImageNet datasets, and successfully achieve an optimized architecture with a reduction of up to 3.8X and 9X in the number of operations and parameters respectively, while trading off less than 1% accuracy. We also apply the method to MobileNet, and achieve 1.7X and 3.9X reduction in the number of operations and parameters respectively, while improving accuracy by almost one percentage point.https://ieeexplore.ieee.org/document/8941144/CNNsefficient deep learningmodel architecturemodel compressionPCAdimensionality reduction
collection DOAJ
language English
format Article
sources DOAJ
author Isha Garg
Priyadarshini Panda
Kaushik Roy
spellingShingle Isha Garg
Priyadarshini Panda
Kaushik Roy
A Low Effort Approach to Structured CNN Design Using PCA
IEEE Access
CNNs
efficient deep learning
model architecture
model compression
PCA
dimensionality reduction
author_facet Isha Garg
Priyadarshini Panda
Kaushik Roy
author_sort Isha Garg
title A Low Effort Approach to Structured CNN Design Using PCA
title_short A Low Effort Approach to Structured CNN Design Using PCA
title_full A Low Effort Approach to Structured CNN Design Using PCA
title_fullStr A Low Effort Approach to Structured CNN Design Using PCA
title_full_unstemmed A Low Effort Approach to Structured CNN Design Using PCA
title_sort low effort approach to structured cnn design using pca
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Deep learning models hold state of the art performance in many fields, yet their design is still based on heuristics or grid search methods that often result in overparametrized networks. This work proposes a method to analyze a trained network and deduce an optimized, compressed architecture that preserves accuracy while keeping computational costs tractable. Model compression is an active field of research that targets the problem of realizing deep learning models in hardware. However, most pruning methodologies tend to be experimental, requiring large compute and time intensive iterations of retraining the entire network. We introduce structure into model design by proposing a single shot analysis of a trained network that serves as a first order, low effort approach to dimensionality reduction, by using PCA (Principal Component Analysis). The proposed method simultaneously analyzes the activations of each layer and considers the dimensionality of the space described by the filters generating these activations. It optimizes the architecture in terms of number of layers, and number of filters per layer without any iterative retraining procedures, making it a viable, low effort technique to design efficient networks. We demonstrate the proposed methodology on AlexNet and VGG style networks on the CIFAR-10, CIFAR-100 and ImageNet datasets, and successfully achieve an optimized architecture with a reduction of up to 3.8X and 9X in the number of operations and parameters respectively, while trading off less than 1% accuracy. We also apply the method to MobileNet, and achieve 1.7X and 3.9X reduction in the number of operations and parameters respectively, while improving accuracy by almost one percentage point.
topic CNNs
efficient deep learning
model architecture
model compression
PCA
dimensionality reduction
url https://ieeexplore.ieee.org/document/8941144/
work_keys_str_mv AT ishagarg aloweffortapproachtostructuredcnndesignusingpca
AT priyadarshinipanda aloweffortapproachtostructuredcnndesignusingpca
AT kaushikroy aloweffortapproachtostructuredcnndesignusingpca
AT ishagarg loweffortapproachtostructuredcnndesignusingpca
AT priyadarshinipanda loweffortapproachtostructuredcnndesignusingpca
AT kaushikroy loweffortapproachtostructuredcnndesignusingpca
_version_ 1724184534801448960