Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models
碩士 === 國立臺灣科技大學 === 電機工程系 === 106 === Deep learning still encounters several issues like overfitting and oversize due to the use of a large number of layers. The huge size greatly constrains performance and portability of deep learning models in limited environments like embedded and IoT devices. In...
Main Author: | |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/2cvk37 |
id |
ndltd-TW-106NTUS5442128 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-106NTUS54421282019-05-16T00:59:40Z http://ndltd.ncl.edu.tw/handle/2cvk37 Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models Brilian Tafjira Nugraha Brilian Tafjira Nugraha 碩士 國立臺灣科技大學 電機工程系 106 Deep learning still encounters several issues like overfitting and oversize due to the use of a large number of layers. The huge size greatly constrains performance and portability of deep learning models in limited environments like embedded and IoT devices. In this study, we reported our analysis of activation and gradient output and weight in each layer of the pre-trained models of VGG-16 and custom AlexNet to measure the efficiency of its layers. The efficiencies are estimated by using our measurements and compared with the manual layer reduction to validate the most relevant method. The method for multiple layer reductions is used for validation. With this found approach, the time of one-layer reduction and re-training processes on both models can save up to 9 folds and 5 folds respectively without significant accuracy reduction. Shun-Feng Su 蘇順豐 2018 學位論文 ; thesis 48 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺灣科技大學 === 電機工程系 === 106 === Deep learning still encounters several issues like overfitting and oversize due to the use of a large number of layers. The huge size greatly constrains performance and portability of deep learning models in limited environments like embedded and IoT devices. In this study, we reported our analysis of activation and gradient output and weight in each layer of the pre-trained models of VGG-16 and custom AlexNet to measure the efficiency of its layers. The efficiencies are estimated by using our measurements and compared with the manual layer reduction to validate the most relevant method. The method for multiple layer reductions is used for validation. With this found approach, the time of one-layer reduction and re-training processes on both models can save up to 9 folds and 5 folds respectively without significant accuracy reduction.
|
author2 |
Shun-Feng Su |
author_facet |
Shun-Feng Su Brilian Tafjira Nugraha Brilian Tafjira Nugraha |
author |
Brilian Tafjira Nugraha Brilian Tafjira Nugraha |
spellingShingle |
Brilian Tafjira Nugraha Brilian Tafjira Nugraha Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
author_sort |
Brilian Tafjira Nugraha |
title |
Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
title_short |
Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
title_full |
Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
title_fullStr |
Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
title_full_unstemmed |
Analysis of Layer Efficiency and Layer Reduction on Pre-trained CNN Models |
title_sort |
analysis of layer efficiency and layer reduction on pre-trained cnn models |
publishDate |
2018 |
url |
http://ndltd.ncl.edu.tw/handle/2cvk37 |
work_keys_str_mv |
AT briliantafjiranugraha analysisoflayerefficiencyandlayerreductiononpretrainedcnnmodels AT briliantafjiranugraha analysisoflayerefficiencyandlayerreductiononpretrainedcnnmodels |
_version_ |
1719172437773385728 |