Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine
碩士 === 國立成功大學 === 電腦與通信工程研究所 === 107 === With the increasing popularity of mobile devices and the effectiveness of deep learning-based algorithms, people try to put deep learning models on mobile devices. However, it is limited by the complexity of computational and software overhead. We propose an...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2019
|
Online Access: | http://ndltd.ncl.edu.tw/handle/x46nq6 |
id |
ndltd-TW-107NCKU5652009 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-107NCKU56520092019-10-25T05:24:18Z http://ndltd.ncl.edu.tw/handle/x46nq6 Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine 深度卷積網路之逐層定點數量化方法與實作YOLOv3推論引擎 Wei-ChungTseng 曾微中 碩士 國立成功大學 電腦與通信工程研究所 107 With the increasing popularity of mobile devices and the effectiveness of deep learning-based algorithms, people try to put deep learning models on mobile devices. However, it is limited by the complexity of computational and software overhead. We propose an efficient framework for inference to fit resource-limited devices with about 1000 times smaller than Tensorflow in code size, and a layer-wised quantization scheme that allows inference computed by fixed-point arithmetic. The fixed-point quantization scheme is more efficient than floating point arithmetic with power consumption reduced to 8% left in cost grained evaluation and reduce model size to 40%~25% left, and keep TOP5 accuracy loss under 1% in Alexnet on ImageNet. Chung-Ho Chen 陳中和 2019 學位論文 ; thesis 70 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立成功大學 === 電腦與通信工程研究所 === 107 === With the increasing popularity of mobile devices and the effectiveness of deep learning-based algorithms, people try to put deep learning models on mobile devices. However, it is limited by the complexity of computational and software overhead.
We propose an efficient framework for inference to fit resource-limited devices with about 1000 times smaller than Tensorflow in code size, and a layer-wised quantization scheme that allows inference computed by fixed-point arithmetic. The fixed-point quantization scheme is more efficient than floating point arithmetic with power consumption reduced to 8% left in cost grained evaluation and reduce model size to 40%~25% left, and keep TOP5 accuracy loss under 1% in Alexnet on ImageNet.
|
author2 |
Chung-Ho Chen |
author_facet |
Chung-Ho Chen Wei-ChungTseng 曾微中 |
author |
Wei-ChungTseng 曾微中 |
spellingShingle |
Wei-ChungTseng 曾微中 Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
author_sort |
Wei-ChungTseng |
title |
Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
title_short |
Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
title_full |
Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
title_fullStr |
Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
title_full_unstemmed |
Layer-wise Fixed Point Quantization for Deep Convolutional Neural Networks and Implementation of YOLOv3 Inference Engine |
title_sort |
layer-wise fixed point quantization for deep convolutional neural networks and implementation of yolov3 inference engine |
publishDate |
2019 |
url |
http://ndltd.ncl.edu.tw/handle/x46nq6 |
work_keys_str_mv |
AT weichungtseng layerwisefixedpointquantizationfordeepconvolutionalneuralnetworksandimplementationofyolov3inferenceengine AT céngwēizhōng layerwisefixedpointquantizationfordeepconvolutionalneuralnetworksandimplementationofyolov3inferenceengine AT weichungtseng shēndùjuǎnjīwǎnglùzhīzhúcéngdìngdiǎnshùliànghuàfāngfǎyǔshízuòyolov3tuīlùnyǐnqíng AT céngwēizhōng shēndùjuǎnjīwǎnglùzhīzhúcéngdìngdiǎnshùliànghuàfāngfǎyǔshízuòyolov3tuīlùnyǐnqíng |
_version_ |
1719277972964245504 |