Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks
碩士 === 國立中正大學 === 電機工程研究所 === 106 === Lowering the computation complexity is essential for deep neural networks (DNN) to be integrated into cost- and power-sensitive embedded systems. This thesis first proposes a design-time normalization technique to convert a floating-point DNN model into a fixed...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2017
|
Online Access: | http://ndltd.ncl.edu.tw/handle/d48y57 |
id |
ndltd-TW-106CCU00442007 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-106CCU004420072019-05-16T00:00:45Z http://ndltd.ncl.edu.tw/handle/d48y57 Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks 適用於嵌入式神經網路之有號二冪次神經元設計及其權重近似方法 LIN, CHI-YOU 林祺祐 碩士 國立中正大學 電機工程研究所 106 Lowering the computation complexity is essential for deep neural networks (DNN) to be integrated into cost- and power-sensitive embedded systems. This thesis first proposes a design-time normalization technique to convert a floating-point DNN model into a fixed-point one, of which the concept is derived from shared exponents in “block floating-point.” Tedious tradeoffs between integer and fractional bits in traditional floating-point to fixed-point conversion are thus eliminated. In addition, fine-grained per-neuron normalization is proposed to further reduce 1~2-bit effective wordlength. Then, this thesis proposes synapse weight approximation with limited signed-power-of-two (SPT) terms, in order to replace the extensive multiplications in conventional DNN with shifts and additions. In our implementations in 28nm CMOS, the maximum performance and silicon area of a neuron has been improved by 36% and 18% simultaneously, and the silicon area can be improved by ~39% for identical performance. Finally, the vector quantization technique in image compression has been applied to select several outstanding weights, where all synapse weights are replaced by these outstanding weights. In our simulations with the MNIST dataset, only 4~16 outstanding weights can achieve accuracy similar to 16-bit full-set weights. In other words, the weight storage can be reduced by 4X to 8X effectively. YEH, CHING-WEI 葉經緯 2017 學位論文 ; thesis 42 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立中正大學 === 電機工程研究所 === 106 === Lowering the computation complexity is essential for deep neural networks (DNN) to be integrated into cost- and power-sensitive embedded systems. This thesis first proposes a design-time normalization technique to convert a floating-point DNN model into a fixed-point one, of which the concept is derived from shared exponents in “block floating-point.” Tedious tradeoffs between integer and fractional bits in traditional floating-point to fixed-point conversion are thus eliminated. In addition, fine-grained per-neuron normalization is proposed to further reduce 1~2-bit effective wordlength. Then, this thesis proposes synapse weight approximation with limited signed-power-of-two (SPT) terms, in order to replace the extensive multiplications in conventional DNN with shifts and additions. In our implementations in 28nm CMOS, the maximum performance and silicon area of a neuron has been improved by 36% and 18% simultaneously, and the silicon area can be improved by ~39% for identical performance. Finally, the vector quantization technique in image compression has been applied to select several outstanding weights, where all synapse weights are replaced by these outstanding weights. In our simulations with the MNIST dataset, only 4~16 outstanding weights can achieve accuracy similar to 16-bit full-set weights. In other words, the weight storage can be reduced by 4X to 8X effectively.
|
author2 |
YEH, CHING-WEI |
author_facet |
YEH, CHING-WEI LIN, CHI-YOU 林祺祐 |
author |
LIN, CHI-YOU 林祺祐 |
spellingShingle |
LIN, CHI-YOU 林祺祐 Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
author_sort |
LIN, CHI-YOU |
title |
Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
title_short |
Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
title_full |
Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
title_fullStr |
Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
title_full_unstemmed |
Signed-Power-of-Two (SPT) Neuron Design and Synapse Weight Approximation for Embedded Neural Networks |
title_sort |
signed-power-of-two (spt) neuron design and synapse weight approximation for embedded neural networks |
publishDate |
2017 |
url |
http://ndltd.ncl.edu.tw/handle/d48y57 |
work_keys_str_mv |
AT linchiyou signedpoweroftwosptneurondesignandsynapseweightapproximationforembeddedneuralnetworks AT línqíyòu signedpoweroftwosptneurondesignandsynapseweightapproximationforembeddedneuralnetworks AT linchiyou shìyòngyúqiànrùshìshénjīngwǎnglùzhīyǒuhàoèrmìcìshénjīngyuánshèjìjíqíquánzhòngjìnshìfāngfǎ AT línqíyòu shìyòngyúqiànrùshìshénjīngwǎnglùzhīyǒuhàoèrmìcìshénjīngyuánshèjìjíqíquánzhòngjìnshìfāngfǎ |
_version_ |
1719158587223179264 |