High Accuracy RRAM-Based Binary Neural Network Training
碩士 === 國立交通大學 === 電子研究所 === 107 === Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity bina...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | en_US |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/g5t559 |
id |
ndltd-TW-107NCTU5428055 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-107NCTU54280552019-05-16T01:40:47Z http://ndltd.ncl.edu.tw/handle/g5t559 High Accuracy RRAM-Based Binary Neural Network Training 高精確度電阻式記憶體二元類神經網路訓練 Lin, Jia-Wei 林佳緯 碩士 國立交通大學 電子研究所 107 Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity binary neural network model. A promising solution, in-memory computing, computes directly in memory with high parallelism and reduces the data transmission time, especially with the emerging new device such as Resistive Random Access Memory (RRAM) due to its small area and low power nature. However, RRAM devices suffers from the non-ideal state switching and variation characteristics, which could seriously affects the performance of neural networks, even with the on-device training. To solve above problem, we first analyze the non-ideal device effect on the network training performance and propose to combine residual network architectures, more channel width, and 3x3 kernel size for better training on non-ideal devices. Furthermore, we systematically determine the hyper-parameters of RRAM-based BNNs with bad probability updating curve. This training strategy makes it get better training results. Besides, device variation due to non-uniformity of the process will make training process unstable and thus get a worse results. To overcome this problem, we propose hybrid RRAM-based BNNs to combine the benefits of both digital computation and RRAM-based computation. Moreover, we use a longer training period with the above determined hyper-parameters to further improve the tolerance on the C2C variation. With the designs mentioned above, we can improve the accuracy of RRAM-based BNN from 57.39% to 83.69% even though the devices are in a worst condition. Chang, Tian-Sheuan 張添烜 2018 學位論文 ; thesis 52 en_US |
collection |
NDLTD |
language |
en_US |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立交通大學 === 電子研究所 === 107 === Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity binary neural network model. A promising solution, in-memory computing, computes directly in memory with high parallelism and reduces the data transmission time, especially with the emerging new device such as Resistive Random Access Memory (RRAM) due to its small area and low power nature. However, RRAM devices suffers from the non-ideal state switching and variation characteristics, which could seriously affects the performance of neural networks, even with the on-device training.
To solve above problem, we first analyze the non-ideal device effect on the network training performance and propose to combine residual network architectures, more channel width, and 3x3 kernel size for better training on non-ideal devices. Furthermore, we systematically determine the hyper-parameters of RRAM-based BNNs with bad probability updating curve. This training strategy makes it get better training results. Besides, device variation due to non-uniformity of the process will make training process unstable and thus get a worse results. To overcome this problem, we propose hybrid RRAM-based BNNs to combine the benefits of both digital computation and RRAM-based computation. Moreover, we use a longer training period with the above determined hyper-parameters to further improve the tolerance on the C2C variation. With the designs mentioned above, we can improve the accuracy of RRAM-based BNN from 57.39% to 83.69% even though the devices are in a worst condition.
|
author2 |
Chang, Tian-Sheuan |
author_facet |
Chang, Tian-Sheuan Lin, Jia-Wei 林佳緯 |
author |
Lin, Jia-Wei 林佳緯 |
spellingShingle |
Lin, Jia-Wei 林佳緯 High Accuracy RRAM-Based Binary Neural Network Training |
author_sort |
Lin, Jia-Wei |
title |
High Accuracy RRAM-Based Binary Neural Network Training |
title_short |
High Accuracy RRAM-Based Binary Neural Network Training |
title_full |
High Accuracy RRAM-Based Binary Neural Network Training |
title_fullStr |
High Accuracy RRAM-Based Binary Neural Network Training |
title_full_unstemmed |
High Accuracy RRAM-Based Binary Neural Network Training |
title_sort |
high accuracy rram-based binary neural network training |
publishDate |
2018 |
url |
http://ndltd.ncl.edu.tw/handle/g5t559 |
work_keys_str_mv |
AT linjiawei highaccuracyrrambasedbinaryneuralnetworktraining AT línjiāwěi highaccuracyrrambasedbinaryneuralnetworktraining AT linjiawei gāojīngquèdùdiànzǔshìjìyìtǐèryuánlèishénjīngwǎnglùxùnliàn AT línjiāwěi gāojīngquèdùdiànzǔshìjìyìtǐèryuánlèishénjīngwǎnglùxùnliàn |
_version_ |
1719178696492843008 |