High Accuracy RRAM-Based Binary Neural Network Training

碩士 === 國立交通大學 === 電子研究所 === 107 === Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity bina...

Full description

Bibliographic Details
Main Authors: Lin, Jia-Wei, 林佳緯
Other Authors: Chang, Tian-Sheuan
Format: Others
Language:en_US
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/g5t559
id ndltd-TW-107NCTU5428055
record_format oai_dc
spelling ndltd-TW-107NCTU54280552019-05-16T01:40:47Z http://ndltd.ncl.edu.tw/handle/g5t559 High Accuracy RRAM-Based Binary Neural Network Training 高精確度電阻式記憶體二元類神經網路訓練 Lin, Jia-Wei 林佳緯 碩士 國立交通大學 電子研究所 107 Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity binary neural network model. A promising solution, in-memory computing, computes directly in memory with high parallelism and reduces the data transmission time, especially with the emerging new device such as Resistive Random Access Memory (RRAM) due to its small area and low power nature. However, RRAM devices suffers from the non-ideal state switching and variation characteristics, which could seriously affects the performance of neural networks, even with the on-device training. To solve above problem, we first analyze the non-ideal device effect on the network training performance and propose to combine residual network architectures, more channel width, and 3x3 kernel size for better training on non-ideal devices. Furthermore, we systematically determine the hyper-parameters of RRAM-based BNNs with bad probability updating curve. This training strategy makes it get better training results. Besides, device variation due to non-uniformity of the process will make training process unstable and thus get a worse results. To overcome this problem, we propose hybrid RRAM-based BNNs to combine the benefits of both digital computation and RRAM-based computation. Moreover, we use a longer training period with the above determined hyper-parameters to further improve the tolerance on the C2C variation. With the designs mentioned above, we can improve the accuracy of RRAM-based BNN from 57.39% to 83.69% even though the devices are in a worst condition. Chang, Tian-Sheuan 張添烜 2018 學位論文 ; thesis 52 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立交通大學 === 電子研究所 === 107 === Deep learning is getting popular at the edge devices to provide better services. However, its high computational complexity and memory bandwidth makes real time execution on the traditional Von Neuman architecture very challenging even with the low complexity binary neural network model. A promising solution, in-memory computing, computes directly in memory with high parallelism and reduces the data transmission time, especially with the emerging new device such as Resistive Random Access Memory (RRAM) due to its small area and low power nature. However, RRAM devices suffers from the non-ideal state switching and variation characteristics, which could seriously affects the performance of neural networks, even with the on-device training. To solve above problem, we first analyze the non-ideal device effect on the network training performance and propose to combine residual network architectures, more channel width, and 3x3 kernel size for better training on non-ideal devices. Furthermore, we systematically determine the hyper-parameters of RRAM-based BNNs with bad probability updating curve. This training strategy makes it get better training results. Besides, device variation due to non-uniformity of the process will make training process unstable and thus get a worse results. To overcome this problem, we propose hybrid RRAM-based BNNs to combine the benefits of both digital computation and RRAM-based computation. Moreover, we use a longer training period with the above determined hyper-parameters to further improve the tolerance on the C2C variation. With the designs mentioned above, we can improve the accuracy of RRAM-based BNN from 57.39% to 83.69% even though the devices are in a worst condition.
author2 Chang, Tian-Sheuan
author_facet Chang, Tian-Sheuan
Lin, Jia-Wei
林佳緯
author Lin, Jia-Wei
林佳緯
spellingShingle Lin, Jia-Wei
林佳緯
High Accuracy RRAM-Based Binary Neural Network Training
author_sort Lin, Jia-Wei
title High Accuracy RRAM-Based Binary Neural Network Training
title_short High Accuracy RRAM-Based Binary Neural Network Training
title_full High Accuracy RRAM-Based Binary Neural Network Training
title_fullStr High Accuracy RRAM-Based Binary Neural Network Training
title_full_unstemmed High Accuracy RRAM-Based Binary Neural Network Training
title_sort high accuracy rram-based binary neural network training
publishDate 2018
url http://ndltd.ncl.edu.tw/handle/g5t559
work_keys_str_mv AT linjiawei highaccuracyrrambasedbinaryneuralnetworktraining
AT línjiāwěi highaccuracyrrambasedbinaryneuralnetworktraining
AT linjiawei gāojīngquèdùdiànzǔshìjìyìtǐèryuánlèishénjīngwǎnglùxùnliàn
AT línjiāwěi gāojīngquèdùdiànzǔshìjìyìtǐèryuánlèishénjīngwǎnglùxùnliàn
_version_ 1719178696492843008