The Generator of Convolutional Neural Network Acceleration Circuit

碩士 === 國立中山大學 === 資訊工程學系研究所 === 107 === Due to the explosively growing number of applications of neural network (NN) based machine learning models, how to design efficient accelerator circuits for NN has become a very hot topic in recent years. The paper first proposed a VLSI architecture of the co...

Full description

Bibliographic Details
Main Authors: Min-Tao Chen, 陳民濤
Other Authors: Yun-Nan Chang
Format: Others
Language:zh-TW
Published: 2019
Online Access:http://ndltd.ncl.edu.tw/handle/982fcv
id ndltd-TW-107NSYS5392031
record_format oai_dc
spelling ndltd-TW-107NSYS53920312019-05-16T01:40:51Z http://ndltd.ncl.edu.tw/handle/982fcv The Generator of Convolutional Neural Network Acceleration Circuit 卷積神經網路加速電路之生成器 Min-Tao Chen 陳民濤 碩士 國立中山大學 資訊工程學系研究所 107 Due to the explosively growing number of applications of neural network (NN) based machine learning models, how to design efficient accelerator circuits for NN has become a very hot topic in recent years. The paper first proposed a VLSI architecture of the convolutional NN (CNN) accelerator based on the use of small-size processing unit (PE) which contains a single multiply-accumulator (MAC). This type of PE can lead to high hardware utilization for wide range of CNN kernel sizes. The other salient feature of the proposed CNN architecture is it also adopts an efficient line buffer design which can also support various filter kernel sizes. Traditional line buffer design consisting of simply shift registers can reduce the access of the next-level memory hierarchy significantly when processing the convolutional operations. However, in order to accommodate the applications of nowadays CNN models which normally contain different filter sizes, this thesis has extended the conventional line buffer design by including some data skipping mechanism. The resulted architecture can generate the required data stream seamlessly for various number of PEs to achieve the maximum hardware utilization. The proposed line-buffer design is thus very suitable for the design of reconfigurable CNN accelerators. The proposed overall CNN circuit has been realized in an SOC (system-on-a-chip) FPGA platform with OS environment. The driver of the circuit has also been implemented, such that this thesis can demonstrate the classification of CIFAR-10 dataset based on our accelerator, and display the classification result on the screen. Finally, based on the proposed CNN accelerator architecture, this thesis proposed a generator which can generate the accelerator according to the given input setting parameters including the number of processing elements (PE) and pooling function being selected. The users can also decide the size of on-chip memory used in the accelerator circuits. The related software environment under the Xilinx Zedboard environment will also be generated. Yun-Nan Chang 張雲南 2019 學位論文 ; thesis 64 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立中山大學 === 資訊工程學系研究所 === 107 === Due to the explosively growing number of applications of neural network (NN) based machine learning models, how to design efficient accelerator circuits for NN has become a very hot topic in recent years. The paper first proposed a VLSI architecture of the convolutional NN (CNN) accelerator based on the use of small-size processing unit (PE) which contains a single multiply-accumulator (MAC). This type of PE can lead to high hardware utilization for wide range of CNN kernel sizes. The other salient feature of the proposed CNN architecture is it also adopts an efficient line buffer design which can also support various filter kernel sizes. Traditional line buffer design consisting of simply shift registers can reduce the access of the next-level memory hierarchy significantly when processing the convolutional operations. However, in order to accommodate the applications of nowadays CNN models which normally contain different filter sizes, this thesis has extended the conventional line buffer design by including some data skipping mechanism. The resulted architecture can generate the required data stream seamlessly for various number of PEs to achieve the maximum hardware utilization. The proposed line-buffer design is thus very suitable for the design of reconfigurable CNN accelerators. The proposed overall CNN circuit has been realized in an SOC (system-on-a-chip) FPGA platform with OS environment. The driver of the circuit has also been implemented, such that this thesis can demonstrate the classification of CIFAR-10 dataset based on our accelerator, and display the classification result on the screen. Finally, based on the proposed CNN accelerator architecture, this thesis proposed a generator which can generate the accelerator according to the given input setting parameters including the number of processing elements (PE) and pooling function being selected. The users can also decide the size of on-chip memory used in the accelerator circuits. The related software environment under the Xilinx Zedboard environment will also be generated.
author2 Yun-Nan Chang
author_facet Yun-Nan Chang
Min-Tao Chen
陳民濤
author Min-Tao Chen
陳民濤
spellingShingle Min-Tao Chen
陳民濤
The Generator of Convolutional Neural Network Acceleration Circuit
author_sort Min-Tao Chen
title The Generator of Convolutional Neural Network Acceleration Circuit
title_short The Generator of Convolutional Neural Network Acceleration Circuit
title_full The Generator of Convolutional Neural Network Acceleration Circuit
title_fullStr The Generator of Convolutional Neural Network Acceleration Circuit
title_full_unstemmed The Generator of Convolutional Neural Network Acceleration Circuit
title_sort generator of convolutional neural network acceleration circuit
publishDate 2019
url http://ndltd.ncl.edu.tw/handle/982fcv
work_keys_str_mv AT mintaochen thegeneratorofconvolutionalneuralnetworkaccelerationcircuit
AT chénmíntāo thegeneratorofconvolutionalneuralnetworkaccelerationcircuit
AT mintaochen juǎnjīshénjīngwǎnglùjiāsùdiànlùzhīshēngchéngqì
AT chénmíntāo juǎnjīshénjīngwǎnglùjiāsùdiànlùzhīshēngchéngqì
AT mintaochen generatorofconvolutionalneuralnetworkaccelerationcircuit
AT chénmíntāo generatorofconvolutionalneuralnetworkaccelerationcircuit
_version_ 1719178949630623744