Deep Neural Network for Mandarin and Non-Mandarin Recognition System
碩士 === 國立臺北科技大學 === 電子工程系研究所 === 104 === The thesis aims to optimize the Deep Neural Network (DNN) to eliminate the noise to reach better speech recognition of Chinese and non-Chinese. For the building model, the Deep Neural Network has prevailed and had a breakthrough in all aspects in the recent y...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Online Access: | http://ndltd.ncl.edu.tw/handle/fj2ejc |
id |
ndltd-TW-104TIT05427123 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-104TIT054271232019-05-15T23:53:22Z http://ndltd.ncl.edu.tw/handle/fj2ejc Deep Neural Network for Mandarin and Non-Mandarin Recognition System 基於深層類神經網路之中文與非中文語言確認系統 Jayo Wong 翁嘉佑 碩士 國立臺北科技大學 電子工程系研究所 104 The thesis aims to optimize the Deep Neural Network (DNN) to eliminate the noise to reach better speech recognition of Chinese and non-Chinese. For the building model, the Deep Neural Network has prevailed and had a breakthrough in all aspects in the recent years. However, it is uncertain whether they can be applied to the speech recognition of Chinese and non-Chinese. The thesis transformed the eigenvectors of Universal Background Model (UBM) into low-dimension eigenvectors with i-Vector. The thesis also experimented the effect of Activation Function (Sigmoid, Hyperbolic Tangent, ReLU, ELU) , Dropout, and Training (Back Propagation, AdaGrad) to optimize the Deep Neural Network (DNN). Then, the results were compared with Linear Discriminant Analysis system (LDA) and Probabilistic Linear Discriminant Analysis system (PLDA). Experimental results on the Chinese and non-Chinese speech recognition program evaluation database show the proposed method gained relative performance 0.0879 in EER value and 1.31% in equal error rate (EER) comparing with a baseline provided by program. Yuan-Fu Liao 廖元甫 學位論文 ; thesis 0 |
collection |
NDLTD |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺北科技大學 === 電子工程系研究所 === 104 === The thesis aims to optimize the Deep Neural Network (DNN) to eliminate the noise to reach better speech recognition of Chinese and non-Chinese. For the building model, the Deep Neural Network has prevailed and had a breakthrough in all aspects in the recent years. However, it is uncertain whether they can be applied to the speech recognition of Chinese and non-Chinese. The thesis transformed the eigenvectors of Universal Background Model (UBM) into low-dimension eigenvectors with i-Vector. The thesis also experimented the effect of Activation Function (Sigmoid, Hyperbolic Tangent, ReLU, ELU) , Dropout, and Training (Back Propagation, AdaGrad) to optimize the Deep Neural Network (DNN). Then, the results were compared with Linear Discriminant Analysis system (LDA) and Probabilistic Linear Discriminant Analysis system (PLDA).
Experimental results on the Chinese and non-Chinese speech recognition program evaluation database show the proposed method gained relative performance 0.0879 in EER value and 1.31% in equal error rate (EER) comparing with a baseline provided by program.
|
author2 |
Yuan-Fu Liao |
author_facet |
Yuan-Fu Liao Jayo Wong 翁嘉佑 |
author |
Jayo Wong 翁嘉佑 |
spellingShingle |
Jayo Wong 翁嘉佑 Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
author_sort |
Jayo Wong |
title |
Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
title_short |
Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
title_full |
Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
title_fullStr |
Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
title_full_unstemmed |
Deep Neural Network for Mandarin and Non-Mandarin Recognition System |
title_sort |
deep neural network for mandarin and non-mandarin recognition system |
url |
http://ndltd.ncl.edu.tw/handle/fj2ejc |
work_keys_str_mv |
AT jayowong deepneuralnetworkformandarinandnonmandarinrecognitionsystem AT wēngjiāyòu deepneuralnetworkformandarinandnonmandarinrecognitionsystem AT jayowong jīyúshēncénglèishénjīngwǎnglùzhīzhōngwényǔfēizhōngwényǔyánquèrènxìtǒng AT wēngjiāyòu jīyúshēncénglèishénjīngwǎnglùzhīzhōngwényǔfēizhōngwényǔyánquèrènxìtǒng |
_version_ |
1719155898952187904 |