Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.

碩士 === 國立臺灣大學 === 電子工程學研究所 === 107 === Most recent studies on deep learning based speech enhancement (SE) focused on improving denoising performance. However, successful SE applications require striking a desirable balance between denoising performance and computational cost in real scenarios. In th...

Full description

Bibliographic Details
Main Authors: Jyun-Yi Wu, 吳俊易
Other Authors: Shao-Yi Chien
Format: Others
Language:en_US
Published: 2019
Online Access:http://ndltd.ncl.edu.tw/handle/rjz956
id ndltd-TW-107NTU05428116
record_format oai_dc
spelling ndltd-TW-107NTU054281162019-11-21T05:34:27Z http://ndltd.ncl.edu.tw/handle/rjz956 Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques. 利用參數修剪與量化技術以精簡語音除噪之深度學習模型 Jyun-Yi Wu 吳俊易 碩士 國立臺灣大學 電子工程學研究所 107 Most recent studies on deep learning based speech enhancement (SE) focused on improving denoising performance. However, successful SE applications require striking a desirable balance between denoising performance and computational cost in real scenarios. In this study, we propose a novel parameter pruning (PP) technique, which removes redundant channels in a neural network. In addition, a parameter quantization (PQ) technique was applied to reduce the size of a neural network by representing weights with fewer cluster centroids. Because the techniques are derived based on different concepts, the PP and PQ can be integrated to provide even more compact SE models. The experimental results show that the PP and PQ techniques produce a compacted SE model with a size of only 10.03% compared to that of the original model, resulting in minor performance losses of 1.43% (from 0.70 to 0.69) for STOI and 3.24% (from 1.85 to 1.79) for PESQ. The promising results suggest that the PP and PQ techniques can be used in a SE system in devices with limited storage and computation resources. Shao-Yi Chien 簡韶逸 2019 學位論文 ; thesis 66 en_US
collection NDLTD
language en_US
format Others
sources NDLTD
description 碩士 === 國立臺灣大學 === 電子工程學研究所 === 107 === Most recent studies on deep learning based speech enhancement (SE) focused on improving denoising performance. However, successful SE applications require striking a desirable balance between denoising performance and computational cost in real scenarios. In this study, we propose a novel parameter pruning (PP) technique, which removes redundant channels in a neural network. In addition, a parameter quantization (PQ) technique was applied to reduce the size of a neural network by representing weights with fewer cluster centroids. Because the techniques are derived based on different concepts, the PP and PQ can be integrated to provide even more compact SE models. The experimental results show that the PP and PQ techniques produce a compacted SE model with a size of only 10.03% compared to that of the original model, resulting in minor performance losses of 1.43% (from 0.70 to 0.69) for STOI and 3.24% (from 1.85 to 1.79) for PESQ. The promising results suggest that the PP and PQ techniques can be used in a SE system in devices with limited storage and computation resources.
author2 Shao-Yi Chien
author_facet Shao-Yi Chien
Jyun-Yi Wu
吳俊易
author Jyun-Yi Wu
吳俊易
spellingShingle Jyun-Yi Wu
吳俊易
Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
author_sort Jyun-Yi Wu
title Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
title_short Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
title_full Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
title_fullStr Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
title_full_unstemmed Increasing Compactness Of Deep Learning Based Speech Enhancement Models With Parameter Pruning And Quantization Techniques.
title_sort increasing compactness of deep learning based speech enhancement models with parameter pruning and quantization techniques.
publishDate 2019
url http://ndltd.ncl.edu.tw/handle/rjz956
work_keys_str_mv AT jyunyiwu increasingcompactnessofdeeplearningbasedspeechenhancementmodelswithparameterpruningandquantizationtechniques
AT wújùnyì increasingcompactnessofdeeplearningbasedspeechenhancementmodelswithparameterpruningandquantizationtechniques
AT jyunyiwu lìyòngcānshùxiūjiǎnyǔliànghuàjìshùyǐjīngjiǎnyǔyīnchúzàozhīshēndùxuéxímóxíng
AT wújùnyì lìyòngcānshùxiūjiǎnyǔliànghuàjìshùyǐjīngjiǎnyǔyīnchúzàozhīshēndùxuéxímóxíng
_version_ 1719294524831825920