Recurrent Neural Network with Attention Mechanism for Language Model
碩士 === 國立臺中科技大學 === 資訊管理系碩士班 === 106 === The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opini...
Main Authors: | , |
---|---|
Other Authors: | |
Format: | Others |
Language: | zh-TW |
Published: |
2018
|
Online Access: | http://ndltd.ncl.edu.tw/handle/76y7wc |
id |
ndltd-TW-106NTTI5396006 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-TW-106NTTI53960062019-10-03T03:40:46Z http://ndltd.ncl.edu.tw/handle/76y7wc Recurrent Neural Network with Attention Mechanism for Language Model 基於深度學習循環類神經結合注意力機制建立語言模型 Tsung-Che Hsieh 謝宗哲 碩士 國立臺中科技大學 資訊管理系碩士班 106 The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opinions of the product, or some market-relevant information. However, some problems that point to "How to get features from the text” must be solved. The model of extracting the text features by using the neural Network method is called neural network language model (NNLM). The features are based on n-gram Model concept, which are the co-occurrence relationship between the vocabulary. The word vectors are important because the sentence vectors or the document vectors still have to understand the relationship between the words, and based on this, this study discussing the word vectors. This study assumes that the words contains "the meaning in sentences" and "the position of grammar". This study uses RNN (recurrent neural network) with attention mechanism to establish a language model. This study uses Penn Treebank (PTB), WikiText-2 (WT2) and NLPCC2017 text dataset. According to these dataset, the proposed models provide the better performance by the perplexity(PPL). 陳牧言 姜琇森 2018 學位論文 ; thesis 63 zh-TW |
collection |
NDLTD |
language |
zh-TW |
format |
Others
|
sources |
NDLTD |
description |
碩士 === 國立臺中科技大學 === 資訊管理系碩士班 === 106 === The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opinions of the product, or some market-relevant information. However, some problems that point to "How to get features from the text” must be solved. The model of extracting the text features by using the neural Network method is called neural network language model (NNLM). The features are based on n-gram Model concept, which are the co-occurrence relationship between the vocabulary. The word vectors are important because the sentence vectors or the document vectors still have to understand the relationship between the words, and based on this, this study discussing the word vectors. This study assumes that the words contains "the meaning in sentences" and "the position of grammar". This study uses RNN (recurrent neural network) with attention mechanism to establish a language model. This study uses Penn Treebank (PTB), WikiText-2 (WT2) and NLPCC2017 text dataset. According to these dataset, the proposed models provide the better performance by the perplexity(PPL).
|
author2 |
陳牧言 |
author_facet |
陳牧言 Tsung-Che Hsieh 謝宗哲 |
author |
Tsung-Che Hsieh 謝宗哲 |
spellingShingle |
Tsung-Che Hsieh 謝宗哲 Recurrent Neural Network with Attention Mechanism for Language Model |
author_sort |
Tsung-Che Hsieh |
title |
Recurrent Neural Network with Attention Mechanism for Language Model |
title_short |
Recurrent Neural Network with Attention Mechanism for Language Model |
title_full |
Recurrent Neural Network with Attention Mechanism for Language Model |
title_fullStr |
Recurrent Neural Network with Attention Mechanism for Language Model |
title_full_unstemmed |
Recurrent Neural Network with Attention Mechanism for Language Model |
title_sort |
recurrent neural network with attention mechanism for language model |
publishDate |
2018 |
url |
http://ndltd.ncl.edu.tw/handle/76y7wc |
work_keys_str_mv |
AT tsungchehsieh recurrentneuralnetworkwithattentionmechanismforlanguagemodel AT xièzōngzhé recurrentneuralnetworkwithattentionmechanismforlanguagemodel AT tsungchehsieh jīyúshēndùxuéxíxúnhuánlèishénjīngjiéhézhùyìlìjīzhìjiànlìyǔyánmóxíng AT xièzōngzhé jīyúshēndùxuéxíxúnhuánlèishénjīngjiéhézhùyìlìjīzhìjiànlìyǔyánmóxíng |
_version_ |
1719259235570679808 |