Recurrent Neural Network with Attention Mechanism for Language Model

碩士 === 國立臺中科技大學 === 資訊管理系碩士班 === 106 === The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opini...

Full description

Bibliographic Details
Main Authors: Tsung-Che Hsieh, 謝宗哲
Other Authors: 陳牧言
Format: Others
Language:zh-TW
Published: 2018
Online Access:http://ndltd.ncl.edu.tw/handle/76y7wc
id ndltd-TW-106NTTI5396006
record_format oai_dc
spelling ndltd-TW-106NTTI53960062019-10-03T03:40:46Z http://ndltd.ncl.edu.tw/handle/76y7wc Recurrent Neural Network with Attention Mechanism for Language Model 基於深度學習循環類神經結合注意力機制建立語言模型 Tsung-Che Hsieh 謝宗哲 碩士 國立臺中科技大學 資訊管理系碩士班 106 The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opinions of the product, or some market-relevant information. However, some problems that point to "How to get features from the text” must be solved. The model of extracting the text features by using the neural Network method is called neural network language model (NNLM). The features are based on n-gram Model concept, which are the co-occurrence relationship between the vocabulary. The word vectors are important because the sentence vectors or the document vectors still have to understand the relationship between the words, and based on this, this study discussing the word vectors. This study assumes that the words contains "the meaning in sentences" and "the position of grammar". This study uses RNN (recurrent neural network) with attention mechanism to establish a language model. This study uses Penn Treebank (PTB), WikiText-2 (WT2) and NLPCC2017 text dataset. According to these dataset, the proposed models provide the better performance by the perplexity(PPL). 陳牧言 姜琇森 2018 學位論文 ; thesis 63 zh-TW
collection NDLTD
language zh-TW
format Others
sources NDLTD
description 碩士 === 國立臺中科技大學 === 資訊管理系碩士班 === 106 === The rapid growth of the Internet promotes the growth of textual data, and people get the information they need from the amount of textual data to solve problems. The textual data may include some potential information like the opinions of the crowd, the opinions of the product, or some market-relevant information. However, some problems that point to "How to get features from the text” must be solved. The model of extracting the text features by using the neural Network method is called neural network language model (NNLM). The features are based on n-gram Model concept, which are the co-occurrence relationship between the vocabulary. The word vectors are important because the sentence vectors or the document vectors still have to understand the relationship between the words, and based on this, this study discussing the word vectors. This study assumes that the words contains "the meaning in sentences" and "the position of grammar". This study uses RNN (recurrent neural network) with attention mechanism to establish a language model. This study uses Penn Treebank (PTB), WikiText-2 (WT2) and NLPCC2017 text dataset. According to these dataset, the proposed models provide the better performance by the perplexity(PPL).
author2 陳牧言
author_facet 陳牧言
Tsung-Che Hsieh
謝宗哲
author Tsung-Che Hsieh
謝宗哲
spellingShingle Tsung-Che Hsieh
謝宗哲
Recurrent Neural Network with Attention Mechanism for Language Model
author_sort Tsung-Che Hsieh
title Recurrent Neural Network with Attention Mechanism for Language Model
title_short Recurrent Neural Network with Attention Mechanism for Language Model
title_full Recurrent Neural Network with Attention Mechanism for Language Model
title_fullStr Recurrent Neural Network with Attention Mechanism for Language Model
title_full_unstemmed Recurrent Neural Network with Attention Mechanism for Language Model
title_sort recurrent neural network with attention mechanism for language model
publishDate 2018
url http://ndltd.ncl.edu.tw/handle/76y7wc
work_keys_str_mv AT tsungchehsieh recurrentneuralnetworkwithattentionmechanismforlanguagemodel
AT xièzōngzhé recurrentneuralnetworkwithattentionmechanismforlanguagemodel
AT tsungchehsieh jīyúshēndùxuéxíxúnhuánlèishénjīngjiéhézhùyìlìjīzhìjiànlìyǔyánmóxíng
AT xièzōngzhé jīyúshēndùxuéxíxúnhuánlèishénjīngjiéhézhùyìlìjīzhìjiànlìyǔyánmóxíng
_version_ 1719259235570679808