R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
Machine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to addres...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2019-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/8651505/ |
id |
doaj-8ec47b5917a041aa9e32b24c00374b4d |
---|---|
record_format |
Article |
spelling |
doaj-8ec47b5917a041aa9e32b24c00374b4d2021-03-29T22:29:09ZengIEEEIEEE Access2169-35362019-01-017277362774510.1109/ACCESS.2019.29015478651505R-Trans: RNN Transformer Network for Chinese Machine Reading ComprehensionShanshan Liu0https://orcid.org/0000-0002-4966-679XSheng Zhang1https://orcid.org/0000-0003-1732-0011Xin Zhang2Hui Wang3Science and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaMachine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to address English MRC. When applying them directly to Chinese documents, the performance often degrades considerably because of some special characteristics of Chinese, the inevitable segmentation errors in particular. In this paper, we present the RNN Transformer network to tackle the Chinese MRC task. To mitigate the influence of incorrect word segmentation and mine sequential information of whole sentences, deep contextualized word representations and bidirectional gated recurrent units networks are adopted in our model. The extensive experiments have been conducted on a very large scale Chinese MRC corpus, viz., the Les MMRC dataset. The results show that the proposed model outperforms the baseline and other prevalent MRC models notably, and established a new state-of-the-art record on the Les MMRC dataset.https://ieeexplore.ieee.org/document/8651505/Contextualized word representationdeep learningmachine reading comprehension |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Shanshan Liu Sheng Zhang Xin Zhang Hui Wang |
spellingShingle |
Shanshan Liu Sheng Zhang Xin Zhang Hui Wang R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension IEEE Access Contextualized word representation deep learning machine reading comprehension |
author_facet |
Shanshan Liu Sheng Zhang Xin Zhang Hui Wang |
author_sort |
Shanshan Liu |
title |
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension |
title_short |
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension |
title_full |
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension |
title_fullStr |
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension |
title_full_unstemmed |
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension |
title_sort |
r-trans: rnn transformer network for chinese machine reading comprehension |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2019-01-01 |
description |
Machine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to address English MRC. When applying them directly to Chinese documents, the performance often degrades considerably because of some special characteristics of Chinese, the inevitable segmentation errors in particular. In this paper, we present the RNN Transformer network to tackle the Chinese MRC task. To mitigate the influence of incorrect word segmentation and mine sequential information of whole sentences, deep contextualized word representations and bidirectional gated recurrent units networks are adopted in our model. The extensive experiments have been conducted on a very large scale Chinese MRC corpus, viz., the Les MMRC dataset. The results show that the proposed model outperforms the baseline and other prevalent MRC models notably, and established a new state-of-the-art record on the Les MMRC dataset. |
topic |
Contextualized word representation deep learning machine reading comprehension |
url |
https://ieeexplore.ieee.org/document/8651505/ |
work_keys_str_mv |
AT shanshanliu rtransrnntransformernetworkforchinesemachinereadingcomprehension AT shengzhang rtransrnntransformernetworkforchinesemachinereadingcomprehension AT xinzhang rtransrnntransformernetworkforchinesemachinereadingcomprehension AT huiwang rtransrnntransformernetworkforchinesemachinereadingcomprehension |
_version_ |
1724191528591556608 |