R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension

Machine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to addres...

Full description

Bibliographic Details
Main Authors: Shanshan Liu, Sheng Zhang, Xin Zhang, Hui Wang
Format: Article
Language:English
Published: IEEE 2019-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/8651505/
id doaj-8ec47b5917a041aa9e32b24c00374b4d
record_format Article
spelling doaj-8ec47b5917a041aa9e32b24c00374b4d2021-03-29T22:29:09ZengIEEEIEEE Access2169-35362019-01-017277362774510.1109/ACCESS.2019.29015478651505R-Trans: RNN Transformer Network for Chinese Machine Reading ComprehensionShanshan Liu0https://orcid.org/0000-0002-4966-679XSheng Zhang1https://orcid.org/0000-0003-1732-0011Xin Zhang2Hui Wang3Science and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaScience and Technology on Information Systems Engineering Laboratory, College of Systems Engineering, National University of Defense Technology, Changsha, ChinaMachine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to address English MRC. When applying them directly to Chinese documents, the performance often degrades considerably because of some special characteristics of Chinese, the inevitable segmentation errors in particular. In this paper, we present the RNN Transformer network to tackle the Chinese MRC task. To mitigate the influence of incorrect word segmentation and mine sequential information of whole sentences, deep contextualized word representations and bidirectional gated recurrent units networks are adopted in our model. The extensive experiments have been conducted on a very large scale Chinese MRC corpus, viz., the Les MMRC dataset. The results show that the proposed model outperforms the baseline and other prevalent MRC models notably, and established a new state-of-the-art record on the Les MMRC dataset.https://ieeexplore.ieee.org/document/8651505/Contextualized word representationdeep learningmachine reading comprehension
collection DOAJ
language English
format Article
sources DOAJ
author Shanshan Liu
Sheng Zhang
Xin Zhang
Hui Wang
spellingShingle Shanshan Liu
Sheng Zhang
Xin Zhang
Hui Wang
R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
IEEE Access
Contextualized word representation
deep learning
machine reading comprehension
author_facet Shanshan Liu
Sheng Zhang
Xin Zhang
Hui Wang
author_sort Shanshan Liu
title R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
title_short R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
title_full R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
title_fullStr R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
title_full_unstemmed R-Trans: RNN Transformer Network for Chinese Machine Reading Comprehension
title_sort r-trans: rnn transformer network for chinese machine reading comprehension
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2019-01-01
description Machine reading comprehension (MRC) has gained increasingly wide attention over the past few years. A variety of benchmark datasets have been released, which triggers the development of quite a few MRC approaches based on deep learning techniques. However, most existing models are designed to address English MRC. When applying them directly to Chinese documents, the performance often degrades considerably because of some special characteristics of Chinese, the inevitable segmentation errors in particular. In this paper, we present the RNN Transformer network to tackle the Chinese MRC task. To mitigate the influence of incorrect word segmentation and mine sequential information of whole sentences, deep contextualized word representations and bidirectional gated recurrent units networks are adopted in our model. The extensive experiments have been conducted on a very large scale Chinese MRC corpus, viz., the Les MMRC dataset. The results show that the proposed model outperforms the baseline and other prevalent MRC models notably, and established a new state-of-the-art record on the Les MMRC dataset.
topic Contextualized word representation
deep learning
machine reading comprehension
url https://ieeexplore.ieee.org/document/8651505/
work_keys_str_mv AT shanshanliu rtransrnntransformernetworkforchinesemachinereadingcomprehension
AT shengzhang rtransrnntransformernetworkforchinesemachinereadingcomprehension
AT xinzhang rtransrnntransformernetworkforchinesemachinereadingcomprehension
AT huiwang rtransrnntransformernetworkforchinesemachinereadingcomprehension
_version_ 1724191528591556608