Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
Currently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover,...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9178269/ |
id |
doaj-69c46e04203c45d0b2c83ddf07fe1716 |
---|---|
record_format |
Article |
spelling |
doaj-69c46e04203c45d0b2c83ddf07fe17162021-03-30T03:32:01ZengIEEEIEEE Access2169-35362020-01-01816200416201110.1109/ACCESS.2020.30195009178269Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse RelationsDan Jiang0https://orcid.org/0000-0001-6686-8054Jin He1https://orcid.org/0000-0002-4142-2045School of Software Engineering, Beijing University of Posts and Telecommunications (BUPT), Beijing, ChinaSchool of Software Engineering, Beijing University of Posts and Telecommunications (BUPT), Beijing, ChinaCurrently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover, the lack of Chinese labeled data makes it more difficult to complete this task with high accuracy. To address this issue, we propose a novel hybrid DRR model combining a pretrained language model, namely bidirectional encoder representations from transformers (BERT), with recurrent neural networks. We use BERT as a text representation and pretraining model. In addition, we apply a tree structure to the implicit DRR in Chinese to produce hierarchical classes. The 19-class F1 score of our proposed method can reach 74.47% on the HIT-CIR Chinese discourse relation corpus. The attained results showed that the use of BERT and the proposed tree structure forms a novel and precise method that can automatically recognize the implicit relations of Chinese discourse.https://ieeexplore.ieee.org/document/9178269/Discourse relation recognitiondeep neural networktree structureBERT |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Dan Jiang Jin He |
spellingShingle |
Dan Jiang Jin He Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations IEEE Access Discourse relation recognition deep neural network tree structure BERT |
author_facet |
Dan Jiang Jin He |
author_sort |
Dan Jiang |
title |
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations |
title_short |
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations |
title_full |
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations |
title_fullStr |
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations |
title_full_unstemmed |
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations |
title_sort |
tree framework with bert word embedding for the recognition of chinese implicit discourse relations |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
Currently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover, the lack of Chinese labeled data makes it more difficult to complete this task with high accuracy. To address this issue, we propose a novel hybrid DRR model combining a pretrained language model, namely bidirectional encoder representations from transformers (BERT), with recurrent neural networks. We use BERT as a text representation and pretraining model. In addition, we apply a tree structure to the implicit DRR in Chinese to produce hierarchical classes. The 19-class F1 score of our proposed method can reach 74.47% on the HIT-CIR Chinese discourse relation corpus. The attained results showed that the use of BERT and the proposed tree structure forms a novel and precise method that can automatically recognize the implicit relations of Chinese discourse. |
topic |
Discourse relation recognition deep neural network tree structure BERT |
url |
https://ieeexplore.ieee.org/document/9178269/ |
work_keys_str_mv |
AT danjiang treeframeworkwithbertwordembeddingfortherecognitionofchineseimplicitdiscourserelations AT jinhe treeframeworkwithbertwordembeddingfortherecognitionofchineseimplicitdiscourserelations |
_version_ |
1724183292826091520 |