Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations

Currently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover,...

Full description

Bibliographic Details
Main Authors: Dan Jiang, Jin He
Format: Article
Language:English
Published: IEEE 2020-01-01
Series:IEEE Access
Subjects:
Online Access:https://ieeexplore.ieee.org/document/9178269/
id doaj-69c46e04203c45d0b2c83ddf07fe1716
record_format Article
spelling doaj-69c46e04203c45d0b2c83ddf07fe17162021-03-30T03:32:01ZengIEEEIEEE Access2169-35362020-01-01816200416201110.1109/ACCESS.2020.30195009178269Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse RelationsDan Jiang0https://orcid.org/0000-0001-6686-8054Jin He1https://orcid.org/0000-0002-4142-2045School of Software Engineering, Beijing University of Posts and Telecommunications (BUPT), Beijing, ChinaSchool of Software Engineering, Beijing University of Posts and Telecommunications (BUPT), Beijing, ChinaCurrently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover, the lack of Chinese labeled data makes it more difficult to complete this task with high accuracy. To address this issue, we propose a novel hybrid DRR model combining a pretrained language model, namely bidirectional encoder representations from transformers (BERT), with recurrent neural networks. We use BERT as a text representation and pretraining model. In addition, we apply a tree structure to the implicit DRR in Chinese to produce hierarchical classes. The 19-class F1 score of our proposed method can reach 74.47% on the HIT-CIR Chinese discourse relation corpus. The attained results showed that the use of BERT and the proposed tree structure forms a novel and precise method that can automatically recognize the implicit relations of Chinese discourse.https://ieeexplore.ieee.org/document/9178269/Discourse relation recognitiondeep neural networktree structureBERT
collection DOAJ
language English
format Article
sources DOAJ
author Dan Jiang
Jin He
spellingShingle Dan Jiang
Jin He
Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
IEEE Access
Discourse relation recognition
deep neural network
tree structure
BERT
author_facet Dan Jiang
Jin He
author_sort Dan Jiang
title Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
title_short Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
title_full Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
title_fullStr Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
title_full_unstemmed Tree Framework With BERT Word Embedding for the Recognition of Chinese Implicit Discourse Relations
title_sort tree framework with bert word embedding for the recognition of chinese implicit discourse relations
publisher IEEE
series IEEE Access
issn 2169-3536
publishDate 2020-01-01
description Currently, discourse relation recognition (DRR), which is not directly marked with connectives, is a challenging task. Traditional approaches for implicit DRR in Chinese have focused on exploring the concepts and features of words; however, these approaches have only yielded slow progress. Moreover, the lack of Chinese labeled data makes it more difficult to complete this task with high accuracy. To address this issue, we propose a novel hybrid DRR model combining a pretrained language model, namely bidirectional encoder representations from transformers (BERT), with recurrent neural networks. We use BERT as a text representation and pretraining model. In addition, we apply a tree structure to the implicit DRR in Chinese to produce hierarchical classes. The 19-class F1 score of our proposed method can reach 74.47% on the HIT-CIR Chinese discourse relation corpus. The attained results showed that the use of BERT and the proposed tree structure forms a novel and precise method that can automatically recognize the implicit relations of Chinese discourse.
topic Discourse relation recognition
deep neural network
tree structure
BERT
url https://ieeexplore.ieee.org/document/9178269/
work_keys_str_mv AT danjiang treeframeworkwithbertwordembeddingfortherecognitionofchineseimplicitdiscourserelations
AT jinhe treeframeworkwithbertwordembeddingfortherecognitionofchineseimplicitdiscourserelations
_version_ 1724183292826091520