Sense classification of shallow discourse relations with focused RNNs.
Understanding the sense of discourse relations between segments of text is essential to truly comprehend any natural language text. Several automated approaches have been suggested, but all rely on external resources, linguistic feature engineering, and their processing pipelines are built from subs...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2018-01-01
|
Series: | PLoS ONE |
Online Access: | http://europepmc.org/articles/PMC6207334?pdf=render |
id |
doaj-594ccc4f87b14c49bb0186c0c72bf90a |
---|---|
record_format |
Article |
spelling |
doaj-594ccc4f87b14c49bb0186c0c72bf90a2020-11-24T22:08:08ZengPublic Library of Science (PLoS)PLoS ONE1932-62032018-01-011310e020605710.1371/journal.pone.0206057Sense classification of shallow discourse relations with focused RNNs.Gregor WeissMarko BajecUnderstanding the sense of discourse relations between segments of text is essential to truly comprehend any natural language text. Several automated approaches have been suggested, but all rely on external resources, linguistic feature engineering, and their processing pipelines are built from substantially different models. In this paper, we introduce a novel system for sense classification of shallow discourse relations (FR system) based on focused recurrent neural networks (RNNs). In contrast to existing systems, FR system consists of a single end-to-end trainable model for handling all types and senses of discourse relations, requires no feature engineering or external resources, is language-independent, and can be applied at the word and even character levels. At its core, we present our novel generalization of the focused RNNs layer, the first multi-dimensional RNN-attention mechanism for constructing text/argument embeddings. The filtering/gating RNN enables downstream RNNs to focus on different aspects of the input sequence and project it into several embedding subspaces. These argument embeddings are then used to perform sense classification. FR system has been evaluated using the official datasets and methodology of CoNLL 2016 Shared Task. It does not fall a lot behind state-of-the-art performance on English, the most researched and supported language, but it outperforms existing best systems by 2.5% overall results on the Chinese blind dataset.http://europepmc.org/articles/PMC6207334?pdf=render |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Gregor Weiss Marko Bajec |
spellingShingle |
Gregor Weiss Marko Bajec Sense classification of shallow discourse relations with focused RNNs. PLoS ONE |
author_facet |
Gregor Weiss Marko Bajec |
author_sort |
Gregor Weiss |
title |
Sense classification of shallow discourse relations with focused RNNs. |
title_short |
Sense classification of shallow discourse relations with focused RNNs. |
title_full |
Sense classification of shallow discourse relations with focused RNNs. |
title_fullStr |
Sense classification of shallow discourse relations with focused RNNs. |
title_full_unstemmed |
Sense classification of shallow discourse relations with focused RNNs. |
title_sort |
sense classification of shallow discourse relations with focused rnns. |
publisher |
Public Library of Science (PLoS) |
series |
PLoS ONE |
issn |
1932-6203 |
publishDate |
2018-01-01 |
description |
Understanding the sense of discourse relations between segments of text is essential to truly comprehend any natural language text. Several automated approaches have been suggested, but all rely on external resources, linguistic feature engineering, and their processing pipelines are built from substantially different models. In this paper, we introduce a novel system for sense classification of shallow discourse relations (FR system) based on focused recurrent neural networks (RNNs). In contrast to existing systems, FR system consists of a single end-to-end trainable model for handling all types and senses of discourse relations, requires no feature engineering or external resources, is language-independent, and can be applied at the word and even character levels. At its core, we present our novel generalization of the focused RNNs layer, the first multi-dimensional RNN-attention mechanism for constructing text/argument embeddings. The filtering/gating RNN enables downstream RNNs to focus on different aspects of the input sequence and project it into several embedding subspaces. These argument embeddings are then used to perform sense classification. FR system has been evaluated using the official datasets and methodology of CoNLL 2016 Shared Task. It does not fall a lot behind state-of-the-art performance on English, the most researched and supported language, but it outperforms existing best systems by 2.5% overall results on the Chinese blind dataset. |
url |
http://europepmc.org/articles/PMC6207334?pdf=render |
work_keys_str_mv |
AT gregorweiss senseclassificationofshallowdiscourserelationswithfocusedrnns AT markobajec senseclassificationofshallowdiscourserelationswithfocusedrnns |
_version_ |
1725817450723803136 |