Improving discourse representations with node hierarchy attention
Long text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coh...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Elsevier
2021-03-01
|
Series: | Machine Learning with Applications |
Subjects: | |
Online Access: | http://www.sciencedirect.com/science/article/pii/S2666827020300153 |
id |
doaj-6981f3f5f775483f8fcb9ab43a2be028 |
---|---|
record_format |
Article |
spelling |
doaj-6981f3f5f775483f8fcb9ab43a2be0282021-05-06T04:25:52ZengElsevierMachine Learning with Applications2666-82702021-03-013100015Improving discourse representations with node hierarchy attentionErfaneh Gharavi0Hadi Veisi1Rupesh Silwal2Matthew S. Gerber3Department of Engineering Systems and Environment, University of Virginia, Charlottesville, USADepartment of New Sciences and Technologies, University of Tehran, Tehran, Iran; Corresponding author.Department of Engineering Systems and Environment, University of Virginia, Charlottesville, USADepartment of Engineering Systems and Environment, University of Virginia, Charlottesville, USALong text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coherent document there exist systematic connections between sentences. Rhetorical structure theory models this connection in a tree structure format. This tree models text span and their relation. The importance of each text span is distinguished by their hierarchy type in the tree named nucleus and satellite. In this paper, we try to enrich text representation by taking into account the contribution of each phrase in the text based on its hierarchy type. We employ a deep recursive neural network as the attention mechanism to improve text representation. Our hypothesis is evaluated in a sentiment analysis framework. In addition, basic recursive neural network and predefined weighting attention consider as benchmarks. Results show that reweighting span vectors via a deeper layer of recursive neural network outperforms predefined scalar and no attention methods.http://www.sciencedirect.com/science/article/pii/S2666827020300153Text representationRecursive neural networkRhetorical structure theoryAttention mechanismSentiment analysisDocument embedding |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Erfaneh Gharavi Hadi Veisi Rupesh Silwal Matthew S. Gerber |
spellingShingle |
Erfaneh Gharavi Hadi Veisi Rupesh Silwal Matthew S. Gerber Improving discourse representations with node hierarchy attention Machine Learning with Applications Text representation Recursive neural network Rhetorical structure theory Attention mechanism Sentiment analysis Document embedding |
author_facet |
Erfaneh Gharavi Hadi Veisi Rupesh Silwal Matthew S. Gerber |
author_sort |
Erfaneh Gharavi |
title |
Improving discourse representations with node hierarchy attention |
title_short |
Improving discourse representations with node hierarchy attention |
title_full |
Improving discourse representations with node hierarchy attention |
title_fullStr |
Improving discourse representations with node hierarchy attention |
title_full_unstemmed |
Improving discourse representations with node hierarchy attention |
title_sort |
improving discourse representations with node hierarchy attention |
publisher |
Elsevier |
series |
Machine Learning with Applications |
issn |
2666-8270 |
publishDate |
2021-03-01 |
description |
Long text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coherent document there exist systematic connections between sentences. Rhetorical structure theory models this connection in a tree structure format. This tree models text span and their relation. The importance of each text span is distinguished by their hierarchy type in the tree named nucleus and satellite. In this paper, we try to enrich text representation by taking into account the contribution of each phrase in the text based on its hierarchy type. We employ a deep recursive neural network as the attention mechanism to improve text representation. Our hypothesis is evaluated in a sentiment analysis framework. In addition, basic recursive neural network and predefined weighting attention consider as benchmarks. Results show that reweighting span vectors via a deeper layer of recursive neural network outperforms predefined scalar and no attention methods. |
topic |
Text representation Recursive neural network Rhetorical structure theory Attention mechanism Sentiment analysis Document embedding |
url |
http://www.sciencedirect.com/science/article/pii/S2666827020300153 |
work_keys_str_mv |
AT erfanehgharavi improvingdiscourserepresentationswithnodehierarchyattention AT hadiveisi improvingdiscourserepresentationswithnodehierarchyattention AT rupeshsilwal improvingdiscourserepresentationswithnodehierarchyattention AT matthewsgerber improvingdiscourserepresentationswithnodehierarchyattention |
_version_ |
1721457280967245824 |