Improving discourse representations with node hierarchy attention

Long text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coh...

Full description

Bibliographic Details
Main Authors: Erfaneh Gharavi, Hadi Veisi, Rupesh Silwal, Matthew S. Gerber
Format: Article
Language:English
Published: Elsevier 2021-03-01
Series:Machine Learning with Applications
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S2666827020300153
id doaj-6981f3f5f775483f8fcb9ab43a2be028
record_format Article
spelling doaj-6981f3f5f775483f8fcb9ab43a2be0282021-05-06T04:25:52ZengElsevierMachine Learning with Applications2666-82702021-03-013100015Improving discourse representations with node hierarchy attentionErfaneh Gharavi0Hadi Veisi1Rupesh Silwal2Matthew S. Gerber3Department of Engineering Systems and Environment, University of Virginia, Charlottesville, USADepartment of New Sciences and Technologies, University of Tehran, Tehran, Iran; Corresponding author.Department of Engineering Systems and Environment, University of Virginia, Charlottesville, USADepartment of Engineering Systems and Environment, University of Virginia, Charlottesville, USALong text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coherent document there exist systematic connections between sentences. Rhetorical structure theory models this connection in a tree structure format. This tree models text span and their relation. The importance of each text span is distinguished by their hierarchy type in the tree named nucleus and satellite. In this paper, we try to enrich text representation by taking into account the contribution of each phrase in the text based on its hierarchy type. We employ a deep recursive neural network as the attention mechanism to improve text representation. Our hypothesis is evaluated in a sentiment analysis framework. In addition, basic recursive neural network and predefined weighting attention consider as benchmarks. Results show that reweighting span vectors via a deeper layer of recursive neural network outperforms predefined scalar and no attention methods.http://www.sciencedirect.com/science/article/pii/S2666827020300153Text representationRecursive neural networkRhetorical structure theoryAttention mechanismSentiment analysisDocument embedding
collection DOAJ
language English
format Article
sources DOAJ
author Erfaneh Gharavi
Hadi Veisi
Rupesh Silwal
Matthew S. Gerber
spellingShingle Erfaneh Gharavi
Hadi Veisi
Rupesh Silwal
Matthew S. Gerber
Improving discourse representations with node hierarchy attention
Machine Learning with Applications
Text representation
Recursive neural network
Rhetorical structure theory
Attention mechanism
Sentiment analysis
Document embedding
author_facet Erfaneh Gharavi
Hadi Veisi
Rupesh Silwal
Matthew S. Gerber
author_sort Erfaneh Gharavi
title Improving discourse representations with node hierarchy attention
title_short Improving discourse representations with node hierarchy attention
title_full Improving discourse representations with node hierarchy attention
title_fullStr Improving discourse representations with node hierarchy attention
title_full_unstemmed Improving discourse representations with node hierarchy attention
title_sort improving discourse representations with node hierarchy attention
publisher Elsevier
series Machine Learning with Applications
issn 2666-8270
publishDate 2021-03-01
description Long text representation for natural language processing tasks has capture researchers’ attention recently. Beyond the sentence, finding a good representation for the text turns to the bag of the words that losses sequence order. Indeed, the text does not pattern in a haphazard way; rather, in a coherent document there exist systematic connections between sentences. Rhetorical structure theory models this connection in a tree structure format. This tree models text span and their relation. The importance of each text span is distinguished by their hierarchy type in the tree named nucleus and satellite. In this paper, we try to enrich text representation by taking into account the contribution of each phrase in the text based on its hierarchy type. We employ a deep recursive neural network as the attention mechanism to improve text representation. Our hypothesis is evaluated in a sentiment analysis framework. In addition, basic recursive neural network and predefined weighting attention consider as benchmarks. Results show that reweighting span vectors via a deeper layer of recursive neural network outperforms predefined scalar and no attention methods.
topic Text representation
Recursive neural network
Rhetorical structure theory
Attention mechanism
Sentiment analysis
Document embedding
url http://www.sciencedirect.com/science/article/pii/S2666827020300153
work_keys_str_mv AT erfanehgharavi improvingdiscourserepresentationswithnodehierarchyattention
AT hadiveisi improvingdiscourserepresentationswithnodehierarchyattention
AT rupeshsilwal improvingdiscourserepresentationswithnodehierarchyattention
AT matthewsgerber improvingdiscourserepresentationswithnodehierarchyattention
_version_ 1721457280967245824