Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach

Participants in an asynchronous conversation (e.g., forum, e-mail) interact with each other at different times, performing certain communicative acts, called speech acts (e.g., question, request). In this article, we propose a hybrid approach to speech act recognition in asynchronous conversations....

Full description

Bibliographic Details
Main Authors: Shafiq Joty, Tasnim Mohiuddin
Format: Article
Language:English
Published: The MIT Press 2018-12-01
Series:Computational Linguistics
Online Access:https://www.mitpressjournals.org/doi/pdf/10.1162/coli_a_00339
id doaj-b39810f4e7d246e4bfa3c63a25b16152
record_format Article
spelling doaj-b39810f4e7d246e4bfa3c63a25b161522020-11-24T21:22:10ZengThe MIT PressComputational Linguistics1530-93122018-12-0144485989410.1162/coli_a_00339coli_a_00339Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF ApproachShafiq Joty0Tasnim Mohiuddin1Nanyang Technological University, School of Computer Science and Engineering. srjoty@ntu.edu.sgNanyang Technological University, School of Computer Science and Engineering. mohi0004@e.ntu.edu.sgParticipants in an asynchronous conversation (e.g., forum, e-mail) interact with each other at different times, performing certain communicative acts, called speech acts (e.g., question, request). In this article, we propose a hybrid approach to speech act recognition in asynchronous conversations. Our approach works in two main steps: a long short-term memory recurrent neural network (LSTM-RNN) first encodes each sentence separately into a task-specific distributed representation, and this is then used in a conditional random field (CRF) model to capture the conversational dependencies between sentences. The LSTM-RNN model uses pretrained word embeddings learned from a large conversational corpus and is trained to classify sentences into speech act types. The CRF model can consider arbitrary graph structures to model conversational dependencies in an asynchronous conversation. In addition, to mitigate the problem of limited annotated data in the asynchronous domains, we adapt the LSTM-RNN model to learn from synchronous conversations (e.g., meetings), using domain adversarial training of neural networks. Empirical evaluation shows the effectiveness of our approach over existing ones: (i) LSTM-RNNs provide better task-specific representations, (ii) conversational word embeddings benefit the LSTM-RNNs more than the off-the-shelf ones, (iii) adversarial training gives better domain-invariant representations, and (iv) the global CRF model improves over local models.https://www.mitpressjournals.org/doi/pdf/10.1162/coli_a_00339
collection DOAJ
language English
format Article
sources DOAJ
author Shafiq Joty
Tasnim Mohiuddin
spellingShingle Shafiq Joty
Tasnim Mohiuddin
Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
Computational Linguistics
author_facet Shafiq Joty
Tasnim Mohiuddin
author_sort Shafiq Joty
title Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
title_short Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
title_full Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
title_fullStr Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
title_full_unstemmed Modeling Speech Acts in Asynchronous Conversations: A Neural-CRF Approach
title_sort modeling speech acts in asynchronous conversations: a neural-crf approach
publisher The MIT Press
series Computational Linguistics
issn 1530-9312
publishDate 2018-12-01
description Participants in an asynchronous conversation (e.g., forum, e-mail) interact with each other at different times, performing certain communicative acts, called speech acts (e.g., question, request). In this article, we propose a hybrid approach to speech act recognition in asynchronous conversations. Our approach works in two main steps: a long short-term memory recurrent neural network (LSTM-RNN) first encodes each sentence separately into a task-specific distributed representation, and this is then used in a conditional random field (CRF) model to capture the conversational dependencies between sentences. The LSTM-RNN model uses pretrained word embeddings learned from a large conversational corpus and is trained to classify sentences into speech act types. The CRF model can consider arbitrary graph structures to model conversational dependencies in an asynchronous conversation. In addition, to mitigate the problem of limited annotated data in the asynchronous domains, we adapt the LSTM-RNN model to learn from synchronous conversations (e.g., meetings), using domain adversarial training of neural networks. Empirical evaluation shows the effectiveness of our approach over existing ones: (i) LSTM-RNNs provide better task-specific representations, (ii) conversational word embeddings benefit the LSTM-RNNs more than the off-the-shelf ones, (iii) adversarial training gives better domain-invariant representations, and (iv) the global CRF model improves over local models.
url https://www.mitpressjournals.org/doi/pdf/10.1162/coli_a_00339
work_keys_str_mv AT shafiqjoty modelingspeechactsinasynchronousconversationsaneuralcrfapproach
AT tasnimmohiuddin modelingspeechactsinasynchronousconversationsaneuralcrfapproach
_version_ 1725997130402758656