Self-Supervised Contextual Data Augmentation for Natural Language Processing
In this paper, we propose a novel data augmentation method with respect to the target context of the data via self-supervised learning. Instead of looking for the exact synonyms of masked words, the proposed method finds words that can replace the original words considering the context. For self-sup...
Main Authors: | , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2019-11-01
|
Series: | Symmetry |
Subjects: | |
Online Access: | https://www.mdpi.com/2073-8994/11/11/1393 |
id |
doaj-558c881d11294cbab46edf7fc9008f74 |
---|---|
record_format |
Article |
spelling |
doaj-558c881d11294cbab46edf7fc9008f742020-11-24T21:50:05ZengMDPI AGSymmetry2073-89942019-11-011111139310.3390/sym11111393sym11111393Self-Supervised Contextual Data Augmentation for Natural Language ProcessingDongju Park0Chang Wook Ahn1Electrical Engineering and Computer Science, Gwangju Institute of Science and Technology, Gwangju 61005, KoreaElectrical Engineering and Computer Science, Gwangju Institute of Science and Technology, Gwangju 61005, KoreaIn this paper, we propose a novel data augmentation method with respect to the target context of the data via self-supervised learning. Instead of looking for the exact synonyms of masked words, the proposed method finds words that can replace the original words considering the context. For self-supervised learning, we can employ the masked language model (MLM), which masks a specific word within a sentence and obtains the original word. The MLM learns the context of a sentence through asymmetrical inputs and outputs. However, without using the existing MLM, we propose a label-masked language model (LMLM) that can include label information for the mask tokens used in the MLM to effectively use the MLM in data with label information. The augmentation method performs self-supervised learning using LMLM and then implements data augmentation through the trained model. We demonstrate that our proposed method improves the classification accuracy of recurrent neural networks and convolutional neural network-based classifiers through several experiments for text classification benchmark datasets, including the Stanford Sentiment Treebank-5 (SST5), the Stanford Sentiment Treebank-2 (SST2), the subjectivity (Subj), the Multi-Perspective Question Answering (MPQA), the Movie Reviews (MR), and the Text Retrieval Conference (TREC) datasets. In addition, since the proposed method does not use external data, it can eliminate the time spent collecting external data, or pre-training using external data.https://www.mdpi.com/2073-8994/11/11/1393data augmentationself-supervised learningnatural language processingtext classification |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Dongju Park Chang Wook Ahn |
spellingShingle |
Dongju Park Chang Wook Ahn Self-Supervised Contextual Data Augmentation for Natural Language Processing Symmetry data augmentation self-supervised learning natural language processing text classification |
author_facet |
Dongju Park Chang Wook Ahn |
author_sort |
Dongju Park |
title |
Self-Supervised Contextual Data Augmentation for Natural Language Processing |
title_short |
Self-Supervised Contextual Data Augmentation for Natural Language Processing |
title_full |
Self-Supervised Contextual Data Augmentation for Natural Language Processing |
title_fullStr |
Self-Supervised Contextual Data Augmentation for Natural Language Processing |
title_full_unstemmed |
Self-Supervised Contextual Data Augmentation for Natural Language Processing |
title_sort |
self-supervised contextual data augmentation for natural language processing |
publisher |
MDPI AG |
series |
Symmetry |
issn |
2073-8994 |
publishDate |
2019-11-01 |
description |
In this paper, we propose a novel data augmentation method with respect to the target context of the data via self-supervised learning. Instead of looking for the exact synonyms of masked words, the proposed method finds words that can replace the original words considering the context. For self-supervised learning, we can employ the masked language model (MLM), which masks a specific word within a sentence and obtains the original word. The MLM learns the context of a sentence through asymmetrical inputs and outputs. However, without using the existing MLM, we propose a label-masked language model (LMLM) that can include label information for the mask tokens used in the MLM to effectively use the MLM in data with label information. The augmentation method performs self-supervised learning using LMLM and then implements data augmentation through the trained model. We demonstrate that our proposed method improves the classification accuracy of recurrent neural networks and convolutional neural network-based classifiers through several experiments for text classification benchmark datasets, including the Stanford Sentiment Treebank-5 (SST5), the Stanford Sentiment Treebank-2 (SST2), the subjectivity (Subj), the Multi-Perspective Question Answering (MPQA), the Movie Reviews (MR), and the Text Retrieval Conference (TREC) datasets. In addition, since the proposed method does not use external data, it can eliminate the time spent collecting external data, or pre-training using external data. |
topic |
data augmentation self-supervised learning natural language processing text classification |
url |
https://www.mdpi.com/2073-8994/11/11/1393 |
work_keys_str_mv |
AT dongjupark selfsupervisedcontextualdataaugmentationfornaturallanguageprocessing AT changwookahn selfsupervisedcontextualdataaugmentationfornaturallanguageprocessing |
_version_ |
1725885458327535616 |