Unsupervised Word Translation with Adversarial Autoencoder

Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and th...

Full description

Bibliographic Details
Main Authors: Mohiuddin, Tasnim, Joty, Shafiq
Format: Article
Language:English
Published: The MIT Press 2020-06-01
Series:Computational Linguistics
Online Access:https://www.mitpressjournals.org/doi/abs/10.1162/coli_a_00374
id doaj-1d799d878fe74f28888282c8d64bd3a3
record_format Article
spelling doaj-1d799d878fe74f28888282c8d64bd3a32020-11-25T03:16:29ZengThe MIT PressComputational Linguistics0891-20171530-93122020-06-0146225728810.1162/coli_a_00374Unsupervised Word Translation with Adversarial AutoencoderMohiuddin, TasnimJoty, Shafiq Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and the associated word translation task without any parallel data by mapping monolingual embeddings to a shared space. However, recent work has shown superior performance for non-adversarial methods in more challenging language pairs. In this article, we investigate adversarial autoencoder for unsupervised word translation and propose two novel extensions to it that yield more stable training and improved results. Our method includes regularization terms to enforce cycle consistency and input reconstruction, and puts the target encoders as an adversary against the corresponding discriminator. We use two types of refinement procedures sequentially after obtaining the trained encoders and mappings from the adversarial training, namely, refinement with Procrustes solution and refinement with symmetric re-weighting. Extensive experimentations with high- and low-resource languages from two different data sets show that our method achieves better performance than existing adversarial and non-adversarial approaches and is also competitive with the supervised system. Along with performing comprehensive ablation studies to understand the contribution of different components of our adversarial model, we also conduct a thorough analysis of the refinement procedures to understand their effects. https://www.mitpressjournals.org/doi/abs/10.1162/coli_a_00374
collection DOAJ
language English
format Article
sources DOAJ
author Mohiuddin, Tasnim
Joty, Shafiq
spellingShingle Mohiuddin, Tasnim
Joty, Shafiq
Unsupervised Word Translation with Adversarial Autoencoder
Computational Linguistics
author_facet Mohiuddin, Tasnim
Joty, Shafiq
author_sort Mohiuddin, Tasnim
title Unsupervised Word Translation with Adversarial Autoencoder
title_short Unsupervised Word Translation with Adversarial Autoencoder
title_full Unsupervised Word Translation with Adversarial Autoencoder
title_fullStr Unsupervised Word Translation with Adversarial Autoencoder
title_full_unstemmed Unsupervised Word Translation with Adversarial Autoencoder
title_sort unsupervised word translation with adversarial autoencoder
publisher The MIT Press
series Computational Linguistics
issn 0891-2017
1530-9312
publishDate 2020-06-01
description Crosslingual word embeddings learned from monolingual embeddings have a crucial role in many downstream tasks, ranging from machine translation to transfer learning. Adversarial training has shown impressive success in learning crosslingual embeddings and the associated word translation task without any parallel data by mapping monolingual embeddings to a shared space. However, recent work has shown superior performance for non-adversarial methods in more challenging language pairs. In this article, we investigate adversarial autoencoder for unsupervised word translation and propose two novel extensions to it that yield more stable training and improved results. Our method includes regularization terms to enforce cycle consistency and input reconstruction, and puts the target encoders as an adversary against the corresponding discriminator. We use two types of refinement procedures sequentially after obtaining the trained encoders and mappings from the adversarial training, namely, refinement with Procrustes solution and refinement with symmetric re-weighting. Extensive experimentations with high- and low-resource languages from two different data sets show that our method achieves better performance than existing adversarial and non-adversarial approaches and is also competitive with the supervised system. Along with performing comprehensive ablation studies to understand the contribution of different components of our adversarial model, we also conduct a thorough analysis of the refinement procedures to understand their effects.
url https://www.mitpressjournals.org/doi/abs/10.1162/coli_a_00374
work_keys_str_mv AT mohiuddintasnim unsupervisedwordtranslationwithadversarialautoencoder
AT jotyshafiq unsupervisedwordtranslationwithadversarialautoencoder
_version_ 1724635891528368128