Changing the Geometry of Representations: <i>α</i>-Embeddings for NLP Tasks

Word embeddings based on a conditional model are commonly used in Natural Language Processing (NLP) tasks to embed the words of a dictionary in a low dimensional linear space. Their computation is based on the maximization of the likelihood of a conditional probability distribution for each word of...

Full description

Bibliographic Details
Main Authors: Riccardo Volpi, Uddhipan Thakur, Luigi Malagò
Format: Article
Language:English
Published: MDPI AG 2021-02-01
Series:Entropy
Subjects:
Online Access:https://www.mdpi.com/1099-4300/23/3/287
Description
Summary:Word embeddings based on a conditional model are commonly used in Natural Language Processing (NLP) tasks to embed the words of a dictionary in a low dimensional linear space. Their computation is based on the maximization of the likelihood of a conditional probability distribution for each word of the dictionary. These distributions form a Riemannian statistical manifold, where word embeddings can be interpreted as vectors in the tangent space of a specific reference measure on the manifold. A novel family of word embeddings, called <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>-embeddings have been recently introduced as deriving from the geometrical deformation of the simplex of probabilities through a parameter <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>, using notions from Information Geometry. After introducing the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>-embeddings, we show how the deformation of the simplex, controlled by <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>, provides an extra handle to increase the performances of several intrinsic and extrinsic tasks in NLP. We test the <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>-embeddings on different tasks with models of increasing complexity, showing that the advantages associated with the use of <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula>-embeddings are present also for models with a large number of parameters. Finally, we show that tuning <inline-formula><math xmlns="http://www.w3.org/1998/Math/MathML" display="inline"><semantics><mi>α</mi></semantics></math></inline-formula> allows for higher performances compared to the use of larger models in which additionally a transformation of the embeddings is learned during training, as experimentally verified in attention models.
ISSN:1099-4300