Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation.
Topic models and neural networks can discover meaningful low-dimensional latent representations of text corpora; as such, they have become a key technology of document representation. However, such models presume all documents are non-discriminatory, resulting in latent representation dependent upon...
Main Authors: | , , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Public Library of Science (PLoS)
2016-01-01
|
Series: | PLoS ONE |
Online Access: | http://europepmc.org/articles/PMC4718658?pdf=render |
id |
doaj-ac1ddb34836b4c74921a69b5e117dde8 |
---|---|
record_format |
Article |
spelling |
doaj-ac1ddb34836b4c74921a69b5e117dde82020-11-25T00:59:48ZengPublic Library of Science (PLoS)PLoS ONE1932-62032016-01-01111e014667210.1371/journal.pone.0146672Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation.Chao WeiSenlin LuoXincheng MaHao RenJi ZhangLimin PanTopic models and neural networks can discover meaningful low-dimensional latent representations of text corpora; as such, they have become a key technology of document representation. However, such models presume all documents are non-discriminatory, resulting in latent representation dependent upon all other documents and an inability to provide discriminative document representation. To address this problem, we propose a semi-supervised manifold-inspired autoencoder to extract meaningful latent representations of documents, taking the local perspective that the latent representation of nearby documents should be correlative. We first determine the discriminative neighbors set with Euclidean distance in observation spaces. Then, the autoencoder is trained by joint minimization of the Bernoulli cross-entropy error between input and output and the sum of the square error between neighbors of input and output. The results of two widely used corpora show that our method yields at least a 15% improvement in document clustering and a nearly 7% improvement in classification tasks compared to comparative methods. The evidence demonstrates that our method can readily capture more discriminative latent representation of new documents. Moreover, some meaningful combinations of words can be efficiently discovered by activating features that promote the comprehensibility of latent representation.http://europepmc.org/articles/PMC4718658?pdf=render |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Chao Wei Senlin Luo Xincheng Ma Hao Ren Ji Zhang Limin Pan |
spellingShingle |
Chao Wei Senlin Luo Xincheng Ma Hao Ren Ji Zhang Limin Pan Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. PLoS ONE |
author_facet |
Chao Wei Senlin Luo Xincheng Ma Hao Ren Ji Zhang Limin Pan |
author_sort |
Chao Wei |
title |
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. |
title_short |
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. |
title_full |
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. |
title_fullStr |
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. |
title_full_unstemmed |
Locally Embedding Autoencoders: A Semi-Supervised Manifold Learning Approach of Document Representation. |
title_sort |
locally embedding autoencoders: a semi-supervised manifold learning approach of document representation. |
publisher |
Public Library of Science (PLoS) |
series |
PLoS ONE |
issn |
1932-6203 |
publishDate |
2016-01-01 |
description |
Topic models and neural networks can discover meaningful low-dimensional latent representations of text corpora; as such, they have become a key technology of document representation. However, such models presume all documents are non-discriminatory, resulting in latent representation dependent upon all other documents and an inability to provide discriminative document representation. To address this problem, we propose a semi-supervised manifold-inspired autoencoder to extract meaningful latent representations of documents, taking the local perspective that the latent representation of nearby documents should be correlative. We first determine the discriminative neighbors set with Euclidean distance in observation spaces. Then, the autoencoder is trained by joint minimization of the Bernoulli cross-entropy error between input and output and the sum of the square error between neighbors of input and output. The results of two widely used corpora show that our method yields at least a 15% improvement in document clustering and a nearly 7% improvement in classification tasks compared to comparative methods. The evidence demonstrates that our method can readily capture more discriminative latent representation of new documents. Moreover, some meaningful combinations of words can be efficiently discovered by activating features that promote the comprehensibility of latent representation. |
url |
http://europepmc.org/articles/PMC4718658?pdf=render |
work_keys_str_mv |
AT chaowei locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation AT senlinluo locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation AT xinchengma locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation AT haoren locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation AT jizhang locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation AT liminpan locallyembeddingautoencodersasemisupervisedmanifoldlearningapproachofdocumentrepresentation |
_version_ |
1725216037959368704 |