Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis
The latent semantic analysis (LSA) is a mathematical/statistical way of discovering hidden concepts between terms and documents or within a document collection (i.e., a large corpus of text). Each document of the corpus and terms are expressed as a vector with elements corresponding to these concept...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
Hindawi Limited
2019-01-01
|
Series: | Scientific Programming |
Online Access: | http://dx.doi.org/10.1155/2019/1095643 |
id |
doaj-c9638530f13643cda7427f29e423a3a2 |
---|---|
record_format |
Article |
spelling |
doaj-c9638530f13643cda7427f29e423a3a22021-07-02T02:36:20ZengHindawi LimitedScientific Programming1058-92441875-919X2019-01-01201910.1155/2019/10956431095643Alternate Low-Rank Matrix Approximation in Latent Semantic AnalysisFahrettin Horasan0Hasan Erbay1Fatih Varçın2Emre Deniz3Computer Engineering Department, Engineering Faculty, Kırıkkale University, Yahşihan, 71450 Kırıkkale, TurkeyComputer Engineering Department, Engineering Faculty, Kırıkkale University, Yahşihan, 71450 Kırıkkale, TurkeyComputer Engineering Department, Engineering Faculty, Kırıkkale University, Yahşihan, 71450 Kırıkkale, TurkeyComputer Engineering Department, Engineering Faculty, Kırıkkale University, Yahşihan, 71450 Kırıkkale, TurkeyThe latent semantic analysis (LSA) is a mathematical/statistical way of discovering hidden concepts between terms and documents or within a document collection (i.e., a large corpus of text). Each document of the corpus and terms are expressed as a vector with elements corresponding to these concepts to form a term-document matrix. Then, the LSA uses a low-rank approximation to the term-document matrix in order to remove irrelevant information, to extract more important relations, and to reduce the computational time. The irrelevant information is called as “noise” and does not have a noteworthy effect on the meaning of the document collection. This is an essential step in the LSA. The singular value decomposition (SVD) has been the main tool obtaining the low-rank approximation in the LSA. Since the document collection is dynamic (i.e., the term-document matrix is subject to repeated updates), we need to renew the approximation. This can be done via recomputing the SVD or updating the SVD. However, the computational time of recomputing or updating the SVD of the term-document matrix is very high when adding new terms and/or documents to preexisting document collection. Therefore, this issue opened the door of using other matrix decompositions for the LSA as ULV- and URV-based decompositions. This study shows that the truncated ULV decomposition (TULVD) is a good alternative to the SVD in the LSA modeling.http://dx.doi.org/10.1155/2019/1095643 |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Fahrettin Horasan Hasan Erbay Fatih Varçın Emre Deniz |
spellingShingle |
Fahrettin Horasan Hasan Erbay Fatih Varçın Emre Deniz Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis Scientific Programming |
author_facet |
Fahrettin Horasan Hasan Erbay Fatih Varçın Emre Deniz |
author_sort |
Fahrettin Horasan |
title |
Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis |
title_short |
Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis |
title_full |
Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis |
title_fullStr |
Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis |
title_full_unstemmed |
Alternate Low-Rank Matrix Approximation in Latent Semantic Analysis |
title_sort |
alternate low-rank matrix approximation in latent semantic analysis |
publisher |
Hindawi Limited |
series |
Scientific Programming |
issn |
1058-9244 1875-919X |
publishDate |
2019-01-01 |
description |
The latent semantic analysis (LSA) is a mathematical/statistical way of discovering hidden concepts between terms and documents or within a document collection (i.e., a large corpus of text). Each document of the corpus and terms are expressed as a vector with elements corresponding to these concepts to form a term-document matrix. Then, the LSA uses a low-rank approximation to the term-document matrix in order to remove irrelevant information, to extract more important relations, and to reduce the computational time. The irrelevant information is called as “noise” and does not have a noteworthy effect on the meaning of the document collection. This is an essential step in the LSA. The singular value decomposition (SVD) has been the main tool obtaining the low-rank approximation in the LSA. Since the document collection is dynamic (i.e., the term-document matrix is subject to repeated updates), we need to renew the approximation. This can be done via recomputing the SVD or updating the SVD. However, the computational time of recomputing or updating the SVD of the term-document matrix is very high when adding new terms and/or documents to preexisting document collection. Therefore, this issue opened the door of using other matrix decompositions for the LSA as ULV- and URV-based decompositions. This study shows that the truncated ULV decomposition (TULVD) is a good alternative to the SVD in the LSA modeling. |
url |
http://dx.doi.org/10.1155/2019/1095643 |
work_keys_str_mv |
AT fahrettinhorasan alternatelowrankmatrixapproximationinlatentsemanticanalysis AT hasanerbay alternatelowrankmatrixapproximationinlatentsemanticanalysis AT fatihvarcın alternatelowrankmatrixapproximationinlatentsemanticanalysis AT emredeniz alternatelowrankmatrixapproximationinlatentsemanticanalysis |
_version_ |
1721343070918672384 |