Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization

The limitations of neighborhood-based Collaborative Filtering (CF) techniques over scalable and sparse data present obstacle for efficient recommendation systems. These techniques show poor accuracy and dismal speed in generating recommendations. Model-based matrix factorization is an alternative ap...

Full description

Bibliographic Details
Main Authors: Sandeep Kumar Raghuwanshi, Rajesh Kumar Pateriya
Format: Article
Language:English
Published: Elsevier 2021-05-01
Series:Journal of King Saud University: Computer and Information Sciences
Subjects:
Online Access:http://www.sciencedirect.com/science/article/pii/S1319157818300636
id doaj-ce546c1973fc4c17aafecdf66012bd13
record_format Article
spelling doaj-ce546c1973fc4c17aafecdf66012bd132021-06-05T06:03:44ZengElsevierJournal of King Saud University: Computer and Information Sciences1319-15782021-05-01334447452Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent OptimizationSandeep Kumar Raghuwanshi0Rajesh Kumar Pateriya1Corresponding author.; Computer Science & Engineering, Maulana Azad National Institute of Technology Bhopal, M.P., IndiaComputer Science & Engineering, Maulana Azad National Institute of Technology Bhopal, M.P., IndiaThe limitations of neighborhood-based Collaborative Filtering (CF) techniques over scalable and sparse data present obstacle for efficient recommendation systems. These techniques show poor accuracy and dismal speed in generating recommendations. Model-based matrix factorization is an alternative approach use to overcome aforementioned limitations of CF.Singular value decomposition (SVD) is widely used technique to get low-rank factors of rating matrix and use Gradient Descent (GD) or Alternative Least Square (ALS) for optimization of its error objective function. Most researchers have focused on the accuracy of predictions but they did not accumulate the convergence rate of learning approach. In this paper, we propose a new filtering technique that implements SVD using Stochastic Gradient Descent (SGD) optimization and provides an accelerated version of SVD for fast convergence of learning parameters with improved classification accuracy. Our proposed method accelerates SVD in the right direction and dampens oscillation by adding a momentum value in parameters updates. To support our claim, we have tested our proposed model against the famed real world datasets (MovieLens100k, FilmTrust and YahooMovie). The proposed Accelerated Singular Value Decomposition (ASVD) outperformed the existing models and achieved higher convergence rate and better classification accuracy.http://www.sciencedirect.com/science/article/pii/S1319157818300636Gradient DescentInformation filteringMatrix factorizationSingular value decompositionStochastic gradient descent
collection DOAJ
language English
format Article
sources DOAJ
author Sandeep Kumar Raghuwanshi
Rajesh Kumar Pateriya
spellingShingle Sandeep Kumar Raghuwanshi
Rajesh Kumar Pateriya
Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
Journal of King Saud University: Computer and Information Sciences
Gradient Descent
Information filtering
Matrix factorization
Singular value decomposition
Stochastic gradient descent
author_facet Sandeep Kumar Raghuwanshi
Rajesh Kumar Pateriya
author_sort Sandeep Kumar Raghuwanshi
title Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
title_short Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
title_full Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
title_fullStr Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
title_full_unstemmed Accelerated Singular Value Decomposition (ASVD) using momentum based Gradient Descent Optimization
title_sort accelerated singular value decomposition (asvd) using momentum based gradient descent optimization
publisher Elsevier
series Journal of King Saud University: Computer and Information Sciences
issn 1319-1578
publishDate 2021-05-01
description The limitations of neighborhood-based Collaborative Filtering (CF) techniques over scalable and sparse data present obstacle for efficient recommendation systems. These techniques show poor accuracy and dismal speed in generating recommendations. Model-based matrix factorization is an alternative approach use to overcome aforementioned limitations of CF.Singular value decomposition (SVD) is widely used technique to get low-rank factors of rating matrix and use Gradient Descent (GD) or Alternative Least Square (ALS) for optimization of its error objective function. Most researchers have focused on the accuracy of predictions but they did not accumulate the convergence rate of learning approach. In this paper, we propose a new filtering technique that implements SVD using Stochastic Gradient Descent (SGD) optimization and provides an accelerated version of SVD for fast convergence of learning parameters with improved classification accuracy. Our proposed method accelerates SVD in the right direction and dampens oscillation by adding a momentum value in parameters updates. To support our claim, we have tested our proposed model against the famed real world datasets (MovieLens100k, FilmTrust and YahooMovie). The proposed Accelerated Singular Value Decomposition (ASVD) outperformed the existing models and achieved higher convergence rate and better classification accuracy.
topic Gradient Descent
Information filtering
Matrix factorization
Singular value decomposition
Stochastic gradient descent
url http://www.sciencedirect.com/science/article/pii/S1319157818300636
work_keys_str_mv AT sandeepkumarraghuwanshi acceleratedsingularvaluedecompositionasvdusingmomentumbasedgradientdescentoptimization
AT rajeshkumarpateriya acceleratedsingularvaluedecompositionasvdusingmomentumbasedgradientdescentoptimization
_version_ 1721397203884310528