Performance evaluation of latent factor models for rating prediction
Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of...
Main Author: | |
---|---|
Other Authors: | |
Language: | English en |
Published: |
2015
|
Subjects: | |
Online Access: | http://hdl.handle.net/1828/6011 |
id |
ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-6011 |
---|---|
record_format |
oai_dc |
spelling |
ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-60112015-04-27T17:12:26Z Performance evaluation of latent factor models for rating prediction Zheng, Lan Wu, Kui Thomo, Alex Recommender systems latent factor models evaluation Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of LFMs over baseline approaches, however, hovers at only low percentage numbers. Therefore, it is time for a better understanding of their real power beyond the overall root mean square error (RMSE), which as it happens, lies at a very compressed range, without providing too much chance for deeper insight. We introduce an experiment based handbook of LFMs and reveal data preprocessing and updating algorithms' power. We perform a detailed experimental study regarding the performance of classical staple LFMs on a classical dataset, Movielens 1M, that sheds light on a much more pronounced excellence of LFMs for particular categories of users and items, for RMSE and other measures. In particular, LFMs exhibit surprising and excellent advantages when handling several difficult user and item categories. By comparing the distributions of test ratings and predicted ratings, we show that the performance of LFMs is influenced by rating distribution. We then propose a method to estimate the performance of LFMs for a given rating dataset. Also, we provide a very simple, open-source library that implements staple LFMs achieving a similar performance as some very recent (2013) developments in LFMs, and at the same time being more transparent than some other libraries in wide use. Graduate 2015-04-24T20:49:57Z 2015-04-24T20:49:57Z 2015 2015-04-24 Thesis http://hdl.handle.net/1828/6011 Chen, Cheng, et al. "Comparing the staples in latent factor models for recommender systems." Proceedings of the 29th Annual ACM Symposium on Applied Computing. ACM, 2014. English en Available to the World Wide Web |
collection |
NDLTD |
language |
English en |
sources |
NDLTD |
topic |
Recommender systems latent factor models evaluation |
spellingShingle |
Recommender systems latent factor models evaluation Zheng, Lan Performance evaluation of latent factor models for rating prediction |
description |
Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of LFMs over baseline approaches, however, hovers at only low percentage numbers. Therefore, it is time for a better understanding of their real power beyond the overall root mean square error (RMSE), which as it happens, lies at a very compressed range, without providing too much chance for deeper insight.
We introduce an experiment based handbook of LFMs and reveal data preprocessing and updating algorithms' power. We perform a detailed experimental study regarding the performance of classical staple LFMs on a classical dataset, Movielens 1M, that sheds light on a much more pronounced excellence of LFMs for particular categories of users and items, for RMSE and other measures. In particular, LFMs exhibit surprising and excellent advantages when handling several difficult user and item categories. By comparing the distributions of test ratings and predicted ratings, we show that the performance of LFMs is influenced by rating distribution. We then propose a method to estimate the performance of LFMs for a given rating dataset. Also, we provide a very simple, open-source library that implements staple LFMs achieving a similar performance as some very recent (2013) developments in LFMs, and at the same time being more transparent than some other libraries in wide use. === Graduate |
author2 |
Wu, Kui |
author_facet |
Wu, Kui Zheng, Lan |
author |
Zheng, Lan |
author_sort |
Zheng, Lan |
title |
Performance evaluation of latent factor models for rating prediction |
title_short |
Performance evaluation of latent factor models for rating prediction |
title_full |
Performance evaluation of latent factor models for rating prediction |
title_fullStr |
Performance evaluation of latent factor models for rating prediction |
title_full_unstemmed |
Performance evaluation of latent factor models for rating prediction |
title_sort |
performance evaluation of latent factor models for rating prediction |
publishDate |
2015 |
url |
http://hdl.handle.net/1828/6011 |
work_keys_str_mv |
AT zhenglan performanceevaluationoflatentfactormodelsforratingprediction |
_version_ |
1716802170768064512 |