Performance evaluation of latent factor models for rating prediction

Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of...

Full description

Bibliographic Details
Main Author: Zheng, Lan
Other Authors: Wu, Kui
Language:English
en
Published: 2015
Subjects:
Online Access:http://hdl.handle.net/1828/6011
id ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-6011
record_format oai_dc
spelling ndltd-uvic.ca-oai-dspace.library.uvic.ca-1828-60112015-04-27T17:12:26Z Performance evaluation of latent factor models for rating prediction Zheng, Lan Wu, Kui Thomo, Alex Recommender systems latent factor models evaluation Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of LFMs over baseline approaches, however, hovers at only low percentage numbers. Therefore, it is time for a better understanding of their real power beyond the overall root mean square error (RMSE), which as it happens, lies at a very compressed range, without providing too much chance for deeper insight. We introduce an experiment based handbook of LFMs and reveal data preprocessing and updating algorithms' power. We perform a detailed experimental study regarding the performance of classical staple LFMs on a classical dataset, Movielens 1M, that sheds light on a much more pronounced excellence of LFMs for particular categories of users and items, for RMSE and other measures. In particular, LFMs exhibit surprising and excellent advantages when handling several difficult user and item categories. By comparing the distributions of test ratings and predicted ratings, we show that the performance of LFMs is influenced by rating distribution. We then propose a method to estimate the performance of LFMs for a given rating dataset. Also, we provide a very simple, open-source library that implements staple LFMs achieving a similar performance as some very recent (2013) developments in LFMs, and at the same time being more transparent than some other libraries in wide use. Graduate 2015-04-24T20:49:57Z 2015-04-24T20:49:57Z 2015 2015-04-24 Thesis http://hdl.handle.net/1828/6011 Chen, Cheng, et al. "Comparing the staples in latent factor models for recommender systems." Proceedings of the 29th Annual ACM Symposium on Applied Computing. ACM, 2014. English en Available to the World Wide Web
collection NDLTD
language English
en
sources NDLTD
topic Recommender systems
latent factor models
evaluation
spellingShingle Recommender systems
latent factor models
evaluation
Zheng, Lan
Performance evaluation of latent factor models for rating prediction
description Since the Netflix Prize competition, latent factor models (LFMs) have become the comparison ``staples'' for many of the recent recommender methods. Meanwhile, it is still unclear to understand the impact of data preprocessing and updating algorithms on LFMs. The performance improvement of LFMs over baseline approaches, however, hovers at only low percentage numbers. Therefore, it is time for a better understanding of their real power beyond the overall root mean square error (RMSE), which as it happens, lies at a very compressed range, without providing too much chance for deeper insight. We introduce an experiment based handbook of LFMs and reveal data preprocessing and updating algorithms' power. We perform a detailed experimental study regarding the performance of classical staple LFMs on a classical dataset, Movielens 1M, that sheds light on a much more pronounced excellence of LFMs for particular categories of users and items, for RMSE and other measures. In particular, LFMs exhibit surprising and excellent advantages when handling several difficult user and item categories. By comparing the distributions of test ratings and predicted ratings, we show that the performance of LFMs is influenced by rating distribution. We then propose a method to estimate the performance of LFMs for a given rating dataset. Also, we provide a very simple, open-source library that implements staple LFMs achieving a similar performance as some very recent (2013) developments in LFMs, and at the same time being more transparent than some other libraries in wide use. === Graduate
author2 Wu, Kui
author_facet Wu, Kui
Zheng, Lan
author Zheng, Lan
author_sort Zheng, Lan
title Performance evaluation of latent factor models for rating prediction
title_short Performance evaluation of latent factor models for rating prediction
title_full Performance evaluation of latent factor models for rating prediction
title_fullStr Performance evaluation of latent factor models for rating prediction
title_full_unstemmed Performance evaluation of latent factor models for rating prediction
title_sort performance evaluation of latent factor models for rating prediction
publishDate 2015
url http://hdl.handle.net/1828/6011
work_keys_str_mv AT zhenglan performanceevaluationoflatentfactormodelsforratingprediction
_version_ 1716802170768064512