TY - GEN
T1 - Rethinking the recommender research ecosystem
T2 - 5th ACM Conference on Recommender Systems, RecSys 2011
AU - Ekstrand, Michael D.
AU - Ludwig, Michael
AU - Konstan, Joseph A
AU - Riedl, John T.
PY - 2011
Y1 - 2011
N2 - Recommender systems research is being slowed by the difficulty of replicating and comparing research results. Published research uses various experimental methodologies and metrics that are difficult to compare. It also often fails to sufficiently document the details of proposed algorithms or the evaluations employed. Researchers waste time reimplementing well-known algorithms, and the new implementations may miss key details from the original algorithm or its subsequent refinements. When proposing new algorithms, researchers should compare them against finely-tuned implementations of the leading prior algorithms using state-of-the-art evaluation methodologies. With few exceptions, published algorithmic improvements in our field should be accompanied by working code in a standard framework, including test harnesses to reproduce the described results. To that end, we present the design and freely distributable source code of LensKit, a flexible platform for reproducible recommender systems research. LensKit provides carefully tuned implementations of the leading collaborative filtering algorithms, APIs for common recommender system use cases, and an evaluation framework for performing reproducible offline evaluations of algorithms. We demonstrate the utility of LensKit by replicating and extending a set of prior comparative studies of recommender algorithms - showing limitations in some of the original results - and by investigating a question recently raised by a leader in the recommender systems community on problems with error-based prediction evaluation.
AB - Recommender systems research is being slowed by the difficulty of replicating and comparing research results. Published research uses various experimental methodologies and metrics that are difficult to compare. It also often fails to sufficiently document the details of proposed algorithms or the evaluations employed. Researchers waste time reimplementing well-known algorithms, and the new implementations may miss key details from the original algorithm or its subsequent refinements. When proposing new algorithms, researchers should compare them against finely-tuned implementations of the leading prior algorithms using state-of-the-art evaluation methodologies. With few exceptions, published algorithmic improvements in our field should be accompanied by working code in a standard framework, including test harnesses to reproduce the described results. To that end, we present the design and freely distributable source code of LensKit, a flexible platform for reproducible recommender systems research. LensKit provides carefully tuned implementations of the leading collaborative filtering algorithms, APIs for common recommender system use cases, and an evaluation framework for performing reproducible offline evaluations of algorithms. We demonstrate the utility of LensKit by replicating and extending a set of prior comparative studies of recommender algorithms - showing limitations in some of the original results - and by investigating a question recently raised by a leader in the recommender systems community on problems with error-based prediction evaluation.
KW - evaluation
KW - implementation
KW - recommender systems
UR - http://www.scopus.com/inward/record.url?scp=82555195664&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=82555195664&partnerID=8YFLogxK
U2 - 10.1145/2043932.2043958
DO - 10.1145/2043932.2043958
M3 - Conference contribution
AN - SCOPUS:82555195664
SN - 9781450306836
T3 - RecSys'11 - Proceedings of the 5th ACM Conference on Recommender Systems
SP - 133
EP - 140
BT - RecSys'11 - Proceedings of the 5th ACM Conference on Recommender Systems
Y2 - 23 October 2011 through 27 October 2011
ER -