Back to Search Start Over

From Variability to Stability: Advancing RecSys Benchmarking Practices

Authors :
Shevchenko, Valeriy
Belousov, Nikita
Vasilev, Alexey
Zholobov, Vladimir
Sosedka, Artyom
Semenova, Natalia
Volodkevich, Anna
Savchenko, Andrey
Zaytsev, Alexey
Source :
KDD 2024: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
Publication Year :
2024

Abstract

In the rapidly evolving domain of Recommender Systems (RecSys), new algorithms frequently claim state-of-the-art performance based on evaluations over a limited set of arbitrarily selected datasets. However, this approach may fail to holistically reflect their effectiveness due to the significant impact of dataset characteristics on algorithm performance. Addressing this deficiency, this paper introduces a novel benchmarking methodology to facilitate a fair and robust comparison of RecSys algorithms, thereby advancing evaluation practices. By utilizing a diverse set of $30$ open datasets, including two introduced in this work, and evaluating $11$ collaborative filtering algorithms across $9$ metrics, we critically examine the influence of dataset characteristics on algorithm performance. We further investigate the feasibility of aggregating outcomes from multiple datasets into a unified ranking. Through rigorous experimental analysis, we validate the reliability of our methodology under the variability of datasets, offering a benchmarking strategy that balances quality and computational demands. This methodology enables a fair yet effective means of evaluating RecSys algorithms, providing valuable guidance for future research endeavors.<br />Comment: 8 pages with 11 figures

Details

Database :
arXiv
Journal :
KDD 2024: Proceedings of the 30th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
Publication Type :
Report
Accession number :
edsarx.2402.09766
Document Type :
Working Paper
Full Text :
https://doi.org/10.1145/3637528.3671655