TY - GEN
T1 - Tuning Word2vec for Large Scale Recommendation Systems
AU - Chamberlain, Benjamin P.
AU - Rossi, Emanuele
AU - Shiebler, Dan
AU - Sedhain, Suvash
AU - Bronstein, Michael M.
N1 - Publisher Copyright:
© 2020 Owner/Author.
PY - 2020/9/22
Y1 - 2020/9/22
N2 - Word2vec is a powerful machine learning tool that emerged from Natural Language Processing (NLP) and is now applied in multiple domains, including recommender systems, forecasting, and network analysis. As Word2vec is often used off the shelf, we address the question of whether the default hyperparameters are suitable for recommender systems. The answer is emphatically no. In this paper, we first elucidate the importance of hyperparameter optimization and show that unconstrained optimization yields an average 221% improvement in hit rate over the default parameters. However, unconstrained optimization leads to hyperparameter settings that are very expensive and not feasible for large scale recommendation tasks. To this end, we demonstrate 138% average improvement in hit rate with a runtime budget-constrained hyperparameter optimization. Furthermore, to make hyperparameter optimization applicable for large scale recommendation problems where the target dataset is too large to search over, we investigate generalizing hyperparameters settings from samples. We show that applying constrained hyperparameter optimization using only a 10% sample of the data still yields a 91% average improvement in hit rate over the default parameters when applied to the full datasets. Finally, we apply hyperparameters learned using our method of constrained optimization on a sample to the Who To Follow recommendation service at Twitter and are able to increase follow rates by 15%.
AB - Word2vec is a powerful machine learning tool that emerged from Natural Language Processing (NLP) and is now applied in multiple domains, including recommender systems, forecasting, and network analysis. As Word2vec is often used off the shelf, we address the question of whether the default hyperparameters are suitable for recommender systems. The answer is emphatically no. In this paper, we first elucidate the importance of hyperparameter optimization and show that unconstrained optimization yields an average 221% improvement in hit rate over the default parameters. However, unconstrained optimization leads to hyperparameter settings that are very expensive and not feasible for large scale recommendation tasks. To this end, we demonstrate 138% average improvement in hit rate with a runtime budget-constrained hyperparameter optimization. Furthermore, to make hyperparameter optimization applicable for large scale recommendation problems where the target dataset is too large to search over, we investigate generalizing hyperparameters settings from samples. We show that applying constrained hyperparameter optimization using only a 10% sample of the data still yields a 91% average improvement in hit rate over the default parameters when applied to the full datasets. Finally, we apply hyperparameters learned using our method of constrained optimization on a sample to the Who To Follow recommendation service at Twitter and are able to increase follow rates by 15%.
KW - Embeddings
KW - Hyperparameter Optimization
KW - Neural Networks
KW - Recommender System Evaluation
UR - http://www.scopus.com/inward/record.url?scp=85092697122&partnerID=8YFLogxK
U2 - 10.1145/3383313.3418486
DO - 10.1145/3383313.3418486
M3 - ???researchoutput.researchoutputtypes.contributiontobookanthology.conference???
AN - SCOPUS:85092697122
T3 - RecSys 2020 - 14th ACM Conference on Recommender Systems
SP - 732
EP - 737
BT - RecSys 2020 - 14th ACM Conference on Recommender Systems
PB - Association for Computing Machinery, Inc
Y2 - 22 September 2020 through 26 September 2020
ER -