Title
Ensemble Recommendations via Thompson Sampling: an Experimental Study within e-Commerce.
Abstract
This work presents an extension of Thompson Sampling bandit policy for orchestrating the collection of base recommendation algorithms for e-commerce. We focus on the problem of item-to-item recommendations, for which multiple behavioral and attribute-based predictors are provided to an ensemble learner. We show how to adapt Thompson Sampling to realistic situations when neither action availability nor reward stationarity is guaranteed. Furthermore, we investigate the effects of priming the sampler with pre-set parameters of reward probability distributions by utilizing the product catalog and/or event history, when such information is available. We report our experimental results based on the analysis of three real-world e-commerce datasets.
Year
DOI
Venue
2018
10.1145/3172944.3172967
IUI
Keywords
Field
DocType
E-commerce Recommender Systems, Streaming Recommendations, Bandit Ensembles, Session-based Recommendations, Thompson Sampling, Reinforcement Learning
Computer science,Thompson sampling,Priming (psychology),Human–computer interaction,Probability distribution,Artificial intelligence,E-commerce,Machine learning,Reinforcement learning
Conference
ISBN
Citations 
PageRank 
978-1-4503-4945-1
0
0.34
References 
Authors
37
4
Name
Order
Citations
PageRank
Björn Brodén161.89
Mikael Hammar216316.22
Bengt J. Nilsson321024.43
Dimitris Paraschakis492.91