Title
Online Multiclass Learning with "Bandit" Feedback under a Confidence-Weighted Approach.
Abstract
Data volume has been increasing explosively in recent years and learning methods are vitally important to extract key information in such mass data. Traditional offline learning requires multiple traversals to the dataset, thus frequently suffering from lack of computational resources. Online learning can benefit in shrinking total time consumed by training model and lowering computational capacity. However they often converge slowly due to memory loss. Considering partial feedback, we uniquely propose online Confidence-Weighted learning in Bandit setting (CWB) for lower cumulative error and higher convergence rate. Specifically, historical information is preserved to adjust the weights of features for speeding up the convergence rate. Moreover, we novelly integrate the random sampling into the confidence-weighted learning, which can balance the exploitation and exploration in bandit setting. Extensive experiments demonstrate efficiency and effectiveness of our proposed scheme.
Year
Venue
Field
2016
IEEE Global Communications Conference
Online learning,Offline learning,Convergence (routing),Online machine learning,Semi-supervised learning,Computer science,Sampling (statistics),Rate of convergence,Generalization error,Artificial intelligence,Machine learning
DocType
ISSN
Citations 
Conference
2334-0983
0
PageRank 
References 
Authors
0.34
0
5
Name
Order
Citations
PageRank
Chaoran Shi100.34
Xiong Wang2705.15
Xiaohua Tian356865.92
Xiaoying Gan434448.16
Xinbing Wang551.11