Title | ||
---|---|---|
Online Multiclass Learning with "Bandit" Feedback under a Confidence-Weighted Approach. |
Abstract | ||
---|---|---|
Data volume has been increasing explosively in recent years and learning methods are vitally important to extract key information in such mass data. Traditional offline learning requires multiple traversals to the dataset, thus frequently suffering from lack of computational resources. Online learning can benefit in shrinking total time consumed by training model and lowering computational capacity. However they often converge slowly due to memory loss. Considering partial feedback, we uniquely propose online Confidence-Weighted learning in Bandit setting (CWB) for lower cumulative error and higher convergence rate. Specifically, historical information is preserved to adjust the weights of features for speeding up the convergence rate. Moreover, we novelly integrate the random sampling into the confidence-weighted learning, which can balance the exploitation and exploration in bandit setting. Extensive experiments demonstrate efficiency and effectiveness of our proposed scheme. |
Year | Venue | Field |
---|---|---|
2016 | IEEE Global Communications Conference | Online learning,Offline learning,Convergence (routing),Online machine learning,Semi-supervised learning,Computer science,Sampling (statistics),Rate of convergence,Generalization error,Artificial intelligence,Machine learning |
DocType | ISSN | Citations |
Conference | 2334-0983 | 0 |
PageRank | References | Authors |
0.34 | 0 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Chaoran Shi | 1 | 0 | 0.34 |
Xiong Wang | 2 | 70 | 5.15 |
Xiaohua Tian | 3 | 568 | 65.92 |
Xiaoying Gan | 4 | 344 | 48.16 |
Xinbing Wang | 5 | 5 | 1.11 |