Abstract | ||
---|---|---|
Focus on unsupervised domain adaptation by subspace learning.Highlight the importance of learning jointly the prediction model and the subspace.Propose the JCSL method.Thorough experimental evaluation against existing subspace adaptive methods.Insight into JCSL with analysis on parameters and domain shift reduction. Domain adaptation aims at adapting the knowledge acquired on a source domain to a new different but related target domain. Several approaches have been proposed for classification tasks in the unsupervised scenario, where no labeled target data are available. Most of the attention has been dedicated to searching a new domain-invariant representation, leaving the definition of the prediction function to a second stage. Here we propose to learn both jointly. Specifically we learn the source subspace that best matches the target subspace while at the same time minimizing a regularized misclassification loss. We provide an alternating optimization technique based on stochastic sub-gradient descent to solve the learning problem and we demonstrate its performance on several domain adaptation tasks. |
Year | DOI | Venue |
---|---|---|
2015 | 10.1016/j.patrec.2015.07.009 | Pattern Recognition Letters |
Keywords | Field | DocType |
Unsupervised domain adaptation,Subspace modeling,Max-margin classifiers | Pattern recognition,Subspace topology,Domain adaptation,Computer science,Random subspace method,Artificial intelligence,Machine learning | Journal |
Volume | Issue | ISSN |
65 | C | 0167-8655 |
Citations | PageRank | References |
11 | 0.50 | 40 |
Authors | ||
3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Basura Fernando | 1 | 775 | 35.60 |
Tatiana Tommasi | 2 | 14 | 1.55 |
Tinne Tuytelaars | 3 | 10161 | 609.66 |