Abstract | ||
---|---|---|
In this work, we target a known problem in representation learning that is: beyond coarse classification, how can we better model fine-grained categorization? To address this problem, we introduce Deep Subclass Linear Discriminant Analysis (DeepSDA), which utilizes intra-class variation and inter-class similarity during training. We could achieve multimodal classification by maximizing the ratio of between-subclass scatter matrix and within-subclass scatter matrix. We maximize the eigenvalues along the discriminative eignevector directions. Hence the deep neural network is able to learn more discriminative representation space and thus has higher class separation in the linearly separable latent space. We show that DeepSDA leads to significant improvements on diverse fine-grained categorization and attribute learning benchmarks. |
Year | DOI | Venue |
---|---|---|
2020 | 10.1109/ICIP40778.2020.9191197 | 2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP) |
Keywords | DocType | ISSN |
subclass linear discriminant analysis, deep learning, multimodal optimization, fine-grained categorization, attribute distribution | Conference | 1522-4880 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
4 |
Name | Order | Citations | PageRank |
---|---|---|---|
abin jose | 1 | 1 | 2.40 |
Shen Yan | 2 | 0 | 0.68 |
Mi Zhang | 3 | 206 | 21.68 |
J. -R. Ohm | 4 | 2812 | 166.79 |