Title
Feature Estimations Based Correlation Distillation for Incremental Image Retrieval
Abstract
Deep learning for fine-grained image retrieval in an incremental context is less investigated. In this paper, we explore this task to realize the model's continuous retrieval ability. That means, the model enables to perform well on new incoming data and reduce forgetting of the knowledge learned on preceding old tasks. For this purpose, we distill semantic correlations knowledge among the representations extracted from the new data only so as to regularize the parameters updates using the teacher-student framework. In particular, for the case of learning multiple tasks sequentially, aside from the correlations distilled from the penultimate model, we estimate the representations for all prior models and further their semantic correlations by using the representations extracted from the new data. To this end, the estimated correlations are used as an additional regularization and further prevent catastrophic forgetting over all previous tasks, and it is unnecessary to save the stream of models trained on these tasks. Extensive experiments demonstrate that the proposed method performs favorably for retaining performance on the already-trained old tasks and achieving good accuracy on the current task when new data are added at once or sequentially.
Year
DOI
Venue
2022
10.1109/TMM.2021.3073279
IEEE TRANSACTIONS ON MULTIMEDIA
Keywords
DocType
Volume
Task analysis, Correlation, Data models, Modeling, Training, Context modeling, Image retrieval, Incremental learning, fine-grained image retrieval, correlations distillation, feature estimation
Journal
24
ISSN
Citations 
PageRank 
1520-9210
0
0.34
References 
Authors
27
6
Name
Order
Citations
PageRank
Wei Chen11711246.70
Yu Liu219825.45
Nan Pu333.41
Wang Weiping433563.84
Li Liu573350.04
Michael S. Lew62742166.02