Title
Generalizing Deep Models for Ultrasound Image Segmentation.
Abstract
Deep models are subject to performance drop when encountering appearance discrepancy, even on congeneric corpus in which objects share the similar structure but only differ slightly in appearance. This performance drop can be observed in automated ultrasound image segmentation. In this paper, we try to address this general problem with a novel online adversarial appearance conversion solution. Our contribution is three-fold. First, different from previous methods which utilize corpus-level training to model a fixed source-target appearance conversion in advance, we only need to model the source corpus and then we can efficiently convert each single testing image in the target corpus on-the-fly. Second, we propose a self-play training strategy to effectively pre-train all the adversarial modules in our framework to capture the appearance and structure distributions of source corpus. Third, we propose to explore a composite appearance and structure constraints distilled from the source corpus to stabilize the online adversarial appearance conversion, thus the pre-trained models can iteratively remove appearance discrepancy in the testing image in a weakly-supervised fashion. We demonstrate our method on segmenting congeneric prenatal ultrasound images. Based on the appearance conversion, we can generalize deep models at-hand well and achieve significant improvement in segmentation without re-training on massive, expensive new annotations.
Year
DOI
Venue
2018
10.1007/978-3-030-00937-3_57
Lecture Notes in Computer Science
Field
DocType
Volume
Computer vision,Market segmentation,Pattern recognition,Computer science,Generalization,Segmentation,Ultrasound image segmentation,Artificial intelligence
Conference
11073
ISSN
Citations 
PageRank 
0302-9743
0
0.34
References 
Authors
5
8
Name
Order
Citations
PageRank
Xin Yang117512.96
Haoran Dou2274.12
ran li3284.97
Xu Wang410315.76
Cheng Bian5264.00
Shengli Li618418.06
Dong Ni736737.37
Pheng-Ann Heng83565280.98