Abstract | ||
---|---|---|
In deep learning (DL)-based multifocus image fusion, effective multiscale feature learning is a key issue to promote fusion performance. In this article, we propose a novel DL model named multiscale feature interactive network (MSFIN), which can segment the source images into focused and defocused regions accurately by sufficient interaction of multiscale features from layers of different depths in the network for multifocus image fusion. Specifically, based on the popular encoder-decoder framework, two functional modules, namely, multiscale feature fusion (MSFF) and coordinate attention upsample (CAU), are designed for interactive multiscale feature learning. Moreover, the weighted binary cross-entropy (WBCE) loss and the multilevel supervision (MLS) strategy are introduced to train the network more effectively. Qualitative and quantitative comparisons with 19 representative multifocus image fusion methods demonstrate that the proposed method can achieve state-of-the-art performance. The code of our method is available at https://github.com/yuliu316316/MSFIN-Fusion. |
Year | DOI | Venue |
---|---|---|
2021 | 10.1109/TIM.2021.3124058 | IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT |
Keywords | DocType | Volume |
Image fusion, Transforms, Feature extraction, Image segmentation, Deep learning, Training, Image reconstruction, Convolutional neural networks (CNNs), deep learning (DL), focus map, multifocus image fusion, multiscale features | Journal | 70 |
ISSN | Citations | PageRank |
0018-9456 | 0 | 0.34 |
References | Authors | |
0 | 4 |
Name | Order | Citations | PageRank |
---|---|---|---|
Yu Liu | 1 | 492 | 30.80 |
Lei Wang | 2 | 65 | 54.21 |
Juan Cheng | 3 | 62 | 11.53 |
Xun Chen | 4 | 458 | 52.73 |