Title
Deep Gradual Multi-Exposure Fusion Via Recurrent Convolutional Network
Abstract
The performance of multi-exposure image fusion (MEF) has been recently improved with deep learning techniques but there are still a couple of problems to be overcome. In this paper, we propose a novel MEF network based on recurrent neural network (RNN). Multi-exposure images have different useful information depending on their exposure levels, and in order to fuse them complementarily, we first extract the local detail and global context features of input source images, and both features are separately combined. A weight map is learned from the local features for effectively fusing according to the importance of each source image. Adopting RNN as a backbone network enables gradual fusion, where more inputs result in further improvement of the fusion gradually. Also, information can be transferred to the deeper level of the network. Experimental results show that the proposed method achieves the reduction of fusion artifacts and improves detail restoration performance, compared to conventional methods.
Year
DOI
Venue
2021
10.1109/ACCESS.2021.3122540
IEEE ACCESS
Keywords
DocType
Volume
Feature extraction, Image fusion, Fuses, Image restoration, Image reconstruction, Brightness, Deep learning, Multi-exposure image fusion, recurrent convolutional network, dilated convolution filter, gradual fusion
Journal
9
ISSN
Citations 
PageRank 
2169-3536
0
0.34
References 
Authors
0
3
Name
Order
Citations
PageRank
Je-Ho Ryu100.68
Jong-Han Kim200.68
Jong-Ok Kim313.07