TY - JOUR
T1 - MFIF-GAN
T2 - A new generative adversarial network for multi-focus image fusion
AU - Wang, Yicheng
AU - Xu, Shuang
AU - Liu, Junmin
AU - Zhao, Zixiang
AU - Zhang, Chunxia
AU - Zhang, Jiangshe
N1 - Publisher Copyright:
© 2021 Elsevier B.V.
PY - 2021/8
Y1 - 2021/8
N2 - Multi-Focus Image Fusion (MFIF) is a promising image enhancement technique to generate all-in-focus images meeting visual needs, and it is a precondition for other computer vision tasks. One emergent research trend in MFIF involves approaches to avoiding a defocus spread effect (DSE) around a focus/defocus boundary (FDB). This study proposes a generative adversarial network for MFIF tasks called MFIF-GAN, to attenuate the DSE by generating focus maps in which the foreground region is correctly larger than corresponding objects. A Squeeze and Excitation residual module is employed in the proposed network. By combining prior knowledge of a training condition, the network is trained on a synthetic dataset based on an α-matte model. In addition, reconstruction and gradient regularization terms are combined in the loss functions to enhance boundary details and improve the quality of fused images. Extensive experiments demonstrate that the MFIF-GAN outperforms eight state-of-the-art (SOTA) methods in visual perception and quantitative analysis, as well as efficiency. Moreover, an edge diffusion and contraction module is proposed to verify that focus maps generated by our method are accurate at the pixel level.
AB - Multi-Focus Image Fusion (MFIF) is a promising image enhancement technique to generate all-in-focus images meeting visual needs, and it is a precondition for other computer vision tasks. One emergent research trend in MFIF involves approaches to avoiding a defocus spread effect (DSE) around a focus/defocus boundary (FDB). This study proposes a generative adversarial network for MFIF tasks called MFIF-GAN, to attenuate the DSE by generating focus maps in which the foreground region is correctly larger than corresponding objects. A Squeeze and Excitation residual module is employed in the proposed network. By combining prior knowledge of a training condition, the network is trained on a synthetic dataset based on an α-matte model. In addition, reconstruction and gradient regularization terms are combined in the loss functions to enhance boundary details and improve the quality of fused images. Extensive experiments demonstrate that the MFIF-GAN outperforms eight state-of-the-art (SOTA) methods in visual perception and quantitative analysis, as well as efficiency. Moreover, an edge diffusion and contraction module is proposed to verify that focus maps generated by our method are accurate at the pixel level.
KW - Deep learning
KW - Defocus spread effect
KW - Generative adversarial network
KW - Multi-focus image fusion
UR - http://www.scopus.com/inward/record.url?scp=85104667913&partnerID=8YFLogxK
U2 - 10.1016/j.image.2021.116295
DO - 10.1016/j.image.2021.116295
M3 - 文章
AN - SCOPUS:85104667913
SN - 0923-5965
VL - 96
JO - Signal Processing: Image Communication
JF - Signal Processing: Image Communication
M1 - 116295
ER -