计算机科学
人工智能
光学(聚焦)
特征提取
卷积神经网络
特征(语言学)
Boosting(机器学习)
编码(内存)
模式识别(心理学)
分割
图像分割
语言学
光学
物理
哲学
作者
Bin Xiao,Bocheng Xu,Xiuli Bi,Weisheng Li
标识
DOI:10.1109/tip.2020.3033158
摘要
The convolutional neural network (CNN)-based multi-focus image fusion methods which learn the focus map from the source images have greatly enhanced fusion performance compared with the traditional methods. However, these methods have not yet reached a satisfactory fusion result, since the convolution operation pays too much attention on the local region and generating the focus map as a local classification (classify each pixel into focus or de-focus classes) problem. In this article, a global-feature encoding U-Net (GEU-Net) is proposed for multi-focus image fusion. In the proposed GEU-Net, the U-Net network is employed for treating the generation of focus map as a global two-class segmentation task, which segments the focused and defocused regions from a global view. For improving the global feature encoding capabilities of U-Net, the global feature pyramid extraction module (GFPE) and global attention connection upsample module (GACU) are introduced to effectively extract and utilize the global semantic and edge information. The perceptual loss is added to the loss function, and a large-scale dataset is constructed for boosting the performance of GEU-Net. Experimental results show that the proposed GEU-Net can achieve superior fusion performance than some state-of-the-art methods in both human visual quality, objective assessment and network complexity.
科研通智能强力驱动
Strongly Powered by AbleSci AI