计算机科学
编码器
卷积(计算机科学)
特征(语言学)
块(置换群论)
人工智能
融合
图像融合
源代码
编码(集合论)
模式识别(心理学)
比例(比率)
图像(数学)
图层(电子)
计算机视觉
人工神经网络
数学
语言学
哲学
物理
几何学
化学
集合(抽象数据类型)
有机化学
量子力学
程序设计语言
操作系统
作者
Zhang Xiong,Zhang Xiao-hui,Hongwei Han,Qingping Hu
标识
DOI:10.1016/j.infrared.2023.104962
摘要
We proposed an infrared and visible image fusion method based on the ResCC module and spatial criss-cross attention models. The proposed method adopts an auto-encoder structure consisting of an encoder network, fusion layers, and a decoder network. The encoder network has a convolution layer and three ResCC blocks with dense connections. Each ResCC block can extract multi-scale features from source images without down-sampling operations and retain as many feature details as possible for image fusion. The fusion layer adopts spatial criss-cross attention models, which can capture contextual information in both horizontal and vertical directions. Attention in these two directions can also reduce the calculation of the attention maps. The decoder network consists of four convolution layers designed to reconstruct images from the feature map. Experiments performed on the public datasets demonstrate that the proposed method obtains better fusion performance on objective and subjective evaluations compared to other advanced fusion methods. The code is available at https://github.com/xiongzhangzzz/ResCCFusion.
科研通智能强力驱动
Strongly Powered by AbleSci AI