鉴别器
计算机科学
人工智能
分割
编码器
计算机视觉
模式识别(心理学)
特征(语言学)
红外线的
生成语法
发电机(电路理论)
图像(数学)
生成对抗网络
图像融合
光学
操作系统
物理
哲学
探测器
电信
功率(物理)
量子力学
语言学
作者
Jilei Hou,Dazhi Zhang,Wei Wu,Jiayi Ma,Huabing Zhou
出处
期刊:Entropy
[Multidisciplinary Digital Publishing Institute]
日期:2021-03-21
卷期号:23 (3): 376-376
被引量:53
摘要
This paper proposes a new generative adversarial network for infrared and visible image fusion based on semantic segmentation (SSGAN), which can consider not only the low-level features of infrared and visible images, but also the high-level semantic information. Source images can be divided into foregrounds and backgrounds by semantic masks. The generator with a dual-encoder-single-decoder framework is used to extract the feature of foregrounds and backgrounds by different encoder paths. Moreover, the discriminator’s input image is designed based on semantic segmentation, which is obtained by combining the foregrounds of the infrared images with the backgrounds of the visible images. Consequently, the prominence of thermal targets in the infrared images and texture details in the visible images can be preserved in the fused images simultaneously. Qualitative and quantitative experiments on publicly available datasets demonstrate that the proposed approach can significantly outperform the state-of-the-art methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI