计算机科学
人工智能
特征(语言学)
转化(遗传学)
模式识别(心理学)
棱锥(几何)
计算机视觉
特征提取
红外线的
图像(数学)
光学
生物化学
化学
物理
基因
哲学
语言学
作者
Sirui Wang,Guiling Sun,Dong Liang,Bowen Zheng
标识
DOI:10.1016/j.infrared.2024.105314
摘要
With the rapid development of computer vision, there is a recent trend for intelligent image understanding based on deep learning. RGB images and infrared images have complementary information in image capture tasks in complex environments due to their different imaging modalities. Therefore, the combination of the two plays an important role in improving video surveillance and target detection capabilities. However, large publicly available infrared image datasets are lacking and acquiring infrared images can be resource-intensive. The lack of samples can lead to a breakdown in the training of deep models. In this paper, to overcome this challenge, we construct a GAN-based visible-infrared image transformation model. The model uses existing visible data to generate infrared images by training an end-to-end generative network. We innovatively propose a lightweight PAS feature extraction module applied to the generator. It enriches the image detail representation of the feature domain from multiple dimensions and greatly improves the model feature representation capability. And the image gradient calculation is used to limit the direction of model optimization. We evaluated our model on three different publicly available datasets, evaluating the quality of the generated images in terms of both visual effects and objective numerical assessments. The experimental results show that the network exhibits excellent results in both qualitative and quantitative evaluation compared to the current state-of-the-art image generation methods.The lightweight feature extraction module also gives PAS-GAN a significant advantage in inference speed.
科研通智能强力驱动
Strongly Powered by AbleSci AI