计算机科学
人工智能
基本事实
图像(数学)
计算机视觉
发电机(电路理论)
航空影像
图像分辨率
特征检测(计算机视觉)
特征(语言学)
图像融合
图像形成
任务(项目管理)
约束(计算机辅助设计)
模式识别(心理学)
图像处理
数学
经济
功率(物理)
管理
哲学
物理
量子力学
语言学
几何学
作者
Yingying Zhu,Shihai Chen,Xiufan Lu,Jianyong Chen
标识
DOI:10.1109/tgrs.2023.3257861
摘要
Cross-view image synthesis aims to synthesize a ground-view image covering the same geographic region for a given single aerial-view image (or vice versa). Existing approaches typically tackle this challenging task by relaxing the single-image constraint and using a ground-truth semantic map as additional input to aid synthesis. However, this is nearly infeasible in practice. In this paper, we investigate how to generate a detail-enriched and structurally accurate ground-level image from only a single aerial-level input image, in which there are no other prior knowledge except for the input image. Towards this goal, we propose a novel Progressive Parallel Generative Adversarial Network (PPGAN) that starts from generating low-resolution outputs and progressively produces ground images at higher resolutions as the network propagates forward. In this manner, our PPGAN decomposes the task into several manageable sub-tasks, which helps to generate detail-enriched and structurally accurate ground images. During progressive generation, the PPGAN employs a parallel generation paradigm that enables the generator to produce multi-resolution images in parallel, thereby avoiding excessive time cost on training. Furthermore, for effective information propagation across multi-resolution images, a feature fusion module (FFM) is devised to mitigate the domain gap between cross-level image features, which enables a balance of detail and structural information synthesis. Additionally, the proposed Channel-Space Attention Selection Module (CSASM) learns the mapping relationship between cross-view images in a larger scale space to enhance the quality of the output image. Quantitative and qualitative experiments demonstrate that, our method requires only one input image without the aid of additional inputs, but is capable of synthesizing detail-enriched and structurally accurate ground images and outperforms the existing state-of-the-art methods on two famous benchmarks.
科研通智能强力驱动
Strongly Powered by AbleSci AI