修补
变压器
计算机科学
利用
高保真
忠诚
标杆管理
人工智能
水准点(测量)
计算机工程
图像(数学)
工程类
电信
计算机安全
大地测量学
营销
电压
地理
电气工程
业务
作者
Wenbo Li,Zhe Lin,Kun Zhou,Lu Qi,Yi Wang,Jiaya Jia
标识
DOI:10.1109/cvpr52688.2022.01049
摘要
Recent studies have shown the importance of modeling long-range interactions in the inpainting problem. To achieve this goal, existing approaches exploit either standalone attention techniques or transformers, but usually under a low resolution in consideration of computational cost. In this paper, we present a novel transformer-based model for large hole inpainting, which unifies the merits of transformers and convolutions to efficiently process high-resolution images. We carefully design each component of our framework to guarantee the high fidelity and diversity of recovered images. Specifically, we customize an inpainting-oriented transformer block, where the attention module aggregates non-local information only from partial valid tokens, indicated by a dynamic mask. Extensive experiments demonstrate the state-of-the-art performance of the new model on multiple benchmark datasets. Code is released at https://github.com/fenglinglwb/MAT.
科研通智能强力驱动
Strongly Powered by AbleSci AI