计算机科学
人工智能
相关性
限制
计算机视觉
模式识别(心理学)
数学
工程类
几何学
机械工程
作者
Niuniu Dong,Yihui Liang,Kun Zou,Wensheng Li,Fujian Feng
摘要
Video matting aims at accurately separating foreground from videos. Recent video matting researches pursue to eliminate auxiliary inputs. However, due to the limited ability of extracting global correlation features, these methods suffer from performance degradation when dealing with complex scenes or natural background videos. To address this challenge, we propose a video matting method called Video Matting Based on Local-Global Features Fusion (VMBLGFF) which can extract both comprehensive global correlation features and local subtle features. VMBLGFF contains two closely connected networks: a transformer network that utilizes window and global attention mechanisms to obtain global correlation features within and cross windows, and a fusion network that integrates local subtle features into the global correlation features to supplement the local detail information which may be overlooked by the attention mechanisms. VMBLGFF alleviates the issue of limiting global correlation features and has been benchmarked on both synthetic and real datasets, and the results demonstrate that VMBLGFF improves the quality of video matting and exhibits good generalization performance.
科研通智能强力驱动
Strongly Powered by AbleSci AI