计算机科学
人工智能
模式识别(心理学)
特征提取
生物识别
步态
卷积(计算机科学)
计算机视觉
块(置换群论)
背景(考古学)
特征(语言学)
显著性(神经科学)
数学
地理
生理学
语言学
哲学
几何学
人工神经网络
生物
考古
作者
Tianhuan Huang,Xianye Ben,Chen Gong,Baochang Zhang,Rui Yan,Qiang Wu
标识
DOI:10.1109/tcsvt.2022.3175959
摘要
Gait recognition can be used in person identification and re-identification by itself or in conjunction with other biometrics. Although gait has both spatial and temporal attributes, and it has been observed that decoupling spatial feature and temporal feature can better exploit the gait feature on the fine-grained level. However, the spatial-temporal correlations of gait video signals are also lost in the decoupling process. Direct 3D convolution approaches can retain such correlations, but they also introduce unnecessary interferences. Instead of common 3D convolution solutions, this paper proposes an integration of decoupling process into a 3D convolution framework for cross-view gait recognition. In particular, a novel block consisting of a Parallel-insight Convolution layer integrated with a Spatial-Temporal Dual-Attention (STDA) unit is proposed as the basic block for global spatial-temporal information extraction. Under the guidance of the STDA unit, this block can well integrate spatial-temporal information extracted by two decoupled models and at the same time retain the spatial-temporal correlations. In addition, a Multi-Scale Salient Feature Extractor is proposed to further exploit the fine-grained features through context awareness extension of part-based features and adaptively aggregating the spatial features. Extensive experiments on three popular gait datasets, namely CASIA-B, OULP and OUMVLP, demonstrate that the proposed method outperforms state-of-the-art methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI