计算机科学
联营
骨架(计算机编程)
接头(建筑物)
异步通信
人工智能
计算机视觉
动作(物理)
动作识别
模式识别(心理学)
计算机网络
班级(哲学)
工程类
物理
建筑工程
程序设计语言
量子力学
作者
Shanaka Ramesh Gunasekara,Wanqing Li,Jie Yang,Philip Ogunbona
标识
DOI:10.1109/tcsvt.2024.3465845
摘要
Deep neural networks for skeleton-based human action recognition (HAR) often utilize traditional averaging or maximum temporal pooling to aggregate features by treating all joints and frames equally. However, this approach can excessively aggregate less discriminative or even indiscriminative features into the final feature vectors for recognition. To address this issue, a novel method called asynchronous joint adaptive temporal pooling (AJTP) is introduced in this paper. The method aims to enhance action recognition by identifying a set of informative joints across the temporal dimension and applying a joint-based and asynchronous motion-preservative pooling rather than conventional frame-based pooling. The effectiveness of the proposed AJTP has been empirically validated by integrating it with popular Graph Convolutional Network (GCN) models on three benchmark datasets: NTU RGB+D 120, PKUMMD, and Kinetic400. The results have shown that a GCN model with AJTP substantially improves performance compared to its counterpart GCN model with conventional temporal pooling techniques. The source code is available at https://github.com/ShanakaRG/AJTP.
科研通智能强力驱动
Strongly Powered by AbleSci AI