对偶(语法数字)
计算机科学
动作识别
人工智能
动作(物理)
融合
模式识别(心理学)
语言学
哲学
物理
量子力学
班级(哲学)
作者
Biaozhang Huang,Xinde Li
标识
DOI:10.1145/3633624.3633635
摘要
In the field of semi-supervised human action recognition, the effective utilization of both labeled and unlabeled data remains a central and challenging pursuit. To address this issue, we present an innovative framework (DSCF-AEC) that combines a Dual-stream Cross Fusion network (DSCF) with an Ambiguous Exclude Contrastive Learning (AEC) module. Specifically, our Dual-stream Cross Fusion network utilizes the ST-GCN as encoder, independently encoding two augmented versions of the joint and bone streams, which are subsequently cross-fused to achieve enhanced representation. To further bolster the performance, we designed the AEC module. This module constructs a memory bank capable of distinguishing reliable positive and negative samples, while ambiguous samples are excluded. This strategic approach ensures that, through contrastive learning, the model is trained solely on meaningful and trustworthy samples. Extensive experiments on NTU RGB+D and NW-UCLA datasets validate the effectiveness of our approach. The results indicate that, our proposed method significantly outperforms other existing methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI