Yao Zhou,Yusong Qin,Song Wang,Zengxi Huang,Dazhi Zhang
标识
DOI:10.1109/prai59366.2023.10332122
摘要
Person Re-Identification (ReID) is a crucial task in computer vision that aims to match persons captured from non-overlapping camera views. In this paper, to alleviate the impacts of background and occlusion, we propose to use instance segmentation and pose estimation methods to create masks for global feature extraction. Furthermore, we divide the pedestrian images into three regions according to pedestrian keypoints, trying to eliminate the alignment errors. This part-based matching strategy also helps to address the occlusion issue. Overall, we construct a deep learning network with three branches, including two global branches and a part-based branch. The two global branches extract global features using segmentation-based mask and the mask derived from pedestrian keypoint heatmaps, respectively. In the end, a weighted fusion strategy is used to combine the global scores and part-based scores for final classification. This network enables us to acquire robust global feature of pedestrians by excluding background and occlusion, and simultaneously address the alignment errors to some extent. Experimental results on these three widely used datasets demonstrate the effectiveness of our method: Specifically, it achieves 64.5% rank-1 accuracy and 54.3% mAP on Occluded-Duke, and 94.4% and 87.1% rank-1 accuracies on Market-1501 and DukeMTMC-reID, respectively.