计算机科学
转化(遗传学)
透视图(图形)
卷积神经网络
人工智能
发电机(电路理论)
对抗制
鉴定(生物学)
集合(抽象数据类型)
图像(数学)
国家(计算机科学)
噪音(视频)
机器学习
模式识别(心理学)
计算机视觉
数据挖掘
算法
量子力学
生物化学
基因
植物
生物
化学
功率(物理)
程序设计语言
物理
作者
Yanbing Chen,Wei Ke,Hong Lin,Chan–Tong Lam,Kai Lv,Hao Sheng,Zhang Xiong
标识
DOI:10.1016/j.jvcir.2021.103432
摘要
Vehicle re-identification (V-ReID) aims at discovering an image of a specific vehicle from a set of images typically captured by different cameras. Vehicles are one of the most important objects in cross-camera target recognition systems, and recognizing them is one of the most difficult tasks due to the subtle differences in the visible characteristics of vehicle rigid objects. Compared to various methods that can improve re-identification accuracy, data augmentation is a more straightforward and effective technique. In this paper, we propose a novel data synthesis method for V-ReID based on local-region perspective transformation, transformation state adversarial learning and a candidate pool. Specifically, we first propose a parameter generator network, which is a lightweight convolutional neural network, to generate the transformation states. Secondly, an adversarial module is designed in our work, it ensures that noise information is added as much as possible while keeping the labeling and structure of the dataset intact. With this adversarial module, we are able to promote the performance of the network and generate more proper and harder training samples. Furthermore, we use a candidate pool to store harder samples for further selection to improve the performance of the model. Our system pays more balanced attention to the features of vehicles. Extensive experiments show that our method significantly boosts the performance of V-ReID on the VeRi-776, VehicleID and VERI-Wild datasets. • A data augmentation method is designed that integrates a deep learning framework. • Local-region perspective transformation is optimized with an adversarial model. • More balanced attention is paid to the overall features of vehicles. • Data augmentation and the recognition model are jointly optimized. • Amplified samples are more efficient due to the automatic learning process. • A candidate pool stores the augmented images through the dynamic learning process. • Datasets are augmented by lifting the difficulty rather than the quantity. • The structure of the original dataset is reserved including the size and labels.
科研通智能强力驱动
Strongly Powered by AbleSci AI