1. PFNet: Part-guided feature-combination network for vehicle re-identification.
- Author
-
Qian, Jiahe and Zhao, Jiandong
- Subjects
COMPARATIVE method ,DEEP learning ,FEATURE extraction ,LEARNING ability ,BIG data - Abstract
With the development of highways, there has been an increase in toll evasion by unscrupulous individuals who employ various means to avoid paying fees. This has made it essential for highway management to focus on effective methods to prevent and manage toll evasion, as well as regulate toll-related operations. However, traditional auditing techniques are heavily reliant on manual toll data and lack sufficient auxiliary data such as video and pictures, making big data analysis and in-depth data analysis difficult. Consequently, these methods can only identify a portion of evasion behavior, with a significant number of potential evasion cases remaining unidentified. To address this issue, we propose utilizing vehicle re-identification to enable cross-camera identification and continuous tracking of vehicles, allowing for accurate calculation of their operating mileage. However, vehicle re-identification using images captured by highway cameras is a challenging task, primarily due to the low resolution of the images and the lack of relevant datasets. To tackle these problems, we construct a large-scale, low-resolution dataset based on images captured by cameras on highways. We also propose a Part-guided Feature-combination Network (PFNet) to analyze the dataset. PFNet adopts a three-stage approach, wherein each stage of the network requires independent training. In the first stage, the Pyramid Scene Parsing Network (PSPNet) is improved to segment the entire vehicle into parts. In the second stage, based on a comparative learning approach, the Siamese Feature Extraction Module (SFEM) module was created to extract features of pairs of vehicle parts. To train the SFEM, pairs of positive samples, which consist of the same vehicle component pairs captured by different cameras, and pairs of negative samples, which consist of different vehicle component pairs are needed. To ensure that the positive sample pairs are closer to each other in the embedding space, and the negative samples are as far away as possible from each other, contrast loss is utilized. Moreover, The SFEM incorporates VGG16 and a Multi-period Convolutional Block Attention Module (MCBAM), which enhances the learning ability of the conventional Convolutional Block Attention Module (CBAM). In the third stage, information propagation among the vehicle parts is achieved using a Multi-layer Perceptron (MLP), which calculates the similarity after combining the features of all parts. Our self-built test set demonstrates that PFNet is effective for continuous monitoring of vehicles on highways under multiple cameras, resulting in improved accuracy of vehicle re-identification, with Top1 accuracy of 93.61% and Top5 accuracy of 99.44%. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF