Next Article in Journal
A Dynamic Bridge Architecture for Efficient Interoperability Between AUTOSAR Adaptive and ROS2
Previous Article in Journal
UltraScanNet: A Mamba-Inspired Hybrid Backbone for Breast Ultrasound Classification
Previous Article in Special Issue
LMEC-YOLOv8: An Enhanced Object Detection Algorithm for UAV Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Analysis of Partitioned Convolutional Model for Vehicle Re-Identification

by
Rajsekhar Kumar Nath
*,† and
Debjani Mitra
Indian Institute of Technology (ISM), Dhanbad 826004, India
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2025, 14(18), 3634; https://doi.org/10.3390/electronics14183634
Submission received: 15 July 2025 / Revised: 19 August 2025 / Accepted: 3 September 2025 / Published: 14 September 2025
(This article belongs to the Special Issue Deep Learning for Computer Vision, 2nd Edition)

Abstract

Local Feature generation for vehicle re-identification is a challenging research area that is not yet well-investigated. The part-based convolutional baseline model with refined part pooling (PCB-RPP) architecture commonly approached in person reidentification problems was experimented over two standard vehicle image datasets (VReId and VehicleId) to establish that RPP over uniform partitions do not work well. To address the limitation, we propose a novel approach, Overlapped-PCB, which overlaps parts of two adjacent parts to generate new parts to train the classifiers. The results are concatenated to generate the feature set and this improves the re-identification accuracy in comparison to the RPP approach. Performance comparison results of extensive testing are also presented using re-ranking and ensembling in the evaluation stage. Our proposed model has been ensembled over three architectures, ResNet50, ResNet101, and ResNext50, to show the extent of performance improvement over existing works. The re-ranking process is shown to be strongly dataset-dependent for which the conventionally used k-reciprocal neighbors method has been improved by augmenting a new simple score-based algorithm for obtaining the best mix of component distances. This can be used as a generalized tool to finetune re-ranking for different datasets.

1. Introduction

Monitoring behavior, activities, and information for the purpose of gaining insights, managing, influencing, or controlling is known as surveillance. CCTV cameras serve as a widely utilized tool for surveillance, having applications in law enforcement, crime scene investigation, and the observation of events and processes through visual data. Surveillance systems leveraging such images and videos can concentrate on objects, individuals, or vehicles to perform tasks such as recognition, identification, re-identification, tracking, or retrieval. This paper specifically addresses the re-identification of vehicles. Vehicle re-identification (VReId) refers to the process of retrieving images of a specific vehicle from a database, based on a query image. VReId finds applications in smart surveillance, traffic management, and fleet management systems [1]. Some of the significant practical challenges faced in VReId include the following: (i) differences in lighting between the query image and the images in the database, (ii) occlusion caused by speed, (iii) varying perspectives from multiple cameras, (iv) low-resolution images, (v) differing camera resolutions, and (vi) lack of available license plate information [2]. Person re-identification, a subject that has been extensively researched, is closely related to VReId. The successful outcomes in person re-identification, predominantly grounded in deep learning techniques, have motivated further investigation into VReId. Most recent works in re-identification adopt the approach of pre-training an existing network architecture on an image classification task utilizing the ImageNet dataset [2,3]. This is because image classification also involves the task of feature extraction to generate discriminative properties. Such networks are referred to as the backbone network in the literature and person re-id or VReId models are built upon such models by addition and deletion of convolutional layers. Any re-identification problem involves two steps. Firstly, some discriminative features are identified in the subject images. Then, the distinctive features are used for finding matches from a collection of such images. It has been observed that various methods which have worked well in person re-identification also work well in the case of VReId. While there exists a lot of visual discriminative features for person reidentification in traffic surveillance data, the discriminative features seem to reduce a lot in the case of VReId which makes the problem quite challenging. The features can be broadly classified into attribute-based and general re-id-based features. Attributes such as color, make, model, and viewpoint serve as global features based on specific characteristics. Local features are certain parts of the vehicle like headlights, mirrors, number plates, backlights, etc. A separate module/network is used to learn attributes. The labels are mapped to learned attributes. This leads to an increase in complexity in such methods. For general re-id feature extraction there exists two approaches, global and local. The global features help in obtaining an overall view of the image and the local ones provide insight into different local information. Another method that has been used quite successfully in person reidentification is learning part-based features, which is essentially a general re-id method of learning local features. In these approaches, the convolutional tensor produced by the backbone model is divided into separate sections to obtain local representations. In this paper, we take inspiration from a high-performing supervised learning method, the part-based convolutional baseline (PCB) used in person re-identification to solve VReId [4]. The principal advantage of PCB over most other re-identification methods is its relative simplicity owing to its single-branch model and its data-independent approach which makes it an easy ready to use model for re-identification tasks. In a supervised learning framework, this technique employs a generalized uniform partitioning of the convolutional tensor generated after the image is processed by a backbone network. Our research thoroughly examines the influence of the backbone network through experiments conducted with multiple advanced pre-trained network architectures. The parts obtained after partitioning ensure that we have localized information from different parts of the image which can have higher discriminative value. We also use a refined part pooling as described in [4] for removing in-part inconsistencies within the parts. Inspired by speech and signal analysis, we also propose a novel overlapped part-based approach for solving part inconsistencies over neighboring parts by using an overlap of the original parts. The approach used is, however, different from the conventional windowing technique used in signal processing, as we utilize a two-step process of first generating overlapped parts and then using the classifier weights of the original parts, with a new set of parts generated through combining the information from neighbouring parts. Finally, in the testing phase, we explore a reranking process using k-reciprocal neighbor encoding during evaluation which uses a composite mixture of two distances between two image features—viz. the Euclidean distance and the Jaccard distance calculated from k-reciprocal nearest features for calculating the results. To the best of our knowledge, we present the first detailed analysis, on two VReId datasets, of the effect of the optimal mix of the two distances for the re-ranking process. We then finally propose an ensemble model based on feature learnings in case of different backbone networks. Experiments are conducted on two large-scale datasets—VeRi and VehicleId—and are shown to have comparable results with the state of the art. Our present work is related closely to the part-based network (PCB) analyzed over person re-identification datasets [4]. The main contributions in this paper can be summarized below:
  • Our work is an exhaustive extension of Sun et al. which uses PCB with refined part pooling (PCB-RPP) over person re-identification (PReid) datasets [4]. To the best of our knowledge, the application of this method to vehicle re-identification (VReid) has not been reported until now. Our analysis revealed that the performance behavior of RPP is different when applied to vehicle re-identification problem. RPP improves performance in the case of the VeRi dataset while it degrades performance in the case of the VehicleId dataset. The results of Sun et al. over three PReid datasets however showed consistent performance for all of them. The reason investigated was that RPP fails to solve within-part inconsistencies over neighboring parts in the case of images in the VehicleId dataset. This motivated further investigations to improve the PCB-RPP architecture.
  • We propose a novel Overlapped Part-based Convolutional Baseline method (OPCB) which can take care of within-part inconsistencies. OPCB overlaps parts of two adjacent parts to generate new parts. The new parts are then trained using the classifiers corresponding to the original parts and the results are concatenated to generate the feature set. This method is found to work consistently well across both VeRi and VehicleId datasets and even outperforms the refined part pooling approach.
  • CNN-based models reported in the literature are usually tested over multiple architectures to identify and compare performance variations if there are any. Our work therefore investigates three residual networks as backbone networks viz. ResNet50, ResNet101, and ResNext50. The main difference lies in the number of layers, with ResNet101 and ResNext50 having higher number of layers than ResNet50. It was found that increasing number of layers or cardinality does not necessarily increase accuracy in solving vehicle re-identification problem. The approach of ensembling different architectures/models is commonly reported in the CNN literature. However, in the VeRid application domain, there are very few such references. In this work, we present a model ensembled over the three architectures which significantly improves re-identification results compared to the state of the art.
  • The widely used re-ranking using k-reciprocal nearest neighbors scheme has been improved and incorporated in our work. This process uses a mix of Euclidean and Jaccard distances. We have augmented an algorithm that uses a score to obtain the best mix. This can be used as a generalized tool to finetune re-ranking for different datasets. Our results demonstrate that this has a significant impact in improving re-identification accuracy.
We have demonstrated the PCB-based model along with overlapped parts and ensembled approach using different backbone networks as a strong model for VReId problem in light of comparable results with the state of the art.
The subsequent sections of the paper are structured in the following manner: Section 2 focuses on the review of the existing literature and the rationale behind the study and Section 3 outlines the methodology employed. In Section 4, the experimental results are examined, analyzed, and contrasted with current standards, while Section 5 summarizes the findings and draws conclusions.

2. Related Work and Motivation

In this section, we review some of the existing approaches used in VReId. We also review the existing datasets used in our work.

2.1. Vehicle Re-Id Datasets

Several vehicle re-identification datasets have been created for use in vehicle re-identification [5,6,7,8,9,10,11]. The majority of existing studies present their findings primarily on the VeRi and VehicleId datasets. Therefore, we utilize the VeRi and VehicleId datasets for our experiments, which are detailed below:
  • VeRi Dataset: The dataset consists of 49,357 images of 776 vehicles taken from 20,138 different camera angles. These cameras are distributed across an area of 1 square kilometer. The images are sourced from unconstrained real-world surveillance footage. Each vehicle is photographed by 20 different cameras. The images vary in terms of viewpoint, resolution, lighting conditions, and occlusion. The training set contains 575 unique vehicle identities, while the remainder are found in the test set. The test set includes both a query set and a gallery set. In the query set, there is only one image per vehicle identity, which is utilized to locate the corresponding identity in the gallery set.
  • VehicleId Dataset: The collection consists of 221,763 pictures of 26,267 vehicles. Each vehicle is depicted from two distinct angles: front and back. The training dataset includes 110,178 images representing 13,134 different vehicle identities. The test set has been divided into three parts. These parts include 800, 1600, and 2400 vehicles, containing 7332, 12,995, and 20,038 images, respectively. For every one of the three parts, the query set consists of a single vehicle image for each identity, while the remainder of the vehicle images are included in the gallery set, where matches need to be identified.

2.2. Vehicle Re-Id Methods

Vehicle re-id methods in the literature may be divided broadly into four categories:
  • Single modal methods comprise a pre-trained backbone network that is used to extract features which are then used for re-identification [3,10].
  • Models with separate convolutional networks for extracting global/local attribute features which are then combined with the output features of a pre-trained network for re-identification [12,13,14]. This two-pronged approach adds to their complexity.
  • CNN-based models for detecting global/local re-id features without using any pre-trained backbone model [15,16,17,18].
  • Part-based models with single [4] or multiple branches [19] for extracting features. Single-branch models are inherently simpler than multi-branch models because they have fewer convolutional layers.
The early works on VReid propose methods based on the fusion of texture-based handcrafted features along with deeply learned features [5,20]. However, they fail to achieve significant improvement in accuracy. Some methods then have used models pre-trained on ImageNet, such as AlexNet [5,21], VGG [6,22], GoogLeNet [20,23], Resnet [24,25], and Mobile Net [26] for feature extraction. However, all these methods are single-modal methods focusing on the simple use of the backbone architecture to extract the features. Wang et al. suggest a combined method that incorporates learning deep features alongside attributes like camera angle, vehicle color, and type [13]. Qian et al. utilize a dual-branch framework in which local characteristics are gathered through horizontal stripes from a backbone layer’s output, merging this with attribute-related information to generate a comprehensive feature [27]. Zhao presented a single-stage Single Shot Detector aimed at pinpointing areas of interest in vehicle images, utilizing deep features as local characteristics [14]. Another work adopts a dual strategy that initially identifies local areas of interest and subsequently aligns them with the overall feature map, guaranteeing a combination of local attribute-based features and global re-identification characteristics [12]. Typically, various types of custom object detectors are employed to extract specific parts or attributes [28]. A significant drawback of this model is the necessity of a separate module for attribute extraction, which substantially increases the complexity. Wang et al. utilized a convolutional neural network to extract region feature vectors from various segmentation outcomes and acquired appearance feature vectors of the target vehicle by combining them with global feature vectors [15]. Liu et al. uses a region-aware model for finding local features [29]. Zhu et al. implemented a novel architecture without a backbone network, utilizing closely connected convolutional units for feature learning. This architecture was later applied with a Siamese network-like structure to enhance feature detection, and also incorporated a modified quadrature pooling layer for feature production [16,17,18]. Peng et al. utilizes a feature learning approach that incorporates a Spatial Transformer Network-based localization model to combine global and local features into a representation for vehicle ReID [30]. However, they do not employ any pre-trained backbones and only propose independent network architectures, thereby affecting the discriminative ability of the models in the form of lower accuracy. Spatio-temporal features of the vehicles in city traffic surveillance scenarios which are also used for traffic prediction have been utilized for solving VReid [31,32]. However, this imposes a limitation for such temporal information to be known for retrieval. Another method that has been used quite successfully in person reidentification and in general image classification is learning part-based features. In a method used for person re-identification where horizontal part-based features are used as local features [4], a refined part pooling is also used to reduce in-part inconsistencies. However, it does not analyze the effect of using different backbones or independently using partitions in other dimensions. Another work used in person re-identification with good accuracy uses a multi-branch mix of part-based and global feature-based networks [3]. Wang et al. employ a dual-branch part-based network that incorporates both horizontal and vertical segmentation, along with an external memory to store the features from each branch [19]. It employs a blend of two losses for metric learning. Such multi-branch networks are naturally more complex than single-branch networks.
The problem of within part inconsistency in the case of part-based methods has been studied in the literature and methods like RPP have been proposed for solving this problem [4]. Works like PAN target the problem of misalignment of parts across vehicle images due to differing views by using a novel Part Alignment Network. The method uses cross-correlation for alignment of vehicle parts [33]. For maintaining consistency in different parts for person re-identification, a model was proposed which aimed at aligning the predicted distributions between each part using KL loss. The parts are trained following the global feature in a multi-stage format [34]. Another work in VReid uses a cross part interaction module to enhance correlation between vehicle parts obtained by rigid partioning to generate local information. The architecture follows a three branch structure with global and part features [35]. In speech and signal analysis problems, windowing is used to analyze sections of the signal or speech data [36]. An overlap approach is used in such cases to ensure continuity and context which is akin to the problem of within part inconsistency in this instant case. Taking inspiration from such methods, we also propose a part-based overlap method for solving within part inconsistency. To the best of our knowledge, single branch part-based models for VReId have not been thoroughly studied. Also, by using some augmentations like refined part pooling, part overlapping, re-ranking, and ensembling, high accuracy can be achieved.

3. Methodology

The framework of the part-based model (PCB) is as illustrated in Figure 1. The steps and methods are detailed below.

3.1. Data Preprocessing

The unprocessed images undergo specific alterations during the pre-processing phase.
  • First, the images are resized to a fixed size (384,192). This is undertaken while keeping the pre-trained backbone model and the input it expects in view.
  • Random Horizontal Flip—This is a technique for augmenting image data where an input image is flipped along the horizontal axis based on a specified probability.
  • Random Erasing—This method involves randomly selecting a portion of an image and removing its pixels, which aids in training the model to be more robust with the provided data.
  • Finally, the images are transformed into tensors and standardized to the same mean (0.485, 0.456, 0.406) and standard deviation (0.229, 0.224, 0.225) that are applied to typical ImageNet images. This standardization is essential for attaining the best outcomes when utilizing backbone architectures that have been pre-trained on ImageNet.

3.2. Backbone Model

Any model that is appropriate for image classification can be utilized as the backbone model. For this instance, we have selected ResNet50, ResNet101, and ResNeXt50, which have been pre-trained on the ImageNet dataset. ResNet or residual network is used in many computer vision applications. It uses the concept of interconnections from higher to lower layers in order to mitigate the problem of vanishing gradients, particularly in very deep networks. The building block of Resnet is defined as follows:
y = F ( x , w i ) + x
Here, x and y are, respectively, the input and output vectors of the layers. The function F ( x , w i ) represents the layers in between where wi are the weights. Here, the output is arrived at by addition of the output of the layers with the original input which helps preserve information from the input layer which in effect is the essence of ResNet. ResNet has been used in similar part-based networks [3,4,25], which inspired us to use it as one of the base models. ResNext [37] is a residual network, a modified form of ResNet with an extra dimension, cardinality. The structure of ResNet and ResNext is shown in Figure 2
This addition of the extra dimension reduces the number of hyperparameters compared to conventional ResNet without compromising on accuracy. In a typical neural network, convolutional layers are usually succeeded by a final pooling layer and then a fully connected layer to produce the outputs. This is also true for our ResNet 50 backbone network. The last two layers consist of a global pooling layer followed by a fully connected layer. We will remove the last two layers of the backbone network to continue constructing our model.

3.3. Part-Based Convolutional Baseline (PCB) Model

On passing an image through the backbone network, we obtain a 3D tensor, T, as shown in Figure 1. The 3D tensor has three axes viz. horizontal, vertical, and channel, and is of the dimensions N x C x H x V where N is the batch size, C is the number of channels, H is the height wise dimension, and V is the width wise dimension. The feature vector is a column vector along the channel axis. Several parts (say, p) are extracted uniformly of equal width and height on the convolutional tensor T by passing it through a 2D pooling layer with size p x 1 which converts the tensor dimension to N x C x p x 1 . p parts are then extracted on the convolutional tensor with each part having dimension N x C x 1 x 1 and the channel dimensions being the same as that of the original convolutional tensor. All the feature vectors in a particular part are pooled into a single column vector g i ( i = 1 , 2 , , p ) . We pool all the column vectors in a same part into a single part-level column vector g i ( i = 1 , 2 , , p ) of dimension 2048. We have used average pooling as the pooling method. The column vector, g, is then transformed into a dimension-reduced column vector, h of dimension 256, using a convolutional layer. Each vector, h, is then fed into a classifier comprising a fully-connected (FC) layer with Softmax activation for identity prediction. For every image, we add the cross-entropy losses over the parts for every training image. The model is sought to be optimized by minimizing the loss. In the testing phase, the concatenation of part vectors g or h is used as the vehicle image descriptor, denoted by G or H, where G = [ g 1 , g 2 , , g p ] or H = [ h 1 , h 2 , , h p ] . In our experiment, we use the vector H due to its small size and hence it is computationally less expensive.

3.4. Refined Part Pooling

Partitioning of the convolutional tensor aims to target attention towards various localized parts of the image which may be specific identifiers of the image. However, in the case of uniform partition, there is a high chance of within part inconsistency, i.e., some portions of a part might actually be more similar to the nearby part. Although these might essentially be outliers, rearranging them to their respective nearest parts might provide respite from such inconsistencies. For this, in the referred paper, they have proposed a refined part pooling method, using the trained parts from the earlier stage, i.e., new parts are arrived at which are closest to the original parts [4]. Initially, the part-based model with uniform parts is trained for convergence. Then, a classifier with p-part categories is used on the convolutional tensor T to generate a new set of parts to replace the original parts. In this process, the other learned layers of the part-based model are kept fixed and only the part classifier is retrained on the training data. This step ensures that the classifier generates parts which are close to the original parts. The entire model is then trained again for a few epochs for fine tuning and optimization.

3.5. Proposed Overlapped Part-Based Convolutional Baseline (OPCB)

Partitioning of the convolutional tensor aims to target attention towards various localized parts of the image which may be specific identifiers of the image. However, in case of uniform partition, there is a high chance of within-part inconsistency, i.e., some portion of a part might actually be more similar to the nearby part. Although these might essentially be outliers, rearranging them to their respective nearest parts might provide respite from such inconsistencies. We aim to achieve this by introducing the concept of overlapped parts. We train a separate set of uniform parts which are essentially overlapped parts corresponding to the original parts. The process is illustrated in Figure 3 and Figure 4 and described below:
Let the original parts be p = { p 0 , p 1 , , p n 1 }.
Now, any overlapped part would essentially be an overlap between p i 1 and p i i > 0 and i < n .
To achieve this, each p i is partitioned into two parts p ( i , 0 ) and p ( i , 1 ) such that p ( i , 0 ) , p ( i , 1 ) ϵ p i .
The new overlapped part q i is formed such that
q i = concat ( p ( i 1 , 1 ) , p ( i , 0 ) )
i > 0 and i < n
q = { q 1 , q 2 , , q ( n 1 ) }
So, if there are n original parts, there must be n − 1 overlapped parts, q. Now, each new part, q i , is composed partly of original parts p ( i 1 ) and p i i > 0 and i < n .
In addition, there are two residual parts, p ( 0 , 0 ) and p ( n 1 , 1 ) , in addition to the n 1 overlapped parts.
So, there are a total of n + 1 new parts, n 1 overlapped and 2 residual parts. Let the original fully connected linear layers/classifiers be c = c 0 , c 1 , , c n 1 which correspond to each of the original parts in p.
Each of the overlapped parts are passed through two fully connected layers/classifiers separately. The two classifiers are the ones corresponding to the original parts of each overlapped part. So, we obtain 2 ( n 1 ) no. of features for n 1 overlapped parts. The two residual parts, p ( 0 , 0 ) and p ( n 1 , 1 ) , are passed through c 0 and c n 1 , respectively, to obtain two more features. So, we have a total of 2 n features which are concatenated pairwise to obtain n final features. The final feature set is given by the following:
m i = c o n c a t ( c i + 1 ( p ( i , 0 ) ) , c i + 1 ( p ( i , 1 ) ) )
where i { 0 , 1 , 5 }
The process is described in the following steps:
  • Step 1: Train the PCB model.
  • Step 2: Generate overlapped and residual parts.
  • Step 3: Train overlapped and residual parts through original classifiers.
  • Step 4: Concatenate the output pairwise to generate final features.

3.6. Generalized Training and Evaluation Procedure

The model is trained on the VeRi and VehicleId datasets. Most of the state-of-the-art methods are trained on these datasets, so it gives us an opportunity for comparison.
For the RPP method, the training is completed in three phases. First, it is trained in the PCB phase for 60 epochs in the case of VeRi dataset and 100 epochs in the case of VehicleId dataset. After that, it is trained in the RPP phase for 5 epochs for learning and realigning the parts. Finally, the overall network is trained for 10 epochs for final optimization. The learning rates are fixed to be different for all the three phases. After that, the evaluation is performed using a query and gallery dataset. The images in the query dataset are used to search for images of the same images in the gallery dataset. The trained model is used to collect the features, and then, using a Euclidean distance, the distance between the probe and gallery features is found out and is utilized for finding the match.
For the overlapped method, the training is performed in two phases. First, it is trained for the part-based network, PCB, as described above and then for the overlapped parts, OPCB, for 20 epochs for both VeRi and VehicleId datasets. The learning rates are fixed to be the same for the two phases as we use the same classifiers as the PCB stage and there is no separate trainable part classifier. After this, the evaluation is performed using a query and gallery dataset. The images in the query dataset are used to search for the same images in the gallery dataset. The features are collected from the trained models of the overlapped phase to generate the feature vector, and then, using a Euclidean distance, the distance between the probe and gallery features is found out and is utilized for finding the match.
Evaluation is performed for Mean Average Precision, Rank-1 accuracy, Rank-5 accuracy, and Rank-10 accuracy. The metrics are defined below: Average precision is given by the following:
A P = k = 1 n p ( k ) g ( k ) / N g
where n is the number of test and N g is the number of reference (ground truth) images while the precision is p(k) at the k th position. g ( k ) represents a function where the value is 1 if a match is found at the k th position; otherwise it is 0. The mean average precision (mAP) is formulated as follows:
m A P = ( q = 1 Q A P ( q ) ) / Q
where Q is the total number queries. Rank measures the similarity (in percentage) of a test to its class, e.g., if test1 corresponds to class1 and is found in top1 results then it is called rank@1; if found in the top5 results then it is called rank@5 and so on.

3.7. Re-Ranking Using k-Reciprocal Nearest Neighbours

During the evaluation stage, given a probe image, a ranked list is generated which contains images in descending order of their similarity to the probe image. In our evaluation mechanism, we normally find the distance between the probe image and the images in the gallery set, which is generally a Euclidean distance, and then use the distance to create an ordered list of similar items according to their distances. Re-ranking is used as a post processing activity in order to increase accuracy. Inspired by [38], we set about creating a new distance which is a combination of the original Euclidean distance and the Jaccard distance, which in turn is calculated from the k-reciprocal features of two images. A new ranking list or a re-ranked list is then created from the new distance and hence the name re-ranking. If two images are in the k-nearest neighbors of each other, then the images can be said to be in a k-reciprocal neighborhood of each other. Assuming a probe person p and a gallery set with N images, G = g i | i = 1 , 2 , N . If  N ( p , k ) is the k-nearest neighbors of the probe p, the k-reciprocal nearest neighbors R ( p , k ) is defined as follows:
R ( p , k ) = { g i ( g i N ( p , k ) ( p N ( g i , k ) }
The k-reciprocal neighbor set of probe and gallery images is encoded in a vector form to form a k-reciprocal feature. A few nearest neighbors of the probe are then added to the feature vector to improve performance. The k-reciprocal features are then used to calculate the Jaccard distance as in Equation (6). The final distance is calculated as the weighted combination of the original distance (Euclidean distance between probe and gallery) and the Jaccard distance. It is subsequently used to acquire the re-ranking list as described in [38]. The final distance in the case of re-ranking is given by the following:
F D = ( 1 λ ) JD + λ × ED
where FD—Final distance, JD—Jaccard Distance and ED—Euclidean Distance, and the value of λ lies between 0 and 1. Jaccard distance is given by,
J D = 1 | ( R ( p ) R ( g i ) ) | / | ( R ( p ) R ( g i ) ) |
where R ( p ) and R ( g i ) indicates the k-reciprocal feature set of the probe and gallery, respectively. A low value of λ will increase the contribution of the Jaccard distance and a high value of λ will increase the contribution of the original Euclidean distance. Selection of an appropriate value of λ is therefore a design choice.

3.8. Proposed Algorithm for Best Distance Mix Calculation

If we consider a parameter p and the corresponding metrics ( m 1 , m 2 , ) , and, if we need to find the best value of p for which we would have the highest values of the metrics, we would calculate an error score, S, for each value of p, such that
S = All metrics ( m 1 , m 2 , ) max ( V metric ) V
where V m e t r i c is the set of values of the metric across all values of p and V is the value of the metric for a particular p. The lowest value of S will correspond to the best value of p for the dataset. The detailed process is described in Algorithm 1.
Algorithm 1 Algorithm for distance mix calculation for a dataset
  • Step 1: Compute values of ( m 1 , m 2 ,…) for different values of p.
  • Step 2: Determine highest accuracy value for each metric corresponding to different values of p, i.e., max( V m e t r i c )
  • Step 3: Calculate (max( V m e t r i c )-V) for each metric for all selected different values of p.
  • Step 4: Calculate the score, S as per Equation (10), of the expression computed in Step 3 for each value of p.
  • Step 5: The smallest value of S corresponds to the ideal value of parameter, p for the dataset.

4. Experimental Results and Analysis

We carried out a number of thorough experiments on the VeRi and VehicleId datasets utilizing PCB, refined part pooling, and OPCB with re-ranking during the evaluation phase, and the results of these experiments are presented in this section. Three backbone networks viz. Resnet-50, Resnet-101, and ResNext 50 were used. The images during both the testing and evaluation phases were resized to 384 × 192. The output tensor after the backbone layers has a channel dimension of 2048. The training batch size is set at 32. The weights are learned by an SGD optimizer using cross entropy loss as the loss criterion. We perform uniform partitioning in horizontal dimension with six parts. In the PCB stage, the learning rate is fixed at 0.1, while during re-ranking and final training it is fixed at 0.01. During OPCB, however, we have fixed the learning rate as the same as during the PCB stage. This is because the same classifiers as used in PCB are reused for OPCB. In OPCB stage, a mixture of triplet loss and cross entropy loss is used as the loss funtion with the ratio of triplet to cross-entropy being 1 : 2 . The re-ranking process was evaluated to find the best final distance with respect to λ using Equation (9). To the best of our knowledge, this type of analysis has not been investigated until now in the literature. We have investigated the performance of an ensemble model using all three backbone networks over both VeRi and VehicleId datasets.

4.1. Distance Calculation for Re-Ranking

Using (5), we experimented with VeRi and VehicleId datasets to find out the best value of λ for which we obtain the best values of the accuracy metrics. We varied the value of λ from 0 to 1 and calculated the re-ranked mAP, Rank-1 accuracy, Rank-5 accuracy, and Rank-10 accuracy for the part-based convolutional model. The results are shown in Table 1. Following the process outlined in Section 3.6, we present the accuracy values for values of λ , which is our parameter in this case and compute the corresponding values of score, S.
From the above table, we can clearly see that the best values of accuracies for the VeRi and VehicleId datasets occur at different values of λ . For the VeRi dataset, the highest values seem to occur at lower values of λ , while for the VehicleId dataset, the highest values occur for higher values of λ . The lowest value of S is found at 0.3 for the VeRi dataset, while it is found at 0.9 for VehicleId dataset.
The major conclusion of the above results is that the effect of the k-reciprocal neighbor feature-based Jaccard distance is significantly greater for the VeRi dataset than VehicleId. The reason for this may be due to the fact that there are as many as 20 camera viewpoints in the VeRi dataset in contrast to only 2 in VehicleId dataset. As a result, with only two viewpoints for diversity, the Euclidean distance is sufficient for the VehicleId dataset, and using a nearest reciprocal neighbor extrapolation does not provide any additional advantages. The result also highlights the strong impact of the type of dataset on the value of λ . For Market1501 dataset, the best value was reported to be 0.3 with six camera viewpoints, which further vindicates our proposition [38].

4.2. Impact of Backbone and Ensembled Model

In the earlier section, the results tabulated were all on the ResNet50 (R50) backbone. Here, we investigate the comparative performance taking two other backbones, Resnet 101 (R101) and ResNext50 (RN50). Table 2 and Table 3, respectively, present the accuracy metrices on VeRi and VehicleId datasets.
From Table 2, it can be observed that PCB along with refined part pooling and re-ranking with Resnet-50 as the backbone gives the best results in terms of mAP. Hence, we find that increasing the number of layers need not necessarily increase the accuracy for the Veri dataset.
From Table 3, it can be observed that PCB with re-ranking gives the best results across all metrics. Also, the backbones, Resnet 50, Resnet 101, and ResNext50, perform at par in the case of PCB, and a clear winner cannot be judged in terms of accuracy. However, ResNet50 has lower computational complexity compared to the other two backbones. It can also be seen that refined part pooling, although quite effective in the case of VeRi dataset, fails to improve performance in the case of the VehicleId dataset. The reason for this is the unique information in parts is less likely to be effective in the case of datasets with a lower number of viewpoints. When the number of viewpoints is large, the feature vector for each part for each vehicle identity is more likely to contain more unique information. This is a major issue in the generalizability of such refinement methods used in person re-identification to vehicle re-identification as the parts are naturally more unique in the case of person images, irrespective of viewpoint, while this is not the case in vehicle images.
During the testing phase, the query and the gallery features are computed separately for all the different backbones for a particular dataset. The distances are then found between the query and the gallery features which are then averaged to arrive at new distances. The new distances are then used for the testing. If we have S b 1 , S b 2 , S b 3 as the sets of distances corresponding to different backbones b1, b2, and b3, then we must have the combined average set of distances
F = a v g ( S b 1 , S b 2 , S b 3 )
The ensembled results on backbones ResNet50, ResNet101, and ResNext50 are similarly computed for the VeRi and VehicleID datasets and are shown in Table 4 and Table 5, respectively. The results show considerable improvement over standalone backbone networks.

4.3. Performance of OPCB

The accuracy results of OPCB for the VeRi and VehicleId datasets are shown in Table 6 and Table 7, respectively. From the tables, the following points can be adduced:
  • OPCB improves PCB consistently across the VeRi and VehicleId datasets. This is in contrast with RPP which can only improve PCB for the VeRi dataset and fails to do so in the case of the VehicleId dataset.
  • OPCB performs at par with RPP in the case of the VeRi dataset and largely improves over RPP and PCB in the case of the VehicleId dataset.
OPCB attempts to bring similar portions of neighbouring parts together to resolve part inconsistency. RPP trains the part classifier for the entire feature set which might bring similar features from distant parts together. However, some contextual information gets lost in the process which is not the case for OPCB which preserves contextuality.

4.4. Comparison with State of the Art

The analysis of our results gives two major takeaways: Our proposed ensembled model with refined part pooling and re-ranking works best on the VeRi dataset, but for the VehicleId dataset, only the re-ranking scheme is better suited. RPP in fact slightly degrades the performance of the latter. Accordingly, we make a comparison with the state-of-the-art results on the VeRi and VehicleId datasets, as shown in Table 8 and Table 9, respectively.
The following facts come to light through the above comparison:
  • In the case of the VeRi dataset, our proposed ensembled part model with horizontal partitioning along with the refined part pooling gives significantly better results compared to all in terms of mAP. The R-1 and R-5 performances are as good as most of the reported works. The proposed OPCB and ensembled model also outperforms transformer-based models like Swin Transformer and TransReid.
  • In the case of the Vehicle-Id dataset, most works have not reported their mean average precision and hence the values could not be compared. But on comparison with the state of the art in terms of Rank-1 and Rank-5 accuracies, it is found that our ensembled model performs at par with most of the reported results and outperforms transformer-based models like TransReid.

5. Conclusions

In this work, we have analyzed the PCB model with refined part pooling for VReId over two standard datasets using three different backbones networks. An ensemble over these three was proposed which was found to outperform significantly in comparison to the use of standalone backbones. A limitation of the refined part pooling identified was that the performance in the case of the VehicleId dataset was not as good as it was for the VeRi dataset. To solve this, we have proposed a novel method, OPCB, which was found to perform consistently across datasets and improves upon PCB and RPP. The scope of re-ranking in improving performance was evaluated and extensively analyzed. The best mix of Jaccard and Euclidean distances was found to be dataset-specific and, therefore, a generalized method for arriving at the best mix was proposed. Finally, the proposed ensembled model with refined part pooling and re-ranking and the OPCB method was found to have a great application potential in VReId systems, which matches closely with recent related works. The study could be expanded by exploring how various factors, such as the number of components, pre-processing settings, and overlap percentage, influence re-identification accuracy. Future research could also examine the problem of feature redundancy in overlapping parts, as well as the efficiency of the computations involved.

Author Contributions

Conceptualization and methodology, R.K.N. and D.M.; software R.K.N.; validation, R.K.N. and D.M.; formal analysis, R.K.N. and D.M.; investigation, R.K.N. and D.M.; data curation, R.K.N. and D.M.; writing—original draft preparation, R.K.N.; writing—review and editing, R.K.N. and D.M.; supervision D.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

We used the publicly available VeRi and VehicleId datasets in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Amiri, A.; Kaya, A.; Keceli, A.S. A Comprehensive Survey on Deep-Learning-based Vehicle Re-Identification: Models, Data Sets and Challenges. arXiv 2024, arXiv:2401.10643. [Google Scholar] [CrossRef]
  2. Wang, H.; Hou, J.; Chen, N. A Survey of Vehicle Re-Identification Based on Deep Learning. IEEE Access 2019, 7, 172443–172469. [Google Scholar] [CrossRef]
  3. Wang, G.; Yuan, Y.; Chen, X.; Li, J.; Zhou, X. Learning Discriminative Features with Multiple Granularities for Person Re-Identification. arXiv 2018, arXiv:1804.01438. [Google Scholar] [CrossRef]
  4. Sun, Y.; Zheng, L.; Li, Y.; Yang, Y.; Tian, Q.; Wang, S. Learning Part-based Convolutional Features for Person Re-Identification. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 902–917. [Google Scholar] [CrossRef] [PubMed]
  5. Liu, X.; Liu, W.; Ma, H.; Fu, H. Large-scale vehicle re-identification in urban surveillance videos, 2016. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Seattle, WA, USA, 11–15 July 2016. [Google Scholar]
  6. Liu, H.; Tian, Y.; Wang, Y.; Pang, L.; Huang, T. Deep Relative Distance Learning: Tell the Difference between Similar Vehicles. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2167–2175. [Google Scholar] [CrossRef]
  7. Kanacı, A.; Zhu, X.; Gong, S. Vehicle Re-identification in Context. In Pattern Recognition; Brox, T., Bruhn, A., Fritz, M., Eds.; Springer: Cham, Switzerland, 2019; pp. 377–390. [Google Scholar]
  8. Lou, Y.; Bai, Y.; Liu, J.; Wang, S.; Duan, L. VERI-Wild: A Large Dataset and a New Method for Vehicle Re-Identification in the Wild. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–19 June 2019; pp. 3230–3238. [Google Scholar] [CrossRef]
  9. Tang, Z.; Naphade, M.; Liu, M.; Yang, X.; Birchfield, S.; Wang, S.; Kumar, R.; Anastasiu, D.; Hwang, J. CityFlow: A City-Scale Benchmark for Multi-Target Multi-Camera Vehicle Tracking and Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Los Alamitos, CA, USA, 15–19 June 2019; pp. 8789–8798. [Google Scholar] [CrossRef]
  10. Yan, K.; Tian, Y.; Wang, Y.; Zeng, W.; Huang, T. Exploiting Multi-grain Ranking Constraints for Precisely Searching Visually-similar Vehicles. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 562–570. [Google Scholar] [CrossRef]
  11. Zhou, Y.; Liu, L.; Shao, L. Vehicle Re-Identification by Deep Hidden Multi-View Inference. IEEE Trans. Image Process. 2018, 27, 3275–3287. [Google Scholar] [CrossRef] [PubMed]
  12. He, B.; Li, J.; Zhao, Y.; Tian, Y. Part-Regularized Near-Duplicate Vehicle Re-Identification. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–19 June 2019; pp. 3992–4000. [Google Scholar] [CrossRef]
  13. Wang, H.; Peng, J.; Chen, D.; Jiang, G.; Zhao, T.; Fu, X. Attribute-Guided Feature Learning Network for Vehicle Reidentification. IEEE MultiMedia 2020, 27, 112–121. [Google Scholar] [CrossRef]
  14. Zhao, Y.; Shen, C.; Wang, H.; Chen, S. Structural Analysis of Attributes for Vehicle Re-Identification and Retrieval. IEEE Trans. Intell. Transp. Syst. 2020, 21, 723–734. [Google Scholar] [CrossRef]
  15. Wang, Z.; Tang, L.; Liu, X.; Yao, Z.; Yi, S.; Shao, J.; Yan, J.; Wang, S.; Li, H.; Wang, X. Orientation Invariant Feature Embedding and Spatial Temporal Regularization for Vehicle Re-identification. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 379–387. [Google Scholar] [CrossRef]
  16. Zhu, J.; Du, Y.; Hu, Y.; Zheng, L.; Cai, C. VRSDNet: Vehicle re-identification with a shortly and densely connected convolutional neural network. Multimed. Tools Appl. 2019, 78, 29043–29057. [Google Scholar] [CrossRef]
  17. Zhu, J.; Zeng, H.; Lei, Z.; Liao, S.; Zheng, L.; Cai, C. A Shortly and Densely Connected Convolutional Neural Network for Vehicle Re-identification. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 3285–3290. [Google Scholar] [CrossRef]
  18. Zhu, J.; Huang, J.; Zeng, H.; Ye, X.; Li, B.; Lei, Z.; Zheng, L. Object Reidentification via Joint Quadruple Decorrelation Directional Deep Networks in Smart Transportation. IEEE Internet Things J. 2020, 7, 2944–2954. [Google Scholar] [CrossRef]
  19. Wang, H.; Peng, J.; Jiang, G.; Xu, F.; Fu, X. Discriminative feature and dictionary learning with part-aware model for vehicle re-identification. Neurocomputing 2021, 438, 55–62. [Google Scholar] [CrossRef]
  20. Liu, X.; Liu, W.; Mei, T.; Ma, H. PROVID: Progressive and Multimodal Vehicle Reidentification for Large-Scale Urban Surveillance. IEEE Trans. Multimed. 2018, 20, 645–658. [Google Scholar] [CrossRef]
  21. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Advances in Neural Information Processing Systems; Pereira, F., Burges, C., Bottou, L., Weinberger, K., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25. [Google Scholar]
  22. Liu, S.; Deng, W. Very deep convolutional neural network based image classification using small training sample size. In Proceedings of the 2015 3rd IAPR Asian Conference on Pattern Recognition (ACPR), Kuala Lumpur, Malaysia, 3–6 November 2015; pp. 730–734. [Google Scholar] [CrossRef]
  23. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Los Alamitos, CA, USA, 11–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  24. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  25. Chen, H.; Lagadec, B.; Bremond, F. Partition and Reunion: A Two-Branch Neural Network for Vehicle Re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Long Beach, CA, USA, 16–20 June 2019. [Google Scholar]
  26. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  27. Qian, J.; Jiang, W.; Luo, H.; Yu, H. Stripe-based and attribute-aware network: A two-branch deep model for vehicle re-identification. Meas. Sci. Technol. 2020, 31, 095401. [Google Scholar] [CrossRef]
  28. Yan, L.; Li, K.; Gao, R.; Wang, C.; Xiong, N. An Intelligent Weighted Object Detector for Feature Extraction to Enrich Global Image Information. Appl. Sci. 2022, 12, 7825. [Google Scholar] [CrossRef]
  29. Liu, X.; Zhang, S.; Huang, Q.; Gao, W. RAM: A Region-Aware Deep Model for Vehicle Re-Identification. In Proceedings of the 2018 IEEE International Conference on Multimedia and Expo (ICME), Los Alamitos, CA, USA, 23–27 July 2018; pp. 1–6. [Google Scholar] [CrossRef]
  30. Peng, J.; Wang, H.; Zhao, T.; Fu, X. Learning multi-region features for vehicle re-identification with context-based ranking method. Neurocomputing 2019, 359, 427–437. [Google Scholar] [CrossRef]
  31. Zhou, Y.; Li, J.; Chen, H.; Wu, Y.; Wu, J.; Chen, L. A spatiotemporal attention mechanism-based model for multi-step citywide passenger demand prediction. Inf. Sci. 2020, 513, 372–385. [Google Scholar] [CrossRef]
  32. Kim, H.G.; Na, Y.; Joe, H.W.; Moon, Y.H.; Cho, Y.J. Vehicle Re-identification with Spatio-temporal Information. In Proceedings of the 2023 14th International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 11–14 October 2023; pp. 1825–1827. [Google Scholar] [CrossRef]
  33. Chen, Y.; Ma, B.; Chang, H. Part alignment network for vehicle re-identification. Neurocomputing 2020, 418, 114–125. [Google Scholar] [CrossRef]
  34. Li, Z.; Lv, J.; Chen, Y.; Yuan, J. Person re-identification with part prediction alignment. Comput. Vis. Image Underst. 2021, 205, 103172. [Google Scholar] [CrossRef]
  35. Pang, X.; Tian, X.; Nie, X.; Yin, Y.; Jiang, G. Vehicle re-identification based on grouping aggregation attention and cross-part interaction. J. Vis. Commun. Image Represent. 2023, 97, 103937. [Google Scholar] [CrossRef]
  36. Bäckström, T.; Räsänen, O.; Zewoudie, A.; Zarazaga, P.P.; Koivusalo, L.; Das, S.; Mellado, E.G.; Mansali, M.B.; Ramos, D.; Kadiri, S.; et al. Introduction to Speech Processing, 2nd ed.; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar] [CrossRef]
  37. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar] [CrossRef]
  38. Zhong, Z.; Zheng, L.; Cao, D.; Li, S. Re-ranking Person Re-identification with k-Reciprocal Encoding. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3652–3661. [Google Scholar] [CrossRef]
  39. Kuma, R.; Weill, E.; Aghdasi, F.; Sriram, P. Vehicle Re-identification: An Efficient Baseline Using Triplet Embedding. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019; pp. 1–9. [Google Scholar] [CrossRef]
  40. Xu, Y.; Jiang, N.; Zhang, L.; Zhou, Z.; Wu, W. Multi-scale Vehicle Re-identification Using Self-adapting Label Smoothing Regularization. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 2117–2121. [Google Scholar] [CrossRef]
  41. Lin, W.; Li, Y.; Yang, X.; Peng, P.; Xing, J. Multi-View Learning for Vehicle Re-Identification. In Proceedings of the 2019 IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 832–837. [Google Scholar] [CrossRef]
  42. Chu, R.; Sun, Y.; Li, Y.; Liu, Z.; Zhang, C.; Wei, Y. Vehicle Re-Identification With Viewpoint-Aware Metric Learning. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8281–8290. [Google Scholar] [CrossRef]
  43. Zhang, X.; Zhang, R.; Cao, J.; Gong, D.; You, M.; Shen, C. Part-Guided Attention Learning for Vehicle Instance Retrieval. IEEE Trans. Intell. Transp. Syst. 2022, 23, 3048–3060. [Google Scholar] [CrossRef]
  44. Li, J.; Yu, C.; Shi, J.; Zhang, C.; Ke, T. Vehicle Re-identification method based on Swin-Transformer network. Array 2022, 16, 100255. [Google Scholar] [CrossRef]
  45. Qian, J.; Pan, M.; Tong, W.; Law, R.; Wu, E.Q. URRNet: A Unified Relational Reasoning Network for Vehicle Re-Identification. IEEE Trans. Veh. Technol. 2023, 72, 11156–11168. [Google Scholar] [CrossRef]
  46. He, S.; Luo, H.; Wang, P.; Wang, F.; Li, H.; Jiang, W. TransReID: Transformer-Based Object Re-Identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 15013–15022. [Google Scholar]
  47. Wu, F.; Yan, S.; Smith, J.S.; Zhang, B. Vehicle re-identification in still images: Application of semi-supervised learning and re-ranking. Signal Process. Image Commun. 2019, 76, 261–271. [Google Scholar] [CrossRef]
  48. Huang, F.; Lv, X.; Zhang, L. Coarse-to-fine sparse self-attention for vehicle re-identification. Knowl.-Based Syst. 2023, 270, 110526. [Google Scholar] [CrossRef]
  49. Quispe, R.; Lan, C.; Zeng, W.; Pedrini, H. AttributeNet: Attribute enhanced vehicle re-identification. Neurocomputing 2021, 465, 84–92. [Google Scholar] [CrossRef]
  50. Pang, X.; Zheng, Y.; Nie, X.; Yin, Y.; Li, X. Multi-axis interactive multidimensional attention network for vehicle re-identification. Image Vis. Comput. 2024, 144, 104972. [Google Scholar] [CrossRef]
  51. Taufique, A.M.N.; Savakis, A. LABNet: Local graph aggregation network with class balanced loss for vehicle re-identification. Neurocomputing 2021, 463, 122–132. [Google Scholar] [CrossRef]
  52. Sun, K.; Pang, X.; Zheng, M.; Nie, X.; Li, X.; Zhou, H.; Yin, Y. Heterogeneous context interaction network for vehicle re-identification. Neural Netw. 2024, 169, 293–306. [Google Scholar] [CrossRef]
  53. Li, B.; Liu, P.; Fu, L.; Li, J.; Fang, J.; Xu, Z.; Yu, H. VehicleGAN: Pair-flexible Pose Guided Image Synthesis for Vehicle Re-identification. In Proceedings of the 2024 IEEE Intelligent Vehicles Symposium (IV), Jeju Island, Republic of Korea, 2–5 June 2024; pp. 447–453. [Google Scholar] [CrossRef]
  54. Bai, L.; Rong, L. Vehicle re-identification with multiple discriminative features based on non-local-attention block. Sci. Rep. 2024, 14, 31386. [Google Scholar] [CrossRef]
Figure 1. Framework of part-based model (PCB).
Figure 1. Framework of part-based model (PCB).
Electronics 14 03634 g001
Figure 2. ResNet (Left) vs. ResNext (Right) with cardinality of 32.
Figure 2. ResNet (Left) vs. ResNext (Right) with cardinality of 32.
Electronics 14 03634 g002
Figure 3. Framework of OPCB network.
Figure 3. Framework of OPCB network.
Electronics 14 03634 g003
Figure 4. Schema of the overalapping procedure.
Figure 4. Schema of the overalapping procedure.
Electronics 14 03634 g004
Table 1. Comparison of re-ranking accuracies. The lowest value of S and corresponding λ value is marked by bold character.
Table 1. Comparison of re-ranking accuracies. The lowest value of S and corresponding λ value is marked by bold character.
Dataset λ mAPRank-1Rank-5Rank-10Score, S
VeRi060.0095.4196.9698.035.84
0.180.9295.6597.2098.450.39
0.380.5095.7797.4498.570.37
0.580.0095.7797.5698.810.41
0.779.3095.7197.6298.750.60
0.977.9995.9597.9198.870.76
176.2495.0597.7998.991.43
VehicleId (Small)082.8880.4591.6492.645.12
0.185.1682.0996.7198.991.29
0.385.8182.8297.4097.911.04
0.586.4483.5497.6398.950.39
0.786.8083.8997.7598.930.18
0.987.1184.2597.7398.820.05
186.5983.6197.6198.790.38
VehicleId (Medium)079.2476.6988.4090.265.64
0.181.1577.8893.3296.662.04
0.381.8178.5594.2097.201.35
0.582.4879.2294.5497.350.89
0.783.2280.0994.5497.390.48
0.984.0581.1794.4597.270.05
183.8480.9894.2097.180.24
VehicleId (Large)077.7475.2087.0989.814.85
0.179.4876.4890.4495.001.96
0.380.0176.9991.2395.721.33
0.580.7777.8291.7395.960.74
0.781.4878.6492.1496.090.23
0.981.8279.0192.3396.070.01
181.2678.3292.1895.950.39
Table 2. Comparison of backbone networks Resnet 50, Resnet 101, and ResNext50 on VeRi dataset.
Table 2. Comparison of backbone networks Resnet 50, Resnet 101, and ResNext50 on VeRi dataset.
ModelmAP (R50)mAP (R101)mAP (RN50)R1 (R50)R1 (R101)R1 (ResNext 50)R5 (R50l)R5 (R101)R5 (ResNext 50)
PCB76.2476.3475.8295.0594.9994.1097.7997.4497.32
PCB-RR80.5080.0479.9795.7695.7196.0797.4396.9697.20
RPP77.0876.1576.4794.9994.4693.7497.9797.3897.38
RPP-RR81.4279.9680.6996.2494.9395.5997.5096.7896.78
Table 3. Comparison of backbone networks Resnet 50, Resnet 101, and ResNext50 on Vehicle Id dataset.
Table 3. Comparison of backbone networks Resnet 50, Resnet 101, and ResNext50 on Vehicle Id dataset.
Dataset VariantModelmAP (R50)mAP (R101)mAP (RNext 50)R1 (R50)R1 (R101)R1 (RNext 50)R5 (R50)R5 (R101)R5 (RNext 50)
SmallPCB86.5987.3886.5783.6184.7083.6397.6197.1496.99
PCB-RR87.1187.8787.20%84.2585.2584.3897.7397.2497.22
RPP83.6184.9185.9580.4381.9883.0495.3395.3196.11
RPP-RR84.1885.7486.6081.0582.9383.8995.7095.7396.34
MediumPCB83.8283.5682.1980.9780.7479.1294.1794.1993.74
PCB-RR84.0584.0382.3181.1781.2779.1794.4594.4594.08
RPP81.4581.6781.5778.6278.9078.5291.7091.8392.90
RPP-RR81.7682.3881.9678.9079.6778.9492.1392.2593.21
LargePCB81.2681.2480.8778.3278.2777.8692.1891.8692.13
PCB-RR81.8281.6181.2379.0178.7178.2592.3392.0992.29
RPP79.0979.6780.3976.2376.9677.4289.3989.2191.31
RPP-RR79.7280.3380.8576.9277.6977.9689.7489.6091.56
Table 4. Performance of ensembled model on VeRi dataset.
Table 4. Performance of ensembled model on VeRi dataset.
ModelmAP (R50)R1 (R50)R5 (R50l)R10 (R50)
Ensembled (PCB)79.8395.2998.0998.92
Ensembled (PCB)-RR83.1397.0297.6198.56
Ensembled—RPP80.3795.9598.0999.04
Ensembled–RPP-RR83.6396.6097.6198.45
Table 5. Performance of ensembled model on VehicleId dataset.
Table 5. Performance of ensembled model on VehicleId dataset.
VariantModelmAPR1R5R10
SmallEnsembled (PCB)88.2485.5298.2199.07
Ensembled (PCB)-RR88.5185.8098.1999.17
Ensembled—RPP87.3484.4997.4198.87
Ensembled–RPP-RR87.8185.0197.7299.02
MediumEnsembled (PCB)84.6381.8295.0897.77
Ensembled (PCB)-RR84.5981.7395.3297.82
Ensembled—RPP84.2481.4994.7197.12
Ensembled–RPP-RR84.5381.8095.0097.20
LargeEnsembled (PCB)82.3379.3293.6596.90
Ensembled (PCB)-RR82.8179.9293.7396.88
Ensembled—RPP82.1579.2792.8896.26
Ensembled–RPP-RR82.5979.8392.8696.35
Table 6. Performance of OPCB on VeRi dataset.
Table 6. Performance of OPCB on VeRi dataset.
ModelmAP (R50)R1 (R50)R5 (R50l)R10 (R50)
PCB (H)76.2495.0597.7998.98
PCB (H)-RR80.5095.7697.4398.56
RPP77.0894.9997.9798.75
RPP-RR81.4296.2497.5098.51
OPCB76.7495.2397.9198.98
OPCB-RR80.7095.5996.9698.27
Table 7. Performance of OPCB on VehicleId dataset.
Table 7. Performance of OPCB on VehicleId dataset.
VariantModelmAP (R50)R1 (R50)R5 (R50)R10 (R50)
SmallPCB86.5983.6197.6198.75
PCB-RR87.1184.2597.7398.82
RPP83.6180.4395.3398.33
RPP-RR84.1881.0595.7097.61
OPCB87.4984.7197.5598.87
OPCB-RR87.9485.2097.8798.98
MediumPCB83.8280.9794.1797.15
PCB-RR84.0581.1794.4597.27
RPP81.4578.6291.7095.18
RPP-RR81.7678.9092.1395.39
OPCB84.0381.294.4297.4
OPCB-RR84.3581.4994.8297.52
LargePCB81.2678.3292.1895.95
PCB-RR81.8279.0192.3396.07
RPP79.0976.2389.3993.48
RPP-RR79.7276.9289.7493.68
OPCB81.8178.992.596.17
OPCB-RR82.5979.8492.8496.23
Table 8. Comparison with state of the art on VeRi dataset. NR—Not Reported.
Table 8. Comparison with state of the art on VeRi dataset. NR—Not Reported.
Method and ReferencemAP (%)R-1 (%)R-5 (%)
Batch sample [39]67.5590.2396.42
SLSR [40]65.1391.24NR
MRL + Softmax Loss [41]78.5094.3098.70
VANet [42]66.3489.7895.99
MRM [14]68.5591.7795.82
Part-regularized near duplicate [12]74.3094.3098.70
PGAN [43]79.3096.5098.30
SAN [14]72.593.397.1
TCPM [23]74.5993.9897.13
Swin Transformer [44]78.697.3NR
URRNet [45]72.293.197.1
TransReid [46]78.296.5NR
SSL + re-ranking [47]69.9089.6995.41
CFSA [48]79.8994.9998.81
Attribute Net, [49]80.197.198.6
URRNet [45]72.293.197.1
MIMA Net [50]79.8994.9998.81
OPCB + RR (Ours)80.7795.5996.96
PCB + RPP + RR (Ensembled) (ours)83.6397.0298.45
Table 9. Comparison with state of the art on VehicleId dataset. NR—Not Reported.
Table 9. Comparison with state of the art on VehicleId dataset. NR—Not Reported.
MethodRank-1Rank-5
Small (%)Medium (%)Large (%)Small (%)Medium (%)Large (%)
MRM [14]76.6474.2070.8692.3488.5484.82
SLSR [40]75.1071.8068.7089.7086.1083.10
Part regularized near duplicate [12]78.4075.0074.2092.3088.3086.40
Batch sample [39]78.8073,4169.3396.1792.5789.45
SAN [14]79.778.475.694.391.388.3
MRL + Softmax Loss [41]84.8080.9078.4096.9094.1092.10
VANet [42]88.1283.1780.3597.2995.1492.97
PGAN [40]NRNR77.8NRNR92.1
LABNet [51]84.0280.1877.2NRNRNR
TransReid [46]83.6NRNR97.1NRNR
HCI-Net [52]83.879.476.496.592.791.2
VehicleGAN [53]83.578.275.796.593.290.6
Attribute Net [49]86.081.979.697.495.192.7
MIMANet [50]83.2880.1477.7296.3193.7191.29
URRNet [45]76.5,73.768.296.592.089.6
MDFENet [54]83.6680.7877.88NRNRNR
OPCB + RR (Ours)85.2081.4979.8497.8794.8292.84
PCB + RR (Ensemble) (ours)85.2581.7279.9297.2495.3293.73
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nath, R.K.; Mitra, D. An Analysis of Partitioned Convolutional Model for Vehicle Re-Identification. Electronics 2025, 14, 3634. https://doi.org/10.3390/electronics14183634

AMA Style

Nath RK, Mitra D. An Analysis of Partitioned Convolutional Model for Vehicle Re-Identification. Electronics. 2025; 14(18):3634. https://doi.org/10.3390/electronics14183634

Chicago/Turabian Style

Nath, Rajsekhar Kumar, and Debjani Mitra. 2025. "An Analysis of Partitioned Convolutional Model for Vehicle Re-Identification" Electronics 14, no. 18: 3634. https://doi.org/10.3390/electronics14183634

APA Style

Nath, R. K., & Mitra, D. (2025). An Analysis of Partitioned Convolutional Model for Vehicle Re-Identification. Electronics, 14(18), 3634. https://doi.org/10.3390/electronics14183634

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop