Previous Article in Journal
Supporting ASD Diagnosis with EEG, ML and Swarm Intelligence: Early Detection of Autism Spectrum Disorder Based on Electroencephalography Analysis by Machine Learning and Swarm Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion

by
Trinh Quoc Nguyen
1,2,*,
Oky Dicky Ardiansyah Prima
1,
Syahid Al Irfan
1,2,
Hindriyanto Dwi Purnomo
3 and
Radius Tanone
3
1
Graduate School of Software and Information Science, Iwate Prefectural University, Takizawa-shi 020-0693, Iwate, Japan
2
CyberCore Co., Ltd., Morioka-shi 020-0045, Iwate, Japan
3
Department of Information Technology, Satya Wacana Christian University, Salatiga 50711, Indonesia
*
Author to whom correspondence should be addressed.
AI Sens. 2025, 1(1), 4; https://doi.org/10.3390/aisens1010004
Submission received: 8 May 2025 / Revised: 4 June 2025 / Accepted: 30 June 2025 / Published: 4 July 2025

Abstract

This study presents CORE-ReID V2, an enhanced framework built upon CORE-ReID V1. The new framework extends its predecessor by addressing unsupervised domain adaptation (UDA) challenges in person ReID and vehicle ReID, with further applicability to object ReID. During pre-training, CycleGAN is employed to synthesize diverse data, bridging image characteristic gaps across different domains. In the fine-tuning, an advanced ensemble fusion mechanism, consisting of the Efficient Channel Attention Block (ECAB) and the Simplified Efficient Channel Attention Block (SECAB), enhances both local and global feature representations while reducing ambiguity in pseudo-labels for target samples. Experimental results on widely used UDA person ReID and vehicle ReID datasets demonstrate that the proposed framework outperforms state-of-the-art methods, achieving top performance in mean average precision (mAP) and Rank-k Accuracy (Top-1, Top-5, Top-10). Moreover, the framework supports lightweight backbones such as ResNet18 and ResNet34, ensuring both scalability and efficiency. Our work not only pushes the boundaries of UDA-based object ReID but also provides a solid foundation for further research and advancements in this domain.

1. Introduction

Object re-identification (ReID) focuses on retrieving specific object instances across diverse viewpoints [1,2,3,4,5] and has gained significant attention within the computer vision community due to its broad range of practical applications. Substantial advancements have been made in both supervised [6,7,8,9] and unsupervised ReID tasks [5,10,11], with most approaches employing backbone models originally designed for generic image classification tasks [12,13].
Unsupervised domain adaptation (UDA) for object ReID aims to transfer knowledge learned from a labeled source domain to accurately measure inter-instance affinities in an unlabeled target domain. Typical ReID tasks, such as person ReID and vehicle ReID, involve source and target domain datasets that do not share identical class identities. State-of-the-art UDA methods [11,14,15,16,17,18] generally adopt a two-stage training paradigm: (1) supervised pre-training on the source domain and (2) unsupervised fine-tuning on the target domain. During the second stage, pseudo-labeling strategies have demonstrated effectiveness in recent works [11,16,17]. These strategies iteratively alternate between generating pseudo-class labels through clustering target domain instances and refining the network by training on these pseudo-classes. This iterative process helps the pre-trained source-domain model gradually capture inter-sample relationships within the target domain, even in the presence of label noise. Compared to fully supervised methods, which rely on large amounts of labeled target data for optimal performance, UDA methods are more scalable and cost-effective but may suffer from reduced label reliability and slower convergence. The key advantage of supervised approaches lies in their access to precise annotations, while the strength of unsupervised methods is their ability to generalize to new domains without manual labeling, making them particularly valuable in real-world ReID deployments.
The integration of global features, which encapsulate coarse semantic information, and local features, which provide fine-grained details, has proven effective in enhancing algorithm performance. CORE-ReID V1 [11] introduced the ensemble fusion framework, which combines global and local features with the Efficient Channel Attention Block (ECAB). ECAB leverages inter-channel relationships to guide the model’s attention toward salient structures within the input image. Although CORE-ReID V1 achieves competitive results in person ReID under the UDA setting, it has three main limitations. Firstly, ECAB is exclusive to local features, leaving global features unenhanced. Secondly, CORE-ReID V1 only supports deep and complex backbone networks, such as ResNet50, ResNet101, and ResNet152, while neglecting shallower architectures such as ResNet18 and ResNet34, which offer computational efficiency and are well-suited for resource-constrained environments. Finally, CORE-ReID V1 is limited to the person ReID task, restricting its applicability to other ReID scenarios.
In this paper, we present CORE-ReID V2, an enhanced version of CORE-ReID V1 that addresses its limitations and introduces several novel contributions. CORE-ReID V2 not only achieves superior performance in UDA person ReID but also extends its applicability to object ReID tasks and supports lightweight backbone networks, such as ResNet18 and ResNet34, making it suitable for real-time systems and mobile devices. Building on the principles of LF2 [17] and CORE-ReID V1, we design a mean-teacher-based framework that iteratively learns multi-view features and refines noisy pseudo-labels through multiple clustering steps. We introduce the ensemble fusion++ module in CORE-ReID V2, which adaptively enhances both local and global features. This module applies ECAB to local features and the Simplified Efficient Channel Attention Block (SECAB) to global features, resulting in a fused representation that provides a more comprehensive feature set. Furthermore, we improve clustering outcomes by incorporating the KMeans++ [19] initialization strategy, which balances randomness and centroid selection to enhance cluster quality. To validate the framework, we pre-train the model on a source domain that integrates camera-aware style-transferred data for person ReID and domain-aware style-transferred data for vehicle ReID. Additionally, we adopt a teacher-student architecture for iterative domain adaptation, which has been used in prior works, such as LF2 [17], Deep Mutual Learning (DML) [20], MMT [15], MEB-Net [21] and mean teacher [22]. In our framework, the teacher network captures global features, while the student network refines diverse local features, both contributing to a more effective pseudo-labeling process. To summarize, the key contributions of CORE-ReID V2 are as follows:
  • Advanced data augmentation techniques: the framework integrates novel data augmentation strategies, such as Local Grayscale Patch Replacement and Random Image-to-Grayscale Conversion for the UDA task. These methods introduce diversity in the training data, enhancing the model’s stability.
  • Dynamic and flexible backbone support: CORE-ReID V2 extends compatibility to smaller backbone architectures, including ResNet18 and ResNet34, without compromising performance. This flexibility allows for deployment in resource-constrained environments while maintaining high accuracy.
  • Expansion to vehicle and further object ReID: unlike its predecessor, which focused solely on person re-identification, CORE-ReID V2 extends its scope to vehicle re-identification and further general object re-identification. This expansion demonstrates its versatility and adaptability across various domains.
  • Introduction of ensemble fusion++: the framework incorporates the SECAB into the global feature extraction pipeline to enhance feature representation by dynamically emphasizing informative channels, thereby improving discrimination between instances.

2. Related Work

Extensive research has been conducted on UDA for object ReID [23,24,25,26,27] and knowledge transfer techniques, such as knowledge distillation, which enable well-trained models to transfer expertise and improve learning in complex domain scenarios [28,29,30,31]. Methods generally fall into two categories: domain translation [32,33,34,35,36,37,38,39,40,41], which aligns visual styles between domains, and pseudo-labeling [11,14,15,16,17,42,43,44,45], which iteratively cluster target samples to generate pseudo-labels. While pseudo-labeling has demonstrated superior performance, both methods face challenges related to domain shifts and noisy labels. Moreover, the fusion of global and local features has proven effective in various tasks, including classification [46,47,48,49,50,51], object detection [52,53,54,55,56,57], and semantic segmentation [58,59,60,61,62,63], by integrating contextual information with fine-grained details. Building on these insights, our work refines the CORE-ReID V1 [11] fusion module to achieve a better balance between global and local features, leading to improved performance across multiple ReID tasks, including person and vehicle ReID.

2.1. UDA for Object ReID

UDA has gained significant attention for its ability to reduce reliance on costly manual annotations. By utilizing labeled data from a source domain, UDA enhances model performance in a target domain without requiring target-specific annotations. Research in object ReID has primarily concentrated on person ReID and vehicle ReID [64]. Existing UDA approaches for ReID can be broadly grouped into two categories: domain translation-based methods and pseudo-label-based methods [65,66].
Domain translation-based methods: These methods align the visual style of labeled source-domain images with that of the target domain. The translated images, along with their original ground-truth labels, are then used for training [37].
Several methods attempt to map source and target distributions to mitigate domain shifts [32,33,34,35,36]. Saenko et al. [32] introduced a domain adaptation technique based on cross-domain transformations by learning a regularized non-linear transformation that brings source-domain points closer to the target domain. In [33], the Geodesic Flow Kernel (GFK) was proposed to address domain shifts by integrating an infinite number of subspaces that capture geometric and statistical changes between the source and target domains. Similarly, Fernando et al. [34] developed a mapping function to align the source subspace with the target subspace for improved adaptation. Correlation Alignment (CORAL) [35] addressed domain shifts by computing the covariance statistics of each domain and applying a whitening and re-coloring linear transformation to align the source feature with the target domain. The Disentanglement Then Reconstruction (DTR) framework [36] enhanced alignment by disentangling the distributions and reconstructing them to ensure consistency across domains.
Another line of research [38,39,40,41] adopts adversarial approaches to learn transformations in the pixel space between domains. Methods such as PixelDA [38], PTGAN [39], and SBSGAN [40] enforce pixel-level constraints to preserve color consistency during domain translation. CoGAN [41] extends this concept by learning joint distributions, such as the joint distribution of color and depth images or face images with varying attributes.
Other methods focus on discovering a domain-invariant feature space to bridge domain gaps [66,67,68,69,70,71,72]. SPGAN [66] and CGAN-TM [69] improve feature-level similarity between translated and original images. The Deep Adaption Network (DAN) [71] employes the Maximum Mean Discrepancy (MMD) [73,74,75] to align feature distributions across domains. Similarly, Ganin et al. [67] and Ajakan et al. [76] introduced a domain confusion loss to encourage the learning of domain-invariant features. Hoffman et al. [70] proposed the Intermediate Domain Module (IDM) to generate intermediate domain representations dynamically by mixing the hidden features of the source and target domains through two domain features. CyCADA [72] combines both pixel-level and feature-level adaptation to improve domain adaptation.
Pseudo-label-based methods: The second category, pseudo-labeling methods [11,14,15,16,17,42,43,44,45], models the relationships between unlabeled target domain data and generated pseudo-labels. Fan et al. [42] proposed the progressive unsupervised learning (PUL) method that alternates between assigning labels to unlabeled samples and optimizing the network using the generated targets. This iterative refinement aligns the model’s representations more closely with the target domain, enhancing adaptation over time. Lin et al. [43] developed a bottom-up clustering framework enhanced by a repelled loss mechanism, which aims to increase the discriminative power of learned features while mitigating intra-cluster variations. Similarly, UDAP [14] proposed a self-training scheme that minimizes loss functions iteratively using clustering-based pseudo-labels. SSG [16], LF2 [17], and CORE-ReID V1 [11] further contributed to this category, which introduces techniques to assign pseudo-labels to both global and local features. Ge et al. [15] proposed Mutual Mean Teaching (MMT), which combines offline hard pseudo-labels and online soft pseudo-labels in an alternating training process, enhancing the model’s ability to adapt to domain shifts. This technique improves the model’s capacity to handle domain shifts by iteratively refining both the pseudo-labels and feature representations throughout training. SpCL [44] advanced this field by using a hybrid memory module that stored centroids of labeled source-domain images alongside un-clustered target instances and target domain clusters. This hybrid memory provides additional supervision to the feature extractor, while minimizing a unified contrastive loss over the three types of stored information. Additionally, Zheng et al. [45] developed the Uncertainty-Guided Noise-Resilient Network (UNRN), which evaluates the reliability of predicted pseudo-labels for target domain samples. By incorporating uncertainty estimates into the training process, UNRN improves performance with noisy annotations, thereby enhancing performance in domain adaptation scenarios.
Pseudo-labeling methods analyze data at different levels of detail, allowing them to capture small differences within the target domain. While these methods have demonstrated strong performance in many recent UDA ReID studies [11,15,44,45], their effectiveness can vary depending on factors such as dataset bias, domain shift severity, clustering quality, and label noise. Domain translation methods, on the other hand, remain valuable for reducing style mismatches and have shown advantages in specific scenarios, especially when high-quality translated images can be generated [77,78,79].

2.2. Knowledge Transfer

Knowledge transfer is a broad concept that refers to using knowledge gained from one model, dataset, or task to improve performance in another [28,29,30,31]. Within this broad scope, knowledge distillation is a specific technique in which a “teacher” model guides a “student” model by transferring soft predictions, features, or intermediate representations [80]. Knowledge distillation techniques help student networks become more accurate and generalize better, as the teacher model’s output implicitly contains rich information about the relationships between training samples and their underlying distribution [81]. For example, Laine and Aila [82] introduced the mean teacher model, which averages model weights across multiple training iterations to guide supervision for unlabeled data. In contrast, Deep Mutual Learning (DML) [20], proposed by Zhang et al., shifts from the traditional teacher-student framework by employing a group of student models that train collaboratively, providing mutual supervision and facilitating the exploration of diverse feature representations. Ge et al. introduced MMT [15], which adopts an alternative training method that uses both offline-refined hard pseudo-labels and online-refined soft pseudo-labels. MEB-Net [21] further builds on this by using three networks (six models in total) to conduct mutual mean teacher training and generate pseudo-labels.
In the context of our method, we adopt a teacher-student architecture aligned with the mean teacher framework [22], where the teacher network is updated through Exponential Moving Average (EMA) of the student’s weights. This form of distillation encourages consistency and stability in the learned representations, particularly useful in UDA, where target domain labels are absent. For clarity, in this work, a domain refers to a specific dataset distribution (e.g., Market-1501 [83] or VehicleID [84]), typically captured under different environmental or camera conditions, while a task refers to the objective of re-identifying objects (e.g., people or vehicles) across domains with different identities and styles. Although the task remains constant (object ReID), our goal is to transfer the knowledge learned from a labeled source domain to an unlabeled target domain under domain shift conditions.

2.3. Feature Fusion

The feature fusion of global and local features has proven highly effective across various computer vision tasks, including classification [46,47,48,49,50,51], object detection [52,53,54,55,56,57], semantic segmentation [58,59,60,61,62,63], and more [85].
In image classification, global features capture the overall structure and appearance, while local features focus on fine-grained details. Combining both types provides complementary information, enhancing the model’s ability to generalize across variations such as pose, lighting, and occlusions. For example, an early approach to feature fusion based on Canonical Correlation Analysis (CCA) was proposed by Sun et al. [46], who applied CCA to extract correlation features between two groups of feature vectors for improved pattern recognition performance. This approach effectively captured discriminative information while reducing feature redundancy, demonstrating significant improvements in recognition rates on datasets such as CENPARMI and Yale Face Database compared to single-feature and traditional fusion methods. Later, Sudha and Ramakrishna [47] studied iris feature fusion with pixel-level methods such as DU-Fusion and showed that combining features from techniques such as 2D-FFT, LBP, and PCA enhances recognition performance. The results on the CASIA dataset confirmed that DU-Fusion outperformed other methods in both verification and identification accuracy. Tian et al. [48] proposed a vehicle model recognition system using an iterative discrimination CNN based on selective multi-convolutional region feature extraction. Their SMCR model combines global and local features to boost classification accuracy. Similarly, Lu et al. [49] introduced a script identification framework that leverages both global CNNs, trained on segmented images, and local CNNs, trained on image patches. He et al. [50] presented a traffic sign recognition approach that integrates global and local features using histograms of oriented gradients (HOG), color histograms, and edge features. Suh et al. [51] employed fusion layers to concatenate global and local features for shipping label image classification, improving image quality verification. These studies demonstrated that fusing global and local features consistently improves classification performance compared to models relying on whole-image analysis alone.
In object detection, global features provide spatial awareness of objects within a scene, while local features capture subtle patterns, such as textures and edges, which are essential for accurate detection under occlusions. Li et al. [52] proposed Feature Fusion Single Shot Multibox Detector (FSSD), an enhanced version of SSD that incorporates a lightweight feature fusion module to better utilize multi-scale features. By concatenating features from different layers and applying down-sampling blocks, FSSD significantly improves detection accuracy with minimal speed loss, outperforming SSD and several state-of-the-art detectors on VOC and COCO benchmarks. Cong et al. [53] proposed an end-to-end co-salient object detection network that uses collaborative learning to enhance inter-image relationships. Their model includes a global correspondence module to extract interactive information across images and a local correspondence module to capture pairwise relationships. Later, Li et al. [54] developed an anchor-free object detector that uses a global-local feature extraction transformer (GLFT) to capture semantic information from both micro- and macro-level perspectives.
In semantic segmentation, the fusion of global and local features improves pixel-level predictions by combining overall scene context with localized information, especially in complex environments. For example, Zhang et al. [58] proposed ExFuse to address the semantic and resolution gap in fusing low-level and high-level features for semantic segmentation. By enriching low-level features with semantic context and high-level features with spatial detail, ExFuse significantly improved fusion effectiveness. Dai et al. [59] introduced Attentional Feature Fusion, a general framework that used multiscale channel attention to improve the fusion of features with inconsistent semantics and scales. By incorporating iterative attention mechanisms, their method addressed bottlenecks in conventional fusion strategies and achieved great performance on CIFAR-100 and ImageNet with fewer parameters, highlighting the effectiveness of attention-based fusion in deep networks. Yang et al. [60] introduced AFNet, which uses a multi-path encoder to extract diverse features, a multi-path attention fusion module, and a fine-grained attention fusion module to combine high-level abstract and low-level spatial features. Tian et al. [61] extended this concept with two encoders to extract both global high-order interactive features and local low-order features. These encoders form the backbone of the global and local feature fusion network (GLFFNet), enabling effective segmentation of remote sensing images through a dual-encoder structure. Later, Zhou et al. [62] proposed a local–global multi-scale fusion network (LGMFNet) for building segmentation in SAR images. LGMFNet includes a dual encoder–decoder structure, with a transformer-based auxiliary encoder complementing the CNN-based primary encoder. The global–local semantic aggregation module (GLSM) is also introduced to bridge the two encoders, enabling semantic guidance across multiple scales through a specialized fusion decoder.
Inspired by these advances, feature fusion techniques have gained traction in domain adaptation for object Re-identification. Self-Similarity Grouping (SSG) [16] is the first approach to applied both global and local features for unsupervised domain adaptation (UDA) in Person ReID. However, SSG faces two challenges: first, using a single network for feature extraction often introduces noisy pseudo-labels; and second, it performs clustering independently on global and local features, potentially assigning multiple inconsistent pseudo-labels to the same sample. To address these limitations, LF2 [17] was proposed to fuse global and local features into a unified representation, reducing noise and improving clustering consistency. Building on this idea, CORE-ReID V1 [11] introduced an ensemble fusion module equipped with the ECAB, which effectively fuses global and local features.

3. Materials and Methods

Despite advancements in domain translation-based methods, these often suffer from a persistent domain gap between translated images and real target domain images, which can adversely impact performance. To address this issue, our approach employs a pseudo-labeling strategy, which enables data analysis at multiple levels of granularity. This method has demonstrated superior performance compared to domain translation-based techniques [11,15,44,45].
While existing pseudo-labeling frameworks such as Deep Mutual Learning (DML) [20], MMT [15], and MEB-Net [21] have proven effective, they suffer from limitations due to their heavy reliance on pseudo-labels generated by the teacher model. These pseudo-labels can be noisy or inaccurate, adversely affecting model training. To mitigate this issue, we utilize a teacher-student network paradigm, where the student network is trained on labeled source-domain data, and the teacher network is iteratively refined using the mean teacher method. Furthermore, we incorporate the ensemble fusion++ module, which enhances feature extraction by adaptively refining both local and global representations, thereby it is expected to produce more stable and reliable pseudo-labels than existing approaches.
In CORE-ReID V1 [11], the Efficient Channel Attention Block (ECAB) was primarily applied to local features, restricting the full potential of the ensemble fusion module. In CORE-ReID V2, we extend and enhance this module to ensure that both global and local features undergo comprehensive optimization. This enhancement results in a more balanced and discriminative feature representation, improving generalization across diverse ReID tasks. Additionally, the improved ensemble fusion++ is not only effective for person ReID but also demonstrates strong domain adaptation capabilities in vehicle ReID, further validating its versatility in object ReID.
This chapter outlines the methodology and materials used in CORE-ReID V2 for unsupervised domain adaptation (UDA) in object ReID. The proposed framework consists of two main stages: (1) pre-training on a labeled source domain and (2) fine-tuning on an unlabeled target domain.

3.1. Overview

3.1.1. CORE-ReID V1 and CORE-ReID V2

CORE-ReID V1: a baseline for unsupervised domain adaptation in person re-identification: CORE-ReID V1 was introduced as a framework to address unsupervised domain adaptation (UDA) in person re-identification (ReID). It effectively tackled domain shifts between camera views by leveraging camera-aware style transfer for synthetic data generation, Random Grayscale Patch Replacement for data augmentation, and K-means clustering for pseudo-labeling. Additionally, the ensemble fusion module with Efficient Channel Attention Block (ECAB) played a crucial role in integrating local and global features, improving the model’s performance in cross-domain scenarios.
Despite its success, CORE-ReID V1 had several limitations:
  • Limited application domain: the framework was specifically designed for Person ReID, restricting its applicability to other ReID tasks such as vehicle ReID and object ReID.
  • Synthetic-data generation challenge: the camera-aware style transfer method relied on predefined camera information, making it ineffective when the number of cameras was unspecified.
  • Inefficient data augmentation: The Random Grayscale Patch Replacement technique only operated locally, limiting its effectiveness in learning color-invariant features.
  • Clustering limitations: the K-means clustering used random centroid initialization, leading to poor centroid placement, slow convergence, high variance in clustering results, and imbalanced cluster sizes.
  • Feature fusion issue: the ECAB module enhanced only local features, neglecting improvements to global representations.
  • Restricted backbone support: the framework exclusively supported deep networks such as ResNet50, ResNet101, and ResNet152, making it computationally expensive and unsuitable for lightweight applications.
CORE-ReID V2: expanding scope, enhancing performance: to overcome these limitations, CORE-ReID V2 is proposed as a enhancement over CORE-ReID V1, expanding its capabilities to vehicle ReID and object ReID while introducing architectural and methodological improvements.
  • Expanded application scope: unlike CORE-ReID V1, which was restricted to person ReID, CORE-ReID V2 extends its applicability to vehicle ReID and object ReID, making it a versatile framework for various ReID tasks.
  • Advanced synthetic data generation: CORE-ReID V2 incorporates both camera-aware style transfer and domain-aware style transfer, allowing effective synthetic data generation even when the number of cameras is unknown.
  • Improved data augmentation: a new grayscale patch-replacement strategy considers both local grayscale transformation and global grayscale conversion, leading to better feature generalization across domains.
  • Enhanced clustering with greedy K-means++: instead of relying on random initialization, CORE-ReID V2 employs greedy K-means++, which selects optimized centroids to improve cluster spread; minimizes redundancy, requiring fewer iterations; enhances stability and consistency, reducing randomness; ensures better centroid distribution, leading to improved clustering performance.
  • Ensemble fusion++ for Comprehensive Feature Enhancement: CORE-ReID V2 introduces ensemble fusion++, which integrates both ECAB and SECAB, ensuring that global features are enhanced alongside local features, leading to a more balanced and comprehensive feature representation.
  • Flexible backbone support: CORE-ReID V2 broadens its applicability by supporting lightweight networks such as ResNet18 and ResNet34, alongside ResNet50, ResNet101, and ResNet152. This allows deployment in computationally constrained environments, such as real-time and edge-based applications.
CORE-ReID V2 represents a substantial advancement over CORE-ReID V1 by expanding its scope beyond Person ReID, improving clustering stability, introducing adaptive feature enhancement mechanisms, and supporting lightweight architectures. Table 1 shows the summary of these improvements.

3.1.2. Problem Definition and Methodology

Problem definition: We represent the customized labeled source-domain data as D S = x S , i ,   y S , i i = 1 N S } , where x S ,   i and y S , i denote the i t h source image and its corresponding ground truth identity label, respectively, and N S is the total number of source images. Similarly, the unlabeled target domain data is denoted as D T = x T , i i = 1 N T } , where x T , i indicates the i t h target image, and N T is the number of target images. Identity labels are unavailable for the images in the target domain dataset, and it is important to note that the identities across the source and target domains do not overlap. The objective of unsupervised domain adaptation (UDA) for object ReID is to transfer knowledge from the source domain S to the target domain T . To accomplish this, we propose the CORE-ReID V2 framework, designed to achieve effective knowledge transfer through a pseudo-label-based method.
Methodology: we adopt a pseudo-label-based approach by dividing the process into two stages: pre-training the model on the source domain using a fully supervised strategy, followed by fine-tuning it on the target domain through an unsupervised learning approach (Figure 1).
Depending on the specific task (person ReID or vehicle ReID), the appropriate dataset is utilized. Our method uses a pair of teacher–student networks. After training the model on a customized labeled source-domain dataset, the parameters of the pre-trained model are copied to both the student and teacher networks as an initialization step for the fine-tuning stage. During fine-tuning, we first train the student model and then optimize the teacher model using the mean teacher method [22]. This is because averaging model weights across multiple training steps generally yields a more accurate model than relying solely on the final weights [86]. Following the mean teacher method, the teacher model uses Exponential Moving Average (EMA) weight parameters of the student model instead of directly sharing weights. This approach is expected to allow the teacher network to aggregate information after every step, rather than every epoch, improving consistency [22]. To minimize computational costs, only the teacher model is used during inference.

3.2. Source-Domain Pre-Training

3.2.1. Image-to-Image Translation

Building on CORE-ReID V1, we employ CycleGAN to generate additional training samples by treating the stylistic variations across different cameras as distinct domains for person ReID task. This involves training image-to-image translation models using CycleGAN for images captured from various camera views within the dataset. Our goal is to train on a source domain S and evaluate the algorithm during the fine-tuning phase on a different target domain T . By incorporating test data into the training set, similar to DGNet++ [87], we can fully leverage the available data in S . In the person ReID task, for a source-domain dataset containing images from C different cameras, we utilize C ( C 1 ) generative models to produce data in both X Y and Y X directions. The final training set is a combination of the original real images and the style-transferred images from both the training and test sets within the source-domain dataset. These style-transferred images retain the labels of their corresponding real images.
In the case of vehicle ReID, due to the simpler nature of vehicle features and the large number of cameras used (some datasets do not provide the number of cameras used), we adopt domain-aware transfer models instead of camera-aware models. As a result, only a single transfer model is needed to generate style-transferred images from the source domain to the target domain (Figure 2).
Figure 3 illustrates two representative examples from both the training and test sets in the Market-1501 and CUHK03 datasets, where image styles have been modified according to camera views. This adjustment showcases our approach to data augmentation, where images are transformed to mimic the visual characteristics associated with each camera’s unique viewpoint and color distribution. By aligning image styles with camera perspectives, this method effectively reduces the inconsistencies in appearance caused by differences in lighting, angles, and color shifts across camera views. This approach helps the model generalize better, thus mitigating overfitting in convolutional neural networks (CNNs). Furthermore, incorporating camera-specific style information allows the model to learn more robust pedestrian features that are less sensitive to variations across different camera setups, leading to enhanced performance in ReID tasks.
Given the simpler nature of vehicle features and the large number of cameras involved (with some datasets not specifying the number of cameras), we aim to utilize labels from the source domain along with target-domain-style-transferred images to create a shared-features domain dataset. This approach involves developing domain-aware style transfer models to bridge the feature gap between the two domains by transforming source-domain images into target-domain-style outputs. Figure 4 shows two examples of input images from the VeRi-776 [1] and VehicleID [84] datasets, with the pixel differences between the input and output images highlighted in pink.

3.2.2. Fully Supervised Pre-Training

Similar to many existing UDA approaches [17] that rely on a model pre-trained on a source dataset, we employ a ResNet-based model, pre-trained on ImageNet as the backbone network. In this setup, the original final fully connected (FC) layer is removed and replaced with two new layers. The first one is a batch normalization layer with either 2048 or 512 features, depending on the specific ResNet architecture. The second layer is an FC layer with M S dimensions, where M S   represents the number of identities (classes) in the source dataset S (Figure 5).
In our training process, we define the number of identities for the full training set in the source domain as
M S , t r a i n = M S , t r a i n o r i g i n a l + M S , t e s t o r i g i n a l ,
where M S , t r a i n o r i g i n a l and M S , t e s t o r i g i n a l represent the number of identities in the original training and test sets of S , respectively. For each labeled image x S , i and its ground truth identity y S , i in the source-domain data D S = x S , i ,   y S , i i = 1 N S } , with N S representing the total number of images, we train the model using both identity classification (cross-entropy) loss L S , I D and triplet loss L S , t r i p l e t . The identity classification loss is applied to the final fully connected (FC) layer, handling the task as a classification problem, while the triplet loss, applied after batch normalization, is used for feature verification. The loss functions are defined as follows:
L S , I D = 1 N s i = 1 N S L c e ( C S f x S , i ,   y S , i ,    
L S , t r i p l e t = 1 N s i = 1 N S max 0 , | | f x s , i f x S , i + | | 2 | | f x s , i f x S , i | | 2 + m ,
where f x S . i is the feature embedding vector of the source image x S , i   extracted from the network with the backbone in Figure 5, L c e is the cross-entropy loss, C S is a learnable classifier in the source domain: f x S . i { 1 , 2 , ,   M S } . | | · | | 2 denotes the L 2 -norm distance, x S , i + and x S , i are the hardest positive and hardest negative feature indices in each mini-batch for the sample x S . i , and m represents the triplet distance margin. Using a balance parameter κ , the total loss for source-domain pre-training is
L S , t o t a l = L S , R e I D = L S , I D + κ L S , t r i p l e t .
The model is expected to achieve strong performance on fully labeled source-domain data, but its performance significantly drops when applied directly to the unlabeled target domain. Before feeding images into the network, we utilize the “Data Adapter” component (Figure 6) to preprocess them by resizing to a specific size depending on the type of object. We then apply several data augmentation techniques, including edge padding, random cropping, and random horizontal flipping. To address color deviation, we incorporate random color dropout through global and local grayscale transformations [88], preserving key information while minimizing overfitting and enhancing the model’s generalization. These approaches specifically balance the model’s weighting of color features and color-independent features, resulting in improved feature robustness in the neural network.
We employ the global grayscale transformation to a training batch with a set probability p g l o b a l , then feed it into the model for training. This process is defined as I = R G B T o G r a y s c a l e ( I ) , where R G B T o G r a y s c a l e ( ) represents the grayscale conversion function using the NTSC formula ( 0.299 × R e d + 0.587 × G r e e n + 0.114 × B l u e ), I denotes the input image, and I is the random grayscale image. This function operates by performing pixel-wise accumulation on the red, green, and blue channels of the original RGB image, resulting in a grayscale output. Importantly, the labels remain consistent between the converted grayscale image and the original. The procedure for local grayscale transformation is outlined in Algorithm 1.
Algorithm 1: Global Grayscale Transformation
Input: Input image I;
   Grayscale transformation probability pglobal
Output :   Randomly   grayscale   image   I .
Initialization :   p t R a n d   0,1 .
1: if   p t p g l o b a l   then
2: I   I .
3:else
4: I R G B T o G r a y s c a l e ( I ) .
5: return   I .
6:end
To enhance model adaptability to significant biases from localized color dropout, we apply a local grayscale transformation to each visible image I in the training batch using the following equation:
I p o s i t i o n = R G B T o G r a y s c a l e   I p o s i t i o n = R G B T o G r a y s c a l e ( R a n d P o s i t i o n I ) ,
where R a n d P o s i t i o n ( ) generates a random rectangular region within the image I . The transformed sample is represented by I . During model training, local grayscale transformation is applied randomly to images in each batch with a probability p l o c a l . This involves selecting a random rectangular region within the image and replacing it with the grayscale pixels of that same region. Consequently, images with mixed grayscale levels are generated, aiding the model in learning with color-variant features without altering object structure. The process includes several parameters: s m i n and s m a x define the minimum and maximum size ratios of the rectangle relative to the full image area; the rectangle’s area S t is computed by sampling from S t R a n d s m i n ,   s m a x × S , where S is the input image area; r t is a coefficient that sets the rectangle’s shape ratio within the interval r l o c a l ,   1 / r l o c a l ; the coordinates x t and y t for the rectangle’s top-left corner are generated randomly. If the rectangle exceeds image boundaries, new coordinates and dimensions are selected. This approach produces images with grayscale sections that vary in intensity without impacting the core structure, allowing the model to learn features invariant to color variations. The full procedure for local grayscale transformation is detailed in Algorithm 2.
Algorithm 2: Local Grayscale Transformation
Input :   Input   image   I ;
                          Grayscale   transformation   probability   p l o c a l ;
                          Area   ratio   range   ( low   to   high )   s m i n   and   s m a x ;
                          Aspect   ratio   r l o c a l .
Output :   Randomly   transformed   image   I .
Initialization :   p t R a n d ( 0,1 ) ;
                                                    W I . s i z e [ 0 ] ,   H I . s i z e [ 1 ] ;
                                                    S   W H .
1: if   p t p l o c a l    then
2:       I   I ;
return   I .
3:else
4:       while   T r u e   do
5:                 S t R a n d s m i n ,   s m a x × S ;
6:                 r t R a n d r l o c a l ,     1 / r l o c a l ;
7:                 W t s t / r t ,     H t S t × r t ;
8:                 x t R a n d 0 , W ,   y t R a n d 0 , H ;
9:                 if   x t + W t W   and   y t + H t H  then
10:                                 I p o s i t i o n ( x t ,   y t ,   x t + W t ,   y t + H t ) ;
11:                                 I p o s i t i o n R G B T o G r a y s c a l e   ( I p o s i t i o n ) ;
12:                                 I   I ;
13:                                 return   I .
14:        end
15:   end
16:end

3.2.3. Implementation Details

To perform camera-aware image-to-image translation for generating synthetic data, we train 30 generative models for the Market-1501 dataset and 2 models for the CUHK03 dataset. These numbers are derived from the formulas 6 × 6 1 = 30 and 2 × 2 1 = 2 , corresponding to the number of camera pairs in each dataset. For domain-aware image-to-image translation, we train 2 generative models (VeRi-776 to VehicleID and reverse). During training, all input images are first resized to 286 × 286 pixels, followed by cropping them to 256 × 256 pixels. We use the Adam optimizer for training all models from scratch, with a batch size of 8. The learning rate is initialized at 0.0002 for the Generator and 0.0001 for the Discriminator. For the first 30 epochs, these rates are kept constant and then linearly decayed to near zero over the subsequent 20 epochs according to a lambda learning rate schedule.
For pre-training, we adopt ResNet101 as the backbone (with support for other ResNet architectures as well). The initial learning rate is set to 0.00035, then reduced by a factor of 0.1 at the 40th and 70th epochs, totaling 350 training epochs with a 10-epoch warmup period. Each training batch consists of 32 identities, with 4 images per identity, resulting in a final batch size of 128. The balance parameter κ for computing the total loss is set to 1. Regarding preprocessing, each image is resized to 256 × 128 pixels for person ReID task and 256 × 256 pixels for vehicle ReID task. The resized images are padded with 10 pixels using edge padding, followed by random cropping back to their original resized dimensions. Additional augmentation techniques include random horizontal flipping, global grayscale transformation p g l o b a l , and local grayscale transformation p l o c a l , applied with probabilities of 0.5, 0.05, and 0.4, respectively. Images are then converted to 32 bit floating-point pixel values normalized to the [ 0,1 ] range. The RGB channels are further normalized by subtracting mean values of [ 0.485 , 0.456 , 0.406 ] and dividing by standard deviations of [ 0.229 , 0.224 , 0.225 ] .

3.3. Target-Domain Fine-Tuning

3.3.1. Overall Algorithm

In this phase, we use the pre-trained model to perform comprehensive optimization. We present our CORE-ReID V2 framework (Figure 7) along with the Efficient Channel Attention Block (ECAB) and the Simplified Efficient Channel Attention Block (SECAB) in ensemble fusion++.
Building upon the strategies utilized in SSG [16], LF2 [17], and CORE-ReID V1 [11], our objective is to enable the model to dynamically integrate both global and local features. This approach allows for feature representations that encompass comprehensive global and detailed local information. To further enhance these feature representations, we incorporate ECAB and SECAB modules during the fusion process. By organizing multiple clusters based on global and fused features, we aim to generate more reliable pseudo-labels, thereby reducing the risk of ambiguous learning.
To refine these pseudo-labels, we implement a teacher-student network pair grounded in the mean-teacher framework. We feed the same unlabeled image from the target domain into both the teacher and student networks. During iteration i , the student network’s parameters, ρ ς , are updated using mean teacher momentum, adjusting them through backpropagation within the target domain training. In parallel, the teacher network’s parameters, ρ τ , are derived as a moving average of the student network’s parameters ρ ς . This is controlled by the temporal momentum coefficient η , which is restricted to the range [ 0,1 ) . The update rule is defined as
ρ τ , i = η ρ τ , i 1 + 1 η ρ ς ,
We employ the K-means clustering algorithm to assign pseudo-labels to the data, using the Euclidean distance as the similarity metric. As a result, each sample x T , i is assigned three pseudo-labels (global, top, and bottom). The target domain dataset is defined as D T = x T , i , y ^ T , i , j i = 1 N T } , where j g l o b a l , t o p , b o t t o m , and N T represents the total number of images in the target dataset T . The pseudo-label y ^ T , i , j 1,2, , M T , j indicates that y ^ T , i , j is derived from the clustering results Y ^ j = y ^ T , i , j i = 1,2 , , N T } . These are obtained using the combined feature with its flipped counterpart x T , i   generated by BMFN, denoted as φ l , l { t o p , b o t t o m } . Here, M T , j stands for the number of distinct identities (classes) in the clustering outcome Y ^ j .
Before computing the loss function, we use BMFN to extract optimized features from the networks f j ς , f j τ , j g l o b a l , t o p , b o t t o m and φ l , l t o p , b o t t o m from the ensemble fusion++. Given an image x T , i in the target dataset, along with its flipped version   x T , i , we extract the feature maps F j m   and the flipped feature maps F j m for j g l o b a l ,   t o p   b o t t o m   and m ς , τ . The BMFN output is computed as follows:
f j m = B M F N F j m , F j m = F j m + F j m 2 | | F j m + F j m 2 | | 2 .
φ l m = B M F N θ l m , θ l m = θ l m + θ l m 2 | | θ l m + θ l m 2 | | 2 ,
After obtaining multiple pseudo-labels, we generate three new target domain datasets to train the student network. The pseudo-labels derived from the local fusion features, denoted as φ l , l { t o p , b o t t o m } , are used to calculate the softmax triplet loss for the corresponding local features f l ς from the student network:
L T , t r i p l e t l = 1 N T i = 1 N T log e | | f l ς x T , i ρ ς f l ς x T , i ρ ς   |       | 2 e | | f l ς x T , i ρ ς f l ς x T , i ρ ς   |       | 2 + e | | f l ς x T , i ρ ς f l ς x T , i + ρ ς   |       | 2 ,
where ρ τ and ρ ς are the parameters of the teacher and student networks, respectively. The optimized local feature from the student network is denoted as f l ς , with l t o p , b o t t o m . Here, x T , i + and x T , i represent the hardest positive and negative samples relative to the anchor image x T , i in the target domain.
In a similar fashion to supervised learning, we utilize the cluster results Y ^ g l o b a l of the globally clustered feature f g l o b a l ς as pseudo-labels to compute the classification loss L T , R e I D g l o b a l and the global triplet loss L T , t r i p l e t g l o b a l . These losses are defined as follows:
L T , I D g l o b a l = 1 N T i = 1 N T L c e ( C T f g l o b a l ς x T , i ,   y ^ T , i , g l o b a l ,
L T , t r i p l e t g l o b a l = 1 N T i = 1 N T max 0 ,   | | f g l o b a l ς x T , i f g l o b a l ς x T , i + | | 2 | | f g l o b a l ς x T , i f g l o b a l ς x T , i | | 2 + m ,
where C T represents the fully connected classification layer of the student network, mapping f g l o b a l ς ( x T , i ) to the set { 1 , 2 , ,   M T , g l o b a l } . The notation | | · | | 2 indicates the L 2 -norm distance.
The total loss is computed by combining the different losses with weighting parameters α ,   β ,   γ :
L T ,   t o t a l = L T ,   R e I D g l o b a l + γ L T , t r i p l e t t o p + δ L T , t r i p l e t b o t t o m = α L T ,   I D g l o b a l + β L T ,   t r i p l e t g l o b a l + γ L T , t r i p l e t t o p + δ L T , t r i p l e t b o t t o m .
During the inference phase, the ensemble fusion++ process is bypassed, using only the optimized teacher network to reduce computational overhead. Specifically, the global feature map from the teacher network is split into two segments, referred to as top and bottom features (which also acts similarly in the student network). These segments undergo global average pooling, after which the two local features and the global feature are concatenated. Finally, L 2 normalization and the BMFN method are applied to obtain the optimal feature representation for inference.

3.3.2. Ensemble Fusion++ Component

To extract the fusion features, we horizontally divide the final global feature map of the student network into two segments (top and bottom), resulting in ς t o p and ς b o t t o m   after applying global average pooling. Unlike the ensemble fusion component in CORE-ReID V1 [11], the final global feature map τ g l o b a l   from the teacher network is further enhanced using the proposed SECAB module. These features ς t o p and ς b o t t o m from the student network, along with τ g l o b a l from the teacher network are then utilized for adaptive feature fusion through the ensemble fusion++ module, which includes learnable parameters.
The inputs ς t o p   and ς b o t t o m   are processed by ECAB, while τ g l o b a l   is processed by SECAB for adaptive fusion. The enhanced attention maps ( ψ t o p and ψ b o t t o m ) generated by ECAB are combined with the output τ g l o b a l through element-wise multiplication, resulting in the ensemble fusion feature maps: τ g l o b a l t o p   and τ g l o b a l b o t t o m . These maps undergo global average pooling (GAP) and batch normalization, yielding the fusion features θ t o p   and θ b o t t o m . These features are then fed into the BMFN for predicting pseudo-labels using clustering algorithms in subsequent steps.
The process within ensemble fusion++ (Figure 8) can be summarized as follows:
τ g l o b a l t o p = ψ t o p τ g l o b a l = E C A B ς t o p τ g l o b a l S E C A B τ g l o b a l ,
τ g l o b a l b o t = ψ b o t t o m τ g l o b a l = E C A B ς b o t t o m τ g l o b a l S E C A B τ g l o b a l ,

3.3.3. SECAB

The importance of attention has been extensively explored in previous literature [89,90,91]. Attention not only guides where to focus but also enhances the representation of relevant features. Inspired by the ECAB [11], we introduce a new component named SECAB, a straightforward yet impactful attention module for feed-forward convolutional neural networks to enhance the global feature. ECAB is designed to refine channel-wise feature representations by using both max-pooling and average-pooling operations, followed by a Shared Multilayer Perceptron (SMP) with ReLU activations. This SMP enhances non-linearity, and models complex inter-channel dependencies more effectively. After generating attention maps from both pooled features, ECAB applies a sigmoid activation and multiplies the result with the sum of the pooled inputs, producing a strongly refined output. In contrast, the SECAB (Simplified ECAB) is a lightweight, GPU-friendly variant specifically tailored for global feature refinement. While it retains the same attention generation pathway (pooling → SMP → sigmoid), it omits the reweighting step and directly outputs the attention map. This reduces computational complexity while still maintaining meaningful channel attention on global features. Figure 9 shows the design of SECAB, while Table 2 describes the comparison between ECAB and SECAB.
Given an intermediate input feature map ς R C × W × H , where C , W , H denote the number of channels, width, and height, respectively. After performing max-pooling and average-pooling, then fit the outputs ς m a x , ς a v g into a Shared Multilayer Perceptron (SMP). We can obtain refined features such as ς S M L P m a x and ς S M L P a v g . The SMP has multiple hidden layers with a reduction rate r and the same expansion rate and uses the activation function ReLU. Sigmoid activation squashes the sum to the range between 0 and 1, producing a channel-wise attention mask. The enhanced attention map ς σ R C × 1 × 1 is calculated as follows:
ς σ = σ ( ς S M L P m a x + ς S M L P a v g ) ,

3.3.4. Greedy K-Means++

In the K-means clustering problem, we are given a set of points N T R d i m in a d i -dimensional space and a specified number of clusters M T , j . The objective is to identify a set of M T , j centroids C = { c 1 ,   c 2 ,   , c M T , j } R d i m that minimizes the total sum of squared distances from each point in N T to its nearest centroid. Specifically, if we define the cost of a point x with respect to a set of centers C as D x , C m i n c C x c 2 , the goal is to find C such that C = M T , j and the cost D N T ,   C x N T D ( x , C ) is minimized. In practice, a simple way to initialize centroids is to select a random subset of N T with the size M T , j . However, this random approach does not guarantee any approximation bounds and can perform poorly in certain cases, such as when M T , j well-separated clusters exist along a single line. Arthur and Vassilvitskii [19] propose a probabilistic seeding method known as K-means++, which improves centroid initialization by favoring points that are far from already-selected centroids, while still maintaining a degree of randomness. Empirical results demonstrate that K-means++ consistently outperforms random seeding on real-world datasets [19].
The greedy K-means++ algorithm [92] refines this process further by eliminating randomness and selecting centroids deterministically to explicitly maximize the spatial spread. The algorithm operates as follows. At each step, it samples l candidate c i + 1 1 , c i + 1 2 , , c i + 1 l from a distribution constructed based on the current centroid configuration. Then, for each candidate c i + 1 j , the algorithm calculates the new cost D ( X ,   C   { c i + 1 j } ) that would result from adding this candidate to the set of centroids. Next, the candidate center that minimizes this cost is selected as the next centroid. In our implementation, l is typically set to 2 + log M T , j . By systematically evaluating multiple candidates at each step, greedy K-means++ ensures better centroid initialization and improved clustering outcomes (Algorithm 3).
Algorithm 3: Greedy K-means++ seeding
Input:   The   number   of   images   in   target   dataset   N T ;
    The   number   of   the   clusters   M T , j ;
    The   number   of   candidate   centers l .
Output:   The   set   of   centers   C .
Initialization:   Uniformly   independently   sample   c 1 1 , c 1 2 , , c 1 l N T .
1: Let   c 1 = a r g m i n c c 1 1 , c 1 2 , , c 1 l D ( X , c ) and   set   C 1 = { c 1 } .
2: for   i 1,2 , , M T , j 1   do
3: Sample   c i + 1 1 , c i + 1 2 , , c i + 1 l N T independently;
4: Sample   x   with   probability   D ( x , C i )   D X , C i ;
5: Let   c i + 1 = a r g m i n c c i 1 , c i 2 , , c i l D ( X ,   C i c ) ;
6: Set   C i + 1 = C i { c i + 1 } .
7: return   C C M T , j

3.3.5. Detailed Implementation

The training process lasts 80 epochs, with each epoch consisting of 400 iterations. A fixed learning rate of 0.00035 for the person ReID task (0.00007 for the vehicle ReID task) is maintained throughout, and the Adam optimizer is employed with a weight decay of 0.0005 to ensure stable convergence. Clustering operations utilize the K-means algorithm with greedy K-means++ initialization, where the maximum number of iterations is capped at 100, striking a balance between computational efficiency and solution accuracy. The mini-batch size is set to 512, allowing for efficient centroid updates without processing the entire dataset. An early stopping criterion is applied, terminating clustering if no improvement in inertia is observed over 50 consecutive mini-batches. To address the issue of empty clusters, a reassignment ratio of 0.05 is used, ensuring toughness in dynamic data distributions. For centroid initialization, 1500 data points are used for global features, while 900 data points are allocated for both top and bottom local features.
In the temporal ensemble regularization process, we follow the common practice from the original mean teacher paper [22] and set the momentum parameter ( η ) to 0.999, which has been shown to work well in practice across various tasks. To balance the contributions of the various components in the loss function, we assign weights as follows: α = 1 ,   λ = 1 ,   γ = 0.5 , and δ = 0.5 . For the ensemble fusion++ module, a reduction ratio and expansion rate ( r ) of 4 are utilized, along with 5 hidden layers ( h ) for both ECAB and SECAB components.
The data adapter resizes the input images to 128 × 256 for the person ReID task and to 256 × 256 for the vehicle ReID task. Edge padding of 10 pixels is applied before randomly cropping the images to their respective dimensions ( 128 × 256 or 256 × 256 ). Data augmentation strategies include random horizontal flipping, global grayscale transformation, local grayscale transformation, and random erasing, applied with probabilities of 0.5, 0.05, 0.4, and 0.5, respectively. These steps ensure a robust and diverse training dataset to improve generalization performance.

4. Results

In this section, we present experimental results, comparing our method against state-of-the-art (SOTA) techniques on widely used datasets for the task of unsupervised domain adaptation (UDA) for object ReID.

4.1. Dataset Description

We evaluated the effectiveness of our proposal by conducting evaluations on three benchmark datasets: Market-1501 [83], CUHK03 [93], and MSMT17 [39] for person ReID; and two benchmark datasets: Veri-776 [1], VehicleID [84], and VERI-Wild [94] for vehicle ReID.
Market-1501 [83] contains 32,668 images of 1501 individuals captured from six different camera views. The training set includes 12,936 images representing 751 identities, while the testing set comprises 3368 query images and 19,732 gallery images, covering the remaining 750 identities.
CUHK03 [93] features 14,097 images of 1467 unique individuals, recorded by six campus cameras, with each identity captured by two cameras. The dataset provides two types of annotations: manually labeled bounding boxes and those generated by an automatic detector. For both training and testing, we utilized the manually annotated bounding boxes. Additionally, we followed a more rigorous testing protocol, proposed in [95], which splits the dataset into 767 identities (7365 images) for training and 700 identities for testing, with 5332 images in the gallery and 1400 images in the query set.
MSMT17 [39] is a large-scale dataset comprising 126,441 bounding boxes of 4101 identities, recorded by 12 outdoor and three indoor cameras (15 cameras total) during three periods of the day (morning, afternoon, and noon) over 4 different days. The training set includes 32,621 images featuring 1041 identities, while the testing set contains 93,820 images representing 3060 identities. The testing set is further divided into 11,659 query images and 82,161 gallery images. Especially, MSMT17 is significantly larger in scale than both Market-1501 and CUHK03.
VeRi-776 [1] was collected from 20 real-world surveillance cameras in an urban area under diverse conditions, such as orientations, illuminations, and occlusions. It comprises over 50,000 images of 776 vehicles, and approximately 9000 trajectories. The dataset provides a variety of labels, including identity annotations, vehicle attributes, and spatiotemporal information. It is divided into two subsets for training and testing: the training set contains 37,778 images of 576 vehicles, while the test set consists of 11,579 images of the remaining 200 vehicles.
VehicleID [84] contains vehicle images captured by real-world cameras during the daytime. Each subject in the dataset has numerous images taken from the front and back, with some images annotated with model information to aid vehicle identification. The training set comprises 110,178 images of 13,134 vehicles. The test set is divided into three sections: Test800, with 6532 query images and 800 gallery images of 800 vehicles; Test1600, with 11,385 query images and 1600 gallery images of 1600 vehicles; and Test2400, with 17,638 query images and 2400 gallery images of 2400 subjects. Following the evaluation protocol of the authors [84], each testing subset is divided into query and gallery sets by randomly selecting one image per subject for the query subset, while the remaining images for each subject form the gallery subset.
VERI-Wild [94] is a large-scale dataset comprising 416,314 vehicle images across 40,671 unique identities. These images were collected using a wide-area surveillance system equipped with 174 cameras, spanning an urban region of over 200 km2. The camera network operated continuously, capturing vehicle footage 24 h a day for an entire month. The dataset is split into a training set and three testing subsets. The training set includes 277,797 images of 30,671 vehicle identities. The testing set is further divided into three parts: Test3000 (small), with 41,816 images, Test5000 (medium), with 69,389 images, and Test10000 (large), with 138,517 images.
The comprehensive overview of the datasets utilized in this document is presented in Table 3.

4.2. Evaluation Metrics

For the cross-domain object ReID task, we utilized Rank- k accuracy (where k { 1 , 5 , 10 } and mean average precision (mAP) to evaluate overall performance on test images.
Rank ratio accuracy (Rank- k ): The ranking process involves comparing the features extracted from a query object image i with all images in the gallery. This comparison results in a list of images sorted in descending order of similarity, with the most similar images appearing at the top. According to the ground truth of the selected dataset, the position within this sorted list where an image corresponds to the same object as the query image determines its rank. The Rank- k metric reflects the algorithm’s accuracy in correctly identifying object images within the top k ranks among the retrieved results for each query:
R a n k k = i = 1 M g t ( i , k ) M
Here, M represents the total number of probe images queried from the gallery, and g t ( i , k ) is a binary function:
g t i , k = 1   i f   t h e r e   a r e   p o s i t i v e   s a m p l e s   i   w i t h i n   t h e   t o p   n   r a n k i n g   r e s u l t s   0   o t h e r w i s e .
Mean average precision (mAP): In object ReID, where models produce a ranked list of images, it is crucial to consider the position of each image within the list. For each probe image, the average precision (AP) is calculated as follows:
A P = j = 1 N p j × g t ( j ) N
where N is the total number of images in the gallery set. The values p ( j ) and g t ( j ) represent the precision at the j -th position in the ranking list and a binary function, respectively. If the probe matches the j -th element, then g t ( j ) = 1 ; otherwise, g t ( j ) = 0 . The mean average precision (mAP) across all probe images is then computed using the A P values:
m A P = i = 1 M A P ( i ) M
Here, M denotes the total number of probe images queried, and A P ( i ) is the average precision calculated for each probe image i .

4.3. Benchmark on Person ReID

Our study began by comparing CORE-ReID V2 with state-of-the-art (SOTA) methods on two domain adaptation tasks: Market ➝ CUHK and CUHK ➝ Market (Table 4). We then expanded the evaluation to include two additional tasks: Market ➝ MSMT and CUHK ➝ MSMT (Table 5). In these comparisons, “Baseline” refers to the CORE-ReID method developed in our previous work, while CORE-ReID V2 represents the framework proposed in this paper. The term “Direct Transfer” indicates that the model is trained on the source domain and directly evaluated on the target domain, without applying any pseudo-labeling strategy. Additionally, CORE-ReID V2 Tiny is a lightweight version utilizing the smaller ResNet18 backbone. The evaluation metrics include mAP (%) and rank (R) at k accuracy (%).
The results highlight that CORE-ReID V2 significantly outperforms existing SOTA methods, demonstrating the effectiveness of our approach. By incorporating the ensemble fusion++ component with ECAB and proposed SECAB, CORE-ReID V2 achieves substantial improvements over the original CORE-ReID V1. Notably, CORE-ReID V2 surpasses PAOA+ by large margins, achieving mAP improvements of 16.1% and 6.6% on the Market ➝ CUHK and CUHK ➝ Market tasks, respectively, even though PAOA+ utilizes additional training data. Additionally, our framework delivers significant enhancements over CACL and PAOA+, achieving mAP gains of 7.6% and 14.7% mAP on the Market ➝ MSMT and CUHK ➝ MSMT tasks, respectively.

4.4. Benchmark on Vehicle ReID

We evaluated CORE-ReID V2 against state-of-the-art methods on the VehicleID ➝ VeRi-776 (Table 6), VehicleID ➝ VERI-Wild (Table 7), and VeRi-776 ➝ VehicleID (Table 8) tasks. “Baseline” refers to the implementation based on CORE-ReID V1 [11] with ensemble fusion component, while CORE-ReID V2 is the proposed algorithm. “Direct Transfer” indicates that the model is trained on the source domain and directly evaluated on the target domain, without any clustering-based pseudo-labeling operation. CORE-ReID V2 Tiny is a lightweight variant using ResNet18. Metrics include mAP (%) and rank (R) at k accuracy (%).
Across multiple domain adaptation scenarios (VeRi-776 ➝ VehicleID, VehicleID ➝ VERI-Wild, and VehicleID ➝ VeRi-776), the proposed CORE-ReID V2 framework demonstrates superior performance compared to existing state-of-the-art (SOTA) methods. Specifically, our evaluation included supervised approaches such as FACT and Mixed Diff+CCL, as well as unsupervised person ReID methods PUL and UDAR. Additionally, we incorporated leading SOTA techniques, including PAL, MMT, SPCL PLM, CSP+FCD, and DMDU, for a comprehensively comparative analysis.
In the VeRi-776 ➝ VehicleID adaptation setting, CORE-ReID V2 achieves 67.0%, 63.02%, and 57.99% mAP across three test modes, surpassing the DMDU method by 5.21%, 6.29%, and 4.02% in each case. For the VehicleID ➝ VERI-Wild scheme, our framework records a mAP of 40.2%, 34.9% and 27.8%, outperforming the ECN method by 5.5%, 4.3%, and 3.1% on the Test3000, Test5000 and Test10000 evaluation modes, respectively. In the VehicleID ➝ VeRi-776 scenario, MGR-GCL and MATNet+DMDU attain 48.73% and 49.25% mAP, respectively. CORE-ReID V2 outperforms all competing methods, achieving a mAP of 49.50% and a Rank-1 accuracy of 80.15%, setting its position as a new SOTA approach.

4.5. Ablation Study

Feature map visualization: To validate our approach, we employed Grad-CAM [127] to visualize feature maps at the global feature level. Key features for each person and vehicle are highlighted using heatmaps, where color intensity indicates importance—blue represents less significant regions, while red denotes the most crucial areas for object re-identification. As illustrated in Figure 10 and Figure 11, the essential features are concentrated on the target person’s body and the vehicle’s structure. Furthermore, the heatmaps exhibit similar distributions between the original and flipped images, reinforcing the performance of our method. This consistency aligns with the accuracy results reported above, further validating the effectiveness of CORE-ReID V2.
In the Market ➝ MSMT and CUHK ➝ MSMT scenarios, the Market ➝ MSMT model demonstrates a slightly superior ability to extract important features. The heatmaps reveal a more concentrated distribution in the middle and lower body regions for both the original and flipped images. This observation may explain the higher accuracy achieved by the Market ➝ MSMT model compared to the CUHK ➝ MSMT model, as reported in Table 5.
K-means clustering settings: we utilized the K-means clustering approach to generate pseudo-labels for the target domain, with parameters varying across different datasets. As shown in Table 9, our framework achieves optimal performance on Market → CUHK, CUHK → Market, Market → MSMT, CUHK → MSMT, VehicleID → VeRi-776, and VeRi-776 → VehicleID Small with cluster settings of 900, 900, 2000, 2000, 500, and 700, respectively.
Figure 12 shows that the performance of our approach varies depending on the dataset pairs and the clustering parameter values ( M T , j ) utilized.
Greedy K-means++ initialization: we enhanced clustering performance by employing the greedy K-means++ initialization strategy, which optimally balances randomness and centroid selection to improve cluster quality. This approach not only strengthens feature learning but also ensures a more stable pseudo-label generation, addressing challenges associated with obscure learning. Table 10 presents the experimental results comparing greedy K-means++ initialization with a random approach.
SECAB configuration: We used SECAB to enhance global features, leveraging attention mechanisms to emphasize crucial features while suppressing unnecessary ones. To validate the effectiveness of SECAB, we conducted an experiment by removing it from our network, as shown in Table 11.
Ensemble fusion++ configuration: We applied ECAB to refine local features by capturing rich inter-channel dependencies and SECAB to enhance global features. This design allows for adaptive feature recalibration at both local and global levels. To assess the effectiveness of this configuration, we conducted a series of experiments with different attention settings, as summarized in Table 12, demonstrating that this combination yields the most consistent performance improvement.
Backbone settings: we assessed the performance of both complex backbones (ResNet50, ResNet101, and ResNet152) and lightweight backbones (ResNet18 and ResNet34) for unsupervised domain adaptation in object ReID. Through extensive experiments and analysis, we gained insights into the impact of backbone architecture on overall performance, as well as its computational efficiency and suitability for resource-constrained environments, as shown in Table 13. Among the tested configurations, ResNet101 delivers the best performance in Market → CUHK, CUHK → Market, VehicleID → VeRi-776, and VeRi-776 → VehicleID scenarios. All experiments were conducted on two machines, each equipped with dual Quadro RTX 8000 GPUs (NVIDIA, Santa Clara, CA, USA).
Figure 13 demonstrates that lightweight backbones, such as ResNet18 and ResNet34, are also supported, highlighting the adaptability of our framework to various backbone architectures.
Computational complexity analysis: One of the key strengths of CORE-ReID V2 lies in its flexibility to support lightweight backbone networks, such as ResNet-18 and ResNet-34, in addition to deeper ones, including ResNet-50, 101, and 152. This flexibility allows users to balance accuracy and computational cost, enabling deployment in real-time or edge-based environments with limited resources. To highlight this advantage, we included a comparative table (Table 14) of commonly used ResNet backbones in terms of parameters and Giga Floating-Point Operations per Second (GFLOPs).
By allowing the use of ResNet-18/34 as backbones, CORE-ReID V2 can achieve substantial reductions in training/inference cost, enabling faster deployment while maintaining competitive UDA performance. Additionally, only the teacher model is used during inference, further reducing runtime complexity. We believe these design choices make CORE-ReID V2 an efficient yet effective UDA solution suitable for practical deployment, especially in low-power or embedded AI systems.

5. Conclusions

In this paper, we introduced CORE-ReID V2, an enhanced framework designed to address limitations in its predecessor, CORE-ReID V1, while extending its applicability to object ReID tasks. By incorporating the novel ensemble fusion++ module, which adaptively enhances both local and global features, and utilizing advanced clustering techniques such as greedy KMeans++ initialization, CORE-ReID V2 achieves superior performance in unsupervised domain adaptation (UDA) for person and vehicle ReID tasks. Furthermore, support for lightweight backbones such as ResNet18 and ResNet34 makes the framework suitable for real-time and resource-constrained applications. Experimental results on widely used UDA person ReID and vehicle ReID datasets demonstrate that CORE-ReID V2 outperforms state-of-the-art methods, showcasing its strength and adaptability. These contributions not only push the boundaries of UDA-based object ReID but also provide a solid foundation for further exploration in this domain.
Despite its advancements, CORE-ReID V2 has several limitations that warrant attention. The scalability of the framework to other datasets, such as BV-Person [128], ENTIRe-ID [129], VRID-1 [130], VRAI [131], Vehicle-Rear [132], and V2I-CARLA [133], remains unexplored, which could introduce challenges in both performance and efficiency. Furthermore, the framework’s primary focus on person and vehicle ReID tasks also limits its exploration of broader object ReID applications, such as animal or product identification. Moreover, its reliance on the quality of pseudo-labels makes it vulnerable to performance degradation in noisy or highly complex scenarios.
To address these limitations, future work will focus on the following directions. First, to enhance scalability, we plan to evaluate and optimize CORE-ReID V2 across a broader range of object ReID datasets, including mentioned datasets above. This will allow us to assess the generalization capability of the framework and adapt it to more diverse environments and conditions. Domain-specific normalization strategies and adaptive sampling techniques may be introduced to maintain performance consistency across different dataset characteristics. Second, although this work focuses on person and vehicle ReID, the proposed framework is general and can be extended to other object-level UDA tasks. The core mechanisms, such as feature aggregation, pseudo-label refinement, and cross-domain consistency learning, are not restricted to a particular object category. We plan to further evaluate CORE-ReID V2 on general object, such as animal, product, or scene-specific ReID, by incorporating domain-specific priors and augmentation strategies to demonstrate this broader applicability. Third, we plan to explore advanced techniques, such as contrastive learning and adversarial regularization, to mitigate the impact of noisy pseudo-labels and improve model performance. Hopefully, these directions aim to unlock the full potential of CORE-ReID V2 and inspire future advancements in UDA for object ReID.

Author Contributions

Writing—original draft, T.Q.N. and O.D.A.P.; conceptualization, T.Q.N., O.D.A.P. and S.A.I.; methodology, T.Q.N. and O.D.A.P.; software, T.Q.N., O.D.A.P. and S.A.I.; validation, T.Q.N., H.D.P. and R.T.; formal analysis, T.Q.N. and O.D.A.P.; investigation, T.Q.N., O.D.A.P. and S.A.I.; resources, T.Q.N., O.D.A.P., H.D.P. and R.T.; data curation, T.Q.N., S.A.I., H.D.P. and R.T.; writing—review and editing, T.Q.N., O.D.A.P., S.A.I., H.D.P. and R.T.; visualization, T.Q.N., H.D.P. and R.T.; supervision, O.D.A.P.; project administration, T.Q.N.; funding acquisition, T.Q.N. and O.D.A.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by a Grant-in-Aid for Scientific Research (KAKENHI) from the Japan Society for the Promotion of Science (JSPS). Grant number: 25K15160.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data (code and models) presented in this study are openly available on GitHub at https://github.com/TrinhQuocNguyen/CORE-ReID-V2 (accessed on 30 June 2025).

Acknowledgments

The authors sincerely appreciate the support of CyberCore Co., Ltd. for providing the necessary training environments. They also extend their gratitude to the anonymous reviewers for their insightful feedback, which has contributed to enhancing the quality of this paper. Additionally, the authors would like to thank Raymond Swannack of the Graduate School of Information and Computer Science, Iwate Prefectural University, for his assistance in proofreading the manuscript.

Conflicts of Interest

Authors Trinh Quoc Nguyen and Syahid Al Irfan were employed by the company CyberCore Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ECABEfficient Channel Attention Block
BMFNBidirectional Mean Feature Normalization
CBAMConvolutional Block Attention Module
CNNConvolutional neural network
CORE-ReID V1Comprehensive optimization and refinement through ensemble fusion in domain adaptation for person re-identification
HHLHetero-Homogeneous Learning
MMFAMulti-task Mid-level Feature Alignment
MMTMutual Mean-Teaching
Object ReIDobject re-identification
SECABSimplified Efficient Channel Attention Block
SOTAState-of-the-srt
SSGSelf-Similarity Grouping
UDAUnsupervised domain adaptation
UNRNUncertainty-Guided Noise-Resilient Network

References

  1. Liu, X.; Liu, W.; Mei, T.; Ma, H. A Deep Learning-Based Approach to Progressive Vehicle Re-identification for Urban Surveillance. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 869–884. [Google Scholar]
  2. Liu, X.; Liu, W.; Mei, T.; Ma, H. PROVID: Progressive and Multimodal Vehicle Reidentification for Large-Scale Urban Surveillance. IEEE Trans. Multimed. 2018, 20, 645–658. [Google Scholar] [CrossRef]
  3. Wang, Z.; He, L.; Tu, X.; Zhao, J.; Gao, X.; Shen, S. Robust Video-Based Person Re-Identification by Hierarchical Mining. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 8179–8191. [Google Scholar] [CrossRef]
  4. Zheng, L.; Yang, Y.; Hauptmann, A.G. Person Re-identification: Past, Present and Future. arXiv 2016, arXiv:1610.02984. [Google Scholar]
  5. Chang, Z.; Zheng, S. Revisiting Multi-Granularity Representation via Group Contrastive Learning for Unsupervised Vehicle Re-identification. arXiv 2024, arXiv:2410.21667. [Google Scholar]
  6. Luo, H.; Jiang, W.; Gu, Y.; Liu, F.; Liao, X.; Lai, S. A Strong Baseline and Batch Normalization Neck for Deep Person Re-Identification. IEEE Trans. Multimed. 2020, 22, 2597–2609. [Google Scholar] [CrossRef]
  7. Sharma, C.; Kapil, S.R.; Chapman, D. Person Re-Identification with a Locally Aware Transformer. arXiv 2021, arXiv:2106.03720. [Google Scholar]
  8. Chen, W.; Xu, X.; Jia, J.; Luo, H.; Wang, Y.; Wang, F. Beyond Appearance: A Semantic Controllable Self-Supervised Learning Framework for Human-Centric Visual Tasks. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 15050–15061. [Google Scholar]
  9. Almeida, E.; Silva, B.; Batista, J. Strength in Diversity: Multi-Branch Representation Learning for Vehicle Re-Identification. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), Bilbao, Spain, 24–28 September 2023; pp. 4690–4696. [Google Scholar]
  10. Li, J.; Gong, X. Prototypical Contrastive Learning-based CLIP Fine-tuning for Object Re-identification. arXiv 2023, arXiv:2310.17218. [Google Scholar]
  11. Nguyen, T.Q.; Prima, O.D.A.; Hotta, K. CORE-ReID: Comprehensive Optimization and Refinement through Ensemble Fusion in Domain Adaptation for Person Re-Identification. Software 2024, 3, 227–249. [Google Scholar] [CrossRef]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  13. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jegou, H. Training Data-efficient Image Transformers & Distillation Through Attention. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 10347–10357. [Google Scholar]
  14. Song, L.; Wang, C.; Zhang, L.; Du, B.; Zhang, Q.; Huang, C.; Wang, X. Unsupervised Domain Adaptive Re-identification: Theory and Practice. Pattern Recognit. 2020, 102, 107173. [Google Scholar] [CrossRef]
  15. Ge, Y.; Chen, D.; Li, H. Mutual Mean-Teaching: Pseudo Label Refinery for Unsupervised Domain Adaptation on Person Re-identification. arXiv 2020, arXiv:2001.01526. [Google Scholar]
  16. Fu, Y.; Wei, Y.; Wang, G.; Zhou, Y.; Shi, H.; Uiuc, U. Self-Similarity Grouping: A Simple Unsupervised Cross Domain Adaptation Approach for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October 2019–2 November 2019; pp. 6111–6120. [Google Scholar]
  17. Ding, J.; Zhou, X. Learning Feature Fusion for Unsupervised Domain Adaptive Person Re-identification. In Proceedings of the 2022 26th International Conference on Pattern Recognition (ICPR), Montreal, QC, Canada, 21–25 August 2022; pp. 2613–2619. [Google Scholar]
  18. Zhou, R.; Wang, Q.; Cao, L.; Xu, J.; Zhu, X.; Xiong, X.; Zhang, H.; Zhong, Y. Dual-Level Viewpoint-Learning for Cross-Domain Vehicle Re-Identification. Electronics 2024, 13, 1823. [Google Scholar] [CrossRef]
  19. Arthur, D.; Vassilvitskii, S. K-means++: The Advantages of Careful Seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, New Orleans, LA, USA, 7–9 January 2007; pp. 1027–1035. [Google Scholar]
  20. Zhang, Y.; Xiang, T.; Hospedales, T.M.; Lu, H. Deep Mutual Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4320–4328. [Google Scholar]
  21. Zhai, Y.; Ye, Q.; Lu, S.; Jia, M.; Ji, R.; Tian, Y. Multiple Expert Brainstorming for Domain Adaptive Person Re-identification. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 594–611. [Google Scholar]
  22. Tarvainen, A.; Valpola, H. Mean Teachers Are Better Role Models: Weight-averaged Consistency Targets Improve Semi-supervised Deep Learning Results. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 1195–1204. [Google Scholar]
  23. Brown, L.M. View independent vehicle/person classification. In Proceedings of the ACM 2nd International Workshop on Video Surveillance & Sensor Networks, New York, NY, USA, 15 October 2004; pp. 114–123. [Google Scholar]
  24. Yue, J.X.S. Mimicking visual searching with integrated top down cues and low-level features. Neurocomput 2014, 133, 1–17. [Google Scholar] [CrossRef]
  25. Wei, L.; Liu, X.; Li, J.; Zhang, S. VP-ReID: Vehicle and Person Re-Identification System. In Proceedings of the 2018 ACM on International Conference on Multimedia Retrieval, Yokohama, Japan, 11–14 June 2018; pp. 501–504. [Google Scholar]
  26. Chen, X.; Sui, H.; Fang, J.; Feng, W.; Zhou, M. Vehicle Re-Identification Using Distance-Based Global and Partial Multi-Regional Feature Learning. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1276–1286. [Google Scholar] [CrossRef]
  27. Lin, X.; Li, R.; Zheng, X.; Peng, P.; Wu, Y.; Huang, F. Aggregating Global and Local Visual Representation for Vehicle Re-IDentification. IEEE Trans. Multimed. 2020, 23, 3968–3977. [Google Scholar] [CrossRef]
  28. Akbarian, S.; Seyyed-Kalantari, L.; Khalvati, F.; Dolatabadi, E. Evaluating Knowledge Transfer in Neural Network for Medical Images. IEEE Access 2023, 11, 85812–85821. [Google Scholar] [CrossRef]
  29. Chen, S.; Wang, Y.; Shi, Y.; Yan, K.; Geng, M.; Tian, Y. Deep Transfer Learning for Person Re-Identification. In Proceedings of the 2018 IEEE Fourth International Conference on Multimedia Big Data (BigMM), Xi’an, China, 13–16 September 2018. [Google Scholar]
  30. Zhao, G.; Zhang, X.; Tang, H.; Shen, J.; Qian, X. Domain-Oriented Knowledge Transfer for Cross-Domain Recommendation. IEEE Trans. Multimed. 2024, 26, 9539–9550. [Google Scholar] [CrossRef]
  31. Wang, Q.; Qian, X.; Li, B.; Fu, Y.; Xue, X. Image-Text-Image Knowledge Transferring for Lifelong Person Re-Identification with Hybrid Clothing States. arXiv 2024, arXiv:2405.16600. [Google Scholar]
  32. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting Visual Category Models to New Domains. In Proceedings of the Computer Vision—ECCV 2010, Berlin/Heidelberg, Germany, 5–11 September 2010; pp. 213–226. [Google Scholar]
  33. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic Flow Kernel for Unsupervised Domain Adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  34. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised Visual Domain Adaptation Using Subspace Alignment. In Proceedings of the 2013 IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 2960–2967. [Google Scholar]
  35. Sun, B.; Feng, J.; Saenko, K. Return of Frustratingly Easy Domain Adaptation. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12 February 2016; pp. 2058–2065. [Google Scholar]
  36. Zhou, L.; Ye, M.; Li, X.; Zhu, C.; Liu, Y.; Li, X. Disentanglement Then Reconstruction: Unsupervised Domain Adaptation by Twice Distribution Alignments. Expert Syst. Appl. 2024, 237, 121498. [Google Scholar] [CrossRef]
  37. Zhang, X.; Ge, Y.; Qiao, Y.; Li, H. Refining Pseudo Labels with Clustering Consensus over Generations for Unsupervised Object Re-identification. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 3435–3444. [Google Scholar]
  38. Bousmalis, K.; Silberman, N.; Dohan, D.; Erhan, D.; Krishnan, D. Unsupervised Pixel–Level Domain Adaptation with Generative Adversarial Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 95–104. [Google Scholar]
  39. Wei, L.; Zhang, S.; Gao, W.; Tian, Q. Person Transfer GAN to Bridge Domain Gap for Person Re-Identification. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 79–88. [Google Scholar]
  40. Huang, Y.; Wu, Q.; Xu, J.; Zhong, Y. SBSGAN: Suppression of Inter-Domain Background Shift for Person Re-Identification. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), 27 October–2 November 2019; pp. 9526–9535. [Google Scholar]
  41. Liu, M.-Y.; Tuzel, O. Coupled Generative Adversarial Networks. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 469–477. [Google Scholar]
  42. Fan, H.; Yan, L.Z.C.; Yang, Y. Unsupervised Person Re-identification: Clustering and Fine-tuning. ACM Trans. Multimed. Comput. Commun. Appl. 2018, 14, 1–18. [Google Scholar] [CrossRef]
  43. Lin, Y.; Dong, X.; Zheng, L.; Yan, Y.; Yang, Y. A Bottom-Up Clustering Approach to Unsupervised Person Re-Identification. In Proceedings of the Thirty-Third AAAI Conference on Artificial Intelligence and Thirty-First Innovative Applications of Artificial Intelligence Conference and Ninth AAAI Symposium on Educational Advances in Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; AAAI Press: Washington, DC, USA, 2019; pp. 8738–8745. [Google Scholar]
  44. Ge, Y.; Zhu, F.; Chen, D.; Zhao, R.; Li, H. Self-paced Contrastive Learning with Hybrid Memory for Domain Adaptive Object Re-ID. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6 December 2020; pp. 11309–11321. [Google Scholar]
  45. Zheng, K.; Lan, C.; Zeng, W.; Zhang, Z.; Zha, Z.-J. Exploiting Sample Uncertainty for Domain Adaptive Person Re-Identification. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence (AAAI), Virtual Event, 2–9 February 2021. [Google Scholar]
  46. Sun, Q.-S.; Zeng, S.-G.; Liu, Y.; Heng, P.-A.; Xia, D.-S. A new method of feature fusion and its application in image recognition. Pattern Recognit. 2005, 38, 2437–2448. [Google Scholar] [CrossRef]
  47. Sudha, D.; Ramakrishna, M. Comparative Study of Features Fusion Techniques. In Proceedings of the 2017 International Conference on Recent Advances in Electronics and Communication Technology (ICRAECT), Bangalore, India, 16–17 March 2017; pp. 235–239. [Google Scholar]
  48. Tian, Y.; Zhang, W.; Zhang, Q.; Lu, G.; Wu, X. Selective Multi-Convolutional Region Feature Extraction based Iterative Discrimination CNN for Fine-Grained Vehicle Model Recognition. In Proceedings of the 2018 24th International Conference on Pattern Recognition (ICPR), Beijing, China, 20–24 August 2018; pp. 3279–3284. [Google Scholar]
  49. Lu, L.; Yi, Y.; Huang, F.; Wang, K.; Wang, Q. Integrating Local CNN and Global CNN for Script Identification in Natural Scene Images. IEEE Access 2019, 7, 52669–52679. [Google Scholar] [CrossRef]
  50. He, Z.; Nan, F.; Li, X.; Lee, S.-J.; Yang, Y. Traffic Sign Recognition by Combining Global and Local Features Based on Semi-supervised Classification. IET Intell. Transp. Syst. 2019, 14, 323–330. [Google Scholar] [CrossRef]
  51. Suh, S.; Lukowicz, P.; Lee, Y.O. Fusion of Global-Local Features for Image Quality Inspection of Shipping Label. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2643–2649. [Google Scholar]
  52. Li, Z.; Yang, L.; Zhou, F. FSSD: Feature Fusion Single Shot Multibox Detector. arXiv 2017, arXiv:1712.00960. [Google Scholar]
  53. Cong, R.; Yang, N.; Li, C.; Fu, H.; Zhao, Y.; Huang, Q. Global-and-Local Collaborative Learning for Co-Salient Object Detection. IEEE Trans. Cybern. 2023, 53, 1920–1931. [Google Scholar] [CrossRef]
  54. Li, T.; Zhang, Z.; Zhu, M.; Cui, Z.; Wei, D. Combining Transformer Global and Local Feature Extraction for Object Detection. Complex Intell. Syst. 2024, 10, 4897–4920. [Google Scholar] [CrossRef]
  55. Wei, J.; Wang, S.; Huang, Q. F3Net: Fusion, Feedback and Focus for Salient Object Detection. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 12321–12328. [Google Scholar]
  56. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object Detection in 20 Years: A Survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  57. Zhuge, M.; Fan, D.-P.; Liu, N.; Zhang, D.; Xu, D.; Shao, L. Salient Object Detection via Integrity Learning. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 3738–3752. [Google Scholar] [CrossRef]
  58. Zhang, Z.; Zhang, X.; Peng, C.; Xue, X.; Sun, J. ExFuse: Enhancing Feature Fusion for Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 269–284. [Google Scholar]
  59. Dai, Y.; Gieseke, F.; Oehmcke, S.; Wu, Y.; Barnard, K. Attentional Feature Fusion. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 3559–3568. [Google Scholar]
  60. Yang, X.; Li, S.; Chen, Z.; Chanussot, J.; Jia, X.; Zhang, B.; Li, B.; Chen, P. An Attention-Fused Network for Semantic Segmentation of Very-High-Resolution Remote Sensing Imagery. ISPRS J. Photogramm. Remote Sens. 2021, 177, 238–262. [Google Scholar] [CrossRef]
  61. Tian, Q.; Zhao, F.; Zhang, Z.; Qu, H. GLFFNet: A Global and Local Features Fusion Network with Biencoder for Remote Sensing Image Segmentation. Appl. Sci. 2023, 13, 8725. [Google Scholar] [CrossRef]
  62. Zhou, X.; Zhou, L.; Zhang, H.; Ji, W.; Zhou, B. Local–Global Multiscale Fusion Network for Semantic Segmentation of Buildings in SAR Imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 7410–7421. [Google Scholar] [CrossRef]
  63. Minaee, S.; Boykov, Y.; Porikli, F.; Plaza, A.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 3523–3542. [Google Scholar] [CrossRef] [PubMed]
  64. He, S.; Luo, H.; Wang, P.; Wang, F.; Li, H.; Jiang, W. TransReID: Transformer-based Object Re-Identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 15013–15022. [Google Scholar]
  65. Chen, Y.; Zhu, X.; Gong, S. Instance-Guided Context Rendering for Cross-Domain Person Re-Identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 232–242. [Google Scholar]
  66. Deng, W.; Zheng, L.; Ye, Q.; Kang, G.; Yang, Y.; Jiao, J. Image-Image Domain Adaptation with Preserved Self-Similarity and Domain-Dissimilarity for Person Re-identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 994–1003. [Google Scholar]
  67. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; March, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  68. Motiian, S.; Piccirilli, M.; Adjeroh, D.A.; Doretto, G. Unified Deep Supervised Domain Adaptation and Generalization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5716–5726. [Google Scholar]
  69. Tang, Y.; Yang, X.; Wang, N.; Song, B.; Gao, X. CGAN-TM: A Novel Domain-to-Domain Transferring Method for Person Re-Identification. IEEE Trans. Image Process. 2020, 29, 5641–5651. [Google Scholar] [CrossRef]
  70. Dai, Y.; Liu, J.; Sun, Y.; Tong, Z.; Zhang, C.; Duan, L.-Y. IDM: An Intermediate Domain Module for Domain Adaptive Person Re-ID. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 11844–11854. [Google Scholar]
  71. Long, M.; Cao, Y.; Wang, J.; Jordan, M. Learning Transferable Features with Deep Adaptation Networks. In Proceedings of the 32nd International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 97–105. [Google Scholar]
  72. Hoffman, J.; Tzeng, E.; Park, T.; Zhu, J.-Y.; Isola, P.; Saenko, K.; Efros, A.; Darrell, T. CyCADA: Cycle-Consistent Adversarial Domain Adaptation. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; pp. 1989–1998. [Google Scholar]
  73. Gretton, A.; Borgwardt, K.M.; Rasch, M.J.; Schölkopf, B.; Smola, A. A Kernel Two-Sample Test. J. Mach. Learn. Res. 2012, 13, 723–773. [Google Scholar]
  74. Gao, R.; Liu, F.; Zhang, J.; Han, B.; Liu, T.; Niu, G.; Sugiyama, M. Maximum Mean Discrepancy Test is Aware of Adversarial Attacks. arXiv 2020, arXiv:2010.11415. [Google Scholar]
  75. Zhang, W.; Ouyang, W.; Li, W.; Xu, D. Collaborative and Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3801–3809. [Google Scholar]
  76. Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M. Domain-Adversarial Neural Networks. arXiv 2014, arXiv:1412.4446. [Google Scholar]
  77. Jiang, Y.; Chen, W.; Sun, X.; Shi, X.; Wang, F.; Li, H. Exploring the Quality of GAN Generated Images for Person Re-Identification. In Proceedings of the 29th ACM International Conference on Multimedia, Chengdu, China, 20–24 October 2021; pp. 4146–4155. [Google Scholar]
  78. Pan, H.; Chen, Y.; He, Y.; Li, X.; He, Z. How Image Generation Helps Visible-to-Infrared Person Re-Identification? arXiv 2022, arXiv:2210.01585. [Google Scholar]
  79. Xie, H.; Luo, H.; Gu, J.; Jiang, W. Unsupervised Domain Adaptive Person Re-Identification via Intermediate Domains. Appl. Sci. 2022, 12, 6990. [Google Scholar] [CrossRef]
  80. Hu, C.; Li, X.; Liu, D.; Wu, H.; Chen, X.; Wang, J.; Liu, X. Teacher-Student Architecture for Knowledge Learning: A Survey. arXiv 2022, arXiv:2210.17332. [Google Scholar]
  81. Passalis, N.; Tefas, A. Learning Deep Representations with Probabilistic Knowledge Transfer. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 268–284. [Google Scholar]
  82. Laine, S.; Aila, T. Temporal Ensembling for Semi-Supervised Learning. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  83. Zheng, L.; Shen, L.; Tian, L.; Wang, S.; Wang, J.; Tian, Q. Scalable Person Re-identification: A Benchmark. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1116–1124. [Google Scholar]
  84. Liu, H.; Tian, Y.; Wang, Y.; Pang, L.; Huang, T. Deep Relative Distance Learning: Tell the Difference between Similar Vehicles. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2167–2175. [Google Scholar]
  85. Ju, Y.; Jia, S.; Cai, J.; Guan, H.; Lyu, S. GLFF: Global and Local Feature Fusion for AI-synthesized Image Detection. IEEE Trans. Multimed. 2024, 26, 4073–4085. [Google Scholar] [CrossRef]
  86. Polyak, B.T.; Juditsky, A. Acceleration of Stochastic Approximation by Averaging. SIAM J. Control Optim. 1992, 30, 838–855. [Google Scholar] [CrossRef]
  87. Zou, Y.; Yang, X.; Yu, Z.; Kumar, B.V.K.V.; Kautz, J. Joint Disentangling and Adaptation for Cross-Domain Person Re-Identification. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 87–104. [Google Scholar]
  88. Gong, Y.; Huang, L.; Chen, L. Eliminate Deviation with Deviation for Data Augmentation and a General Multi-modal Data Learning Method. arXiv 2021, arXiv:2101.08533. [Google Scholar]
  89. Bahdanau, D.; Cho, K.; Bengio, Y. Neural Machine Translation by Jointly Learning to Align and Translate. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  90. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  91. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  92. Grunau, C.; Özüdoğru, A.A.; Rozhoň, V.; Tětek, J. A Nearly Tight Analysis of Greedy K-means++. In Proceedings of the 2023 Annual ACM-SIAM Symposium on Discrete Algorithms (SODA), Florence, Italy, 22–25 January 2023; pp. 1012–1070. [Google Scholar]
  93. Li, W.; Zhao, R.; Xiao, T.; Wang, X. DeepReID: Deep Filter Pairing Neural Network for Person Re-identification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 152–159. [Google Scholar]
  94. Lou, Y.; Bai, Y.; Liu, J.; Wang, S.; Duan, L. VERI-Wild: A Large Dataset and a New Method for Vehicle Re-Identification in the Wild. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3230–3238. [Google Scholar]
  95. Zhong, Z.; Zheng, L.; Cao, D.; Li, S. Re-ranking Person Re-identification with k-Reciprocal Encoding. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3652–3661. [Google Scholar]
  96. Jin, X.; Lan, C.; Zeng, W.; Chen, Z.; Zhang, L. Style Normalization and Restitution for Generalizable Person Re-Identification. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 3140–3149. [Google Scholar]
  97. Liao, S.; Shao, L. Interpretable and Generalizable Person Re-identification with Query-Adaptive Convolution and Temporal Lifting. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020. [Google Scholar]
  98. Zhao, Y.; Zhong, Z.; Yang, F.; Luo, Z.; Lin, Y.; Li, S.; Sebe, N. Learning to Generalize Unseen Domains via Memory-based Multi-Source Meta-Learning for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  99. Choi, S.; Kim, T.; Jeong, M.; Park, H.; Kim, C. Meta Batch-Instance Normalization for Generalizable Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  100. Yang, F.; Zhong, Z.; Luo, Z.; Li, S.; Sebe, N. Federated and Generalized Person Re-identification through Domain and Feature Hallucinating. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  101. Xu, B.; Liang, J.; He, L.; Sun, Z. Mimic Embedding via Adaptive Aggregation: Learning Generalizable Person Re-identification. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 372–388. [Google Scholar]
  102. Zhang, P.; Dou, H.; Yu, Y.; Li, X. Adaptive Cross-Domain Learning for Generalizable Person Re-Identification. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  103. Li, H.; Wang, Y.; Zhu, L.; Wang, W.; Yin, K.; Li, Y.; Yin, G. Weakly Supervised Cross-Domain Person Re-Identification Algorithm Based on Small Sample Learning. Electronics 2023, 12, 4186. [Google Scholar] [CrossRef]
  104. Mao, Y.; Huafeng, L.; Yafei, Z. Unsupervised Domain Adaptation for Cross-Regional Scenes Person Re-identification. J. Shanghai Jiao Tong Univ. 2023, 1–26. [Google Scholar] [CrossRef]
  105. Sun, J.; Li, Y.; Chen, L.; Chen, H.; Peng, W. Multiple Integration Model for Single-source Domain Generalizable Person Re-identification. J. Vis. Commun. Image Represent 2024, 98, 104037. [Google Scholar] [CrossRef]
  106. Li, Q.; Gong, S. Mitigate Domain Shift by Primary-Auxiliary Objectives Association for Generalizing Person ReID. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 393–402. [Google Scholar]
  107. Zhao, F.; Liao, S.; Xie, G.-S.; Zhao, J.; Zhang, K.; Shao, L. Unsupervised Domain Adaptation with Noise Resistible Mutual-Training for Person Re-identification. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 526–544. [Google Scholar]
  108. Dai, Y.; Liu, J.; Bai, Y.; Tong, Z.; Duan, L.-Y. Dual-Refinement: Joint Label and Feature Refinement for Unsupervised Domain Adaptive Person Re-Identification. IEEE Trans. Image Process. 2020, 30, 7815–7829. [Google Scholar] [CrossRef]
  109. Bai, Z.; Wang, Z.; Wang, J.; Hu, D.; Ding, E. Unsupervised Multi-Source Domain Adaptation for Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12914–12923. [Google Scholar]
  110. Wang, Y.; Liang, X.; Liao, S. Cloning Outfits from Real-World Images to 3D Characters for Generalizable Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4890–4899. [Google Scholar]
  111. Tay, C.-P.; Yap, K.-H. Collaborative learning mutual network for domain adaptation in person re-identification. Neural Comput. Appl. 2022, 34, 12211–12222. [Google Scholar] [CrossRef]
  112. Zhou, H.; Kong, J.; Jiang, M.; Liu, T. Heterogeneous dual network with feature consistency for domain adaptation person re-identification. Int. J. Mach. Learn. Cyber 2023, 14, 1951–1965. [Google Scholar] [CrossRef]
  113. Yun, X.; Wang, Q.; Cheng, X.; Song, K.; Sun, Y. Discrepant mutual learning fusion network for unsupervised domain adaptation on person re-identification. Appl. Intell. 2023, 53, 2951–2966. [Google Scholar] [CrossRef]
  114. Lee, G.; Lee, S.; Kim, D.; Shin, Y.; Yoon, Y.; Ham, B. Camera-Driven Representation Learning for Unsupervised Domain Adaptive Person Re-identification. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 1453–11462. [Google Scholar]
  115. Rami, H.; Giraldo, J.H.; Winckler, N.; Lathuilière, S. Source-Guided Similarity Preservation for Online Person Re-Identification. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2024; pp. 1700–1709. [Google Scholar]
  116. Zhang, B.; Wu, D.; Lu, X.; Li, Y.; Gu, Y.; Li, J.; Wang, J. A domain generalized person re-identification algorithm based on meta-bond domain alignment. J. Vis. Commun. Image Represent 2024, 98, 2913–2933. [Google Scholar] [CrossRef]
  117. Tian, Q.; Cheng, Y.; He, S.; Sun, J. Unsupervised multi-source domain adaptation for person re-identification via feature fusion and pseudo-label refinement. Comput. Electr. Eng 2024, 113, 109029. [Google Scholar] [CrossRef]
  118. Bashir, R.M.S.; Shahzad, M.; Fraz, M.M. VR-PROUD: Vehicle Re-identification using Progressive Unsupervised Deep Architecture. Pattern Recognit. 2019, 90, 52–65. [Google Scholar] [CrossRef]
  119. Zhong, Z.; Zheng, L.; Luo, Z.; Li, S.; Yang, Y. Invariance Matters: Exemplar Memory for Domain Adaptive Person Re-Identification. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 598–607. [Google Scholar]
  120. Peng, J.; Wang, Y.; Wang, H.; Zhang, Z.; Fu, X.; Wang, M. Unsupervised Vehicle Re-identification with Progressive Adaptation. In Proceedings of the International Joint Conference on Artificial Intelligence, Yokohama, Japan, 7–15 January 2021; pp. 913–919. [Google Scholar]
  121. Wang, H.; Peng, J.; Jiang, G.; Fu, X. Learning Multiple Semantic Knowledge For Cross-Domain Unsupervised Vehicle Re-Identification. In Proceedings of the IEEE International Conference on Multimedia and Expo, Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  122. Wang, Y.; Peng, J.; Wang, H.; Wang, M. Progressive learning with multi-scale attention network for cross-domain vehicle re-identification. Sci. China Inf. Sci. 2022, 65, 160103. [Google Scholar] [CrossRef]
  123. Zhang, F.; Zhang, L.; Zhang, H.; Ma, Y. Image-to-image domain adaptation for vehicle re-identification. Multimed. Tools Appl. 2023, 82, 40559–40584. [Google Scholar] [CrossRef]
  124. Zhan, G.; Wang, Q.; Min, W.; Han, Q.; Zhao, H.; Wei, Z. Unsupervised Vehicle Re-Identification Based on Cross-Style Semi-Supervised Pre-Training and Feature Cross-Division. Electronics 2023, 12, 2931. [Google Scholar] [CrossRef]
  125. Zhang, X.; Ling, Y.; Li, K.; Zhou, W.S.Z. Multimodality Adaptive Transformer and Mutual Learning for Unsupervised Domain Adaptation Vehicle Re-Identification. IEEE Trans. Intell. Transport. Syst. 2024, 25, 20215–20226. [Google Scholar] [CrossRef]
  126. Ding, Y.; Fan, H.; Xu, M.; Yang, Y. Adaptive Exploration for Unsupervised Person Re-identification. ACM Trans. Multimed. Comput. Commun. Appl. 2020, 16, 1–19. [Google Scholar] [CrossRef]
  127. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  128. Yan, C.; Pang, G.; Wang, L.; Jiao, J.; Feng, X.; Shen, C.; Li, J. BV-Person: A Large-Scale Dataset for Bird-View Person Re-Identification. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 10943–10952. [Google Scholar]
  129. Yildiz, S.; Kasim, A.N. ENTIRe-ID: An Extensive and Diverse Dataset for Person Re-Identification. In Proceedings of the 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition (FG), Istanbul, Turkiye, 27–31 May 2024; pp. 1–5. [Google Scholar]
  130. Li, X.; Yuan, M.; Jiang, Q.; Li, G. VRID-1: A Basic Vehicle Re-identification Dataset for Similar Vehicles. In Proceedings of the 2017 IEEE 20th International Conference on Intelligent Transportation Systems (ITSC), Yokohama, Japan, 16–19 October 2017; pp. 1–8. [Google Scholar]
  131. Wang, P.; Jiao, B.; Yang, L.Y.Y.; Zhang, S.; Wei, W. Vehicle Re-identification in Aerial Imagery: Dataset and Approach. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October 2019–2 November 2019; pp. 460–469. [Google Scholar]
  132. Oliveira, I.O.D.; Laroca, R.; Menotti, D.; Fonseca, K.V.O.; Minetto, R. Vehicle-Rear: A New Dataset to Explore Feature Fusion for Vehicle Identification Using Convolutional Neural Networks. IEEE Access 2021, 9, 101065–101077. [Google Scholar] [CrossRef]
  133. Wang, H.; Yuan, X.; Cai, Y.; Chen, L.; Li, Y. V2I-CARLA: A Novel Dataset and a Method for Vehicle Reidentification-Based V2I Environment. IEEE Trans. Instrum. Meas. 2022, 71, 1–9. [Google Scholar] [CrossRef]
Figure 1. The overall method proposed in this study. First, the model is trained on a customized labeled source-domain dataset, after which the parameters of the pre-trained model are transferred to both the student and teacher networks as an initialization step for the next stage. During fine-tuning, the student model is trained, and the teacher model is updated through the mean teacher method. To optimize computational efficiency, only the teacher model is employed for inference.
Figure 1. The overall method proposed in this study. First, the model is trained on a customized labeled source-domain dataset, after which the parameters of the pre-trained model are transferred to both the student and teacher networks as an initialization step for the next stage. During fine-tuning, the student model is trained, and the teacher model is updated through the mean teacher method. To optimize computational efficiency, only the teacher model is employed for inference.
Aisens 01 00004 g001
Figure 2. Our process for creating a complete training set for the source domain is as follows: in the person ReID task, we first combine the training set (represented by green icons) and the test set (represented by dark green icons) from the source dataset to create a comprehensive set of real images. This combined set is then used to train a camera-aware style transfer model, which generates style-transferred images (blue icons for the training set and dark blue icons for the test set) that reflect the stylistic characteristics of the target cameras. The final training set for the source domain is formed by merging the real images (green and dark green icons) with the style-transferred images (blue and dark blue icons). For vehicle ReID, due to the simpler nature of vehicle features and the extensive number of cameras involved (with some datasets not specifying the number of cameras), we use domain-aware transfer models instead of camera-aware models. These models generate style-transferred images (orange icons for the training set and dark orange icons for the test set) that capture the target domain’s style. The final source-domain training set is then constructed by integrating the real images (green and dark green icons) with the style-transferred images (orange and dark orange icons).
Figure 2. Our process for creating a complete training set for the source domain is as follows: in the person ReID task, we first combine the training set (represented by green icons) and the test set (represented by dark green icons) from the source dataset to create a comprehensive set of real images. This combined set is then used to train a camera-aware style transfer model, which generates style-transferred images (blue icons for the training set and dark blue icons for the test set) that reflect the stylistic characteristics of the target cameras. The final training set for the source domain is formed by merging the real images (green and dark green icons) with the style-transferred images (blue and dark blue icons). For vehicle ReID, due to the simpler nature of vehicle features and the extensive number of cameras involved (with some datasets not specifying the number of cameras), we use domain-aware transfer models instead of camera-aware models. These models generate style-transferred images (orange icons for the training set and dark orange icons for the test set) that capture the target domain’s style. The final source-domain training set is then constructed by integrating the real images (green and dark green icons) with the style-transferred images (orange and dark orange icons).
Aisens 01 00004 g002
Figure 3. The camera-aware style-transferred samples from Market-1501 and CUHK03 datasets. Each original image, captured by a specific camera, was transformed to match the styles of the other five cameras in Market-1501 and one camera in CUHK03, covering both training and test data. By applying the style transfer models, these transformations produce style-transferred images as outputs based on the real input images.
Figure 3. The camera-aware style-transferred samples from Market-1501 and CUHK03 datasets. Each original image, captured by a specific camera, was transformed to match the styles of the other five cameras in Market-1501 and one camera in CUHK03, covering both training and test data. By applying the style transfer models, these transformations produce style-transferred images as outputs based on the real input images.
Aisens 01 00004 g003
Figure 4. The domain-aware style-transferred samples from VeRi-776 [1] and VehicleID [84] datasets. Each original real image from the source-domain dataset was transformed to match the styles of the target domain dataset. By applying domain-aware style transfer models, the model trained in the pre-training stage will be able to capture the style and features of the target domain.
Figure 4. The domain-aware style-transferred samples from VeRi-776 [1] and VehicleID [84] datasets. Each original real image from the source-domain dataset was transformed to match the styles of the target domain dataset. By applying domain-aware style transfer models, the model trained in the pre-training stage will be able to capture the style and features of the target domain.
Aisens 01 00004 g004
Figure 5. The comprehensive training process employed during the fully supervised pre-training stage. A ResNet-based model, adaptable to various backbone sizes (from ResNet18 and 34 to ResNet50, 101, and 150), serves as the backbone architecture within our training framework.
Figure 5. The comprehensive training process employed during the fully supervised pre-training stage. A ResNet-based model, adaptable to various backbone sizes (from ResNet18 and 34 to ResNet50, 101, and 150), serves as the backbone architecture within our training framework.
Aisens 01 00004 g005
Figure 6. Data adapter component. The transformations of random flipping, random global grayscale, random local grayscale and random erasing will be controlled by probability parameters. In addition, random erasing is only applied in the fine-tuning stage.
Figure 6. Data adapter component. The transformations of random flipping, random global grayscale, random local grayscale and random erasing will be controlled by probability parameters. In addition, random erasing is only applied in the fine-tuning stage.
Aisens 01 00004 g006
Figure 7. The comprehensive overview of our CORE-ReID V2 framework. The data adapter will pre-process the data depending on the type of object. We integrate local and global features using the enhanced ensemble fusion++ component. Specifically, the Efficient Channel Attention Block (ECAB) and Simplified Efficient Channel Attention Block (SECAB) are utilized to boost local and global feature extraction, respectively. By using Bi-directional Mean Feature Normalization (BMFN), the framework effectively merges features from the original image x T , i and its horizontally flipped counterpart x T , i , generating a fused feature φ l , l { t o p , b o t t o m } . The student network is trained in a supervised manner using pseudo-labels, while the teacher network is updated through a mean teacher approach, which computes the temporal average of the student network’s weights. Especially, the flipped image features are processed identically to the original image’s features until they reach the BMFN stage, ensuring consistent feature fusion.
Figure 7. The comprehensive overview of our CORE-ReID V2 framework. The data adapter will pre-process the data depending on the type of object. We integrate local and global features using the enhanced ensemble fusion++ component. Specifically, the Efficient Channel Attention Block (ECAB) and Simplified Efficient Channel Attention Block (SECAB) are utilized to boost local and global feature extraction, respectively. By using Bi-directional Mean Feature Normalization (BMFN), the framework effectively merges features from the original image x T , i and its horizontally flipped counterpart x T , i , generating a fused feature φ l , l { t o p , b o t t o m } . The student network is trained in a supervised manner using pseudo-labels, while the teacher network is updated through a mean teacher approach, which computes the temporal average of the student network’s weights. Especially, the flipped image features are processed identically to the original image’s features until they reach the BMFN stage, ensuring consistent feature fusion.
Aisens 01 00004 g007
Figure 8. The comparison between ensemble fusion in CORE-ReID V1 [11] and the proposed ensemble fusion++ component. The ς t o p   and ς b o t t o m features are passed through the ECAB and the τ g l o b a l feature is passed via the SECAB to produce the channel attention maps by exploiting the inter-channel relationship of features, which helps to enhance the features.
Figure 8. The comparison between ensemble fusion in CORE-ReID V1 [11] and the proposed ensemble fusion++ component. The ς t o p   and ς b o t t o m features are passed through the ECAB and the τ g l o b a l feature is passed via the SECAB to produce the channel attention maps by exploiting the inter-channel relationship of features, which helps to enhance the features.
Aisens 01 00004 g008
Figure 9. ECAB and SECAB designs. The structure of our SECAB is similar to that of ECAB [11] but simpler. The module only takes the Shared Multilayer Perceptron into account. It has odd h hidden layers, where the first h 1 2 layers are reduced in size with the reduction rate r , and the last h 1 2 layers will be expanded with the same rate r .
Figure 9. ECAB and SECAB designs. The structure of our SECAB is similar to that of ECAB [11] but simpler. The module only takes the Shared Multilayer Perceptron into account. It has odd h hidden layers, where the first h 1 2 layers are reduced in size with the reduction rate r , and the last h 1 2 layers will be expanded with the same rate r .
Aisens 01 00004 g009
Figure 10. Feature maps visualization using Grad-CAM [127]. (ad) Illustrate the feature maps of those pairs on Market ➝ CUHK, CUHK ➝ Market, Market ➝ MSMT, and CUHK ➝ MSMT, respectively. Warmer colors (red, yellow) in the heatmap indicate regions the model focused on, while cooler colors (blue) indicate less important areas.
Figure 10. Feature maps visualization using Grad-CAM [127]. (ad) Illustrate the feature maps of those pairs on Market ➝ CUHK, CUHK ➝ Market, Market ➝ MSMT, and CUHK ➝ MSMT, respectively. Warmer colors (red, yellow) in the heatmap indicate regions the model focused on, while cooler colors (blue) indicate less important areas.
Aisens 01 00004 g010
Figure 11. Feature maps visualization using Grad-CAM [127]. (a,b) Illustrate the feature maps of those pairs on VehicleID ➝ VeRi-776 and VeRi-776 ➝ VehicleID respectively. Warmer colors (red, yellow) in the heatmap indicate regions the model focused on, while cooler colors (blue) indicate less important areas.
Figure 11. Feature maps visualization using Grad-CAM [127]. (a,b) Illustrate the feature maps of those pairs on VehicleID ➝ VeRi-776 and VeRi-776 ➝ VehicleID respectively. Warmer colors (red, yellow) in the heatmap indicate regions the model focused on, while cooler colors (blue) indicate less important areas.
Aisens 01 00004 g011
Figure 12. Impact of clustering parameter M T ,   j . Results on (a) Market → CUHK, (b) CUHK → Market, (c) Market → MSMT, (d) CUHK → MSMT, (e) VehicleID → VeRi-776, and (f) VeRi-776 → VehicleID Small.
Figure 12. Impact of clustering parameter M T ,   j . Results on (a) Market → CUHK, (b) CUHK → Market, (c) Market → MSMT, (d) CUHK → MSMT, (e) VehicleID → VeRi-776, and (f) VeRi-776 → VehicleID Small.
Aisens 01 00004 g012
Figure 13. Impact of the backbone settings. Results on (a) Market → CUHK, (b) CUHK → Market, (c) VehicleID → VeRi-776, and (d) VeRi-776 → VehicleID Small shows that ResNet101 backbone yields the best overall results.
Figure 13. Impact of the backbone settings. Results on (a) Market → CUHK, (b) CUHK → Market, (c) VehicleID → VeRi-776, and (d) VeRi-776 → VehicleID Small shows that ResNet101 backbone yields the best overall results.
Aisens 01 00004 g013
Table 1. Summary of the main advancement of CORE-ReID V2 over CORE-ReID V1.
Table 1. Summary of the main advancement of CORE-ReID V2 over CORE-ReID V1.
CategoryCORE-ReID V1CORE-ReID V2
Current StatusDrawbacks/Issues
Applied DomainPerson ReIDOnly support person ReID.Expansion from person ReID to vehicle ReID and further object ReID.
Synthetic Data GenerationCamera-aware style transferDo not work in case the number of cameras is not specified.Camera-aware style transfer and domain-aware style transfer (in the case where the number of cameras is not specified).
Data AugmentationRandom gray scale patch replacementOnly replace a random gray scale patch in the image locally.Locally gray scale patch replacement and global gray scale conversion.
K-means ClusteringRandom initializationProblems from random initialization:
(1) Poor centroid placement;
(2) Slow convergence;
(3) Stuck in local minima;
(4) High variance in results;
(5) Imbalanced cluster sizes.
Greedy K-means++ initialization helps
(1) Select centroids with optimized spread;
(2) Minimize redundancy, requiring fewer iterations;
(3) Improve initialization stability;
(4) Reduce randomness and to provide consistent clusters;
(5) Ensure better centroid distribution.
Ensemble FusionEnsemble fusion with ECABOnly the local features are enhanced in the ensemble fusion.Ensemble fusion++ (with ECAB and SECAB) helps enhance both local and global features.
Supported BackbonesResNet50, 101, 152Do not support small backbones such as ResNet18, 34.ResNet18, 34, 50, 101, 152
Table 2. Comparison between ECAB and SECAB.
Table 2. Comparison between ECAB and SECAB.
AspectECABSECAB
Target UseLocal feature vectorsGlobal feature map
PoolingAdaptive max + avg poolingAdaptive max + avg pooling
Attention CoreShared Multilayer PerceptronSame Shared Multilayer Perceptron
Output ProcessingAttention map ×
(max + avg feature)
Attention map only
Residual Information Fusion (Later)With refined global featuresWith original global features
Computational CostHigher (due to residual and additional element-wise operations)Lower (no fusion step, lightweight on GPU)
Deployment StageLocal-level features refinementGlobal-level features refinement
Used in Ensemble FusionYesNo
Used in Ensemble Fusion++YesYes
Table 3. Details of datasets used in this manuscript.
Table 3. Details of datasets used in this manuscript.
CategoryDatasetCamerasTraining Set
(ID/Image)
Test Set (ID/Image)
GalleryQuery
Person ReIDMarket-15016751/12,936750/19,732750/3368
CUHK032767/7365700/5332700/1400
MSMT17151401/32,6213060/82,1613060/11,659
Vehicle ReIDVeRi-77620576/37,778200/11,579200/1678
VehicleID-13,134/110,178Test800: 800/800Test800: 800/6532
Test1600: 1600/1600Test1600: 1600/11,395
Test2400: 2400/2400Test2400: 2400/17,638
VERI-Wild17430,671/277,794Test3000: 3000/38,816Test3000: 3000/3000
Test5000: 5000/64,389Test5000: 5000/5000
Test10000: 10,000/128,517Test10000: 10,000/10,000
Table 4. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods (Acc %) on Market-1501 and CUHK03 datasets. Bold values represent the best results, while underlined values indicate the second-best performance. a indicates that the method uses multiple source datasets.
Table 4. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods (Acc %) on Market-1501 and CUHK03 datasets. Bold values represent the best results, while underlined values indicate the second-best performance. a indicates that the method uses multiple source datasets.
Market ➝ CUHKCUHK ➝ Market
MethodReferencemAPR-1R-5R-10mAPR-1R-5R-10
SNR a [96]CVPR 202017.517.1--52.477.8--
UDAR [14]PR 202020.920.3--56.677.1--
QAConv50 a [97]ECCV 202032.933.3--66.585.0--
M3L a [98]CVPR 202135.736.5--62.482.7--
MetaBIN a [99]CVPR 202143.043.1--67.284.5--
DFH-Baseline [100]CVPR 202210.211.2--13.231.1--
DFH a [100]CVPR 202227.230.5--31.356.5--
META a [101]ECCV 202247.146.2--76.590.5--
ACL a [102]ECCV 202249.450.1--76.890.6--
RCFA [103]Electronics 202317.718.533.643.434.563.378.883.9
CRS [104]JSJTU 2023----65.382.593.095.9
MTI [105]JVCIR 202416.316.2------
PAOA+ a [106]WACV 202450.350.9--77.991.4--
Baseline (CORE-ReID) [11]Software 202462.961.079.687.283.693.697.398.7
Direct TransferOurs23.924.640.348.935.563.377.883.2
CORE-ReID V2 Tiny (ResNet18)Ours33.031.948.959.160.383.491.894.7
CORE-ReID V2Ours66.466.983.488.984.593.997.698.7
Table 5. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods (Acc %) from Market-1501 and CUHK03 source datasets to the target domain MSMT17 dataset. Bold values represent the best results, while underlined values indicate the second-best performance. a Denotes that the method uses multiple source datasets, b indicates that the implementation is based on the author’s code.
Table 5. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods (Acc %) from Market-1501 and CUHK03 source datasets to the target domain MSMT17 dataset. Bold values represent the best results, while underlined values indicate the second-best performance. a Denotes that the method uses multiple source datasets, b indicates that the implementation is based on the author’s code.
Market ➝ MSMTCUHK ➝ MSMT
MethodReferencemAPR-1R-5R-10mAPR-1R-5R-10
NRMT [107]ECCV 202019.843.756.562.2----
DG-Net++ [87]ECCV 202022.148.4------
MMT [15]ICLR 202022.952.5--13.5 b30.9 b44.4 b51.1 b
UDAR [14]PR 202012.030.5--11.329.6--
Dual-Refinement [108]ArXiv 202025.153.366.171.5----
SNR a [96]CVPR 2020----7.722.0--
QAConv50 a [97]ECCV 2020----17.646.6--
M3L a [98]CVPR 2021----17.438.6--
MetaBIN a [99]CVPR 2021----18.841.2--
RDSBN [109]CVPR 202130.961.273.177.4----
ClonedPerson [110]CVPR 202214.641.0--13.442.3--
META a [101]ECCV 2022----24.452.1--
ACL a [102]ECCV 2022----21.747.3--
CLM-Net [111]NCA 202229.056.669.074.3----
CRS [104]JSJTU 202322.943.656.362.722.242.555.762.4
HDNet [112]IJMLC 2023 25.953.466.472.1----
DDNet [113]AI 202328.559.372.176.8----
CaCL [114]ICCV 202336.566.675.380.1----
PAOA+ a [106]WACV 2024----26.052.8--
OUDA [115]WACV 202420.246.1------
M-BDA [116]VCIR 202426.751.464.368.7----
UMDA [117]VCIR 202432.762.472.778.4----
Baseline (CORE-ReID) [11]Software 202441.969.580.384.440.467.379.083.1
Direct TransferOurs11.730.242.948.035.563.377.882.7
CORE-ReID V2 Tiny (ResNet18)Ours35.864.776.680.818.844.257.162.3
CORE-ReID V2Ours44.171.382.486.040.768.779.783.4
Table 6. Experimental results of the CORE-ReID V2 framework and SOTA methods on VehicleID ➝ VeRi-776. Bold values represent the best results, while underline values indicate the second-best performance.
Table 6. Experimental results of the CORE-ReID V2 framework and SOTA methods on VehicleID ➝ VeRi-776. Bold values represent the best results, while underline values indicate the second-best performance.
VehicleID ➝ VeRi-776
MethodReferencemAPR-1R-5R-10
FACT [1]ECCV 201618.7552.2172.88-
PUL [42]ACM 201817.0655.2467.34-
SPGAN [66]CVPR 201816.457.470.075.6
VR-PROUD [118]PR 201922.7555.7870.02-
ECN [119]CVPR 201920.0657.4170.53-
MMT [15]ICLR 202035.374.682.6-
SPCL [44]NIPS 202038.980.486.8-
PAL [120]IJCAI 202042.0468.1779.91-
UDAR [14]PR 202035.8076.9085.8089.00
ML [121]ICME 202136.9077.8085.50-
PLM [122]Sci.China 202247.3777.5987.00-
VDAF [123]MTA 202324.8646.3255.17-
CSP+FCD [124]Elec 202345.6074.3083.70-
MGR-GCL [5]ArXiv 202448.7379.2987.95-
MATNet+DMDU [125]ArXiv 202449.2579.1388.97-
Baseline Ours47.7078.1286.2388.14
Direct TransferOurs22.7162.0471.7976.32
CORE-ReID V2 Tiny (ResNet18)Ours40.1773.0081.4185.40
CORE-ReID V2Ours49.5080.1589.0590.29
Table 7. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods on VehicleID ➝ VERI-Wild. Bold values represent the best results, while underlined values indicate the second-best performance.
Table 7. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods on VehicleID ➝ VERI-Wild. Bold values represent the best results, while underlined values indicate the second-best performance.
VehicleID ➝ VERI-Wild
Test3000Test5000Test10000
MethodReferencemAPR-1R-5R-10mAPR-1R-5R-10mAPR-1R-5R-10
SPGAN [66]CVPR 201824.159.176.2-21.655.074.5-17.547.466.1-
ECN [119]CVPR 201934.773.488.8-30.668.684.6-24.761.078.2-
MMT [15]ICLR 202027.755.677.4-23.647.771.5-18.040.265.0-
SPCL [44]NIPS 202025.148.872.8-21.542.066.1-16.632.755.7-
UDAR [14]PR 202030.068.485.3-26.262.581.8-20.853.773.9-
AE [126]CCA 202029.967.088.5-26.261.881.5-20.953.173.7-
DLVL [18]Elec 202431.459.980.7-27.351.974.9-21.741.865.8-
Baseline Ours39.875.289.391.634.569.681.788.726.861.179.681.3
Direct TransferOurs20.948.264.370.718.944.360.966.915.638.053.359.8
CORE-ReID V2 Tiny (ResNet18)Ours28.656.574.980.223.152.170.678.419.948.166.374.6
CORE-ReID V2Ours40.276.690.292.134.970.286.289.327.862.179.882.3
Table 8. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods on VeRi-776 ➝ VehicleID. Bold values represent the best results, while underlined values indicate the second-best performance.
Table 8. Experimental results of the proposed CORE-ReID V2 framework and SOTA methods on VeRi-776 ➝ VehicleID. Bold values represent the best results, while underlined values indicate the second-best performance.
VeRi-776 ➝ VehicleID
Test800
VeRi-776 ➝ VehicleID
Test1600
VeRi-776 ➝ VehicleID
Test2400
MethodReferencemAPR-1R-5R-10mAPR-1R-5R-10mAPR-1R-5R-10
FACT [1]ECCV 2016-49.5367.96--44.6364.19--39.9160.49-
Mixed Diff+CCL [84]CVPR 2016-49.0073.50--42.8066.80--38.2061.60-
PUL [42]ACM 201843.9040.0356.03-37.6833.8349.72-34.7130.9047.18-
PAL [120]IJCAI 202053.5050.2564.91-48.0544.2560.95-45.1441.0859.12-
UDAR [14]PR 202059.6054.0066.1072.0155.3048.1064.1070.2052.9045.2062.6069.14
ML [121]ICME 202161.6054.8069.20-48.7040.3057.70-45.0036.5054.10-
PLM [122]Sci.China 202254.8551.2367.11-49.4145.4063.37-46.0041.7360.94-
CSP+FCD [124]Elec 202351.9054.4067.40-46.5052.7065.60-42.7045.9060.30-
VDAF [123]MTA 2023-----47.0364.86--43.6961.76-
MGR-GCL [5]ArXiv 202455.2452.3875.29-50.5645.8867.65-47.5942.8364.36-
DMDU [125]TITS 202461.8355.6168.25-56.7353.2863.56-53.9747.5961.85-
Baseline Ours64.2856.1674.5581.1560.0251.8471.6278.0856.1547.8566.8975.27
Direct TransferOurs61.2853.5069.8176.1357.2348.5767.0573.7752.3144.0461.0868.60
CORE-ReID V2 Tiny (ResNet18)Ours63.8755.1873.4381.1159.6950.0570.8877.7555.1445.9965.0773.54
CORE-ReID V2Ours67.0458.3276.5184.3263.0253.4974.3681.8557.9948.6268.3077.11
Table 9. Experimental results on different settings of number of pseudo identities in K-means clustering algorithm for both person and vehicle ReID tasks. Bold values represent the best results.
Table 9. Experimental results on different settings of number of pseudo identities in K-means clustering algorithm for both person and vehicle ReID tasks. Bold values represent the best results.
Person ReIDMarket → CUHKCUHK → Market
Number of ClustersmAPR-1R-5R-10mAPR-1R-5R-10
Ours   ( M T , j = 500 ) 44.443.265.376.469.486.894.996.7
Ours   ( M T , j = 700 ) 57.859.176.183.681.792.797.198.1
Ours   ( M T , j = 900 ) 66.466.983.488.984.593.997.698.7
Person ReIDMarketMSMTCUHKMSMT
Number of ClustersmAPR-1R-5R-10mAPR-1R-5R-10
Ours   ( M T , j = 2000 ) 44.171.382.486.040.6868.6679.7483.36
Ours   ( M T , j = 2500 ) 41.168.980.584.238.9167.2678.9782.80
Ours   ( M T , j = 3000 ) 38.967.279.083.235.864.776.680.8
Vechile ReIDVehicleIDVeRi-776VeRi-776VehicleID Small
Number of ClustersmAPR-1R-5R-10mAPR-1R-5R-10
Ours   ( M T , j = 500 ) 49.5080.1589.0590.2966.6058.2075.9083.70
Ours   ( M T , j = 700 ) 49.6379.1486.6589.6967.0458.3276.5184.32
Ours   ( M T , j = 900 ) 48.6179.0286.2989.1566.7057.5077.6084.20
Table 10. Experimental results using the greedy K-means++ initialization and random approach. The clustering parameter values M T , j are carried out from the study of K-means clustering settings. Bold values represent better results.
Table 10. Experimental results using the greedy K-means++ initialization and random approach. The clustering parameter values M T , j are carried out from the study of K-means clustering settings. Bold values represent better results.
Person ReID Market     CUHK   ( M T , j = 900 ) CUHK     Market   ( M T , j = 900 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (Random)63.663.880.987.883.893.697.498.6
Ours (Greedy Initialization)66.466.983.488.984.593.997.698.7
Person ReID Market     MSMT   ( M T , j = 2000 ) CUHK     MSMT   ( M T , j = 2000 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (Random)42.269.780.284.940.567.678.883.1
Ours (Greedy Initialization)44.171.382.486.040.768.779.783.4
Vehicle ReID VehicleID     VeRi - 776   ( M T , j = 500 ) VeRi - 776     VehicleID   Small   ( M T , j = 700 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (Random)47.7278.2386.5688.2665.7956.1475.9583.56
Ours (Greedy Initialization)49.5080.1589.0590.2967.0458.3276.5184.32
Table 11. Experimental results validating the effectiveness of SECAB in our proposed framework. The clustering parameter values ( M T , j ) are derived from the study of K-means clustering settings. Bold values represent better results.
Table 11. Experimental results validating the effectiveness of SECAB in our proposed framework. The clustering parameter values ( M T , j ) are derived from the study of K-means clustering settings. Bold values represent better results.
Person ReID Market     CUHK   ( M T , j = 900 ) CUHK     Market   ( M T , j = 900 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (without SECAB)65.065.182.687.683.993.797.498.6
Ours (with SECAB)66.466.983.488.984.593.997.698.7
Person ReID Market     MSMT   ( M T , j = 2000 ) CUHK     MSMT   ( M T , j = 2000 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (without SECAB)43.270.381.885.240.568.079.283.1
Ours (with SECAB)44.171.382.486.040.768.779.783.4
Vehicle ReID VehicleID     VeRi - 776   ( M T , j = 500 ) VeRi - 776     VehicleID   Small   ( M T , j = 700 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (without SECAB)48.0378.9287.6188.9365.1457.0275.5682.97
Ours (with SECAB)49.5080.1589.0590.2967.0458.3276.5184.32
Table 12. Experimental results with different attention settings in ensemble fusion++ component. The clustering parameter values ( M T , j ) are derived from the study of K-means clustering settings. Bold values represent the best results.
Table 12. Experimental results with different attention settings in ensemble fusion++ component. The clustering parameter values ( M T , j ) are derived from the study of K-means clustering settings. Bold values represent the best results.
Person ReID Market     CUHK   ( M T , j = 900 ) CUHK     Market   ( M T , j = 900 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (Top-local: ECAB, Bottom-Local: ECAB, Global: ECAB)64.364.981.684.383.292.697.398.6
Ours (Top-local: SECAB, Bottom-Local: SECAB, Global: SECAB)62.363.280.282.881.089.694.595.1
Ours (Top-local: ECAB, Bottom-Local: ECAB, Global: SECAB)66.466.983.488.984.593.997.698.7
Table 13. Experimental results on different settings of ResNet backbones in Market → CUHK, CUHK → Market scenarios. Bold values represent the best results.
Table 13. Experimental results on different settings of ResNet backbones in Market → CUHK, CUHK → Market scenarios. Bold values represent the best results.
Person ReID Market     CUHK   ( M T , j = 900 ) CUHK     Market   ( M T , j = 900 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (ResNet18)33.031.948.959.160.383.491.894.7
Ours (ResNet34)38.838.455.964.764.485.993.795.4
Ours (ResNet50)64.964.181.387.983.793.897.698.5
Ours (ResNet101)66.466.983.488.984.593.997.698.7
Ours (ResNet152)65.265.182.187.983.593.297.598.1
Vehicle ReIDVehicleID → VeRi-776  ( M T , j = 500 ) VeRi-776 → VehicleID Small  ( M T , j = 700 )
MethodmAPR-1R-5R-10mAPR-1R-5R-10
Ours (ResNet18)40.1773.0081.4185.4063.8755.1873.4381.11
Ours (ResNet34)46.6275.9283.7387.4963.8054.8073.6080.30
Ours (ResNet50)48.1178.8486.7189.8167.0258.3077.0083.90
Ours (ResNet101)49.5080.1589.0590.2967.0458.3276.5184.32
Ours (ResNet152)48.0778.2686.7389.9666.9758.2376.4983.86
Table 14. Evaluated using single-image inference on a Quadro RTX 8000 GPU with a batch size of 1, CORE-ReID V2 with ResNet18 and ResNet34 demonstrates significantly lower computational cost (fewer GFLOPs and parameters) and faster processing speed (higher frames per second) compared to deeper models.
Table 14. Evaluated using single-image inference on a Quadro RTX 8000 GPU with a batch size of 1, CORE-ReID V2 with ResNet18 and ResNet34 demonstrates significantly lower computational cost (fewer GFLOPs and parameters) and faster processing speed (higher frames per second) compared to deeper models.
CORE-ReID V2 with BackboneParameters
(Millions)
GFLOPs
(per Image)
Image SizeFPS (Using 1 Quadro RTX 8000 GPU)
ResNet-1812.97 M1.18128 × 256254
ResNet-3423.08 M2.35128 × 256185
ResNet-5046.62 M5.10128 × 256144
ResNet-10165.61 M7.58128 × 25687
ResNet-15281.26 M10.61128 × 25661
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nguyen, T.Q.; Prima, O.D.A.; Irfan, S.A.; Purnomo, H.D.; Tanone, R. CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion. AI Sens. 2025, 1, 4. https://doi.org/10.3390/aisens1010004

AMA Style

Nguyen TQ, Prima ODA, Irfan SA, Purnomo HD, Tanone R. CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion. AI Sensors. 2025; 1(1):4. https://doi.org/10.3390/aisens1010004

Chicago/Turabian Style

Nguyen, Trinh Quoc, Oky Dicky Ardiansyah Prima, Syahid Al Irfan, Hindriyanto Dwi Purnomo, and Radius Tanone. 2025. "CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion" AI Sensors 1, no. 1: 4. https://doi.org/10.3390/aisens1010004

APA Style

Nguyen, T. Q., Prima, O. D. A., Irfan, S. A., Purnomo, H. D., & Tanone, R. (2025). CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion. AI Sensors, 1(1), 4. https://doi.org/10.3390/aisens1010004

Article Metrics

Back to TopTop