Next Article in Journal
Smart Contracts, Blockchain, and Health Policies: Past, Present, and Future
Previous Article in Journal
MoCap-Impute: A Comprehensive Benchmark and Comparative Analysis of Imputation Methods for IMU-Based Motion Capture Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cross-Domain Residual Learning for Shared Representation Discovery

1
School of Artificial Intelligence, Ningbo Polytechnic University, Ningbo 315800, China
2
School of Information Science and Technology, Hangzhou Normal University, Hangzhou 311121, China
3
Ningbo University Health Science Center, Ningbo University, Ningbo 315211, China
*
Author to whom correspondence should be addressed.
Information 2025, 16(10), 852; https://doi.org/10.3390/info16100852
Submission received: 11 June 2025 / Revised: 26 July 2025 / Accepted: 28 July 2025 / Published: 2 October 2025
(This article belongs to the Special Issue Machine Learning in Image Processing and Computer Vision)

Abstract

In order to solve the problem of inconsistent data distribution in machine learning, domain adaptation based on feature representation methods extracts features from the source domain, and transfers them to the target domain for classification. The existing feature representation-based methods mainly solve the problem of inconsistent feature distribution between the source domain data and the target domain data, but only a few methods analyze the correlation of cross-domain features between the original space and shared latent space, which reduces the performance of domain adaptation. To this end, we propose a domain adaptation method with a residual module, the main ideas of which are as follows: (1) transfer the source domain data features to the target domain data through the shared latent space to achieve features sharing; (2) build a cross-domain residual learning model using the latent feature space as the residual connection of the original feature space, which improves the propagation efficiency of features; (3) use a regular feature space to sparse feature representation, which can improve the robustness of the model; and (4) give an optimization algorithm, and the experiments on the public visual datasets (Office31, Office-Caltech, Office-Home, PIE, MNIST-UPS, COIL20) results show that our method achieved 92.7% accuracy on Office-Caltech and 83.2% on PIE and achieved the highest recognition accuracy in three datasets, which verify the effectiveness of the method.

1. Introduction

In machine learning, it is necessary to have complete and sufficient training data for parameter tuning to achieve good performance. Therefore, the amount of data and the distribution of data features become an important factor. Traditional machine learning assumes that the training data and test data satisfy the condition of being independent and identically distributed. Under this assumption, most models work well. However, in practical applications, it is usually difficult to obtain a large amount of annotated data for a specific task, and data annotation is time-consuming and laborious, which leads to insufficient data in machine learning models. Moreover, there is a problem of inconsistent distribution of data between different data domains, which is called sample selection bias [1] or covariance displacement [2,3]. The above issues will reduce the robustness and reliability of traditional machine learning models, leading to a decrease in the performance of the model.
To avoid repetitive data annotation and improve the performance of the model, currently, domain adaptation (DA) methods that can achieve knowledge transfer between different domains, solving the problems of different data distributions and lack of train data in different domains effectively [4,5,6]. Domain adaptation can be mainly divided into instance-based methods, feature representation-based methods, and classifier-based methods. This paper focuses on the feature representation-based methods, which assumes that the relevant domain data has shared implicit features [7]. That is, the marginal distribution between data in different domains is matched [8]. Then, feature space information is used to align the domain data in the feature space in a mapping way to find the shared features between different domains. Therefore, a reliable machine learning model can be constructed using source domain data with sufficient labels [9]. By transferring source and target domain data features, common features between different domain data are utilized to expand the features of target domain data, reducing the strict constraint requirements of the model on data and improving the performance of the model [10].
The initial domain adaptation methods based on feature representation use feature representation methods to transform source domain data and target domain data into latent space, and construct a classifier through the latent feature representation of the target domain. The principle of these methods is summarized in Figure 1a, which employs feature extraction to construct a feature invariant latent space and then design a classifier for the target domain [11]. Gheisari et al. [12] demonstrated that minimizing classification error while maximizing manifold consistency in a shared space led to improved classification accuracy on target domains compared to non-adaptive baselines. Zheng et al. [13] showed that finding a dimensionality reduction technique minimizing distribution distance in the latent space effectively enabled feature transfer, resulting in measurable performance gains. Blitzer et al. [14] established that modeling feature correlations across domains via structural correspondence learning identified pivot features crucial for cross-domain discrimination, enhancing adaptation robustness. Jiang et al. [15] extended this concept effectively to multi-view data, showing that incorporating latent space features across views significantly addressed the unique challenges posed by multi-perspective domain differences. Xu et al. [16] leveraged distributionally robust optimization for feature extraction under weak supervision, demonstrating enhanced robustness to distributional uncertainties and achieving competitive results.
Later methods incorporate the original space information, which can guide target domain classifier construction by combining it with latent space features. This joint information is then utilized to build the classifier [17,18], as shown in Figure 1b. Dong et al. [19] proved that embedding a shared low-dimensional latent space into an SVM (support vector machine) framework, constrained by source-target latent space alignment, substantially improved SVM performance in domain adaptation tasks compared to standard SVMs. Yao et al. [20] successfully expanded this concept to the more complex multi-source adaptation scenario, showing that constraining the predicted label matrix across sources further enhanced adaptation effectiveness and robustness. Zhang et al. [21] demonstrated that combining latent space features with target pseudo-labels effectively mined richer domain-invariant information, yielding state-of-the-art results on several benchmarks. However, the above methods do not consider the distribution of latent space features from the perspective of distribution consistency, and the differences in distribution will reduce the effectiveness of shared features, thereby reducing the performance of domain adaptation.
Inspired by Residual Network [22], we propose a shared latent space domain adaptation with residual model (LRDA) to make full use of the relationship between original feature and latent space feature in domains, and address the problem of mismatched distribution of latent space features in domain adaptation. Specifically, mapping the source and target domain data into a shared latent feature space, mining common features between source and target domains, to achieve the goal of fully utilizing the source domain data to enhance the feature expression of the target domain data. Subsequently, minimizing the distribution differences between the shared feature space of the source and target domain, ensuring latent space features have a more consistent feature distribution, and reducing model performance degradation caused by inconsistent feature distribution. Finally, by combining the original spatial features with the the latent space features in both source and target domain, a residual model is formed to reduce the difficulty of fitting the model and obtain a better classifier. Following the above description, the principle of LRDA is shown in Figure 1c.
The main contributions of this paper are:
1.
We introduce a shared latent space by source domain feature and target domain feature, constructing a latent space domain adaptation method. Measuring the differences in latent space feature distribution, further constraining the consistent distribution of shared feature spaces in the domain to align the feature distribution in the shared latent space.
2.
We build a residual model using the original feature space and latent feature space, optimizing residual function to reduce the difficulty of feature transfer and improve model performance.
3.
We adopt the l 2 , 1 -norm for feature selection to sparsely represent the original feature, increasing the robustness of the model for outliers and noise naturally existing in dataset.
4.
Experiments verify that our method has better performance and can effectively recognize shallow and deep features, effectively improve the performance of cross-domain visual recognition tasks.

2. LRDA

2.1. Notation & Definition

We denote with small and capital letters respectively column vectors and matrices. We denote X s = x 1 s , x 2 s , , x n s s R d × n s is the matrix of source domain data x i s , where d means d-dimension feature and n s is the number of source domain samples. We denote Y s = y 1 s , y 2 s , , y n s s R n s × c is source domain label matrix, where y i s = { 0 , 1 } 1 × c is one-hot label for source domain data, and c is the number of categories. Source domain is denoted as D s = x i s , y i s X s × Y s : i = 1 , 2 , , n s . In the same way, given D t = x j t , y j t X t × Y t : j = 1 , 2 , , n t as target domain, X t = x 1 t , x 2 t , , x n t t R d × n t and Y t = y 1 t , y 2 t , , y n t t R n t × c as target domain data matrix and target domain label matrix respectively. It is worth noting that the target domain data usually lacks complete label, but source and target domains share the same c-cardinality label set.
Assuming the distribution of the data is known, then the marginal probability distribution of the source domain data and the target domain data can be represented by P ( X s ) and Q ( X t ) respectively. The goal of domain adaptation is to design a robust model that can predict the labels of target domain data using samples from the source domain. As a matter of fact, D s and D t are neither identical nor uncorrelated, so their edge probability distributions are different. For domain adaptation problems, the source domain data and target domain data come from different domains. Therefore, it is assumed that P ( X s ) Q ( X t ) . However, the distribution of category conditions for the source and target domain data is consistent, i.e., P ( X s | Y s ) = Q ( X t | Y t ) .

2.2. Formulation of LRDA

We design an objective function for latent space residual domain adaptation, which can simultaneously achieve the following goals: (1) finding a shared latent space to ensure that there are more common feature between the source domain data and the target domain data; (2) regularization the models of source domain and target domain to reduce the complexity of model structure; and (3) constraining in the latent space, the distribution of the source and target domains is consistent. Due to the lack of sufficient label in the target domain, we denote a label matrix F = f 1 t , f 2 t , , f n t t R n t × c corresponding to the target domain samples X t . If the sample has a label, set f i t = y i t ; otherwise f i t is a pseudo label. Therefore, we can express our model using the following objective functions:
R = α L f s X s , Y s + L f t X t , F + β Ω r f s + Ω r f t + γ Ω d ( X s , X t ) ,
where f s and f t are source and target domain classifiers, respectively. L ( · ) comprises regression losses in the source and target domains. Ω r is the regularization term of the source domain and the target domain to reduce the generalization error of model. Ω d is used to align the distribution of data between the source and target domains. α , β , γ are regularization hyper parameter to adjust the weight.
Definition 1
(Residual model). Assuming data is x, the objective function of the residual model can be expressed as f ( x ) = x + F ( x ) , where F ( x ) = f ( x ) x is residual function. The structure of the residual model is shown in Figure 2a.
Ref. [22] proved that it is less difficult to optimize residuals F ( x ) than optimization functions f ( x ) . Therefore, the objective function can be transformed into the following: let model F ( x ) approximate the residual function f ( x ) x and use x + F ( x ) to express the objective function.
Inspired by the residual model, we use the feature of the source and target domains to construct a shared latent feature space to constrain the classifier to integrate the common feature of the source and target domain. The latent feature space and the original feature space are constructed as a residual model, reducing the difficulty of fitting the model and improving the performance, as shown in Figure 2b. Specifically, the classifier of the shared feature space consists of two parts: the decision function of the original data and the decision function of the shared feature, which can be represented as follows:
f s ( X s ) = W s T X s + V s T θ T X s f t ( X t ) = W t T X t + V t T θ T X t ,
where W s and W t are the classification model vectors of the source domain and the target domain in the original space. θ R r × d comprises shared latent spatial transformation matrices used to map the source domain and the target domain to the same feature space. r is the dimension of the shared feature space, V s and V t are the classification model vectors of θ X s and θ X t , respectively. The source domain and target domain information are shared by θ .
Then the regression loss of the residual model classifier is:
L f s X s , Y s = W s T X s + V s T θ T X s Y s 2 L f t X t , F = W t T X t + V t T θ T X s F 2 .
In order to improve the robustness of the model, we conduct sparse regression on the feature extracted from the source domain and the target domain [23]. We use l 2 , 1 -norms [24] to sparse the classification model vectors of the original space to obtain the function of Ω r :
Ω r f s = W s 2 , 1 + W s + θ V s 2 Ω r f t = W t 2 , 1 + W t + θ V t 2 .
Due to the source and target domain commonly being P ( X s ) Q ( X t ) , that is, inconsistent distribution between source and target domain. If simply applying a source domain classifier to target domain, it may lead to deterioration of model performance due to misaligned domain distribution. How to minimize the distribution gap between different domains is a key issue for DA. Therefore, we use the MMD (maximum mean discrepancy) [25] to calculate the average vector distance of data in the reproducing kernel Hilbert space, so as to measure the distribution difference between different domains. Therefore, minimizing Equation (5) can align the feature distribution differences in the shared space, and the transformed source and target domains have common latent space distribution feature:
Ω d ( X s , X t ) = 1 n s θ T X s 1 n t θ T X t 2 = t r ( θ T X D X T θ ) s . t . θ T θ = I r × r ,
where D i , j = 1 n s 2 , w h e n x i , x j X s 1 n t 2 , w h e n x i , x j X t 1 n s n t , o t h e r s , X = [ X s , X t ] , r is the dimension of the shared hidden space according to matrix theory. By minimizing Equation (5), the feature in latent space has a similarly distribution, also ensuring both discriminant and domain invariance of the model.

2.3. Final Formulation

In summary, we propose a latent space domain adaptation with residual model. The algorithm schematic is shown in Figure 1c. Combining Equations (3)–(5), the objective function is as follows:
R 1 = arg min W s , W t , V s , V t , F , θ + α X s T W s + X s T V s θ Y s 2 + X s T W t + X s T V t θ F 2 + β W s 2 , 1 + W t 2 , 1 + W s + θ V s 2 + W t + θ V t 2 + γ t r ( θ T X D X T θ ) . s . t . θ T θ = I r × r F F T = I r × r
The objective function is solved by minimize the W s , W t , V s , V t , F and θ , and we apply these parameters to the decision functions f s ( X s ) and f t ( X t ) . Thus we fuse decision functions linearly as final decision functions:
m = max m y i t = φ f s x i t + ( 1 φ ) f t x i t m .
where φ 0 , 1 is hyper parameter to adjust the weight between source and the target domain classifiers. To simplify, we set φ = 0.5 .
For the convenience of subsequent solutions, this paper converts Equation (6) to
R 2 = arg min U s , U t , V s , V t , F , θ + α X s T U s Y s 2 + X t T U t F 2 + β U s 2 + U t 2 + U s θ V s 2 , 1 + U t θ V t 2 , 1 + γ t r ( θ T X D X T θ ) ,
where U s = W s + θ V s , U t = W t + θ V t , are auxiliary variables for simplify the calculation.

3. Optimization Algorithm

To facilitate optimization, we refer to parameter alternation optimization methods in fuzzy clustering; we divide six variables into four groups, namely { U s , U t }, { V s , V t }, {F}, and { θ }. We iteratively optimize these sets of parameters until the variables converge or the objective function is less than the threshold. Next, we optimize each part of the functions in Equation (8).

3.1. Optimization Procedure

3.1.1. Optimize V s , V t by Fixing U s , U t , F and θ

When V s = θ T U s and V t = θ T U t , the objective function R 2 can obtain the minimum value regarding the variable V s and V s .
Proof. 
Assuming that the variables U s , U t , F and θ are known, Equation (8) can be transformed into:
R 3 = arg min V s , V t β U s θ V s 2 , 1 + U t θ V t 2 , 1 .
Let:
L 1 = arg min V s β t r U s θ V s T M s U s θ V s , L 2 = arg min V t β t r U t θ V t T M t U t θ V t .
In terms of the definition in Nie et al. [26], for matrix A R n × d , we can derive A 2 , 1 = 2 t r A T M A , where M i i = 1 2 A i , : 2 . By solving the derivative of Equation (10) w.r.t. V s , V t and letting it equal to zero:
L 1 V s = 0 V s = θ T U s , L 2 V t = 0 V t = θ T U t .
   □

3.1.2. Optimize U s , U t by Fixing V s , V t , F and θ

After obtaining the optimal value obtained from Section 3.1.1, when U s = A s 1 B s and U t = A t 1 B t , the objective function R 2 can obtain the minimum value regarding the variable U s and U s .
Proof. 
It is obvious that Equation (8) with fixed V s , V t , F and θ is equivalent to
R 4 = arg min U s , U t + α X s T U s Y s 2 + X t T U t F 2 + β U s 2 + U t 2 + U s θ V s 2 , 1 + U t θ V t 2 , 1 .
Let
L 3 = arg min U s α X s T U s Y s 2 + β U s 2 + U s θ V s 2 , 1 , L 4 = arg min U t α X t T U t F 2 + β U t 2 + U t θ V t 2 , 1 .
by solving the derivative of Equation (13) w.r.t. V s , V t and letting it equal zero. We have
U s = A s 1 B s , U t = A t 1 B t .
where A s = α X s X s T + β + β M s , B s = α X s Y s + β M s θ V s , A t = α X t X t T + β + β M t and B t = α X t F + β M t θ V .    □

3.1.3. Optimize F by Fixing U s , U t , V s , V t and θ

After obtaining the optimal values from Section 3.1.1 and Section 3.1.2, the objective function R 2 can obtain the minimum value when F = X t T U t
Proof. 
Assuming known variables U s , U t , V s , V t and θ , then Equation (8) can be simplified to be:
R 5 = arg min F α X t T U t F 2 .
By setting the derivative of Equation (14) w.r.t. F as 0, the prediction labels F are obtained by F = X t T U t .    □

3.1.4. Optimize θ by Fixing U s , U t , V s , V t and F

When the variable U s , U t , V s , V t and F obtain the optimal solution obtained from Section 3.1.1, Section 3.1.2 and Section 3.1.3, the solution of θ can be transformed into a singular value decomposition (SVD) problem for the matrix O.
Proof. 
Then with fixed U s , U t and F, we can get the optimal θ by solving
R 6 = arg min V s , V t , θ β U s θ V s 2 , 1 + U t θ V t 2 , 1 + γ t r ( θ T X D X T θ )
To solve the minimum value of Equation (16), let V s = θ T U s , V t = θ T U t and then Equation (16) can be rewritten as solving the generalized eigen-decomposition problem:
R 7 = arg min θ 2 β t r θ T U U T N θ θ T N U U T θ + γ t r ( θ T X D X T θ ) = arg max θ t r θ T O O T θ , s . t . θ T θ = I r × r
where U = U s , U t . Let O O T = 2 β U U T N + 2 β N U U T γ X D X T . According to matrix theory, the solution of the above equation can be obtained by singular value decomposition of the matrix O. Let O = E Σ G T , by descending the diagonal elements in the diagonal matrix Σ , the first r rows of E T are the optimal solutions for θ .    □

3.2. Algorithm Description

According to the above optimization rules, the method adopts iterative learning strategy for parameter optimization to achieve parameter update and optimization. The algorithm description is shown in Algorithm 1.
Algorithm 1 Latent Space Domain Adaptation with Residual Model
Require: Source domain data and labels X s × Y s , target source domain data and labels X t × Y t , regularization hyper-parameter α , β , γ , shared feature dimension r, the maximal iteration number N.
  1:
Set v = 0, set U s , U t , V s , V t , θ as uniform distribution matrix in U [ 0 , 1 ] , set threshold ε of objective function
  2:
while  v > N or R 2 < ε  do
  3:
   Compute V s = θ T U s and V t = θ T U t
  4:
   Compute A s = α X s X s T + β + β M s and B s = α X s Y s + β M s θ V s
  5:
   Compute A t = α X t X t T + β + β M t and B t = α X t F + β M t θ V t
  6:
   Compute U = A 1 B and get U s , U t
  7:
   Compute F = X t T U t
  8:
  Compute O O T = 2 β U U T N + 2 β N U U T γ X D X T , let O = E Σ G T , by descending the diagonal elements in the diagonal matrix Σ , the first r rows of E T are the optimal solutions for θ .
  9:
   Calculate the objective function R 2 according to Equation (8).
10:
  Let v = v + 1
11:
end while

4. Experiment and Analysis

In order to verify the effectiveness of LRDA, we experiment on 6 visual databases as benchmark domain databases, and compare with existing related methods.

4.1. Datasets

Table 1 visually presents detailed information about the six datasets used in the experiment. Next, we will introduce each dataset.
1.
Office31 [27] contains images from Amazon (A), DSLR (D), and Webcam (W). There are 31 categories. Referring to [28], we use the characteristics of fine-tuning on AlexNet-FC7.
2.
Office-Caltech contains 10 categories of images between Caltech-256 (C) [29] and Office31, and each type contains more than 80 images. Among them, the number of images from sources C, A, W, and D are 1123, 958, 295, and 157. Office-Caltech provides SURF features and DeCAF features [30].
3.
Office-Home [31] includes the four following domains: art (sketching, painting, decoration, and other forms, A), clipart (collection of clipping images, Cl), products (object images without a background, Pr), and real-world (object images taken with ordinary cameras, RW). Each domain consists of 65 categories.
4.
PIE [32] involves the pose, illumination, and expressions of faces, including 68 people with 13 attitudes. Referring to [33], we select C05 (left), C07 (upward), C09 (down), C27 (front), and C29 (right).
5.
MNIST-UPS is composed of handwritten digital datasets MNIST [34] and UPS [35]. The experiments in this article are consistent with [35]; we select 2000 and 1800 images from MNIST and UPS, and the size of each image is 16 × 16.
6.
COIL20 [36] contains 20 objects, each rotating 360 ° horizontally and capturing an image every 5 ° , obtaining 72 images per object. According to the direction of the shooting target, the database is divided into two subsets, namely COIL1 and COIL2. Specifically, COIL1 includes all images captured along the [0; 85] and [180; 265] directions, while COIL2 includes the remaining directions.

4.2. Benchmark Methods & Experimental Settings

The proposed algorithm can be applied to both shallow and deep models, and it is a universal optimization paradigm. Therefore, classic traditional algorithms and domain adaptation methods with good performance in recent years have been selected for comparison to verify the effectiveness of the algorithm. Specifically, we compare LRDA with shallow and deep unsupervised domain adaptation methods. The shallow methods are: SVM, GFK [37], JDA [33], DIP [38], CDDA [39], JGSA [30], SA [40], CORAL [28], ATI [41], DICE [42]. The deep methods are: DDC [43], DAN [44], DANN [45], DRCN [46], RTN [47], WDAN [48], JAN [49], ADDA [50], AutoDIAL [47], CAN [51], SFDA [9], GVB-GD [52], GSDA [53], SRDC [54], and DGA-DA [55]. In addition, we rerun public code of JGSA [30] and CORAL [28] and replicate the code of DGA-DA. The original experimental results of other methods were obtained from corresponding papers. We run the CORAL source code in LIBSVM and selected the optimal variant DICESVM method for its performance.
Parameter Setting: In the method of LRDA, there are three hyper parameters need to be set, which are used to balance the importance of residual classifiers, feature selection, and feature distribution alignment. Therefore, these three parameters have a crucial impact on the final performance of the algorithm. How to set is still an unresolved issue in machine learning. Referring to the work of Tao et al. [56], we use grid search strategy to solve this problem. Specifically, we adjust hyper parameters within the range of grids { 10 6 , 10 5 ,…, 10 5 , 10 6 } to find the parameter combination with the highest accuracy. The iteration limit was fixed at 100. At the same time, when the result of the objective function is less than the threshold ε = 0.01 , the model training is completed and optimization is stopped. In order to achieve optimal results, all experimental algorithm parameters will be searched and adjusted.
Performance Evaluation: Due to target domain data without label, standard cross validation cannot be performed. To address this issue, we conduct p-fold cross validation on labeled source domain data, which trains using p-1 fold source domain data and target domain data, then calculates the average accuracy of all category data on 1-fold source domain data. The optimal parameters are obtained when the accuracy is highest. We repeated each experiment three times and calculated the average accuracy.

4.3. Results of Cross-Domain Recognition

In this section, we compare LRDA with other unsupervised adaptive peer methods. In the experimental results, the left side of → represents the source domain and the right side represents the target domain.

4.3.1. Results and on the Office31 Dataset

The comparison between LRDA and shallow domain adaptation method on the Office31 dataset is shown in Table 2. LRDA can basically achieve the optimal recognition accuracy in subtasks. The results of deep domain adaptation methods are shown in Table 3. It is worth noting that whether compared with shallow models or deep models, the recognition accuracies of LRDA in tasks A→D and W→D are slightly lower than comparison methods. The reason is LRDA needs a certain amount of target domain data to provide information for model decision-making, resulting in a greater preference for source domain data in the model. Overall, LRDA generally achieves good performance on both shallow and deep models, achieving 76.6% and 77.8% respectively, indicating its good applicability. In the comparison of deep models, LRDA does not have obvious advantages. However, deep models need a lot of extra training images and computing resource. This is not available in many practical applications and advanced computing platforms.

4.3.2. Results on the Office-Caltech Dataset

This section conducts cross-domain recognition experiments on Office-Caltech, with reference settings [8,9]. The results with SURF and DeCAF6 features are shown in Table 4 and Table 5. The LRDA achieves the highest recognition accuracy among 10 cross-domain recognition tasks with SURF features, only on C→D and W→D tasks are slightly lower than DICE and DGA-DA. In addition, LRDA achieves the highest average accuracy, demonstrating the effectiveness of the method. In experiments with DeCAF6 features, accuracy of LRDA is above 85% in all settings and achieves the best results in average accuracy particularly. Comparing with RTN based on the deep learning model AlexNet, LRDA achieves higher accuracy in all five experimental settings, with a higher accuracy of 5.4% in the D→C compared to RTN, resulting in significant performance improvement. This is due to the significant distribution differences between images of different concepts, it is difficult to align the fields. LRDA can further explore common features between different fields, thereby enhancing the ability of discriminators.

4.3.3. Results on the Office-Home Dataset

Table 6 shows the cross-domain recognition results on Office-Home dataset with ResNet101-P5 features. We fine-tune the parameters on the ResNet101 and obtain the results by inputting the extracted fifth feature pooling layer into the unsupervised domain for adaptive learning. Notice that LRDA achieves the best recognition performance in five domain recognition tasks, with an average recognition accuracy rate 0.2% lower than DICE. We consider that DICE excludes the features of different types of data while incorporating the distribution of similar data. Table 7 shows the cross-domain recognition results on Office-Home with VGG-F7 features. This experiment compares both shallow and deep models, and LRDA is the second-best performing method in terms of average accuracy, being only inferior to JAN. Counter-intuitively, LRDA achieves better result even than deep methods (e.g., DAN and DANN). In addition, the performance of GSDA, GVB-GD, SRDC, and SFDA methods is superior to that of LRDA methods in some tasks. The results indicate that above methods consider the differences between different categories of data or maintain the distribution structure of intra class data when extracting domain features, thereby improving the performance of the above methods. Nevertheless, LRDA still has good performance, especially achieving better performance than deep methods, indicating that shallow methods have advantages in domain adaptation tasks, reducing algorithm complexity and improving model efficiency while ensuring algorithm accuracy.

4.3.4. Results on the PIE Dataset

In the task of cross-domain facial recognition, this experiment uses two data preprocessing methods ( l 2 normalization and z-score standardization) [42]. The experimental settings divide into two groups, and the recognition accuracies are shown in Table 8 and Table 9. In the experiment using the l 2 normalization (Table 8), we can clearly find that LRDA achieves the highest recognition accuracy in four tasks compared to shallow and deep methods. Moreover, LRDA has the highest average accuracy. In the experiment of z-score standardization, LRDA achieves recognition performance comparable to DICE and far superior to other recognition methods. Affected by the features in the background area in facial data, LRDA focuses on how to effectively convert features between the source and target domains, possibly aligning interference features in the background area. However, the recognition accuracy of LRDA on the PIE dataset is still comparable.

4.3.5. Results on the MNIST-USPS & COIL20 Dataset

In order to further demonstrate the advantages of the proposed method in cross-domain adaptation tasks of digital images, Table 10 shows the cross-domain recognition performance on the MNIST-UPS dataset and COIL20. As can be seen from the first three rows, LRDA achieves the highest average accuracy, as well as U→M task. Nevertheless, the performance of LRDA is slightly lower than that of DGA-DA and DICE on the M→U task, possibly due to the significant differences between these two fields. In the M→U task, the feature distribution is difficult to close, and the feature projection methods of DGA-DA and DICE are more suitable for this type of data. The comparison results in last three rows of Table 9 indicate that the performance of LRDA is close to 100%, only slightly lower than DGA-DA, with a recognition accuracy reduction of approximately 0.5%. This is due to the exclusion of heterogeneous data by the DGA-DA method. Overall, LRDA is still significantly superior to other methods.

4.4. Convergence Analysis

In this section, we explore the convergence of the algorithm. LRDA adapts iterative learning strategy for parameter optimization to achieve parameter update optimization. Experiments use a computer with a memory of 64GB and an Intel i5-9600k CPU, use MATLAB (R2024a) as experimental platform. Figure 3 shows the convergence curve of LRDA. Three sets of experiments use Amazon (A), DSLR (D) and Webcam (W) as the source and target domains in Office31 dataset. The result shows that after 40–50 iterations of optimization, the algorithm achieves convergence in all three sets of experiments, indicating that the objective function of the model can be minimized through parameter iterative learning. It proves that the LRDA method has good convergence and can be optimized through parameter iteration.

4.5. Complexity Comparison

The domain adaptation method requires the extraction of features from two data domains, which lead to an increase in computation time. Table 11 compares the computational time of 8 domain adaptation methods on the Office13 dataset, with the experimental setup being the same as Section 4.4. The experimental result indicates that because of LRDA’s need for iterative optimization, its time cost is relatively high, which also occurs in DICE, CDDA, and CORAL. Because of its closed form, the computation time of GFK, DIP, and SA are relatively low. JDA needs feature decomposition of data; its computation time is increasing with the increase in data volume. However, according to the experiment in Section 4.3, although LRDA requires a longer computation time, its performance is higher than other methods.

5. Conclusions

In this study, we explore the mismatched distribution of latent space features and neglect the association between the original space and latent space. In response to these problems, we propose a latent space domain adaptation with residual model. LRDA reduces the difference in feature distribution in the shared latent space to enhance the feature expression. The objective function constructs the original feature space and the shared feature space as residual model to facilitate parameter optimization, improve model performance. The experiments compare LRDA with the-state-of-art shallow and deep domain adaptation algorithms on six open datasets, verify the proposed method is comparable.
From a broader perspective of cross-domain learning, this study provides a new solution to the multi-domain adaptive problem. Although the current model is mainly aimed at two domain adaptive scenarios, its residual modeling idea can be extended to multi-domain transfer learning, which is expected to realize knowledge transfer from multiple source domains to target domains by constructing hierarchical residual connections or designing multi spatial alignment mechanisms. In addition, in multi-modal data scenarios (such as vision text cross-modal learning), this method can provide the possibility of knowledge transfer between heterogeneous modes through mode-specific feature space decoupling and shared potential space modeling. It should be noted that due to the need to calculate the inter domain decision function, this method increases the time cost of the algorithm, and the balance between accuracy and time consumption is still a contradiction to be solved. In follow-up studies, we will further examine (1) how to batch train adaptive models in the field of large-scale data to reduce the time cost; (2) explore the automatic optimization scheme of super parameter settings (such as the meta learning strategy) to reduce the burden of grid search; and (3) expand the universality of the model in multi-domain migration and multi-modal scenarios, as well as promote cross-domain learning to more complex practical applications.

Author Contributions

Conceptualization, B.Z. and J.P.; data curation, B.Z.; formal analysis, B.Z. and J.P.; funding acquisition, B.Z. and J.P.; investigation, B.Z. and F.Y.; methodology, B.Z. and F.Y.; project administration, B.Z. and J.P.; resources, J.P. and Z.Z.; software, B.Z. and Z.Z.; supervision, F.Y.; validation, J.P.; visualization, Z.Z. and J.P.; writing—original draft, B.Z.; writing—review and editing, B.Z., J.P. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the key projects of Ningbo Education Science Planning in 2025 no. 2025YZD023 and Ningbo Natural Science Foundation no. 2023J242.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study is openly available in Office31, Caltech-256, Office-Home, PIE, MNIST-UPS and COIL20, reference number [27,29,31,32,34,35,36].

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Huang, J.; Smola, A.; Gretton, A.; Borgwardt, K.; Scholkopf, B. Correcting sample selection bias by unlabeled data. In Advances in Neural Information Processing Systems 19: Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2007. [Google Scholar]
  2. Sugiyama, M.; Storkey, A.J. Mixture regression for covariate shift. In Advances in Neural Information Processing Systems 19: Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2007. [Google Scholar]
  3. Sugiyama, M.; Nakajima, S.; Kashima, H.; Buenau, P.; Kawanabe, M. Direct importance estimation with model selection and its application to covariate shift adaptation. In Advances in Neural Information Processing Systems 20, Proceedings of the Twenty-First Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–6 December 2007; Curran Associates, Inc.: Red Hook, NY, USA, 2008. [Google Scholar]
  4. Ghifary, M.; Balduzzi, D.; Kleijn, W.B.; Zhang, M. Scatter component analysis: A unified framework for domain adaptation and domain generalization. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 39, 1414–1430. [Google Scholar] [CrossRef]
  5. Evgeniou, T.; Micchelli, C.A.; Pontil, M.; Shawe-Taylor, J. Learning multiple tasks with kernel methods. J. Mach. Learn. Res. 2005, 4, 615–637. [Google Scholar]
  6. Duan, L.; Tsang, I.W.; Xu, D. Domain transfer multiple kernellearning. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 465–479. [Google Scholar] [CrossRef]
  7. Wang, S.; Wang, B.; Zhang, Z.; Heidari, A.A.; Chen, H. Class-aware sample reweighting optimal transport for multi-source domain adaptation. Neurocomputing 2024, 523, 213–223. [Google Scholar] [CrossRef]
  8. Rostami, M.; Rostami, M.; Bose, D.; Narayanan, S.; Galstyan, A. Domain adaptation for sentiment analysis using robust internal representations. In Findings of the Association for Computational Linguistics: EMNLP; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 11484–11498. [Google Scholar]
  9. Ding, N.; Xu, Y.; Tang, Y.; Xu, C.; Wang, Y.; Tao, D. Source-free domain adaptation via distribution estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 7212–7222. [Google Scholar]
  10. Ge, C.; Huang, R.; Xie, M.; Lai, Z.; Song, S.; Li, S.; Huang, G. Domain adaptation via prompt learning. IEEE Trans. Neural Netw. Learn. Syst. 2023, 36, 1160–1170. [Google Scholar] [CrossRef] [PubMed]
  11. Kay, J.; Haucke, T.; Stathatos, S.; Deng, S.; Young, E.; Perona, P.; Beery, S.; Van Horn, G. Align and distill: Unifying and improving domain adaptive object detection. arXiv 2024, arXiv:2403.12029. [Google Scholar] [CrossRef]
  12. Gheisari, M.; Baghshah, M.S. Unsupervised domain adaptation via representation learning and adaptive classifier learning. Neurocomputing 2015, 165, 300–311. [Google Scholar] [CrossRef]
  13. Zheng, V.W.; Pan, S.J.; Yang, Q.; Pan, J.J. Transferring Multi-device Localization Models using Latent Multi-task Learning. In Proceedings of the 23rd National Conference on Artificial Intelligence, Chicago, IL, USA, 13–17 July 2008; pp. 1427–1432. [Google Scholar]
  14. Blitzer, J.; McDonald, R.; Pereira, F. Domain adaptation with structural correspondence learning. In Proceedings of the 2006 Conference on Empirical Methods in Natural Language Processing, Sydney, Australia, 22–23 July 2006; pp. 120–128. [Google Scholar]
  15. Jiang, L.; Hauptmann, A.G.; Xiang, G. Leveraging high-level and low-level features for multimedia event detection. In Proceedings of the 20th ACM international conference on Multimedia, Nara, Japan, 29 October–2 November 2012; pp. 449–458. [Google Scholar]
  16. Xu, H.; Guo, H.; Yi, L.; Ling, C.; Wang, B.; Yi, G. Revisiting Source-Free Domain Adaptation: A New Perspective via Uncertainty Control. In Proceedings of the The Thirteenth International Conference on Learning Representations, Vienna, Austria, 7–11 May 2024. [Google Scholar]
  17. Tao, J.; Xu, H. Discovering domain-invariant subspace for depression recognition by jointly exploiting appearance and dynamics feature representations. IEEE Access 2019, 7, 186417–186436. [Google Scholar] [CrossRef]
  18. Wu, Y.; Li, Z.; Wang, C.; Zheng, H.; Zhao, S.; Li, B.; Tao, D. Domain re-modulation for few-shot generative domain adaptation. Adv. Neural Inf. Process. Syst. 2024, 36, 57099–57124. [Google Scholar]
  19. Aimei, D.; Shitong, W. A Shared Latent Subspace Transfer Learning Algorithm Using SVM. Acta Autom. Sin. 2014, 40, 2276–2287. [Google Scholar]
  20. Yao, Z.; Tao, J. Multi-source adaptation multi-label classification framework via joint sparse feature selection and shared subspace learning. Comput. Eng. Appl. 2017, 53, 88–96. [Google Scholar]
  21. Zhang, Y.; Tao, J.; Yan, L. Domain-Invariant Label Propagation With Adaptive Graph Regularization. IEEE Access 2024, 12, 190728–190745. [Google Scholar] [CrossRef]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  23. Shi, X.; Guo, Z.; Lai, Z.; Yang, Y.; Bao, Z.; Zhang, D. A framework of joint graph embedding and sparse regression for dimensionality reduction. IEEE Trans. Image Process. 2015, 24, 1341–1355. [Google Scholar] [CrossRef] [PubMed]
  24. Ma, Z.; Nie, F.; Yang, Y.; Uijlings, J.R.R.; Sebe, N. Web image annotation via subspace-sparsity collaborated feature selection. IEEE Trans. Multimed. 2012, 14, 1021–1030. [Google Scholar] [CrossRef]
  25. Gretton, A.; Borgwardt, K.M.; Rasch, M.J.; Bernhard, S.; Smola, A.J. A kernel method for the two-sample-problem. In Advances in Neural Information Processing Systems 19: Proceedings of the Twentieth Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 4–7 December 2006; MIT Press: Cambridge, MA, USA, 2007. [Google Scholar]
  26. Nie, F.; Huang, H.; Cai, X.; Ding, C.H.Q. Efficient and Robust Feature Selection via Joint L2,1-Norms Minimization. In Proceedings of the 24th International Conference on Neural Information Processing Systems, Vancouver, BC, USA, 6–9 December 2010; pp. 1813–1821. [Google Scholar]
  27. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. In Computer VisionECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010, Proceedings, Part IV 11; Springer: Berlin/Heidelberg, Germany, 2010; pp. 213–226. [Google Scholar]
  28. Sun, B.; Feng, J.; Saenko, K. Return of frustratingly easy domain adaptation. In Proceedings of the Thirtieth AAAI Conference on Artificial Intelligence, Phoenix, AZ, USA, 12–17 February 2016. [Google Scholar]
  29. Griffin, G.; Holub, A.; Perona, P. Caltech-256 Object Category Dataset; California Institute of Technology: Pasadena, CA, USA, 2007. [Google Scholar]
  30. Zhang, J.; Li, W.; Ogunbona, P. Joint geometrical and statistical alignment for visual domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1859–1867. [Google Scholar]
  31. Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
  32. Sim, T.; Baker, S.; Bsat, M. The CMU pose, illumination and expression database. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 1615–1618. [Google Scholar] [CrossRef]
  33. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer feature learning with joint distribution adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar]
  34. LeCun, Y.; Bottou, L. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  35. Hull, J.J. A database for handwritten text recognition research. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 550–554. [Google Scholar] [CrossRef]
  36. Nene, S.A.; Nayar, S.K.; Murase, H. Columbia Object Image Library (Coil-20). Technical Report CUCS-005-96. 1996. Available online: https://www.cs.columbia.edu/CAVE/software/softlib/coil-20.php (accessed on 27 July 2025).
  37. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2066–2073. [Google Scholar]
  38. Baktashmotlagh, M.; Harandi, M.T.; Lovell, B.C.; Salzmann, M. Unsupervised domain adaptation by domain invariant projection. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 769–776. [Google Scholar]
  39. Luo, L.; Wang, X.; Hu, S.; Wang, C.; Tang, Y.; Chen, L. Close yet distinctive domain adaptation. arXiv 2017, arXiv:1704.04235. [Google Scholar] [CrossRef]
  40. Fernando, B.; Habrard, A.; Sebban, M.; Tuytelaars, T. Unsupervised visual domain adaptation using subspace alignment. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2014; pp. 2960–2967. [Google Scholar]
  41. Busto, P.P.; Gall, J. Open set domain adaptation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 754–763. [Google Scholar]
  42. Liang, J.; He, R.; Sun, Z.; Tan, T. Aggregating randomized clustering-promoting invariant projections for domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 41, 1027–1042. [Google Scholar] [CrossRef]
  43. Tzeng, E.; Hoffman, J.; Zhang, N.; Saenko, K.; Darrell, T. Deep domain confusion: Maximizing for domain invariance. arXiv 2014, arXiv:1412.3474. [Google Scholar] [CrossRef]
  44. Long, M.; Wang, J. Learning transferable features with deep adaptation networks. In Proceedings of the International Conference on Machine Learning, PMLR, Lille, France, 6–11 July 2015; pp. 97–105. [Google Scholar]
  45. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2030–2096. [Google Scholar]
  46. Ghifary, M.; Kleijn, W.B.; Zhang, M.; Balduzzi, D.; Li, W. Deep reconstruction-classification networks for unsupervised domain adaptation. In Computer VisionECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016, Proceedings, Part IV 14; Springer: Cham, Switzerland, 2016; pp. 597–613. [Google Scholar]
  47. Long, M.; Wang, J.; Jordan, M.I. Unsupervised domain adaptation with residual transfer networks. In Proceedings of the Advances in Neural Information Processing Systems 29: Annual Conference on Neural Information Processing Systems 2016, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  48. Yan, H.; Ding, Y.; Li, P.; Wang, Q.; Xu, Y.; Zuo, W. Mind the class weight bias: Weighted maximum mean discrepancy for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2272–2281. [Google Scholar]
  49. Long, M.; Zhu, H.; Wang, J.; Jordan, M.I. Deep transfer learning with joint adaptation networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 2208–2217. [Google Scholar]
  50. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7167–7176. [Google Scholar]
  51. Kang, G.; Jiang, L.; Wei, Y.; Yang, Y.; Hauptmann, A. Contrastive adaptation network for single-and multisource domain adaptation. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 1793–1804. [Google Scholar] [CrossRef] [PubMed]
  52. Cui, S.; Wang, S.; Zhuo, J.; Su, C.; Tian, Q. Gradually vanishing bridge for adversarial domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 12455–12464. [Google Scholar]
  53. Hu, L.; Kan, M.; Shan, S.; Chen, X. Unsupervised domain adaptation with hierarchical gradient synchronization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4043–4052. [Google Scholar]
  54. Tang, H.; Chen, K.; Jia, K. Unsupervised domain adaptation via structurally regularized deep clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8725–8735. [Google Scholar]
  55. Luo, L.; Chen, L.; Hu, S.; Lu, Y.; Wang, X. Discriminative and geometry aware unsupervised domain adaptation. IEEE Trans. Cybern. 2020, 50, 3914–3927. [Google Scholar] [CrossRef]
  56. Tao, J.; Dan, Y.; Zhou, D.; He, S. Robust latent multi-source adaptation for encephalogram based emotion recognition. Front. Neurosci. 2022, 16, 850–906. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison diagram with the existing latent space domain adaptive algorithm. LRDA achieves feature reuse by sharing latent space transfer features, constructs a cross-domain residual learning model with latent space as residual connections to improve feature propagation efficiency, and introduces feature space normalization to generate sparse representations to enhance model robustness.
Figure 1. Comparison diagram with the existing latent space domain adaptive algorithm. LRDA achieves feature reuse by sharing latent space transfer features, constructs a cross-domain residual learning model with latent space as residual connections to improve feature propagation efficiency, and introduces feature space normalization to generate sparse representations to enhance model robustness.
Information 16 00852 g001
Figure 2. Schematic diagram of residual model. (a) Residual network, (b) latent space residual model.
Figure 2. Schematic diagram of residual model. (a) Residual network, (b) latent space residual model.
Information 16 00852 g002
Figure 3. Convergence curve of LRDA algorithm.
Figure 3. Convergence curve of LRDA algorithm.
Information 16 00852 g003
Table 1. Details of six domain adaptation datasets.
Table 1. Details of six domain adaptation datasets.
Database Subset Image FeatureClasses
Office31AmazonA2817 31
DSLRD498AlexNet-FC7 (4096)
WebcamW795
Office-CaltechAmazonA958 10
CaltechC1123SURF (800)
DSLRD157DeCAF6 (4096)
WebcamW295
Office-HomeArtA2421 65
ClipartCl4379ResNet101-P5
ProductPr4428VGG-FC7
Real-WorldRw4357
PIEC05P13332 68
C06P21629
C07P31632Pixel (1024)
C08P43329
C09P51632
MNIST-USPSMNISTM2000VGG-FC710
USPSU1800
COIL20COIL1C1720VGG-FC720
COIL2C2720
Table 2. Comparison of results with shallow DA methods on the Office31.
Table 2. Comparison of results with shallow DA methods on the Office31.
DATASVMGFKJDADIPCDDAJGSASACORALATIDICELRDA
A→D57.858.266.55664.167.559.460.470.368.567.8
A→W56.959.468.851.965.262.357.75768.772.574.5
D→A47.245.956.3445555.647.247.655.358.159
D→W95.895.697.795.397.298.195.196.29597.298.4
W→A45.543.853.542.353.85246.546.356.960.360.8
W→D98.698.699.698.899.899.8999998.710098.9
Avg.67.066.973.764.772.572.567.567.874.276.176.6
Table 3. Comparison of results with deep DA methods on the Office31.
Table 3. Comparison of results with deep DA methods on the Office31.
DATADDCDANDANNDRCNRTNWDANJANADDAAutoDIALCANGVB-GDGSDASRDCSFDALRDA
A→D64.46772.366.87164.571.872.473.674.67574.875.87674.4
A→W61.868.57368.773.366.874.975.175.579.374.875.775.774.276.8
D→A52.15453.45650.553.858.358.858.160.153.453.556.756.658.7
D→W959696.496.496.895.996.69796.69798.799.199.298.599.1
W→A52.253.151.254.95152.75557.359.458.753.754.957.155.558.8
W→D98.59999.29999.698.799.599.699.59710010010099.899.5
Avg.70.672.974.373.673.772.17676.777.177.875.976.377.476.877.9
Table 4. Comparison of results on Office-Caltech with SURF feature.
Table 4. Comparison of results on Office-Caltech with SURF feature.
DATAGFKJDADIPOTGLCDDAJGSASACORALDICEDGA-DALRDA
A→C43.639.44034.639.641.544.345.144.338.045.2
A→D44.639.536.938.938.245.236.339.552.242.052.1
A→W45.138.035.93746.145.138.344.453.252.254.1
C→A51.644.840.844.249.453.154.854.353.850.354.6
C→D43.345.245.244.549.748.445.236.351.652.950.9
C→W44.141.737.338.938.648.544.438.654.945.855.2
D→A39.133.131.537.232.838.739.437.742.637.143.7
D→C31.731.530.632.433.730.334.333.834.233.534.5
D→W85.489.583.781.189.893.285.184.784.191.291.2
W→A31.832.827.639.436.740.836.335.939.532.440.5
W→C34.331.228.836.03233.633.233.738.631.238.7
W→D87.989.291.784.091.188.583.486.687.391.788.7
Avg.48.546.344.245.748.150.647.947.653.049.854.1
Table 5. Comparison of results on Office-Caltech with DeCAF6 feature.
Table 5. Comparison of results on Office-Caltech with DeCAF6 feature.
DATAGFKJDADIPOTGLCDDAJGSADGA-DASACORALDICEJDOTATIRTNLRDA
A→C82.18578.985.585.88586.78583.687.685.286.587.886.8
A→D82.286.680.98586.685.489.287.384.791.187.992.892.991.5
A→W73.283.169.883.181.484.886.476.674.288.184.888.793.892.1
C→A92.19189.892.291.691.892.392.291.293.491.593.893.293.5
C→D92.487.984.787.391.792.492.488.587.995.589.889.693.993.4
C→W84.482.472.284.288.885.189.88280.795.388.893.696.696.9
D→A88.49185.392.392.192.392.485.683.892.588.193.493.693.2
D→C80.385.175.584.186.385.886.576.971.688.584.385.983.488.8
D→W9999.798.696.39998.69996.997.69996.698.998.698.4
W→A84.391.172.490.690.491.490.783.672.191.190.793.692.792.9
W→C76.585.370.381.585.584.785.674.367.48882.686.384.885.4
W→D10010099.496.310010010099.410010098.1100100100
Avg.86.289.181.588.289.989.890.985.782.992.58991.992.692.7
Table 6. Comparison of results with ResNet101-P5 features on the Office-Home.
Table 6. Comparison of results with ResNet101-P5 features on the Office-Home.
DATAGFKJDADIPCDDAJGSADGA-DACNNSACORALDICELRDA
Ar→Cl35.840.535.540.840.840.83636.736.342.642.5
Ar→Pr54.458.954.357.758.257.753.754.754.161.160.9
Ar→Rw6567.564.966.367.566.364.665.365.368.368.6
Cl→Ar39.240.839.541.340.841.339.239.939.243.343.1
Cl→Pr48.451.94851.75251.748.948.347.954.354.7
Cl→Rw51.855.251.953.954.653.951.751.451.557.157.5
Pr→Ar42.345.141.946.145.346.141.241.441.548.348.4
Pr→Cl32.533.332.135.433.535.432.93332.735.935.5
Pr→Rw64.167.263.76666.46663.264.464.169.268.8
Rw→Ar58.158.858.259.158.559.157.957.857.760.259.3
Rw→Cl39.544.239.245.343.845.339.54039.546.246.0
Rw→Pr69.672.469.671.672.471.668.969.569.373.572.7
Avg.50.15349.952.952.852.949.850.249.95554.8
Table 7. Comparison of results with VGG-F7 features on the Office-Home.
Table 7. Comparison of results with VGG-F7 features on the Office-Home.
DATAGSDAGVB-GDSRDCSFDAJANCDDAJGSADICEDANDANNLRDA
Ar→Cl61.35752.359.745.942.641.943.543.645.644.4
Ar→Pr76.174.776.379.561.262.363.164.25759.364.2
Ar→Rw79.479.88182.468.969.270.170.567 970.170.6
Cl→Ar65.464.669.569.750.446.346.546.745.84746.1
Cl→Pr73.374.176.278.659.752.954.154.656.558.556.9
Cl→Rw74.374.67879.26157.85959.460.460.959.5
Pr→Ar6565.268.766.145.850.550.451.84446.150.7
Pr→Cl53.255.153.857.243.4434243.943.643.744.1
Pr→Rw808181.782.670.369.670.571.867.768.569.8
Rw→Ar72.274.676.373.963.963.162.364.163.163.263.7
Rw→Cl60.659.757.160.852.447.946.24851.551.852.4
Rw→Pr83.184.38585.576.873.873.174.374.376.874.7
Avg.70.370.471.372.958.356.656.657.756.357.658.1
Table 8. Recognition accuracies on the PIE Dataset with l 2 regularization.
Table 8. Recognition accuracies on the PIE Dataset with l 2 regularization.
DATADIPOTGLCDDAJGSASACORALDGA-DAGFKJDACDDADICELRDA
P1→P229.959.476.362.232.831.876.442.276.177.585.183.6
P1→P332.858.772.36034.531.972.553.774.17786.986.4
P1→P436.7092.180.643.541.892.169.892.890.395.295.9
P1→P512.748.460.745.122.519.960.843.970.867.871.872.7
P2→P125.861.97768.227.726.67743.279.877.376.877.1
P2→P353.464.477.564.937.33577.55480.281.179.178.6
P2→P450.1087.177.658.559.787.169.190.488.593.593.6
P2→P529.552.764.352.327.125.963.642.668.370.472.371.6
P3→P122.757.980.862.929.125.180.851.678.381.280.180.7
P3→P236.364.772.260.33736.572.25281.28277.780.4
P3→P445.8084.77154.85484.772.192.391.795.194.8
Avg.34.242.676.864.136.835.376.854.080.480.483.283.2
Table 9. Recognition accuracies on the PIE Dataset with z-score standardization.
Table 9. Recognition accuracies on the PIE Dataset with z-score standardization.
DATADIPOTGLCDDAJGSASACORALDGA-DAGFKJDACDDADICELRDA
P3→P520.252.864.351.230.52664.550.970.579.978.178.6
P4→P131.4093.684.452.448.393.472.695.189.796.896.1
P4→P267.5093.283.57069.793.275.894.794.896.695.7
P4→P376.8092.280.872.772.792.280.992.492.194.594.0
P4→P536.507465.948.648.57461.180.885.190.489.8
P5→P114.245.768.153.534.53267.745.364.167.379.478.8
P5→P229.351.365.157.530.930.465.438.974.274.571.472.2
P5→P331.752.670.554.331.932.671.647.775.379.382.782.3
P5→P426.3079.762.345.144.579.759.381.780.389.588.9
Avg.37.122.577.965.946.345.078.059.281.082.386.686.3
Table 10. Recognition accuracies on the MNIST-USPS and COIL20 dataset.
Table 10. Recognition accuracies on the MNIST-USPS and COIL20 dataset.
DATAGFKSAJDADIPCDDAILSJGSADGA-DACORALDICECPDMLRDA
M→U68.667.870.667.176.271.280.482.335.879.782.180.6
U→M50.148.86046.362.154.968.270.836.459.868.871.1
MU-Avg.59.358.365.356.769.16374.376.536.169.875.475.8
C1→C286.486.894.784.691.586.995.499.682.192.598.599.2
C2→C1858593.58493.986.993.999.781.894.498.799.3
C1C2-Avg.85.785.994.184.392.786.994.799.781.993.598.699.3
Table 11. Computation time(s) of each domain adaptation method on Office31.
Table 11. Computation time(s) of each domain adaptation method on Office31.
TaskJDADIPSAGFKDICECDDACORALLRDA
A→D71.265.174.777.6126.3116.664.2114.8
D→W45.554.459.672371.3396.448.492.8
W→A147.1220.3135.5234.7240.9319.1238.2225.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, B.; Pan, J.; Zhang, Z.; Yang, F. Cross-Domain Residual Learning for Shared Representation Discovery. Information 2025, 16, 852. https://doi.org/10.3390/info16100852

AMA Style

Zhao B, Pan J, Zhang Z, Yang F. Cross-Domain Residual Learning for Shared Representation Discovery. Information. 2025; 16(10):852. https://doi.org/10.3390/info16100852

Chicago/Turabian Style

Zhao, Baoqi, Jie Pan, Zhijie Zhang, and Fang Yang. 2025. "Cross-Domain Residual Learning for Shared Representation Discovery" Information 16, no. 10: 852. https://doi.org/10.3390/info16100852

APA Style

Zhao, B., Pan, J., Zhang, Z., & Yang, F. (2025). Cross-Domain Residual Learning for Shared Representation Discovery. Information, 16(10), 852. https://doi.org/10.3390/info16100852

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop