Next Article in Journal
Updating the Landweber Iteration Method for Solving Inverse Problems
Next Article in Special Issue
An Improved Matting-SfM Algorithm for 3D Reconstruction of Self-Rotating Objects
Previous Article in Journal
SMoCo: A Powerful and Efficient Method Based on Self-Supervised Learning for Fault Diagnosis of Aero-Engine Bearing under Limited Data
Previous Article in Special Issue
Stability of Switched Systems with Time-Varying Delays under State-Dependent Switching
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Multi-Source Domain Adaptation Method with Dempster–Shafer Evidence Theory for Cross-Domain Classification

School of Software Engineering, South China University of Technology (SCUT), Guangzhou 510006, China
*
Authors to whom correspondence should be addressed.
Mathematics 2022, 10(15), 2797; https://doi.org/10.3390/math10152797
Submission received: 23 June 2022 / Revised: 2 August 2022 / Accepted: 4 August 2022 / Published: 6 August 2022

Abstract

:
In this era of big data, Multi-source Domain Adaptation (MDA) becomes more and more popular and is employed to make full use of available source data collected from several different, but related domains. Although multiple source domains provide much information, the processing of domain shifts becomes more challenging, especially in learning a common domain-invariant representation for all domains. Moreover, it is counter-intuitive to treat multiple source domains equally as most existing MDA algorithms do. Therefore, the domain-specific distribution for each source–target domain pair is aligned, respectively. Nevertheless, it is hard to combine adaptation outputs from different domain-specific classifiers effectively, because of ambiguity on the category boundary. Subjective Logic (SL) is introduced to measure the uncertainty (credibility) of each domain-specific classifier, so that MDA could be bridged with DST. Due to the advantage of information fusion, Dempster–Shafer evidence Theory (DST) is utilized to reduce the category boundary ambiguity and output reasonable decisions by combining adaptation outputs based on uncertainty. Finally, extensive comparative experiments on three popular benchmark datasets for cross-domain image classification are conducted to evaluate the performance of the proposed method via various aspects.

1. Introduction

Recently, Deep Learning (DL) has made remarkable advances in various fields  [1,2,3,4,5,6,7], especially in classification [8,9,10]. Despite excellent results, the success of deep methods highly relies on: (1) large-scale labeled data for supervised learning and (2) the training and test data meeting the requirement of being Independently Identically Distributed (IID). However, annotation is time-consuming and unaffordable in practice. If a model is trained on a dataset (known as the source domain), but tested on another non-IID dataset (known as the target domain), domain shifts occur and tend to severely degrade the performance of the learned model [11,12]. Therefore, it is necessary to develop models that are trained on the given labeled datasets, but that can generalize well to a non-IID unlabeled dataset.
Domain Adaptation (DA) aims to learn a discriminative model by reducing domain shifts between training and test distributions [13]. DA transfers the given labeled source domain knowledge to tackle the task to the different, but related target domain by learning domain-invariant representation between domains. Most approaches focus on Single-source Domain Adaptation (SDA), where the labeled data from only one single source domain are considered. Many achievements have emerged in this decade [14,15,16,17,18]. For example, DDC [14] adds an adaptation layer to the pre-trained AlexNet model to confuse the feature representation between the single source domain and the target domain. DSAN [16] proposes a novel fine-grained metric function to align the distribution of the single source domain and the target domain. Most of them learn to map the data from both domains into a common feature space to learn domain-invariant representations by minimizing domain distribution discrepancy, so that the source classifier could then be directly applied to target instances.
However, in practice, it is very likely to obtain multiple available source domains, while SDA is not up to employing those source data adequately. Hence, more challenging, Multi-source Domain Adaptation (MDA) is developed to utilize labeled data from multiple source domains with different distributions and has attracted extensive attention these days  [19,20,21]. The most straightforward way is to combine all source domains into one single source domain and, then, directly apply SDA methods to align distributions. Due to the dataset expansion, the methods might improve the performance. However, the improvements might not be sufficient; the more accurate ways are supposed to explore to make full use of source domains.
With the spurt of progress in DL and SDA today, MDA has been gradually developed. However, there are two typical issues with most techniques [22,23,24,25,26,27,28]. (1) Firstly, it is more challenging to learn a common domain-invariant representation for all domains in MDA, because the damages of domain shifts cannot be eliminated even in SDA. Thereby, MDA is processed by aligning the domain-specific distribution for each source–target domain pair. (2) Secondly, multiple source domains are treated as equivalents. However, the benefits of each source domain to the target domain tasks are diverse in reality. The final output should be closer to the adaptation output of the source–target domain pairs with higher credibility. Some studies [29,30] add extra neural network components to measure the credibility (i.e., transferability). In this research study, we employed Subjective Logic (SL) [31] to obtain the uncertainty of every source domain without any addition of the neural network. Regarding source–target domain pairs as witnesses with different credibility (uncertainty), we introduced Dempster–Shafer evidence Theory (DST) to combine all domain-specific adaptation outputs.
As an uncertainty reasoning method, DST can effectively and reliably deal with uncertainty. It relies on Basic Probability Assignment Functions (BPAFs) to measure the initial degree of belief in the occurrence of an event, which is similar to the concept of the “probability” of a random event in probability theory. To generate BPAFs, DST is bridged with MDA and DL by subjective logic.
Our contributions are summarized as follows:
  • A novel multi-source domain adaptation method with Dempster–Shafer evidence theory is proposed. We provide an effective cross-domain classification solution without any addition of the neural network.
  • There are few studies combining multi-source domain adaptation and Dempster–Shafer evidence theory as of yet. We explored this kind of research early. In our work, DST is employed to fuse all domain-specific adaptation results and output the final credible results.
  • The effectiveness of our cross-domain classification method is verified by conducting comprehensive experiments on three well-known benchmarks. The experimental results prove that the proposed method has better performance than other compared approaches.
The rest of this paper is organized as follows. Section 2 reviews the related work. In Section 3, the preliminaries are given. Section 4 describes the proposed method in detail. A series of experiments is reported in Section 5 and discussed in Section 6. Finally, Section 7 summarizes this research study.

2. Related Work

2.1. Single-Source Domain Adaptation

Single-source Domain Adaptation (SDA) is bound up with multi-source domain adaptation. SDA aims to generalize a model learned from a labeled source domain to a related unlabeled target domain with a different data distribution by reducing the domain shift. SDA can be roughly divided into three categories according to different alignment strategies. (1) Discrepancy-based approaches utilize different metric schemas to explicitly measure the distance between the source and target domains and diminish the domain shift. Commonly used discrepancy metrics for domain adaptation include Maximum Mean Discrepancy (MMD) [32,33,34], moment matching [35,36], Kullback–Leibler (KL) divergence [37], correlation alignment [38,39], and mixture distance [40]. (2) Adversarial-based approaches align different data distributions by confusing a well-trained discriminator (domain classifier). Many methods [41,42,43,44,45,46] are based on Generative Adversarial Networks (GANs), which align different data distributions by implicitly learning the metric function (i.e., domain discriminator) between the source and target domains. (3) Reconstruction-based approaches assume that reconstructing the target domain from a latent representation by using the source task model can help learn domain-invariant representations. The reconstruction is usually obtained via an auto-decoder [47,48,49] or a GAN discriminator [50,51,52].
In our work, the first kind of approach was chosen and the most widely used discrepancy MMD was employed to align the distributions.

2.2. Multi-Source Domain Adaptation

In practice, available source data often come from several different, but related domains. Multi-source Domain Adaptation (MDA) is developed to make full use of these data. However, multiple source domain data provide much information, but challenge the processing of domain shifts. (1) Based on the assumption that the target domain distribution can be approximated by mixing the source domain distribution [53,54], some MDA methods focus on the weighted combination of source domains. For example, Sun and Shi [22] designed a method to weight the source domain classifiers based on the Bayesian learning principle. Xu et al. [23] proposed a voting method for multiple classifiers, which is based on the output of domain discriminators. (2) In addition, some methods are devised to map all source domains and the target domain to a unified feature space. For instance, MDAN [24] aligns the distribution of source domains with the target domain through multiple domain discriminators. M 3 SDA [25] employs moment matching to align the source–target and source-source domains in a common feature space. HoMM [26] exploits the high-order statistics for domain alignment in a reproducing kernel Hilbert space. (3) Some other methods are based on reconstruction [27,28], which reconstruct multiple source domains into an intermediate single source domain and then directly carry out SDA.
Sadly, the damages of domain shifts cannot be eliminated in SDA. It is more difficult to learn a common domain-invariant representation for all domains in MDA. Following MFSAN [55], the domain-specific distribution and classifier alignment architecture for cross-domain classification has proceeded. However, MFSAN treats every source domain equally. This is counter-intuitive because different source domains help the target task differently. Thus, regarding source–target domain pairs as witnesses with different credibility (uncertainty), DST is employed to combine all domain-specific adaptation results. Specifically, the uncertainty is captured, and BPAFs are generated by using subjective logic.

2.3. Dempster–Shafer Evidence Theory

Dempster–Shafer evidence Theory (DST) was first introduced in the 1960s. Based on the investigation of statistical problems, Arthur P. Dempster introduced the concept of upper and lower probabilities and their combining rules [56]. Then, the form of probability that does not satisfy additivity was defined for the first time [57]. Later, Glenn Shafer reinterpreted the upper and lower probabilities based on the belief function and developed the theory into a general framework for modeling epistemic uncertainty [58]. DST allows beliefs from different sources to be fused with various operators to obtain new beliefs considering all available evidence [59]. Currently, generating the belief function through DL has proven to be successful and efficient [60]. These unique characteristics make DST particularly suitable for information fusion [61,62]. Similar to information fusion, the idea of our MDA method is to combine evidence from multiple sources.

3. Preliminaries

3.1. Unsupervised Multi-Source Domain Adaptation

In this research study, the unsupervised MDA problem is investigated. Let D s = { D s i } i = 1 N denote a collection of N available datasets of source domains, and each labeled source dataset D s i = { ( X s i ( j ) , y s i ( j ) ) } j = 1 n s i with n s i samples is sufficient to train a source domain distribution model. Meanwhile, a target dataset D t = { X t ( j ) } j = 1 n t with n t samples drawn from the target domain D t has no labels to support training a reasonable distribution model. With given D s D t , the general goal of this problem is to train a cross-domain classifier f θ ( x ) , which has a low target risk ϵ t = E x D t [ f θ ( x ) y t ] .
The domain-specific distribution and classifier alignment architecture in MFSAN [55] has proceeded to cross-domain classification. Thus, the domain adaptation model involves the source domain task loss L s , the domain adaptation loss L d , and the classifier constraint loss L r . As shown in (1), λ and γ are trade-off parameters.
L = L s + λ L d + γ L r

3.2. Maximum Mean Discrepancy

Maximum mean discrepancy, inspired by the two-sample test in statistics [63,64], is the most widely used discrepancy to align the distributions in domain adaptation. In general, MMD is interpreted as the maximum value (upper bound) of the expectation difference between two distributions mapped by any function f in a predefined function field F , which is an arbitrary vector in the unit sphere (i.e., f < 1 ) of the reproducing Hilbert space:
MMD [ F , p , q ] : = sup f F E p [ f ( x ) ] E q [ f ( y ) ]
In practice, an estimate of the MMD compares the square distance between the empirical kernel mean embeddings as (3). H is the Reproducing Kernel Hilbert Space (RKHS) endowed with a characteristic kernel k. k means k x s , x t = ϕ x s , ϕ x t , where · , · represents the inner product of vectors and ϕ ( · ) denotes some feature map to map the original samples to the RKHS H .
MMD 2 [ F , X s , X t ] = 1 n s x i D s ϕ x i 1 n t x j D t ϕ x j H 2

3.3. Basic Concepts of DST

The Basic Probability Assignment Function (BPAF) is the fundamental unit of DST, which expresses the initial degree of belief in the proposition. Let Θ be a frame of discernment, which specifies the proposition range. The function m : 2 Θ [ 0 , 1 ] becomes the BPAF when it satisfies (4). If  m ( A ) > 0 , m ( A ) is also called the belief mass, and A is named the focal element.
m ( ) = 0 A Θ m ( A ) = 1
Dempster’s rule ⊕ is at the core of DST, as it provides algorithmic rules for combining two pieces of evidence, as shown in (5). Besides, Dempster’s rule is invoked N 1 times to combine N sets of evidence.
m 1 ( X ) m 2 ( X ) = 0 , X = 1 1 K A i B j = X m 1 A i m 2 B j , X
The definition of conflict factor K, shown in (6), reflects the degree of conflict between m 1 and m 2 , whereby 1 / ( 1 K ) represents the normalization factor. Obviously, Dempster’s rule tries to fuse shared parts from different sources and ignores conflicting beliefs.
K = A i B j = m 1 A i m 2 B j

3.4. Dirichlet Distribution

The Dirichlet distribution is involved in SL, which bridges DL, MDA, and DST. In the context of multi-class classification, SL converts the outputs (from DL and MDA) of the neural networks into the concentration parameter of the Dirichlet distribution and associates it with the belief masses (for DST). Accordingly, DST could combine multi-source evidence after BPAFs are obtained and output the final decision.
If the probability density function of multivariate continuous random variable θ = { θ 1 , θ 2 , , θ k } is (7):
p ( θ α ) = Γ i = 1 k α i i = 1 k Γ α i i = 1 k θ i α i 1
where i = 1 k θ i = 1 , θ i 0 , α i > 0 , i = 1 , 2 , , k , and  Γ ( · ) is the Gamma function. Then, the random variable θ is said to obey the Dirichlet distribution with concentration parameter α and denoted as  θ D i r ( α ) .
Dirichlet distribution θ exists on the  ( k 1 ) -dimensional simplex, as shown in Figure 1.
The most important property of the Dirichlet distribution is that it is the conjugate prior to the multinomial distribution. If θ follows the Dirichlet distribution, its prior probability distribution is p ( θ | α ) = D i r ( θ | α ) and posterior probability distribution is p ( θ | D , α ) = D i r ( θ | α + n ) , where D is the given simplex and n = ( n 1 , n 2 , , n k ) is the observation count of the multinomial distribution. The concentration parameters α = { α 1 , α 2 , , α k } of the Dirichlet distribution as a priori distribution are also called the hyperparameters of the posterior distribution. Hence, it is convenient to obtain the posterior distribution from the prior distribution.

4. Research Methodology

Following the two-stage alignment framework in MFSAN [55], a novel Multi-source domain Adaptation Network with Dempster–Shafer evidence theory (MAN-DS) for cross-domain classification is proposed. MAN-DS aims to train a model based on multi-source domain labeled samples and adapts to classify target instances with different distributions. As shown in Figure 2, the MAN-DS framework consists of four key components, i.e., common feature extractor, domain-specific feature extractor, domain-specific classifier, and Dempster’s combination. Different source domains are extracted into different feature spaces, and then, the distribution alignment of each pair of source and target domains and the output alignment of every source classifier are imposed. Domain-specific adaptation outputs are combined by Dempster’s rule in the end. Besides, the  s o f t m a x layer of the classifier is replaced with an activation layer (e.g., ReLU).

4.1. Common Feature Extractor

The damages of domain shifts cannot be eliminated in SDA, so it is more difficult to learn a common domain-invariant representation for all domains in MDA. To address this problem, the easiest way is to train multiple networks to map each source–target domain pair into a specific feature space. However, this would take too much time and space. Thus, the feature extractor is divided into two parts. The first part extracts common features, and the second part extracts domain-specific features (see the next section). In the first part, a common convolutional neural subnetwork f ( · ) is used to automatically map samples in all domains from the original feature space into a common feature space.

4.2. Domain-Specific Feature Extractor

Now, we come to the second part where domain-specific features are extracted by different extractors. For each pair of source and target domains, a specific subnetwork h i ( · ) aims to map f ( x s i ) and f ( x t ) into the same domain-specific feature space. The objective of domain adaptation is to find a domain-invariant representation between domains. In other words, an  h i ( · ) is desired, which makes the distribution discrepancy between h i ( f ( x s i ) ) and h i ( f ( x t ) ) as small as possible. There are many explicit or implicit methods to achieve this goal. Here, the most widely used MMD is employed to reduce the distribution discrepancy between domains. The MMD loss is reformulated as:
L m m d = 1 N i = 1 N MMD 2 [ F , h i ( f ( X s i ) ) , h i ( f ( X t ) ) ]

4.3. Domain-Specific Classifier

Traditionally, a series of  s o f t m a x classifiers c i ( · ) is employed to classify the source domain samples after extracting domain-specific invariant features, respectively. However, the use of the exponent in the  s o f t m a x function leads to the probability of the predicted category being inflated. It was replaced with an activation function (e.g., RELU) to ensure that the network outputs non-negative values in this research study. The multi-classification problem is a multinomial distribution fitting problem. As the conjugate prior, the Dirichlet distribution is convenient to obtain the posterior distribution from the prior distribution.
Subjective logic [31] defines a theoretical framework for obtaining the probabilities of different classes and the overall uncertainty of the multi-classification problem based on the evidence collected from the data. SL provides an additional mass function, which allows the model to distinguish between a lack of evidence. In our model, SL provides the degree of overall uncertainty of each source, which is important for final decisions to some extent.
For the K-classification problem, the nonnegative-activated output e = ( e 1 , e 2 , , e k ) of the last fully connected layer of the classifier refers to evidence and is closely related to the concentration parameters α = ( α 1 , α 2 , , α + k ) of the Dirichlet distribution, as shown in the following:
α k = e k + 1 , k = 1 , 2 , , K
With subjective logic, for each pair of the source–target domain, the probability b k ( i ) for the kth category and the overall uncertainty u ( i ) are calculated by:
b k ( i ) = e k ( i ) S ( i ) = α k ( i ) 1 S ( i ) u ( i ) = K S ( i )
where S ( i ) = k = 1 K ( e k ( i ) + 1 ) = k = 1 K ( α k ( i ) ) is the Dirichlet strength. Obviously, u ( i ) + k = 1 K b k ( i ) = 1 . Correspondingly, the less total evidence observed, the greater the total uncertainty is. The mean of the corresponding Dirichlet distribution P ^ s i for the probability p ^ i ( k ) is computed as  p ^ i ( k ) = α i ( i ) S ( i ) .
In addition, Figure 3 demonstrates the process of the outputs of multiple domain-specific classifiers in detail. The evidence of each source is obtained using neural networks (Step ①). According to subjective logic [31], the obtained evidence parameterizes the Dirichlet distribution (Step ②) to induce the classification probability and uncertainty (Step ③). The classification probability and overall uncertainty are inferred by combining the belief masses of multiple sources based on Dempster’s rule (Step ④). Dempster’s combining is discussed in Section 4.4.
Source domain task loss L c l s is calculated here. To adapt to the Dirichlet distribution [65], the cross-entropy function is formulated as (11).
L a c e α ( i ) = k = 1 K y i j log p j k 1 B α j k = 1 K p j k α j k ( i ) 1 d p j = k = 1 K y j k ψ S ( i ) ψ α j k ( i )
where ψ ( · ) is the digamma function, the parameter α i of the Dirichlet distribution and forming the multinomial opinions D ( p i α i ) , where p i is the category assignment probabilities on a simplex, and  p j k is the predicted probability of the  j t h sample for category k.
The above loss function ensures that more evidence is generated for the correct label of each sample than for other classes, but there is no guarantee that less evidence is generated for the incorrect label. That is, in MAN-DS, the expected evidence of incorrect labels shrinks to 0 [66]. To this end, the following KL divergence term is introduced:
K L D p j α ˜ j D p j 1 = log Γ k = 1 K α ˜ j k Γ ( K ) k = 1 K Γ α ˜ j k + k = 1 K α ˜ j k 1 ψ α ˜ j k ψ r = 1 K α ˜ j r
Therefore, given parameter α j of the Dirichlet distribution for each sample j, the loss is:
L α ( i ) = j = 1 n s i L α j = j = 1 n s i L a c e α j + ρ K L D p j α ˜ j D p j 1
where ρ > 0 is a balance factor. In practice, ρ increases slowly from zero to 1 to avoid paying too much attention to the KL divergence term in the early stage of learning.
That is, the classification loss is formulated as:
L c l s = i N L α ( i )

4.4. Dempster’s Combination

With subjective logic, there is an FoD Θ = { 1 , 2 , , K } and K + 1 focal elements { { 1 } , { 2 } , , { K } , Θ } with belief mass { b 1 , b 2 , , b k , u } in every source–target domain pair. To fuse these adaptation outputs from N sources, only call Dempster’s rule (defined in (5)) N 1 times as:
m ( b k ) = m 1 ( b k ) m 2 ( b k ) , , m N 1 ( b k )
In addition, the prediction results of multiple classifiers for the same target sample should be consistent. Dempster’s combination could help to avoid ambiguity and large uncertainty on the category boundary, which is demonstrated in Figure 4.
Moreover, the Manhattan distance is used to measure the difference among the classifiers to achieve this goal, as well. Denote e ( i ) = e 1 ( i ) , e 2 ( i ) , , e k ( i ) , e ( i ) = α ( i ) 1 = b ( i ) S ( i ) as the final output of the ith source–target domain pair. The loss-of-label Manhattan distance is formulated as:
L d i s t = 1 N i N | e ( i ) m ( e ) |

4.5. Objective Function and Algorithm

The overall objective function of the proposed model is formulated as (17).
arg min f , h , c ( L c l s + γ L m m d + λ L d i s c )
In detail, L c l s is minimized to accomplish the source domain task; L m m d is minimized to reduce the domain shifts between each source domain and the target domain; L d i s c is a consistent regular term and minimized to constrain the outputs of domain-specific classifiers. In addition, γ and λ are trade-off parameters; refer to (1).
The algorithm of [d=Z.]MAN-DSour method is summarized in Algorithm 1, and it can be trained by the standard back-propagation.
Algorithm 1:The algorithm of the proposed method
Input :
source domain data { D s i } i = 1 N , target domain data D t , the number of training iterations T, and batch size M;
Output :
model parameters;
1:
Initialize the parameters of  f ( · ) , g ( · ) , h i ( · ) , c i ( · ) ;
2:
for t = 1 , , T do
3:
   Randomly sample a batch of  { ( x s i ( j ) , y s i ( j ) ) } j = 1 M from  D s i , respectively;
4:
   Randomly sample a batch of  { x t ( j ) } j = 1 M from  D t ;
5:
   Extract common features f ( x s i ( j ) ) and f ( x t ( j ) ) ;
6:
   Extract domain-specific features h i ( f ( x s i ( j ) ) ) and h i ( f ( x t ( j ) ) ) ;
7:
   Compute L m m d with h i ( f ( x s i ( j ) ) ) and h i ( f ( x t ( j ) ) ) by (8);
8:
   Obtain c i ( h i ( f ( x s i ( j ) ) ) ) for classification and compute L c l s by (14);
9:
   Obtain c i ( h i ( f ( x t ( j ) ) ) ) , and combine them by (5)
10:
   Compute L d i s t by (16);
11:
   Update parameters by (17).
12:
end for

5. Experiment

The effectiveness of our cross-domain classification method was verified by conducting comprehensive experiments on three well-known benchmarks: ImageCLEF-DA, Office-31, and Office-Home.

5.1. Data Preparation

ImageCLEF-DA [67] is a benchmark dataset for the ImageCLEF 2014 domain adaptation challenge, which is organized by selecting the 12 common categories shared by the following three public datasets, each considered as a domain: Caltech-256(C), ImageNet ILSVRC 2012(I), and Pascal VOC 2012 (P). There are 50 images in each category and 600 images in each domain. All domain combinations were used, and three transfer tasks were built: C, IP; C,PI; I,PC.
Office-31 [68] is a benchmark for domain adaptation, comprising 4110 images in 31 classes collected from three distinct domains: Amazon (A), which contains images downloaded from amazon.com, Webcam (W), and DSLR (D), which contains images taken by a web camera and digital SLR camera with different photographic settings. The images in each domain are unbalanced. To enable unbiased evaluation, all methods were evaluated on all three transfer tasks: A, WD; A,WD; W,DA.
Office-Home [69] consists of 15,588 images, larger than Office-31 and ImageCLEF-DA. It consists of images from 4 different domains: Artistic images (A), Clip Art (C), Product images (P), and Real-World images (R). For each domain, the dataset contains images of 65 object categories collected in the office and home settings. All domain combinations were used, and four transfer tasks were built:: A, P, RC; A, P, CR; A, R, CP; P, R, CA.

5.2. Compared Method

There is a small amount of MDA work based on a domain-specific distribution and classifier alignment architecture. To verify the effectiveness of our MDSAN model, the Multiple Feature Spaces Adaptation Network (MFSAN) [55] was introduced as the multi-source baseline. In addition, the proposed method was compared with ResNet [70], Deep Domain Confusion (DDC) [14], the Deep Adaptation Network (DAN) [71], Deep CORAL (DCORAL) [72], and Reverse Gradient (RevGrad) [73].
There are several comparative standards for different purposes. (1) Source combine: all source domains are combined into a traditional single-source vs. target setting; (2) Single best: the best single source transfer results among the multiple candidate source domains with SDA methods; (3) Multi-source: the results of MDA methods. The first standard is to verify whether multiple source domains are beneficial for the target task or whether the simple combination of source domains will lead to negative transfer. In addition, the second standard evaluates whether the best SDA method could be further improved by introducing other source domains. The third standard demonstrates the effectiveness of the proposed approach.
Furthermore, ablation experiments were performed to verify the effectiveness of DST for adaptation outputs’ fusion. This variant is denoted as V 1 , which simply averages the outputs in the end. In addition, variant V 2 does not consider L m m d , and variant V 3 ignores L d i s t .

5.3. Implementation Details

All methods were implemented based on the PyTorch framework and deployed and testified on the same device. For a fair comparison, the same data pre-processing routines and model architecture were utilized in all experiments. The pre-trained ResNet50 [70] was employed as the common feature extractor, where the fine-tuning strategy was used to save time. For all domain-specific feature extractors, the same structure ( c o n v ( 1 × 1 ) , c o n v ( 3 × 3 ) , c o n v ( 1 × 1 ) ) was utilized. At the end of the neural network, the channels were reduced to 256, like DDC [14]. According to subjective logic, the s o f t m a x layer was replaced with s o f t p l u s to activate the outputs and avoid negative values. The optimization method was mini-batch stochastic gradient descent with a momentum of 0.9 . The learning rate was gradually decreased by η p = η 0 ( 1 + α ) β , where p is the training progress linearly changing from 0 to 1, and η 0 = 0.01 , α = 10 , β = 0.75 . This would optimize to promote convergence and low error on the source domain. As for the hyperparameters, γ = ρ = 100 λ was simply set. They were changed from 0 to 1 by a progressive schedule γ p = 2 exp ( θ p ) 1 , ( θ = 10 ) , instead of fixing them throughout the experiments.

5.4. Experimental Results

MAN-DS was compared with the above-mentioned methods on three datasets, and the average results of five repeated experiments are reported in Table 1, Table 2 and Table 3, respectively. The maximum accuracy in a transfer task is marked in bold.

6. Discussion

6.1. Result Observations

From these experimental results, insightful observations are given:
  • The results of Source combine were better than Single best, which shows that the knowledge of the multi-source domain is useful to the target task. That is, the multi-source domains have transferability. Combining sources into a single source is helpful in most domain adaptation methods. The performance improvement might be attributed to the data enrichment.
  • MAN-DS outperformed all compared methods on most transfer tasks in all three datasets, especially in the Office-Home dataset. The results indicate that it is beneficial to learn the domain-invariant representation and align the distribution in each pair of the source and target domain with considering domain-specific category boundaries. Besides, DST alleviates the ambiguity and uncertainty of the prediction and promotes classification accuracy successfully.
  • Comparing MAN-DS with the variant V 1 , the only difference is that the proposed method employs DST to fuse the adaptation outputs, while V 1 averages them simply. Although DST was applied in L d i s t to align domain-specific boundaries, the proposed method still has an improvement over V 1 . Thus, DST is excellent to tackle the ambiguity and uncertainty of the prediction.
  • Comparing MAN-DS with the variant V 2 , the only difference is that V 2 does not consider L m m d . The experimental results show that MMD helps domain adaptation very little. Meanwhile, the proposed L d i s t and Dempster’s combination rule could also help to align the distribution to some extent.
  • Comparing MAN-DS with the variant V 3 , the only difference is that V 3 ignores L d i s t . There is little difference in the experimental results, which indicates that DST is powerful to handle the prediction conflicts on the category boundaries.

6.2. Ablation Experiment

Ablation experiments were implemented by conducting V 1 , V 2 , and V 3 , as shown in Table 1, Table 2 and Table 3. The encouraging results show that every component of MAN-DS is positive to improve performance.
To further verify the effectiveness of the DST fusion strategy, supplementary experiments were carried out where S i is the isth domain-specific classifier, as reported in Table 4. The maximum accuracy in a transfer task is marked in bold.

6.3. Feature Visualization

Feature visualization is demonstrated in Figure 5. The category boundaries of the domain-specific classifier on the task D,WA learned by MAN-DS and MFSAN are visualized by using t-SNE embeddings. It is clear that MAN-DS is more effective in dealing with prediction conflicts, in which DST is effective.

6.4. Parameter Sensitivity

Parameter sensitivity was tested by sampling the trade-off parameter (where γ = ρ = 100 λ for simplicity) values in { 0.01 , 0.02 , 0.05 , 0.1 , 0.2 , 0.5 , 1 , 2 } . To study the parameters’ sensitivity, the experiments were implemented on task D,WA and A,CR, and the results are shown in Figure 6. As observed, the accuracy increases with the increase of γ and reaches a peak at γ = 1 , then decreases. The proposed method MAN-DS can keep a relatively stable result in the range of ( 0.1 , 2 ) of γ , which is higher than the baseline. Generally, MAN-DS is not sensitive to changes in the parameters in a certain range. Hence, setting γ to ( 0.1 , 2 ) is recommended to achieve better performance. In the reported experiment, the parameters { γ , ρ , λ } were set to { 1 , 1 , 0.01 } , respectively.

6.5. Computational Complexity

The FLoating point OPerations (FLOPs) were used to measure the operation times of forward propagation in neural network; the smaller the FLOPs, the faster the computation speed is. In addition, the smaller the number of PARAMeters (PARAMs) in the neural network, the smaller the size of the model is. Table 5 shows the FLOPs and PARAMs of MAN-DS, MFSAN, and ResNet50. Compared with ResNet50, the small increase of computational complexity mainly comes from the component of domain-specific feature extractors and classifiers. Compared with the baseline MFSAN, MAN-DS improves the accuracy without increasing the computational complexity.
Moreover, Dempster’s combination does not increase the computational complexity of the algorithm. For the K-classification task, MAN-DS always obtains K + 1 instead of 2 K focal elements, which is { 1 , 2 , , K , Θ } . That is, the computational complexity caused by Dempster’s combination is not O ( 2 n ) , but O ( n ) .

7. Conclusions

The core of MDA is making full use of available source data collected from several different, but related domains. However, it becomes difficult and challenging due to the multiple domain shifts. Following the domain-specific alignment architecture, this study proposed a novel multi-source domain adaptation network combing Dempster–Shafer evidence theory for cross-domain image classification to reduce multiple domain shifts and enhance transfer accuracy. In addition, SL and the Dirichlet distribution were employed to bridge MDA with DST.
To evaluate the effectiveness of the proposed method, three popular benchmark datasets were used and ten transfer tasks were devised to train and validate MAN-DS. Extensive experiments demonstrated that MAN-DS outperforms its competitors in cross-domain image classification. The insightful conclusions are as follows:
  • MAN-DS achieved good accuracy in all ten transfer tasks of three datasets. On the Office-Home dataset, MAN-DS even improved the average adaptation accuracy to 76.20 % , which is about 2% higher than the best baseline.
  • Feature visualization shows that MAN-DS could alleviate boundary conflicts to some extent, due to effective DST.
  • MAN-DS is not sensitive to changes in parameters in a certain range γ ( 0.1 , 2 ) , generally.
  • MAN-DS improved accuracy without increasing computational complexity. Compared with the baseline MFSAN, the FLOPs and PARAMs of MAN-DS were 4.23 G and 25.88 M, which are close to the 4.12 G and 25.56 M of ResNet. Especially, MAN-DS reduced the computational overhead of the outputs’ combination from O ( 2 n ) to O ( n ) .
  • Ablation experiments indicated that every component of MAN-DS is positive to improve performance.
  • The encouraging results show that SL could effectively bridge MDA with DST.
  • This research study empirically demonstrates DST could reduce the category boundary ambiguity, so as to mitigate the negative impact of multiple domain shifts.
In this research study, the original and unimproved Dempster’s rule was used. In the future, the combination rules will be optimized based on the improved information entropy method to take more evidence information into account. Besides, more effective MDA and DST bridging methods will be investigated.

Author Contributions

Conceptualization, M.H. and C.Z.; methodology, M.H. and C.Z.; software, C.Z.; validation, M.H.; formal analysis, M.H.; investigation, M.H. and C.Z.; resources, M.H.; data curation, M.H.; writing—original draft preparation, C.Z.; writing—review and editing, M.H. and C.Z.; visualization, C.Z.; supervision, M.H.; project administration, M.H.; funding acquisition, M.H. All authors have read and agreed to the published version of the manuscript.

Funding

The work described in this paper was partially funded by two Natural Science Foundation of Guangdong Province Projects of Grant Number 2021A1515011496 and Grant Number 2022A1515011370.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to thank the Editor and the Reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
DADomain Adaptation
SDASingle-source Domain Adaptation
MDAMulti-source Domain Adaptation
SLSubjective Logic
DLDeep Learning
DSTDempster–Shafer evidence Theory
BPAFBasic Probability Assignment Function
MMDMaximum Mean Discrepancy
FLOPsFLoating point OPerations
PARAMsPARAMeters

References

  1. Dai, Z.; Cai, B.; Lin, Y.; Chen, J. Up-detr: Unsupervised pre-training for object detection with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 1601–1610. [Google Scholar] [CrossRef]
  2. Xu, J.; Zhou, H.; Gan, C.; Zheng, Z.; Li, L. Vocabulary learning via optimal transport for neural machine translation. In Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Conference on Natural Language Processing, Virtual, 1–6 August 2021; Volume 1, pp. 7361–7373. [Google Scholar] [CrossRef]
  3. Zhou, H.; Zhang, S.; Peng, J.; Zhang, S.; Li, J.; Xiong, H.; Zhang, W. Informer: Beyond efficient transformer for long sequence time-series forecasting. In Proceedings of the Thirty-Fifth AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021. [Google Scholar]
  4. Huang, M.; Cheng, C.; De Luca, G. Remote Sensing Data Detection Based on Multiscale Fusion and Attention Mechanism. Mob. Inf. Syst. 2021, 2. [Google Scholar] [CrossRef]
  5. Yu, Y.; Rashidi, M.; Samali, B.; Mohammadi, M.; Nguyen, T.N.; Zhou, X. Crack detection of concrete structures using deep convolutional neural networks optimized by enhanced chicken swarm algorithm. Struct. Health Monit. 2022. [Google Scholar] [CrossRef]
  6. Lyu, Z.; Yu, Y.; Samali, B.; Rashidi, M.; Mohammadi, M.; Nguyen, T.N.; Nguyen, A. Back-Propagation Neural Network Optimized by K-Fold Cross-Validation for Prediction of Torsional Strength of Reinforced Concrete Beam. Materials 2022, 15, 1477. [Google Scholar] [CrossRef]
  7. Liu, J.; Mohammadi, M.; Zhan, Y.; Zheng, P.; Rashidi, M.; Mehrabi, P. Utilizing Artificial Intelligence to Predict the Superplasticizer Demand of Self-Consolidating Concrete Incorporating Pumice, Slag, and Fly Ash Powders. Materials 2021, 14, 6792. [Google Scholar] [CrossRef]
  8. Gou, J.; He, X.; Lu, J.; Ma, H.; Ou, W.; Yuan, Y. A Class-Specific Mean Vector-Based Weighted Competitive and Collaborative Representation Method for Classification. Neural Netw. 2022, 150, 12–27. [Google Scholar] [CrossRef]
  9. Gou, J.; Yuan, X.; Du, L.; Xia, S.; Yi, Z. Hierarchical Graph Augmented Deep Collaborative Dictionary Learning for Classification. IEEE Trans. Intell. Transp. Syst. 2022. [Google Scholar] [CrossRef]
  10. Gou, J.; Qiu, W.; Yi, Z.; Xu, Y.; Mao, Q.; Zhan, Y. A Local Mean Representation-Based K-Nearest Neighbor Classifier. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–25. [Google Scholar] [CrossRef]
  11. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  12. Gou, J.; Yu, B.; Maybank, S.J.; Tao, D. Knowledge distillation: A survey. Int. J. Comput. Vis. 2021, 129, 1789–1819. [Google Scholar] [CrossRef]
  13. Zhuang, F.; Qi, Z.; Duan, K.; Xi, D.; Zhu, Y.; Zhu, H.; Xiong, H.; He, Q. A Comprehensive Survey on Transfer Learning. Proc. IEEE 2021, 109, 43–76. [Google Scholar] [CrossRef]
  14. Tzeng, E.; Hoffman, J.; Zhang, N.; Saenko, K.; Darrell, T. Deep Domain Confusion: Maximizing for Domain Invariance. arXiv 2014, arXiv:1412.3474. [Google Scholar]
  15. Chen, S.; Harandi, M.; Jin, X.; Yang, X. Domain Adaptation by Joint Distribution Invariant Projections. IEEE Trans. Image Process. 2020, 29, 8264–8277. [Google Scholar] [CrossRef]
  16. Zhu, Y.; Zhuang, F.; Wang, J.; Ke, G.; Chen, J.; Bian, J.; Xiong, H.; He, Q. Deep Subdomain Adaptation Network for Image Classification. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1713–1722. [Google Scholar] [CrossRef]
  17. Qiu, Z.; Zhang, Y.; Lin, H.; Niu, S.; Liu, Y.; Du, Q.; Tan, M. Source-free Domain Adaptation via Avatar Prototype Generation and Adaptation. In Proceedings of the 30th International Joint Conference on Artificial Intelligence, Virtual, 19–26 August 2021. [Google Scholar] [CrossRef]
  18. Chen, S.; Hong, Z.; Harandi, M.; Yang, X. Domain Neural Adaptation. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–12. [Google Scholar] [CrossRef]
  19. Liu, J.; Li, J.; Lu, K. Coupled local–global adaptation for multi-source transfer learning. Neurocomputing 2018, 275, 247–254. [Google Scholar] [CrossRef] [Green Version]
  20. Yin, Y.; Yang, Z.; Hu, H.; Wu, X. Universal multi-Source domain adaptation for image classification. Pattern Recognit. 2022, 121, 108238. [Google Scholar] [CrossRef]
  21. Renchunzi, X.; Pratama, M. Automatic online multi-source domain adaptation. Inf. Sci. 2022, 582, 480–494. [Google Scholar] [CrossRef]
  22. Sun, S.L.; Shi, H.L. Bayesian multi-source domain adaptation. In Proceedings of the 2013 International Conference on Machine Learning and Cybernetics, Tianjin, China, 14–17 July 2013; Volume 1, pp. 24–28. [Google Scholar] [CrossRef]
  23. Chen, Z.; Wei, P.; Zhuang, J.; Li, G.; Lin, L. Deep CockTail Networks A Universal Framework for Visual Multi-source Domain Adaptation. Int. J. Comput. Vis. 2021, 129, 2328–2351. [Google Scholar] [CrossRef]
  24. Zhao, H.; Zhang, S.; Wu, G.; Costeira, J.A.P.; Moura, J.M.F.; Gordon, G.J. Adversarial multiple source domain adaptation. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Red Hook, NY, USA, 3–8 December 2018; Curran Associates Inc.: Red Hook, NY, USA, 2018; pp. 8568–8579. [Google Scholar]
  25. Peng, X.; Bai, Q.; Xia, X.; Huang, Z.; Saenko, K.; Wang, B. Moment matching for multi-source domain adaptation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV 2019), Seoul, Korea, 28–27 October 2019; IEEE Computer Soc.: Los Alamitos, CA, USA, 2019; pp. 1406–1415. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, C.; Fu, Z.; Chen, Z.; Jin, S.; Cheng, Z.; Jin, X.; Hua, X.S. HoMM: Higher-Order Moment Matching for Unsupervised Domain Adaptation. Proc. AAAI Conf. Artif. Intell. 2020, 34, 3422–3429. [Google Scholar] [CrossRef]
  27. Lin, C.; Zhao, S.; Meng, L.; Chua, T.S. Multi-Source Domain Adaptation for Visual Sentiment Classification. Proc. AAAI Conf. Artif. Intell. 2020, 34, 2661–2668. [Google Scholar] [CrossRef]
  28. Fernandes Montesuma, E.; Mboula, F. Wasserstein barycenter for multi-source domain adaptation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16780–16788. [Google Scholar] [CrossRef]
  29. Zuo, Y.; Yao, H.; Xu, C. Attention-Based Multi-Source Domain Adaptation. IEEE Trans. Image Process. 2021, 30, 3793–3803. [Google Scholar] [CrossRef]
  30. Zhang, D.; Ye, M.; Liu, Y.; Xiong, L.; Zhou, L. Multi-source unsupervised domain adaptation for object detection. Inf. Fusion 2022, 78, 138–148. [Google Scholar] [CrossRef]
  31. Jøsang, A. Subjective Logic: A Formalism for Reasoning Under Uncertainty; Springer: Berlin/Heidelberg, Germany, 2016; Volume 3. [Google Scholar]
  32. Chen, Y.; Song, S.; Li, S.; Wu, C. A Graph Embedding Framework for Maximum Mean Discrepancy-Based Domain Adaptation Algorithms. IEEE Trans. Image Process. 2020, 29, 199–213. [Google Scholar] [CrossRef]
  33. Yan, H.; Li, Z.; Wang, Q.; Li, P.; Xu, Y.; Zuo, W. Weighted and Class-Specific Maximum Mean Discrepancy for Unsupervised Domain Adaptation. IEEE Trans. Multimed. 2020, 22, 2420–2433. [Google Scholar] [CrossRef]
  34. Liu, W.; Li, J.; Liu, B.; Guan, W.; Zhou, Y.; Xu, C. Unified Cross-domain Classification via Geometric and Statistical Adaptations. Pattern Recognit. 2021, 110. [Google Scholar] [CrossRef]
  35. Zhen, Z.; Wang, M.; Yan, H.; Nehorai, A. Aligning infinite-dimensional covariance matrices in reproducing Kernel Hilbert spaces for domain adaptation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  36. Han, C.; Lei, Y.; Xie, Y.; Zhou, D.; Gong, M. Learning smooth representations with generalized softmax for unsupervised domain adaptation. Inf. Sci. 2021, 544, 415–426. [Google Scholar] [CrossRef]
  37. Li, J.; Luo, P.; Lin, F.; Chen, B. Conversational model adaptation via KL divergence regularization. In Proceedings of the Thirty-second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 5213–5219. [Google Scholar]
  38. Sun, B.; Kate, S. Deep CORAL: Correlation Alignment for Deep Domain Adaptation; Springer International Publishing: Berlin/Heidelberg, Germany, 2016. [Google Scholar]
  39. Chen, C.; Jiang, B.; Cheng, Z.; Jin, X. Joint Domain Matching and Classification for cross-domain adaptation via ELM. Neurocomputing 2019, 349, 314–325. [Google Scholar] [CrossRef]
  40. Han, G.; Ramakanth, P.M.B. Multi-source domain adaptation for text classification via distanceNet-bandits. Proc. AAAI Conf. Artif. Intell. 2020, 34, 7830–7838. [Google Scholar]
  41. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; Marchand, M.; Lempitsky, V. Domain-adversarial training of neural networks. J. Mach. Learn. Res. 2016, 17, 2030–2096. [Google Scholar]
  42. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial discriminative domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 15–17 June 2017; pp. 2962–2971. [Google Scholar] [CrossRef] [Green Version]
  43. Yu, C.; Wang, J.; Chen, Y.; Huang, M. Transfer learning with dynamic adversarial adaptation network. Proceedigns of the 2019 IEEE International Conference on Data Mining (ICDM), Beijing, China, 8–14 November 2019; pp. 778–786. [Google Scholar]
  44. Feng, C.; He, Z.; Wang, J.; Lin, Q.; Zhu, Z.; Lu, J.; Xie, S. Domain adaptation with SBADA-GAN and Mean Teacher. Neurocomputing 2020, 396, 577–586. [Google Scholar] [CrossRef]
  45. Chen, W.; Hu, H. Generative attention adversarial classification network for unsupervised domain adaptation. Pattern Recognit. 2020, 107. [Google Scholar] [CrossRef]
  46. Kang, Q.; Yao, S.; Zhou, M.; Zhang, K.; Abusorrah, A. Effective Visual Domain Adaptation via Generative Adversarial Distribution Matching. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 3919–3929. [Google Scholar] [CrossRef]
  47. Ghifary, M.; Kleijn, W.B.; Zhang, M.; Balduzzi, D.; Li, W. Deep reconstruction-classification networks for unsupervised domain adaptation. In Computer Vision—ECCV 2016, PT IV; Lecture Notes in Computer Science; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Springer: Cham, Switzerlands, 2016; Volume 9908, pp. 597–613. [Google Scholar] [CrossRef] [Green Version]
  48. Jiang, B.; Chen, C.; Jin, X. Unsupervised domain adaptation with target reconstruction and label confusion in the common subspace. Neural Comput. Appl. 2020, 32, 4743–4756. [Google Scholar] [CrossRef]
  49. Wang, S.; Zhang, L.; Zuo, W.; Zhang, B. Class-Specific Reconstruction Transfer Learning for Visual Recognition Across Domains. IEEE Trans. Image Process. 2020, 29, 2424–2438. [Google Scholar] [CrossRef]
  50. Zhu, J.Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2242–2251. [Google Scholar] [CrossRef] [Green Version]
  51. Yi, Z.; Zhang, H.; Tan, P.; Gong, M. DualGAN: Unsupervised Dual Learning for Image-to-Image Translation. In Proceedings of theIEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2868–2876. [Google Scholar] [CrossRef] [Green Version]
  52. Kim, T.; Cha, M.; Kim, H.; Lee, J.K.; Kim, J. Learning to discover cross-domain relations with generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70. [Google Scholar]
  53. Blitzer, J.; Crammer, K.; Kulesza, A.; Pereira, F.; Wortman, J. Learning bounds for domain adaptation. In Proceedings of the 20th International Conference on Neural Information Processing Systems, Daegu, Korea, 3–7 November 2007; Curran Associates Inc.: Red Hook, NY, USA, 2007; pp. 129–136. [Google Scholar]
  54. Ben-David, S.; Blitzer, J.; Crammer, K.; Kulesza, A.; Vaughan, P.J.W. A theory of learning from different domains. Mach. Learn. 2010, 79, 151–175. [Google Scholar] [CrossRef] [Green Version]
  55. Zhu, Y.; Zhuang, F.; Wang, D. Aligning Domain-Specific Distribution and Classifier for Cross-Domain Classification from Multiple Sources. Proc. AAAI Conf. Artif. Intell. 2019, 33, 5989–5996. [Google Scholar] [CrossRef]
  56. Dempster, A.P. Upper and lower probabilities generated by a random closed interval. Ann. Math. Stat. 1967, 39, 957–966. [Google Scholar] [CrossRef]
  57. Dempster, A.P. A generalization of Bayesian inference. J. R. Stat. Soc. Ser. 1968, 30, 205–232. [Google Scholar] [CrossRef]
  58. Shafer, G. A mathematical theory of evidence. In A Mathematical Theory of Evidence; Princeton University Press: Princeton, NJ, USA, 1976. [Google Scholar]
  59. Jøsang, A.; Hankin, R. Interpretation and fusion of hyper opinions in subjective logic. In Proceedings of the 2012 15th International Conference on Information Fusion, Chicago, IL, USA, 5–8 July 2012; pp. 1225–1232. [Google Scholar]
  60. Tong, Z.; Xu, P.; Denœux, T. An evidential classifier based on Dempster–Shafer theory and deep learning. Neurocomputing 2021, 450, 275–293. [Google Scholar] [CrossRef]
  61. Huang, M.; Liu, Z. Research on mechanical fault prediction method based on multifeature fusion of vibration sensing data. Sensors 2019, 20, 6. [Google Scholar] [CrossRef] [Green Version]
  62. Huang, M.; Liu, Z.; Tao, Y. Mechanical fault diagnosis and prediction in IoT based on multi-source sensing data fusion. Simul. Model. Pract. Theory 2020, 102, 101981. [Google Scholar] [CrossRef]
  63. Gretton, A.; Borgwardt, K.M.; Rasch, M.J.; Schoelkopf, B.; Smola, A. A Kernel Two-Sample Test. J. Mach. Learn. Res. 2012, 13, 723–773. [Google Scholar]
  64. Gretton, A.; Sriperumbudur, B.; Sejdinovic, D.; Strathmann, H.; Kenji, F. Optimal kernel choice for large-scale two-sample tests. Adv. Neural Inf. Process. Syst. 2012, 25, 1205–1213. [Google Scholar]
  65. Sensoy, M.; Kaplan, L.; Kandemir, M. Evidential deep learning to quantify classification uncertainty. Adv. Neural Inf. Process. Syst. 2018, 31, 3179–3189. [Google Scholar]
  66. Han, Z.; Zhang, C.; Fu, H.; Zhou, J.T. Trusted Multi-View Classification with Dynamic Evidential Fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022. [Google Scholar] [CrossRef]
  67. CLEF. ImageCLEF-DA. Available online: https://www.imageclef.org/2014/adaptation (accessed on 24 May 2022).
  68. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. European Conference on Computer Vision, Heraklion, Greece, 5–11 September 2010; Springer: Berlin/Heidelberg, Germany; pp. 1225–1232. [Google Scholar]
  69. Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
  70. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–1232. [Google Scholar] [CrossRef] [Green Version]
  71. Long, M.; Cao, Y.; Cao, Z.; Wang, J.; Jordan, M.I. Transferable Representation Learning with Deep Adaptation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 3071–3085. [Google Scholar] [CrossRef]
  72. Sun, B.; Saenko, K. Deep CORAL: Correlation alignment for deep domain adaptation. In Computer Vision—ECCV 2016 Workshops; Hua, G., Jégou, H., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 443–450. [Google Scholar] [CrossRef] [Green Version]
  73. Ganin, Y.; Lempitsky, V. Unsupervised Domain Adaptation by Backpropagation. arXiv 2014, arXiv:1409.7495v2. [Google Scholar]
Figure 1. Visualization of Dirichlet distribution, where θ = { θ 1 , θ 2 , θ 3 } and θ 1 , θ 2 , θ 3 0 , θ 1 + θ 2 + θ 3 = 1 . (a) α = ( 10 , 1 , 1 ) ; (b) α = ( 1.001 , 1.001 , 1.001 ) ; (c) α = ( 10 , 10 , 10 ) . Bright yellow represents high probability, and dark blue represents low probability. In the multi-classification problem, each vertex is regarded as a category.
Figure 1. Visualization of Dirichlet distribution, where θ = { θ 1 , θ 2 , θ 3 } and θ 1 , θ 2 , θ 3 0 , θ 1 + θ 2 + θ 3 = 1 . (a) α = ( 10 , 1 , 1 ) ; (b) α = ( 1.001 , 1.001 , 1.001 ) ; (c) α = ( 10 , 10 , 10 ) . Bright yellow represents high probability, and dark blue represents low probability. In the multi-classification problem, each vertex is regarded as a category.
Mathematics 10 02797 g001
Figure 2. The overall structure of MAN-DS.
Figure 2. The overall structure of MAN-DS.
Mathematics 10 02797 g002
Figure 3. The process of combining the outputs of multiple domain-specific classifiers.
Figure 3. The process of combining the outputs of multiple domain-specific classifiers.
Mathematics 10 02797 g003
Figure 4. The demonstration the prediction conflict of domain-specific classifiers.
Figure 4. The demonstration the prediction conflict of domain-specific classifiers.
Mathematics 10 02797 g004
Figure 5. Domain-specific classifier feature visualization.
Figure 5. Domain-specific classifier feature visualization.
Mathematics 10 02797 g005
Figure 6. Accuracy with respect to γ = ρ = 100 λ .
Figure 6. Accuracy with respect to γ = ρ = 100 λ .
Mathematics 10 02797 g006
Table 1. Performance comparison of classification accuracy (%) on Office-31 dataset.
Table 1. Performance comparison of classification accuracy (%) on Office-31 dataset.
StandardsMethodA,W→DA,D→WW,D→AAverage
ResNet 99.33 96.50 61.87 85.90
DDC 99.33 95.80 67.33 87.49
Single BestDAN 99.43 97.61 66.70 87.91
DCORAL 99.53 98.20 65.20 87.64
RevGrad 99.27 96.67 68.53 88.16
DAN 99.57 97.50 67.73 88.27
Source CombineDCORAL 99.33 98.00 67.83 88.39
RevGrad 99.73 97.67 67.77 88.39
MFSAN 99.33 98.67 71.50 89.83
V 1 99.79 98.50 67.02 88.44
Multi-Source V 2 99.73 98.74 66.02 88.16
V 3 99.79 98.86 73.87 90.84
MAN-DS100.00 99 . 12 74 . 16 91 . 09
Table 2. Performance comparison of classification accuracy (%) on Image-CLEF dataset.
Table 2. Performance comparison of classification accuracy (%) on Image-CLEF dataset.
StandardsMethodC,P→II,P→CI,C→PAverage
ResNet 74.83 91.53 83.90 83.42
DDC 74.37 91.33 85.33 83.68
Single BestDAN 75.10 93.33 86.13 84.85
DCORAL 76.67 93.43 88.33 86.14
RevGrad 75.07 94.00 87.07 85.38
DAN 77.67 93.00 91.70 87.46
Source CombineDCORAL 77.73 93.20 91.33 87.42
RevGrad 78.00 93.03 91.87 87.63
MFSAN 79.17 94.50 93 . 33 89.00
V 1 77.67 95.50 92.83 88.67
Multi-Source V 2 77.16 93.50 91.33 87.33
V 3 79 . 56 94.50 91.87 88.40
MAN-DS 79.00 95.67 93.17 89 . 28
Table 3. Performance comparison of classification accuracy (%) on Office-Home dataset.
Table 3. Performance comparison of classification accuracy (%) on Office-Home dataset.
StandardsMethodC,P,R→AA,P,R→CA,C,R→PA,C,P→RAverage
ResNet 65.28 48.54 77.56 74.55 66.48
DDC 64.13 50.22 78.42 75.00 66.94
Single BestDAN 69.07 56.46 79.63 74.65 69.95
DCORAL 66.56 55.15 81.38 76.32 69.85
RevGrad 67.58 55.88 80.32 75.86 69.91
DAN 69.07 59.40 78.41 82.50 72.35
Source CombineDCORAL 68.24 57.62 79.67 83.24 72.19
RevGrad 67.88 57.22 79.52 82.74 71.84
MFSAN 72.86 62.34 80.32 81.86 74.35
V 1 74.32 62.12 82.31 83.13 75.47
Multi-Source V 2 72.86 62.34 80.32 81.86 74.35
V 3 74.12 63.56 82.52 82.74 75.74
MAN-DS74.50 64 . 44 82 . 56 83 . 29 76 . 20
Table 4. Classification accuracy (%) with and without DST fusion strategy on Office-Home dataset.
Table 4. Classification accuracy (%) with and without DST fusion strategy on Office-Home dataset.
MethodC,P,R→AA,P,R→CA,C,R→PA,C,P→R
S 1 72.56 59.48 80.33 80.59
S 2 65.58 61.54 75.54 75.78
S 3 71.39 60.56 79.87 82.36
DST74.50 64 . 44 82 . 56 83 . 29
Table 5. FLOPs and PARAMs.
Table 5. FLOPs and PARAMs.
MethodFLOPsPARAMs
MAN-DS 4.23 G 25.88 M
MFSAN 4.23 G 25.88 M
ResNet50 4.12 G 25.56 M
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Huang, M.; Zhang, C. A Novel Multi-Source Domain Adaptation Method with Dempster–Shafer Evidence Theory for Cross-Domain Classification. Mathematics 2022, 10, 2797. https://doi.org/10.3390/math10152797

AMA Style

Huang M, Zhang C. A Novel Multi-Source Domain Adaptation Method with Dempster–Shafer Evidence Theory for Cross-Domain Classification. Mathematics. 2022; 10(15):2797. https://doi.org/10.3390/math10152797

Chicago/Turabian Style

Huang, Min, and Chang Zhang. 2022. "A Novel Multi-Source Domain Adaptation Method with Dempster–Shafer Evidence Theory for Cross-Domain Classification" Mathematics 10, no. 15: 2797. https://doi.org/10.3390/math10152797

APA Style

Huang, M., & Zhang, C. (2022). A Novel Multi-Source Domain Adaptation Method with Dempster–Shafer Evidence Theory for Cross-Domain Classification. Mathematics, 10(15), 2797. https://doi.org/10.3390/math10152797

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop