Next Article in Journal
Partial Energy Transfer Model of Lamb Waves Scattering in Materially Isotropic Waveguides
Previous Article in Journal
Unconstrained Estimation of Multitype Car Rental Demand
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fuzzy Graph Learning Regularized Sparse Filtering for Visual Domain Adaptation

School of Electronics and Information, Northwestern Polytechnical University, Xi’an 710129, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2021, 11(10), 4503; https://doi.org/10.3390/app11104503
Submission received: 2 April 2021 / Revised: 11 May 2021 / Accepted: 13 May 2021 / Published: 14 May 2021
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Distribution mismatch can be easily found in multi-sensor systems, which may be caused by different shoot angles, weather conditions and so on. Domain adaptation aims to build robust classifiers using the knowledge from a well-labeled source domain, while applied on a related but different target domain. Pseudo labeling is a prevalent technique for class-wise distribution alignment. Therefore, numerous efforts have been spent on alleviating the issue of mislabeling. In this paper, unlike existing selective hard labeling works, we propose a fuzzy labeling based graph learning framework for matching conditional distribution. Specifically, we construct the cross-domain affinity graph by considering the fuzzy label matrix of target samples. In order to solve the problem of representation shrinkage, the paradigm of sparse filtering is introduced. Finally, a unified optimization method based on gradient descent is proposed. Extensive experiments show that our method achieves comparable or superior performance when compared to state-of-the-art works.

1. Introduction

Visual recognition is a challenging subject and has attracted increasing attention of researchers over the world. Despite the success have achieved, the performance of leading models may degrade when applied on a cross-domain scenario, including but not limited to image classification [1], semantic segmentation [2] and person re-identification [3]. Such degradation stems from distribution mismatch, which is very common in real life. For images drawn from different sources, there are many factors leading to mismatches in distribution, including background, shooting angle, illumination and so on. Domain adaptation (DA) [4] aims at leveraging the rich knowledge (e.g., labels) in source domain to build classifiers and apply in a different but related target domain (with less or no labels), which has achieved dramatic success in computer vision [5]. Relevant researches can be divided into two main categories according to whether there are labels existing in target domain: Unsupervised domain adaptation (UDA), which means that there are no labels at all; Semi-supervised domain adaptation (SSDA), which supposes that part of labels in target domain are visible. In this paper, we focus on UDA problems. It is worth emphasizing that UDA follows the setting of transductive learning, data from both domains is available during training.
Mainstream UDA works focus on learning domain-invariant representations, they map original input into feature space with the goal of reducing domain discrepancy. Moment-based statistics are often employed to measure how the two domains (or distributions) like each other, such as Maximum Mean Discrepancy (MMD) [6] and CORrelation ALignmen (CORAL) [7]. As a practical example, Transfer Component Analysis (TCA) [8] attempted to minimize MMD by projecting features to a reproducing kernel Hilbert space (RKHS). CORAL matched the second-order statistics of two domains’ distributions by means of linear transformation. Furthermore, the combination of moment-based statistics and deep neural networks also shows excellent performance. Another kind of practical DA solution is pseudo labels-based method, which assigns pseudo labels to target samples so that class-wise adaptation is possible, also made a lot of advances [9]. The last one is adversarial training, both feature extractor and domain discriminator are constructed in the beginning, then they are updated asynchronously during training. Adversarial training can be applied as moment-based statistics, the principle difference is that a learning-based indicator is employed to measure domain discrepancy rather than handcrafted [10,11].
Though previous DA approaches have shown superior performance in some respects, there are two major drawbacks as follows: Firstly, moment-based and adversarial training-based DA methods are able to reduce distribution discrepancy between domains, but they can hardly guarantee that samples belonging to the same category between two domains are similar. For instance, studies indicated that the class weight bias would lead to significant performance reduction [12]. Secondly, for pseudo label-based DA methods, wrongly-chosen labels are inevitable, which will ruin further analysis. Some efforts have been paid on the selection of labels, which adopt an incremental framework. The underlying philosophy is that the classifier is weak in the beginning, so less labels should be picked, as the training process goes, the classifier obtains more stronger and more pseudo labels can be thought of high confidence. However, it has not fundamentally solved the problem of wrongly-chosen labels. In this paper, we attempt to handle this situation from another view: to reduce the negative effects of mislabeling. Specifically, we utilize soft-labeling to construct affinity graph, rather than hard-labeling. In order to keep the discrimination power of the learned representation, we regard the fuzzy graph objective as a regularization item and integrate it into the paradigm of sparse filtering [13]. Finally, a unified optimization framework based on gradient descent is proposed. The contributions of this paper are summarized as follows.
  • We propose a novel solution for UDA problems, which is able to learn both discriminative and domain-shared representations simultaneously. Extensive experiments on several real-world datasets demonstrate its superiority to existing works.
  • Different from previous hard-labeling methods, we design a fuzzy graph regularization based on soft-labeling. Specifically, it attempts to describe cross-domain affinity by means of probability matrix.
  • In order to deal with the problem of representation shrinkage, we combine the proposed fuzzy graph regularization with unsupervised sparse filtering, so that the two objective will be antagonistic and converge to a compromise.
The rest of this paper is organized as follows. Section 2 gives the background knowledge on existing DA works, especially the highly-correlated pseudo label methods. Then sparse filtering is introduced. In Section 3, we describe our method in detail. Section 4 and Section 5 present the experiments and some empirical analysis. Finally, Section 6 concludes the paper and provides some ideas for future researches.

2. Related Works

In this section, we first give a formal definition of domain adaptation, then conclude existing pseudo label based DA methods. For the sake of better understanding, we introduce sparse filtering subsequently.

2.1. Domain Adaptation

Domain adaptation is a branch of transfer learning, which deals with the scenario where training and testing data have different distributions, some literatures call it dataset bias or covariate shift. Specifically, there are two important points in domain adaptation. Matching marginal distributions of source and target domains cannot guarantee the separability of target samples, so pseudo label based DA methods become more popular. Overall, it can be divided into two categories according to whether the method has selection strategies on pseudo labels.

2.1.1. Pseudo Label without Selection

During training, it assigns pseudo labels on the whole target domain without considering the confidence. Class-wise distribution alignment is therefore feasible by utilizing the pseudo labeled target samples together with labeled source samples. After iterative learning, the pseudo labels are expected to be progressively more accurate. Obviously, this expectation is not always held, especially when the initial prediction is weak. As an extension of TCA, Joint Distribution Adaptation (JDA) [14] aligned conditional and marginal distribution simultaneously using pseudo labels, then an iterative training scheme is employed to obtain more accurate labels. Inspired by this, a series of studies tried to modify classical feature extraction methods (e.g., linear discriminant analysis and locality preserving projections [15,16]) by means of pseudo labels.

2.1.2. Pseudo Label with Selection

To alleviate the negative effects caused by wrongly-chosen labels, it selects a subset of target samples with corresponding pseudo labels. Specifically, both labels and their confidences need to be considered and only the selected target samples can be combined with source samples for training. An easy-to-hard training scheme is widely used for DA problems, which adopts an increasing number of target samples. CAPLS is an classical method of selective pseudo label, which uses softmax to denote the classify confidence [17]. Further, SPL considers structure information for more reliable labels [18].

2.2. Sparse Filtering

Sparse filtering is an unsupervised feature extraction algorithm. Unlike auto-encoder and sparse coding, it does not seek to model the data explicitly, but attempts to obtain ideal features from the perspective of so-called good features. First of all, it gives three requirements on the question of What is good features.
  • Population Sparsity. Each example should be represented by only a few active features. Ideally, a sample should be a vector which contains lots of zeros and few non-zero values.
  • Lifetime Sparsity. Good features should be distinguishable, therefore, a feature is only allowed to be activated in few samples. On the contrary, if a feature is activated by all the samples, we cannot classify samples according to this feature, so it is not a good feature.
  • High Dispersal. It requires each feature to have similar statistical properties across all samples, which seems to be useless. The authors indicate that this can prevent feature degradation, such as similar features across samples.
Ngiam et al. pointed out that we can obtain the features that meet all the requirements by optimizing population sparsity and high dispersal [13]. Suppose now we have n samples, each with m-dimensional features that can be written as x = x 1 , x 2 , x n R m . The learning paradigm of sparse filtering is as follows:
(1) Linear feature extraction and non-linear activation. Let f j ( i ) represent the jth feature of the ith sample, f j ( i ) = x ( i ) w j . Then we can use some activation functions to make it more expressive, such as soft-absolute function.
f j ( i ) = ϵ + ( x ( i ) w j ) 2 x ( i ) w j
(2) Solving high dispersal. Each feature is divided by the l 2 norm of the feature on all samples.
f ˜ j = f j f j 2
Remember that the requirement of high dispersal is that the statistical properties of each feature are similar. This step forces the sum of the squares of all features to be 1 roughly.
(3) Solving population sparsity. Each sample is divided by its own l 2 norm. Then we can get the objective function.
m i n i m i z e i = 1 n f ˜ ( i ) f ˜ ( i ) 2 1
This objective function can be easily understood through an example. Imagine that we have different expressions of a sample, ( 0.6 , 0.8 ) T and ( 1 , 0 ) T . We say ( 1 , 0 ) T is a sparser representation since it has more zero, so it has a smaller loss, 1 1 2 + 0 2 + 0 1 2 + 0 2 = 1 0.6 0 . 6 2 + 0 . 8 2 + 0.8 0 . 6 2 + 0 . 8 2 = 1.4 .

3. Methodology

In this section, we give the mathematical definition of DA problem, describe each component of the proposed method in detail and analyze the complexity.

3.1. Problem Definition and Notations

Given labeled source domain D s = { ( x s 1 , y s 1 ) , ( x s 2 , y s 2 ) , , ( x s n s , y s n s ) } and unlabeled target domain D t = { ( x t 1 ) , ( x t 2 ) , , ( x t n t ) } , the goal is to build classifier and make predictions on target data. Specifically, we study homogeneous DA, which means that the source and target data have the same dimensions, x s , x t R m . The target labels y t are only available for evaluating the algorithm. Table 1 introduces some necessary notations and descriptions used in this paper.
Domain: a domain, D , consists of two parts, data X and the corresponding distribution p ( X ) , so D = { X , p ( X ) } . Generally, we have D s and D t to represent the labeled source and unlabeled target domain, respectively. When it refers to domain adaptation, we have p ( X s ) p ( X t ) . It is worth emphasizing to obtain the explicit distribution p ( X ) is often very hard, especially for high-dimensional data, so an alternative solution is utilizing some statistics, i.e., mean or variance, to estimate the distribution.
Task: a task, T , is the specific application of domain adaptation, which contains two parts, label space Y and the mapping function f ( · ) , so T = Y , f ( · ) . Naturally, given data X , we have Y = f ( X ) . For standard domain adaptation problems, the tasks are the same, T s = T t .

3.2. Fuzzy Graph Learning for Domain Alignment

How to reduce the domain discrepancy is a key problem for DA, moment-based and adversarial training-based strategies cannot guarantee the separability of target samples, because they do not exploit the structural information existing across different classes. Therefore, we propose fuzzy graph learning for class-wise alignment. The underlying idea is: samples belonging to the same class (no matter which domain it comes from) should be close to each other. Since there are no labels for target samples, we must adopt pseudo labels as a substitution. Obviously, the quality of pseudo labels will greatly influence the learned classifier/representation, it may cause a catastrophic accumulation of errors if not handled properly.
For our method, we propose fuzzy pseudo labels which employ a probability matrix to update the classifier. In Figure 1, we give a graphical illustration to explain how soft label works. The class centroids are calculated by labeled source samples and the decision boundary can be obtained by means of the distance to each centroid. For an unreliable training sample which located in the intersection of decision boundary (also can be described as it has similar distance to each centroid), the probabilities of it belonging to any category is extremely close, e.g., 24%, 25%, 26%. In such situation, assigning the label with biggest probability is risky. Like the figure in the middle shows, the updated centroid would have a big influence to decision boundaries. When we adopt soft labels, the case is shown in the last figure, each centroid would move a distance to the training sample according to corresponding probability value. Intuitively, soft labels have significant less effect on decision boundaries than hard.
Specifically, suppose we have the learned representation x s ^ = f ( x s ) R k × n s and x t ^ = f ( x t ) R k × n t (we will introduce the transformation f ( · ) later), along with the source labels y s [ 1 , 2 , , c ] . We use x = [ x s ^ , x t ^ ] R k × ( n s + n t ) to denote the combination of two domains’ features. The objective of fuzzy graph learning can be written as:
L F G L = 1 2 ( n s + n t ) i = 1 n s + n t j = 1 n s + n t ( x i x j ) 2 W i j
where W R ( n s + n t ) × ( n s + n t ) is the affinity graph where W i j [ 0 , 1 ] denotes the similarity between x i and x j , and W can be defined as follows:
W = W s s W s t W t s W t t
Fuzzy connections only exist in cross-domain cases, so we describe how to construct the affinity graph separately:
Construction of W s s : For labeled source domain, we adopt the supervised graph, that is, the edges exist only between samples with the same label. Ideally, this can help obtain discriminative representations.
W i j s s = 1 , i f Y i = Y j 0 , o t h e r w i s e
Construction of W t t : For unlabeled target domain, we encourage target samples to be close to their neighborhood samples, thus a cut-off is necessary. In this paper, we exploit the top-k (k = 10 in this paper) nearest neighbors for target domain.
W i j t t = 1 , i f X j N 10 ( X i ) 0 , o t h e r w i s e
where N 10 ( X i ) means the aggregation of X i ’s top-10 nearest neighbors in target domain.
Construction of W s t and W t s : To build the fuzzy graph to model the cross-domain similarity, we first compute the class centroids x ¯ of source samples, then the probability p ( y | x ^ t i ) of target sample x ^ t i belongs to class y can be computed based on the Euclidean distance to each centroid. Finally, W s t indicates the similarity from source to target samples, then for a certain source sample x ^ s i and target sample x ^ t j , if they belongs to the same category, they should have a strong connection and vice versa. Since we utilize soft labels here, it can be interpreted as: if the target sample x ^ t j has a big probability of possessing the same label of the source sample (noted that the source sample x ^ s i and its label y s i are all available), they should have a strong connection. The pseudo code of cross-domain affinity graph construction is given in Algorithm 1, and we can obtain W t s by W t s = W s t T directly.
Algorithm 1: Cross-domain affinity graph ( W s t ) construction.
Applsci 11 04503 i001

3.3. Sparse Filtering for Discriminative Feature Learning

Fuzzy graph learning is able to reduce the domain discrepancy, but it is an ill-posed problem in nature. For instance, if we choose two constant function, k 1 ( x ) = a , k 2 ( x ) = b , w h e r e a b , they both can minimize the objective L F G L . For the intuitive example shown in Figure 1, we can see that unreliable samples lead to representation shrinkage. Consequently, the learned representation will lose the discriminability, hence an additional penalty is indispensable. In this paper, we adopt the objective of sparse filtering for two reasons: Firstly, it requires lifetime sparsity, which projects the feature onto a L2-ball, this can effectively alleviate the feature contraction. Secondly, it allows us to exploit the potential discriminant structures in unlabeled target domain.
Corresponding to the optimization of sparse filtering, the transformation function k ( · ) of our method can be summarized as follows: Given original input x s , x t R n s , n t × m , the goal is to find optimal W R m × k , then obtain new representations x ^ s , x ^ t R n s , n t × k for downstream applications.
1. Linear transformation and non-non-linear activation. We use the soft absolute function as activation function in this paper.
x ^ s 1 = | x s W + ϵ | x ^ t 1 = | x t W + ϵ |
where ϵ denotes a small number, such as 1e-5.
2. L2 column normalization Observing the form of current x ^ , each row represents a sample, and each column represents a feature. This measure enforces the L2 norm of each feature to be 1 thus high dispersal can be achieved. Noted that we do the within-domain normalization since it is proven to be powerful for DA problems.
x ^ s 2 = x ^ s 1 M s c o l , ( M s c o l ) i j = 1 i = 1 n s ( x ^ s 1 ) i j 2 x ^ t 2 = x ^ t 1 M t c o l , ( M t c o l ) i j = 1 i = 1 n t ( x ^ t 1 ) i j 2
which the symbol ∘ represents Hadamard product. Here we use Hadamard product to simplify gradient derivation.
3. L2 row normalization. Just like the previous step, this measure enforce the L2 norm of each sample to be 1 to project data onto L2-ball.
x ^ s 3 = x ^ s 2 M s r o w , ( M s r o w ) i j = 1 j = 1 k ( x ^ s 2 ) i j 2 x ^ t 3 = x ^ t 2 M t r o w , ( M t r o w ) i j = 1 j = 1 k ( x ^ t 2 ) i j 2
For the sake of simplification, we integrate the mapping from x s , x s to x ^ s 3 , x ^ t 3 as a single step k ( · ) , so we can obtain the final representation by x ^ s = k ( x s ) = x ^ s 3 , x ^ t = k ( x t ) = x ^ t 3 . Correspondingly, the objective of sparse filtering can be represented as:
L S F = 1 n s | | x ^ s | | 1 + 1 n t | | x ^ t | | 1
The total objective of our method can be written as:
L = L S F + λ L F G L
where λ is hyper-parameter to balance two objectives.

3.4. Optimization

Intuitively, our method can be formally defined as a weighted sum of two objectives, so we analyze the two parts separately.

3.4.1. Gradient of L F G L x ^

Thinking about the combination of two domains’ features, x ^ = [ x ^ s , x ^ t ] the objective can be reformulated as:
L F G L = 1 2 ( n s + n t ) i = 1 n s + n t j = 1 n s + n t ( x ^ i x ^ j ) 2 W i j = 1 2 ( n s + n t ) ( i = 1 n s + n t x ^ i T D i i x ^ i i = 1 n s + n t j = 1 n s + n t x ^ i T W i j x ^ j ) = 1 2 ( n s + n t ) t r a c e ( x ^ T L x ^ )
where D i i = j = 1 n s + n t W i j denotes the degree of x ^ i , and  L = D W is called Laplacian graph.
Naturally, we can compute the gradient by:
L F G L x ^ = 1 n s + n t L x ^ = [ L F G L x ^ s L F G L x ^ t ]

3.4.2. Gradient of L S F

Noted that we have x ^ 0 because of the activation. So the gradient of sparse filtering is simply a matrix full of 1.
L S F x ^ = 1 = [ L S F x ^ s L S F x ^ t ]
Here we use 1 to represent the 1 matrix with certain dimension, e.g.,  L S F x ^ s = 1 R n s × k and L S F x ^ t = 1 R n t × k .

3.4.3. Gradient of k ( · )

To propagate the gradient back to the input layer, we need x ^ x to apply the chain rule. There are some differences for two domains since we adopt within-domain normalization. To obviate misunderstanding, we give the specific form for two domains.
x ^ s W = x s T ( M s r o w M s c o l s i g n ( x s W ) ) x ^ t W = x t T ( M t r o w M t c o l s i g n ( x t W ) ) x ^ W = [ x ^ s W x ^ t W ]

3.4.4. Unified Optimization Based on Gradient Descent

Given above gradients, we are able to update W by gradient descent from random initialization. The training process is similar to train a neural network, some tricks that help converge can also be applied here, such as initializing with normal distribution and utilizing a decreasing step. The specific training scheme is shown in Algorithm 2.
Algorithm 2: Optimization of the proposed method.
Applsci 11 04503 i002

4. Experiments

In this section, we introduce two widely-used DA datasets, then the proposed method is compared with state-of-the-art works. In addition, we provide parameter sensitivity analysis to show the robustness of the proposed method.

4.1. Datasets

ImageCLEF https://www.imageclef.org/2014/adaptation (accessed on 14 May 2021). ImageCLEF is an online DA competition, which contains three domains: Caltech256 (C) [19], ImageNet (I) [20] and PASCAL (P) [21]. There are twelve kinds of objects, and fifty images of each object. Some pictures can be found in Figure 2. In our experiments, we adopt the ResNet50 (input dimension m = 2048 ) features of images, and six tasks, e.g., C I , C P , are constructed.
Office-Caltech10 [22]: Office-Caltech10 contains ten object categories drawn from 4 image domains: Amazon (A), Webcam (W), DSLR (D), and Caltech256 (C). There are 8–151 samples per category per domain, and 2533 images in total. In our experiments, we adopt the Decaf6 (input dimension m = 4096 ) features of images as input for testing algorithms, twelve cross-domain tasks, e.g., C A , C W , C D , are constructed.

4.2. Experimental Setting

Since there are no deep architectures in the proposed method, we compare our approach to several state-of-the-art shallow methods to evaluate its effectiveness.
  • Nearest Neighbor(NN) [23]. NN is served as a baseline model to check whether the learned representations really work for DA problems.
  • Joint Distribution Alignment(JDA) [14]. JDA [Long et al. ICCV2013] adopts pseudo labels to align the conditional distributions of two domains.
  • Correlation Alignment(CORAL) [7]. CORAL [Sun et al. AAAI2016] obtains transferable representations by aligning the second-order statistics of distributions.
  • Confidence-Aware Pseudo Label Selection(CAPLS) [17]. CAPLS [Wang et al. IJCNN2019] uses a selective pseudo labeling procedure to obtain more reliable labels.
  • Modified A-distance Sparse Filtering(MASF) [24]. MASF [Han et al. Pattern Recognit.2020] employs an L2 constraint combining sparse filtering to learn both domain-shared and discriminative representations.
  • Selective Pseudo Labeling(SPL) [18]. SPL [Wang et al. AAAI2020] is also a selective pseudo labeling strategy based on structured prediction.
Follow the experimental setting of JDA and MASF, we set the subspace dimension k = 100 . For JDA, we set the regularization coefficient λ = 1 and the number of iterations T = 10 . For CAPLS, we set the number of iteration T = 10 . For MASF, we set the regularization coefficient α = 1 e 3 . For SPL, we set the number of iterations T = 11 . It is worth emphasizing that the input features are extracted by deep networks without fine-tuning and not any pre-processing strategy is applied in the experiments.

4.3. Results

In this section, we report the accuracy of the proposed method (abbreviated as ‘FGLSF’ for Fuzzy Graph Learning regularized Sparse Filtering) and other state-of-the-art works, the results are shown on Table 2. From experimental results, we have the following observations:
  • FGLSF vs. NN. According to the results, FGLSF is significantly better than NN. NN cannot handle the domain discrepancy, thus results in unsatisfying performance. On the other hand, it indicates that our method is able to learn transferable representations.
  • FGLSF vs. CORAL, JDA. FGLSF is superior to CORAL and JDA. This two methods are classical distribution matching works, but they have limited considerations on the discrimination of learned representations.
  • FGLSF vs. MASF. MASF is another framework based on sparse filtering, which adopts a modified A distance for domain alignment. Compared to our method, it cannot deal with the problem of conditional distribution matching.
  • FGLSF vs. CAPLS, SPL. Objectively speaking, our method FGLSF has comparable performance when compared state-of-the-art selective hard labeling method, only a 0.3% improvement is gained. It shows that the proposed fuzzy graph regularization is also valid for alleviating the negative effects caused by wrong labeling.
Another interesting finding is that the proposed method works well in ImageCLEF, achieves best performance on all subtasks, but when it refers to Office-Caltech10, it seems that SPL works better. For these two datasets, the biggest difference is the class weights, ImageCLEF is an absolutely class-balanced dataset, which means that each object has the same number of pictures, but Office-Caltech10 has class weight bias, which makes it a more complicated problem. When we infer the pseudo labels, the object who has more examples is expected to be dominant, and vice versa. For the proposed method, we only use source labels for obtaining pseudo labels, so it may easily influenced by the class weight bias. SPL considers both source labels and manifold structures, thus achieves better performance on Office-Caltech10.

4.4. Parameter Sensitivity Analysis

For our method, there is one hyper-parameter to be adjusted, λ indicates the weight of fuzzy graph learning. We run FGLSF varying with values of parameter λ which are sampled from {0, 0.001, 0.01, 0.1, 0.2} on three tasks, e.g., C W , W C and P C . If λ is too small, the model would fail to align distributions of two domains. And as it getting larger, the problem of representation shrinkage is emerging. The experimental results shown on Figure 3 are consistent with the analysis, the accuracy first rises and then falls as λ gets bigger.

5. Discussion

To prove the effectiveness of the proposed method, we conduct two sets of comparative experiments: 1. How the accuracy changes during training? 2. How about using selective hard labeling?

5.1. Does Iterative Learning Help Improve the Model?

To answer the question “Does iterative learning help improve the model”, we report the performance of 18 tasks during training, whose results are shown on Figure 4. Specifically, we divide the full training process to 10 parts, and record current accuracy. Obviously, the model becomes more precise as training goes. On the other hand, we can also see some fluctuations, which suggests that our approach cannot completely eliminate the effects of mislabeling.

5.2. Fuzzy Labeling versus Selective Hard Labeling

Figure 5 shows the performance of two labeling strategies. Intuitively, the proposed fuzzy labeling and state-of-the-art selective hard labeling have similar performance, e.g., selective hard labeling has better performance on tasks {7, 8} while fuzzy labeling is superior on tasks {5, 15}. This finding confirms that the proposed fuzzy labeling is able to alleviate the issue of wrong-labels.

6. Conclusions

In this paper, we propose a novel DA solution based on fuzzy graph learning, which aims to learn both discriminative and domain-shared representations simultaneously. When compared to data augmentation methods, it considers how to learn an supervised model under dataset bias, aims at minimizing the performance degradation brought by domain discrepancy. The main contribution is that we propose another view for the problem of misselected pseudo labels, from selecting labels to minimizing the negative effects of mislabeling. When compared to existing methods, the proposed method adopts a fuzzy labeling framework, which uses label matrix rather than a single selected label for target samples. To avoid trivial solutions, we combine sparse filtering and fuzzy graph learning to solve the problem of representation shrinkage. The experimental results verify that our method outperforms many SOTA shallow methods; the increases of mean accuracy are 0.3–6.6%. In the future, we plan to study how to adjust the hyper-parameter λ according to task similarity.

Author Contributions

Conceptualization, L.M. and D.Z.; methodology, L.M. and D.Z.; software, L.M. and X.L.; validation, L.M. and Q.L.; formal analysis, Q.L.; investigation, L.M.; resources, Y.Z.; data curation, Y.Z.; writing—original draft preparation, L.M.; supervision, D.Z.; project administration, D.Z.; funding acquisition, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 62076204).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2010; pp. 213–226. [Google Scholar]
  2. Patel, V.M.; Gopalan, R.; Li, R.; Chellappa, R. Visual Domain Adaptation: A survey of recent advances. IEEE Signal Process. Mag. 2015, 32, 53–69. [Google Scholar] [CrossRef]
  3. Li, X.; Grandvalet, Y.; Davoine, F.; Cheng, J.; Cui, Y.; Zhang, H.; Belongie, S.; Tsai, Y.; Yang, M. Transfer learning in computer vision tasks: Remember where you come from. Image Vis. Comput. 2020, 93, 103853. [Google Scholar] [CrossRef]
  4. Pan, S.J.; Yang, Q. A Survey on Transfer Learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  5. Wang, M.; Deng, W. Deep visual domain adaptation: A survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef] [Green Version]
  6. Smola, A.J.; Gretton, A.; Song, L.; Scholkopf, B. A Hilbert Space Embedding for Distributions. In Algorithmic Learning Theory; Springer: Berlin/Heidelberg, Germany, 2007; pp. 13–31. [Google Scholar]
  7. Sun, B.; Feng, J.; Saenko, K. Return of frustratingly easy domain adaptation. In Proceedings of the National Conference on Artificial Intelligence, Phoenix, AZ, USA, 2–9 February 2016; pp. 2058–2065. [Google Scholar]
  8. Pan, S.J.; Tsang, I.W.; Kwok, J.T.; Yang, Q. Domain Adaptation via Transfer Component Analysis. IEEE Trans. Neural Netw. Learn. Syst. 2011, 22, 199–210. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Zhang, W.; Ouyang, W.; Li, W.; Xu, D. Collaborative and Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the Computer Vsion and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3801–3809. [Google Scholar]
  10. Tzeng, E.; Hoffman, J.; Saenko, K.; Darrell, T. Adversarial Discriminative Domain Adaptation. In Proceedings of the Computer Vsion and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2962–2971. [Google Scholar]
  11. Ganin, Y.; Lempitsky, V. Unsupervised Domain Adaptation by Backpropagation. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 1180–1189. [Google Scholar]
  12. Yan, H.; Ding, Y.; Li, P.; Wang, Q.; Xu, Y.; Zuo, W. Mind the Class Weight Bias: Weighted Maximum Mean Discrepancy for Unsupervised Domain Adaptation. In Proceedings of the Computer Vsion and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 945–954. [Google Scholar]
  13. Ngiam, J.; Chen, Z.; Bhaskar, S.A.; Koh, P.W.; Ng, A.Y. Sparse Filtering. In Proceedings of the Neural Information Processing Systems, Granada, Spain, 12–15 December 2011; pp. 1125–1133. [Google Scholar]
  14. Long, M.; Wang, J.; Ding, G.; Sun, J.; Yu, P.S. Transfer Feature Learning with Joint Distribution Adaptation. In Proceedings of the International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 2200–2207. [Google Scholar]
  15. He, X.; Niyogi, P. Locality Preserving Projections. In Proceedings of the Neural Information Processing Systems, Vancouver, BC, Canada, 8–13 December 2003; pp. 153–160. [Google Scholar]
  16. Sanodiya, R.K.; Mathew, J. A novel unsupervised Globality-Locality Preserving Projections in transfer learning. Image Vis. Comput. 2019, 90, 103802. [Google Scholar] [CrossRef]
  17. Wang, Q.; Bu, P.; Breckon, T.P. Unifying Unsupervised Domain Adaptation and Zero-Shot Visual Recognition. In Proceedings of the International Joint Conference on Neural Network, Budapest, Hungary, 14–19 July 2019; pp. 1–8. [Google Scholar]
  18. Wang, Q.; Breckon, T.P. Unsupervised Domain Adaptation via Structured Prediction Based Selective Pseudo-Labeling. In Proceedings of the National Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 1–10. [Google Scholar]
  19. CaltechAUTHORS. Caltech-256 Object Category Dataset; Technical Report. Pasadena, CA, USA, 2007. Available online: http://www.vision.caltech.edu/Image_Datasets/Caltech256 (accessed on 14 May 2021).
  20. Deng, J.; Dong, W.; Socher, R.; Li, L.; Li, K.; Feifei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the Computer Vsion and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  21. Everingham, M.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The PASCAL Visual Object Classes Challenge 2007 (VOC2007) Results. Available online: http://www.pascal-network.org/challenges/VOC/voc2007/workshop/index.html (accessed on 14 May 2021).
  22. Gong, B.; Shi, Y.; Sha, F.; Grauman, K. Geodesic flow kernel for unsupervised domain adaptation. In Proceedings of the Computer Vsion and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2066–2073. [Google Scholar]
  23. Cover, T.M.; Hart, P.E. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  24. Han, C.; Lei, Y.; Xie, Y.; Zhou, D.; Gong, M. Visual Domain Adaptation Based on Modified A Distance and Sparse Filtering. Pattern Recognit. 2020, 104, 107254. [Google Scholar] [CrossRef]
Figure 1. Unreliable training samples and its influence on decision boundary.
Figure 1. Unreliable training samples and its influence on decision boundary.
Applsci 11 04503 g001
Figure 2. Objects in different domain.
Figure 2. Objects in different domain.
Applsci 11 04503 g002
Figure 3. Performance on three subtasks with different λ . X-axis: tasks. Y-axis: accuracy.
Figure 3. Performance on three subtasks with different λ . X-axis: tasks. Y-axis: accuracy.
Applsci 11 04503 g003
Figure 4. Performance during training. X-axis: training process. Y-axis: accuracy.
Figure 4. Performance during training. X-axis: training process. Y-axis: accuracy.
Applsci 11 04503 g004
Figure 5. Performance of two labeling strategies. X-axis: tasks. Y-axis: accuracy.
Figure 5. Performance of two labeling strategies. X-axis: tasks. Y-axis: accuracy.
Applsci 11 04503 g005
Table 1. Notations and descriptions used in this paper.
Table 1. Notations and descriptions used in this paper.
NotationsDescription
D s / D t source/target domain
x s / x t original source/target domain data
y s / y t source/target domain label
n s / n t number of source/target samples
m / k original/transformed feature dimension
x s ^ / x t ^ source/target features
f ( · ) mapping function, X ^ = f ( X )
Wthe transformation matrix to be solved
L F G L objective function for fuzzy graph learning
L S F objective function for sparse filtering
λ the balance factors between two objectives
Table 2. Performance (accuracy %) on Office-Caltech10 (No. 1–12) and ImageCLEF (No. 13–18).
Table 2. Performance (accuracy %) on Office-Caltech10 (No. 1–12) and ImageCLEF (No. 13–18).
No.TaskNNJDACORALCAPLSMASFSPLJANDANFGLSF
1C→ A85.6989.7792.0090.9090.8192.8091.9092.0093.32
2C→ W66.1083.7380.0088.8387.4685.0885.4090.6084.75
3C→ D74.5286.6284.7090.0889.8191.7288.8089.3087.90
4A→ C70.3582.2883.2080.6687.3681.3985.0084.1085.31
5A→ W57.2978.6474.6080.6981.0284.0786.1091.8088.14
6A→ D64.9780.2584.1089.4586.6290.4589.0091.7087.26
7W→ C60.3783.5375.5086.6285.0474.0078.0081.2074.27
8W→ A62.5390.1981.2091.3891.3491.9684.9092.1089.25
9W→ D98.73100.00100.00100.0099.36100.00100.00100.0097.45
10D→ C52.0985.1376.8088.0585.7588.5181.1080.3083.35
11D→ A62.7391.4485.5092.3290.4093.3289.5090.0091.75
12D→ W89.1598.9899.3098.6698.98100.0098.2098.5095.59
13C→ I85.1692.0083.0091.0089.8390.8389.5089.5094.00
14C→ P69.1675.5071.5077.3372.8378.1774.2075.8079.70
15I→ C91.1692.3388.6694.1793.1794.3394.7094.2097.00
16I→ P73.1677.0073.6675.8076.8377.5076.8078.2080.71
17P→ C81.3383.8372.5090.6785.3391.3391.7089.2095.50
18P→ I74.5079.1672.3385.0080.8385.8388.0087.5093.00
19AVG73.2886.1382.1488.4287.3888.4087.3888.6688.79
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Min, L.; Zhou, D.; Li, X.; Lv, Q.; Zhi, Y. Fuzzy Graph Learning Regularized Sparse Filtering for Visual Domain Adaptation. Appl. Sci. 2021, 11, 4503. https://doi.org/10.3390/app11104503

AMA Style

Min L, Zhou D, Li X, Lv Q, Zhi Y. Fuzzy Graph Learning Regularized Sparse Filtering for Visual Domain Adaptation. Applied Sciences. 2021; 11(10):4503. https://doi.org/10.3390/app11104503

Chicago/Turabian Style

Min, Lingtong, Deyun Zhou, Xiaoyang Li, Qinyi Lv, and Yuanjie Zhi. 2021. "Fuzzy Graph Learning Regularized Sparse Filtering for Visual Domain Adaptation" Applied Sciences 11, no. 10: 4503. https://doi.org/10.3390/app11104503

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop