Next Article in Journal
Characterization and Reliability Analysis of Enhancement-Mode PEALD AlN/LPCVD SiNx GaN MISFET with In Situ H2/N2 Plasma Pretreatment
Previous Article in Journal
Efficient Hardware Implementation of Elliptic-Curve Diffie–Hellman Ephemeral on Curve25519
Previous Article in Special Issue
CPL-Net: A Malware Detection Network Based on Parallel CNN and LSTM Feature Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Optimal Transport-Embedded Neural Network for Fairness Transfer Problem

1
State Grid Hubei Electric Power Research Institute, Wuhan 430077, China
2
State Grid Xiangyang Power Supply Company, Xiangyang 441000, China
3
School of Automation, Wuhan University of Technology, Wuhan 430070, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(21), 4481; https://doi.org/10.3390/electronics12214481
Submission received: 6 September 2023 / Revised: 11 October 2023 / Accepted: 17 October 2023 / Published: 31 October 2023
(This article belongs to the Special Issue Recent Advances and Related Technologies in Neuromorphic Computing)

Abstract

:
Research on neuromorphic computing has gained popularity in recent years. In particular, regularized embedded neural systems have been applied in several significant real-world situations, such as recommendation systems and transfer learning. This paper deals with the fairness transfer learning problem, which has been insufficiently explored. In fairness transfer settings, the source domain has limit-tagged training samples, which may lead to performance degradation in the target domain. To solve such problems, a linear data-augmentation-based optimal transport-embedded neural network is proposed in this paper. It can augment the source samples to make the distribution of the source domain balanced and can align the source and target distributions simultaneously. Moreover, the distribution of the augmented data by mixup is limited to a certain bound that can avoid the abnormal samples generated. The effectiveness of the proposed method has been demonstrated in several transfer learning tests, including regression and classification. In 1-shot and 3-shot classification tasks on the Office dataset, our method’s accuracy is 4.8 and 3.9% better, respectively, than the second-best model. Additionally, our model’s performance is about 2–3 percentage points superior to the second-best model in the OfficeHome dataset. It is simple yet effective, making it perfect for low-power edge AI applications.

1. Introduction

Neural network-based methods have achieved significant advances in building decision algorithms and have been applied in various real-world applications. They have been used widely in some sensitive areas that usually possess a large number of samples, such as control, classification, prediction, and other tasks [1,2]. Machine-learning methods are good at mining increasingly abstract distributed feature representations from original input data, and these representations have good generalization ability. However, the performance of machine-learning methods relies heavily on the quality and the number of training samples. The decision rule learns on the training set and is applied on the test set under the assumption that the training and the test samples from different domains follow the same underlying distribution [3]. The optimization of most machine-learning methods breaks down in the small-data regime, where only very few labeled examples are available for training. This indicates that the decision boundary of the learned model is highly influenced by the training set.
The availability of massive data with fully labeled information is crucial since data collection and annotation are very expensive and time-consuming for some specific objects. This has motivated researchers to produce novel algorithms or learning models that are trained with few examples (or only one example) and have ideal performance on the test set. In the past decade, few-shot learning has attracted much researcher interest, which is a type of machine-learning problem where the training dataset contains limited information. Many few-shot learning methods are proposed to solve such problems, and they can be subdivided into two broad categories: data-augmentation methods and meta-learning methods [4]. The former is usually based on the Generative Adversarial Network (GAN), which focuses on synthesizing the training data using some specific conditions. GANs contain two sub-networks: a generator and a discriminator. The generator creates samples to deceive the discriminator, while the discriminator determines if a sample is created or real [5]. The latter is based on an episodic training strategy that uses a few examples of each episode from the base class set to mimic the test scenario.
Traditional machine-learning problems assume that the feature space and data distribution of the training set and the test set are the same. In some real applications, samples from specific domains are expensive and difficult to collect. Thus, it is of significant importance to create a high-performance learning model trained with data from easily obtained domains. This methodology is referred to as transfer learning and has been widely studied. It aims to reduce the marginal mismatch in feature space between different domains and transfer information from well-labeled source domain samples to unlabeled target samples. The existing transfer learning models can be categorized into two groups: discrepancy-based methods and GAN-based methods [6]. However, they are all based on the concept of reducing cross-domain gaps by aligning the domain distribution so that models derived from the source domain can be applied directly to the target domain. Inspired by the dynamic control theory [7], more and more machine-learning methods have focused on developing a model that can be explained. In this paper, we propose an optimal transport model that can be theoretically analyzed.
For optimal transport problems, this is the most effective method to perform the transformation of one mass distribution into another mass distribution. The basic types of these problems include the Monge transmission problem, Kantorovich transmission problem, and Kantorovich dual transmission problem [8]. At present, the optimal transmission problem has been applied to many fields, such as transfer learning [9,10], image processing [11], sequence pattern analysis [12], and so on. Courty et al. [13] assumed that there was a nonlinear transformation between the joint distribution of the source domain and the target domain, which could be estimated using the optimal transport method, and proposed a solution model named JDOT to recover the estimated target distribution by optimizing the coupling matrix and classifier simultaneously. On this basis, Damodaran et al. [14] proposed a new deep-learning framework, DeepJDOT, to obtain better classification results through a neural network model.
In this paper, a novel optimal transport-based neural system is proposed to solve the fairness transfer learning problem. Imbalanced or long-tailed distributions are quite normal in real-world scenarios, and transfer learning is much more difficult to solve than in traditional machine-learning settings. Thus, fairness transfer learning is a more challenging and piratical task for imbalanced data, as is commonly encountered in real-world applications. It aims to learn a classifier from only a few examples in the source domain and transfer the knowledge to a novel target domain [15,16]. This paradigm is pretty similar to human behaviors that transfer learned experience to new tasks. This occurs instead of adapting a Generative Adversarial Network to augment the samples from the few-shot source samples. Our proposed method aims to solve the fairness transfer learning problem with a linear data-augmentation-based optimal transport model. It is much easier in the training process and can constrain the distribution of augmented data in a certain bound. Optimal transport-embedded neural systems are an option for solving transfer learning problems. The mixup method is adopted in this paper. It trains the optimal transport model with convex combinations of pairs of examples and their labels. In addition, the conditional distributions of the mixup source domain and target domain are aligned by the optimal transport model. The learning model and the coupling matrix for distribution alignment are optimized simultaneously. Some synthetic examples and widely used transfer learning tasks demonstrate the efficiency of the proposed model, and the theoretic analysis is provided.
The rest of this paper is organized as follows. Section 2 provides the preliminaries. Section 3 reformulates the proposed equivalent problem with the mixup augmentation mechanism, and derives closed-form solutions via the stochastic gradient descent method, and theoretic analysis is provided. Section 4 exhibits a series of experimental results. Finally, Section 5 concludes this paper.

2. Definitions of Fairness Transfer Learning

A domain D is defined with a feature space X and the corresponding marginal distribution P ( X ) , where X = [ x 1 , x 2 , , x n ] X , n is the number of training samples. The learning task T on the domain D is to learn a classifier f that can project the features X X to the corresponding labels Y Y . As for transfer learning, given a source domain D S = { ( x s 1 , y s 1 ) , ( x s 2 , y s 2 ) , , ( x s n , y s n ) } and a target domain D T = { ( x t 1 , y t 1 ) , ( x t 2 , y t 2 ) , , ( x t n , y t n ) } , where the two margin distributions are different, P ( X S ) P ( X T ) . Usually in transfer learning { X s D S } { X t D T } , but Y s and Y t share the same class information. If D S = D T , the problem becomes a traditional machine-learning problem.
The corresponding tasks of the two domains are denoted as T S and T T . It has been demonstrated that a classifier trained with the samples from the source domain will not perform optimally on the target domain if the two marginal distributions are different [17]. Therefore, the goal of transfer learning is to improve the transferability and generalization of the classifier in task T T using the knowledge from the source domain set D S and the task T S .
In the fairness learning scenario, the alignment and the separation of probability distributions are difficult due to the lack of training data. The fairness transfer learning setting can be divided into two categories. One class is with sufficient well-labeled samples in the source domain and a few labeled samples in the target domain. Some work has been done to solve such a problem [18,19]. Most of them focus on extending adversarial learning to exploit the label information of target samples. The other class is the fairness source domain transfer learning problem, which is more challenging than the first class, for it aims to classify the unlabeled target samples with a few labeled source samples. Little work has been done for transfer learning under such fairness settings.
In this paper, we study the problem of fairness source domain transfer learning and define the settings as follows.
Assume the training samples in the source domain can be categorized into two sub-sets: the minority set D S m (the classes with limit samples) and the default set D S d (the classes with sufficient samples). The minority set and the default set have no visual features or class information overlapped. In addition, we want to train a classifier f on the imbalanced source domain and generalize the f to have a good performance on the unlabeled target domain. If the average loss of minority classes and the default classes are directly used to train the classifier, that may lead to bias towards default classes and unsatisfied classification results on the minority classes.
Denote the joint distributions probability over features X and domain D S m / D S d of the classifier f as P ( f ( X S m ) = True | D S m ) and P ( f ( X S d ) = True | D S d ) , respectively. The ( f ( · ) = True | D S i ) , denotes the correctly classified samples in the specific source sub-domain, where i = { D S m , D S d } . Then the balanced error rate with respect to the joint distribution of the feature X can be defined as follows
B E R ( f , X , D S ) = P ( f ( X ) = True | D S m ) + 1 P ( f ( X ) = True | D S d ) 2
where B E R ( f , X , D S ) is the misclassification error of f when the classes of the fairness domain and default domain are equally like, i.e., P ( D S m ) = P ( D S d ) = 1 / 2 , which means the probability of correctly classified results is the same across the groups. And according to [20], the classifier f has disparate impact at level τ , if and only if
B E R ( f , X , D S ) 1 2 P ( f ( X ) = True | D S m ) 2 ( 1 τ 1 ) .
That indicates that, in principle, we can modify the classifier or the input data to eliminate possible classifier-related differences.

3. Linear Data Augmentation Based Optimal Transport Model

To tackle the problem mentioned above, we focus on changing the data distribution of the fairness source domain to ensure the classifier trained from the modified source domain would be fair overall classes.
First, we introduce the baseline of the proposed model. The optimal transport problem is to seek a transformation T that aligns source distribution P ( X S ) to target distribution P ( X T ) , defined as follows
T 0 = arg min T d ( x , T ( x ) ) d P ( X S ) ( x ) , s . t . T P ( X S ) = P ( X T ) ,
where T P ( X S ) is the image map of P ( X S ) to P ( X T ) by T . When T 0 exists, it is called an optimal transport map.
Then the original optimal transport problem (1) can be relaxed to Kantorovitch problem [21], which aims to find a transport plan over the two distributions
γ 0 = arg min γ d ( x 1 , x 2 ) d γ ( x 1 , x 2 ) ,
where γ ( P ( X S ) , P ( X T ) ) , and ( P ( X S ) , P ( X T ) ) = { γ | p + γ = P ( X S ) , p γ = P ( X T ) } , p + and p are the two marginal projections of the joint distributions.
As mentioned in [13], the changes in marginal and conditional distributions are all taken into consideration. It seeks a map T that can align the joint distributions of source and target domains. Following the Kantovorich formulation, then we have
γ 0 = arg min γ D ( x i , y i ; x j , y j ) d γ ( x i , y i ; x j , y j ) ,
where D ( x i , y i ; x j , y j ) = d ( x i , x j ) + L ( y i , y j ) which calculates the distances of the features and the discrepancy of the labels. ( x i , y i ) is the sample from the source domain and ( x j , y j ) the sample from the target domain. The label information y j of the target sample is unknown, but it can be obtained by the classifier y j = f ( x i ) .
Our goal is to train a classifier f from the source domain that can perform well on the target domain, which can optimally match the labels of source domain samples with the features of the target domain in the transport plan. Thus, the joint distribution optimal transport problem in discrete form can be formulated as
min γ i , j D ( x i , y i ; x j , y j ) γ i j .
Additionally, a regularization term is added to the classifier, and f is to be updated while learning the optimal coupling matrix γ :
min γ , f i , j γ i j ( d ( x i , x j ) + L ( y i , f ( x j ) ) + δ Ω ( f ) ,
where Ω is the constraint on f, and L is continuous and differentiable with respect to its second variable. δ is the trade-off parameter.
To tackle the fairness transfer learning problem, most methods utilize GANs to augment the samples to have a balanced data distribution. However, there is a more simple and efficient method proposed for linear data augmentation, named mixup [22]. The mixup is to pair similar samples in the training set, formalized by the Vicinal Risk Minimization principle. It assumes that the examples in the vicinity share the label information and does not model proximity in different classes of examples. Mixup interpolates the training samples as follows
x m s = λ x i s + ( 1 λ ) x j s , y m s = λ y i s + ( 1 λ ) y j s ,
where ( x i s , y i s ) and ( x j s , y j s ) are the training samples randomly selected from the fairness source domain. The distributions of the two new sub-source domains are denoted as μ 1 s and μ 2 s , respectively. The labels are usually set as one-hot labels, and λ [ 0 , 1 ] is the mixup parameter. Mixup extends the source distribution by incorporating feature vectors. It can be simply implemented and introduces minimal computation overhead.
Based on the spirit of JDOT, which is formulated as Equation (5), we combine the mixup data obtained from Equation (6) with the optimal transport model, and the model becomes:
min γ , f i , j γ i j ( d ( x i m , x j t ) + L ( y i m , y j t ) ) + δ Ω ( f ) ,
where ( x i m , y i m ) and ( x j t , y j t ) are the samples from the mixup source domain and the target domain, respectively. δ is the penalty parameter.
The parameters in (7) can be optimized alternatively. The optimization process of the proposed mixup optimal transport model for the fairness transfer learning problem is stated as follows:
Step 1: Given the fairness source samples, augment the data by Equation (6) and obtain the mixup source domain.
Step 2: Fix the parameters of f in Equation (7), optimize γ with simplex flow algorithm.
Step 3: Fix the transformation matrix γ in Equation (7) and update the parameters of f by stochastic gradient method.
Step 4: Check the convergence condition; if satisfied, end the optimization process.
Remark 1.
The mixup procedure randomly changes the distribution of the original source domain by selecting the sample pairs and doing the interpolation. The degree of mixup is dominated by the mixup parameter λ, which can imply the total variation distance of the two conditional distributions. According to [23], the total distance variation (TDV) of the two probabilities distributions P 1 , P 2 can be calculated as
d T D V ( P 1 , P 2 ) = min π ( P 1 , P 2 ) π ( x y ) .
Let M be the target variable in distribution μ i s , where s { 1 , 2 } . T i is the transformation that push each source distribution μ i towards the target distribution μ t , i.e., T μ i = μ t . Set R i = T i 1 and R i ( M ) follows the original source distribution μ s . Then we have:
d T D V ( μ 1 s , μ 2 s ) P ( λ M + ( 1 λ ) R 1 ( M ) λ M + ( 1 λ ) R 2 ( M ) ) = 1 P ( λ M + ( 1 λ ) R 1 ( M ) = λ M + ( 1 λ ) R 2 ( M ) ) 1 λ .
This bound ensures the distribution of the two mixup source domains is a constraint in a constraint. If λ = 0 , it leaves the original distributions unchanged.

4. Experiment

In the following experiments, we test the proposed method on regression tasks, simple toy classification tasks, digital transfer classification tasks, and object transfer classification tasks, respectively. For the fairness transfer learning problem in this paper, the ResNet-50 is utilized as the backbone of the proposed model. The optimal transport embedded neural network is trained using stochastic gradient descent (SGD) with a batch size of 32 and a learning rate of 0.001. The SGD optimizer uses a momentum of 0.9 and a weight decay of 0.001. The bottleneck dimension for the features is set to 2048. The δ is set as 1. The mixup parameter λ follows the beta distribution B e t a ( α , α ) and we set the α as 2.

4.1. Regression Examples

First, we test the proposed model in a simple transfer learning regression problem under the traditional settings that the source domain has sufficient samples for training. The distributions of the two domains are shown in Figure 1a. The blue dots are the source samples, and the orange dots are the target samples. Each domain obtains 200 samples. The two domains have a similar amplitude in y, but different distributions in x. The (x, y) in the figures represent the samples’ corresponding coordinates after embedding. If the regression model learned from the source domain is applied directly to the target domain, the result is not satisfactory. However, the regression model learned from the JDOT model can be generalized to the source domain and target domain as well. When it comes to fairness transfer learning settings, the number of training samples in the source domain is reduced from 200 to 40. The fairness data distributions and the learned models are demonstrated in Figure 1b.
Then, we adapt the traditional optimal transport method JDOT and our proposed method MixupJDOT to the fairness transfer regression problem. First, we augment the source domain by mixup, randomly select the paired training samples, and obtain the new mixup samples by Equation (3). Then, we adapt our proposed method to the fairness transfer regression problem. Figure 2a presents the three regression models, the blue line and the green line are the regression models learned on the source domain and target domain, respectively, and the orange line is the transfer regression model learned from the JDOT model. Figure 2b demonstrates the results of MixupJDOT. Compare the results in Figure 2, it is easy to observe that the optimal transport model is less efficient in dealing with the fairness optimal transport problem, for there is a lack of samples that can align the source distribution and target distribution effectively.
The new source samples are visualized in Figure 3a. The orange dots are the original source samples, and the blue dots are the augmented samples. The distribution of the new source domain has changed but still is a constraint in a certain bound. Then, we applied the mixup optimal transport embedded neural network on the new source domain and the target domain. The alignment results of the two distributions based on the mixup optimal transport model are shown in Figure 3b, which demonstrates the optimal matrix on the two distributions, and we can see that the augmented samples play an important role in the alignment. The mixup optimal transport-based regression model has better generalization ability.

4.2. Classification Examples

Then, we test our proposed model on fairness transfer learning classification tasks. Samples are generated following three Gaussian distributions: one of the categories only contains 30 instances, and the other two categories have 100 instances in each category. The visualization of the source domain and target domain is presented in Figure 4. Figure 4a is the distribution of the source domain, the class represented by the purple dots is the minority set, and the other two classes in yellow and green dots are the default set. Figure 4b is the distribution of the target domain. From the figures, we can see that the samples in the two domains lie in quite different distributions. If the model trained on the source domain is directly utilized in the target domain, the classification results are not competitive at all.
The classification results of the baseline JDOT and the proposed mixup optimal transport model are reported in Figure 5. Figure 5a is the classification result of JDOT, and Figure 5b is the result of the proposed mixup optimal transport model. The classifier trained with the mixup source domain has a larger and more correct classification area, which implies that the mixup term has a significant contribution to the fairness classification problem compared with the baseline JDOT. The purple area in Figure 5b is larger than that in Figure 5a, and the classifier bound in Figure 5b is more accurate.

4.3. Digital Classification

Then, we estimate the proposed optimal transport-embedded neural network and some domain adaptation methods in digit classification tasks on MNIST and USPS datasets. Both datasets have a 10-class classification problem and contain pictures of numbers from 0 to 9. The MNIST dataset contains 60,000 samples for training and 10,000 samples for testing. The USPS dataset has 7291 samples for training and 2007 samples for testing. The images are in grayscale.
In the transfer classification tasks, experiments are under (5-way, 10-shot) setting, which means 5 classes are selected as the minority set, and the other 5 classes are the default set. The minority set only contains 10 samples for each class, and the classes in the default set have sufficient training samples. In the experiments, only the training samples are utilized for such transfer tasks. For example, in the MNIST→USPS transfer tasks, the 60,000 training samples in the MNIST dataset are used for training, and the 7291 training samples in the USPS dataset are for testing. And we conduct the two transfer learning tasks in this subsection, MNIST→USPS, and USPS→MNIST tasks. We compared with the classifier trained on the source domain and applied it on the target domain without any alignment strategy, denoted as “CLF” in the results table, also compared with the domain adaptation methods DANN [24], MCD [25].
The comparison results are shown in Table 1. The domain adaptation methods DANN and MCD are slightly higher than the simple classifier trained on the source domain, but the results are not ideal. And the results of the proposed mixup optimal transport model are much higher than the other domain adaptation methods. Compared with the second-highest method in Table 1, the proposed method is almost 13 % higher in MNIST→USPS task and about 4 % higher in USPS→MNIST task. The comparison results prove that the proposed mixup optimal transport model can improve the fairness transfer learning problem.

4.4. Domain Adaptation

Moreover, we evaluate the model on two well-known datasets, namely Office [27] and OfficeHome [28]. The Office dataset is a real-world dataset. Amazon, DSLR, and Webcam are 3 of the 31 classes it has. A demonstration of the samples from the different domains is presented in Figure 6. In this dataset, experiments are carried out with 1-shot and 3-shot source labels per class. OfficeHome dataset has 65 classes in 4 domains (Art, Clipart, Product, and Real). According to the widely used settings [29], we examine the settings with 3% and 6% labeled source photos per class.
The foundation of the studies is the ResNet-50 pre-trained on ImageNet [30]. We employ SGD with 64 batches, a learning rate of 0.01, and a momentum of 0.9. The proposed method is compared with several state-of-the-art methods on the fairness domain adaptation problem (few-shot unsupervised domain adaptation). The classifier that trained on the source domain and tested on the target domain is denoted as CLF in the following tables. Also, the proposed method are compared with MME [31], CDAN [32], CAN [33], and CDS [29].
The experiment results of the above methods on Office and OfficeHome datasets under different few-shot settings are presented in Table 2, Table 3, Table 4, and Table 5, respectively. We can see that the proposed MixupJDOT outperforms all the other methods in all the benchmarks. Compared with the second-best method, the proposed method makes large improvements: 4.8% and 3.9% on the Office dataset, 2.4% and 3.3% on the OfficeHome dataset. The results demonstrate the efficiency of the proposed method in fairness transfer learning scenarios.

5. Conclusions

In this paper, a novel mixup optimal transport embedded neural network is constructed based on the joint distribution optimal transport theory. The proposed neural network targets the fairness transfer learning problem, and the mixup mechanism is adapted to augment the training source samples on convex combinations of pairs of examples and their labels. Experiments on regression and classification results verified that the proposed methods could significantly improve performance under fairness settings. In fairness transfer learning tasks using 3% labeled source samples per class on the OfficeHome dataset, which is the most challenging benchmark, the experimental findings demonstrate that our method’s accuracy is 2.4% greater than that of the second-best mode. On the OfficeHome dataset, where transfer learning tasks 6% identified source samples per class, our method’s accuracy is 3.3% higher than the second-best mode. However, there are still some drawbacks to the proposed method. Only data regression and image classification are employed in this paper’s application, and there is only one data modality. In future work, we will try to extend the framework to a more complicated model for several widely used benchmarks and real-world applications and utilize the model to power edge AI applications.

Author Contributions

The authors confirm contribution to the paper as follows: study conception and design: M.X. and Y.Z.; data collection: Z.L.; analysis and interpretation of results: Z.L. and Q.L. Author; draft manuscript preparation: M.X. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the fundamental research funds for the State Grid Hubei Electric Power Co., Ltd. Key Technology Project, No. 52153222000F and 5215D0220001.

Data Availability Statement

The access to the data used in the study can be downloaded from the website in the reference. And the code of this study will be uploaded to the Github after the paper being accepted.

Acknowledgments

The authors wish to express their appreciation to the reviewers for their helpful suggestions which greatly improved the presentation of this paper.

Conflicts of Interest

The authors declare that they have no conflict of interest to report regarding the present study.

References

  1. Shekhar, S.; Schrater, P.; Vatsavai, R.; Wu, W.; Chawla, S. Spatial contextual classification and prediction models for mining geospatial data. IEEE Trans. Multimed. 2002, 4, 174–188. [Google Scholar] [CrossRef]
  2. Xu, B.; Zeng, Z.; Lian, C.; Ding, Z. Few-shot domain adaptation via mixup optimal transport. IEEE Trans. Image Process. 2022, 31, 2518–2528. [Google Scholar] [CrossRef] [PubMed]
  3. Kong, L.; Zuo, Y. Smooth depth contours characterize the underlying distribution. J. Multivar. Anal. 2010, 101, 2222–2226. [Google Scholar] [CrossRef]
  4. Ni, R.; Goldblum, M.; Sharaf, A.; Kong, K.; Goldstein, T. Data augmentation for meta-learning. In Proceedings of the International Conference on Machine Learning. PMLR, Virtual, 18–24 July 2021; pp. 8152–8161. [Google Scholar]
  5. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the Advances in Neural Information Processing Systems 27 (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar]
  6. Wang, M.; Deng, W. Deep visual domain adaptation: A survey. Neurocomputing 2018, 312, 135–153. [Google Scholar] [CrossRef]
  7. Wu, A.; Zeng, Z. Observer design and performance for discrete-time uncertain fuzzy-logic systems. IEEE Trans. Cybern. 2020, 51, 2398–2408. [Google Scholar] [CrossRef] [PubMed]
  8. Villani, C. Optimal Transport: Old and New; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008; Volume 338. [Google Scholar]
  9. Perrot, M.; Courty, N.; Flamary, R.; Habrard, A. Mapping estimation for discrete optimal transport. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 4204–4212. [Google Scholar]
  10. Courty, N.; Flamary, R.; Tuia, D. Domain adaptation with regularized optimal transport. In Proceedings of the in Machine Learning and Knowledge Discovery in Databases: European Conference, ECML PKDD 2014, Nancy, France, 15–19 September 2014; Proceedings, Part I 14. Springer: Berlin/Heidelberg, Germany, 2014; pp. 274–289. [Google Scholar]
  11. Rubner, Y.; Tomasi, C.; Guibas, L.J. The earth mover’s distance as a metric for image retrieval. Int. J. Comput. Vis. 2000, 40, 99–121. [Google Scholar] [CrossRef]
  12. Rabin, J.; Peyré, G.; Delon, J.; Bernot, M. Wasserstein barycenter and its application to texture mixing. In Proceedings of the Scale Space and Variational Methods in Computer Vision: Third International Conference, SSVM 2011, Ein-Gedi, Israel, 29 May–2 June 2011; Revised Selected Papers 3. Springer: Berlin/Heidelberg, Germany, 2011; pp. 435–446. [Google Scholar]
  13. Courty, N.; Flamary, R.; Habrard, A.; Rakotomamonjy, A. Joint distribution optimal transportation for domain adaptation. arXiv 2017, arXiv:1705.08848. [Google Scholar]
  14. Damodaran, B.B.; Kellenberger, B.; Flamary, R.; Tuia, D.; Courty, N. Deepjdot: Deep joint distribution optimal transport for unsupervised domain adaptation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 447–463. [Google Scholar]
  15. Qi, H.; Brown, M.; Lowe, D.G. Low-shot learning with imprinted weights. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 5822–5830. [Google Scholar]
  16. Qiao, S.; Liu, C.; Shen, W.; Yuille, A.L. Few-shot image recognition by predicting parameters from activations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7229–7238. [Google Scholar]
  17. Shimodaira, H. Improving predictive inference under covariate shift by weighting the log-likelihood function. J. Stat. Plan. Inference 2000, 90, 227–244. [Google Scholar] [CrossRef]
  18. Motiian, S.; Jones, Q.; Iranmanesh, S.M.; Doretto, G. Few-shot adversarial domain adaptation. arXiv 2017, arXiv:1711.02536. [Google Scholar]
  19. Zhao, A.; Ding, M.; Lu, Z.; Xiang, T.; Niu, Y.; Guan, J.; Wen, J.R. Domain-adaptive few-shot learning. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Virtual, 5–9 January 2021; pp. 1390–1399. [Google Scholar]
  20. Gordaliza, P.; Del Barrio, E.; Fabrice, G.; Loubes, J.M. Obtaining fairness using optimal transport theory. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 2357–2365. [Google Scholar]
  21. Kantorovich, L.V. On the translocation of masses. J. Math. Sci. 2006, 133, 1381–1382. [Google Scholar] [CrossRef]
  22. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  23. Massart, P. Concentration Inequalities and Model Selection: Ecole d’Eté de Probabilités de Saint-Flour XXXIII-2003; Springer: Berlin/Heidelberg, Germany, 2007. [Google Scholar]
  24. Ganin, Y.; Lempitsky, V. Unsupervised domain adaptation by backpropagation. In Proceedings of the International Conference on Machine Learning, Lille, France, 6–11 July 2015; pp. 1180–1189. [Google Scholar]
  25. Saito, K.; Watanabe, K.; Ushiku, Y.; Harada, T. Maximum classifier discrepancy for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3723–3732. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  27. Saenko, K.; Kulis, B.; Fritz, M.; Darrell, T. Adapting visual category models to new domains. In Proceedings of the Computer Vision—ECCV 2010: 11th European Conference on Computer Vision, Heraklion, Crete, Greece, 5–11 September 2010; Proceedings, Part IV 11. Springer: Berlin/Heidelberg, Germany, 2010; pp. 213–226. [Google Scholar]
  28. Venkateswara, H.; Eusebio, J.; Chakraborty, S.; Panchanathan, S. Deep hashing network for unsupervised domain adaptation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 5018–5027. [Google Scholar]
  29. Kim, D.; Saito, K.; Oh, T.H.; Plummer, B.A.; Sclaroff, S.; Saenko, K. Cross-domain self-supervised learning for domain adaptation with few source labels. arXiv 2020, arXiv:2003.08264. [Google Scholar]
  30. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  31. Saito, K.; Kim, D.; Sclaroff, S.; Darrell, T.; Saenko, K. Semi-supervised domain adaptation via minimax entropy. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 16–20 June 2019; pp. 8050–8058. [Google Scholar]
  32. Long, M.; Cao, Z.; Wang, J.; Jordan, M.I. Conditional adversarial domain adaptation. In Proceedings of the Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  33. Kang, G.; Jiang, L.; Yang, Y.; Hauptmann, A.G. Contrastive adaptation network for unsupervised domain adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4893–4902. [Google Scholar]
Figure 1. Visualization of (a) the distributions of the source domain and target domain under traditional transfer learning setting; (b) the distributions of the fairness source domain and target domain under normal transfer learning setting.
Figure 1. Visualization of (a) the distributions of the source domain and target domain under traditional transfer learning setting; (b) the distributions of the fairness source domain and target domain under normal transfer learning setting.
Electronics 12 04481 g001
Figure 2. Visualization of (a) the regression model trained on source domain, target domain, and with JDOT model; (b) the regression model trained on mixup source domain, target domain, and with MixupJDOT model, respectively. The blue dots present the source samples, and the red dots present the target samples. The lines of different colors denote the different regression models.
Figure 2. Visualization of (a) the regression model trained on source domain, target domain, and with JDOT model; (b) the regression model trained on mixup source domain, target domain, and with MixupJDOT model, respectively. The blue dots present the source samples, and the red dots present the target samples. The lines of different colors denote the different regression models.
Electronics 12 04481 g002
Figure 3. Visualization of (a) the mixup source domain; (b) the optimal transport matrix of alignment between the two distributions. The black lines are the alignments between samples from different domains.
Figure 3. Visualization of (a) the mixup source domain; (b) the optimal transport matrix of alignment between the two distributions. The black lines are the alignments between samples from different domains.
Electronics 12 04481 g003
Figure 4. Visualization of (a) fairness source domain; (b) target domain. The green and yellow dots are the samples from the normal classes, and the purple dots are the samples from the class which has fewer samples.
Figure 4. Visualization of (a) fairness source domain; (b) target domain. The green and yellow dots are the samples from the normal classes, and the purple dots are the samples from the class which has fewer samples.
Electronics 12 04481 g004
Figure 5. Classification results of (a) JDOT; (b) mixup optimal transport model. The different color areas present the classification results of each class.
Figure 5. Classification results of (a) JDOT; (b) mixup optimal transport model. The different color areas present the classification results of each class.
Electronics 12 04481 g005
Figure 6. Some examples from the Office dataset.
Figure 6. Some examples from the Office dataset.
Electronics 12 04481 g006
Table 1. Accuracy (%) of fairness transfer learning tasks (5-way, 10-shot) on MNIST and USPS datasets.
Table 1. Accuracy (%) of fairness transfer learning tasks (5-way, 10-shot) on MNIST and USPS datasets.
MethodCLF [26]DANN [24]MCD [25]Ours
MNIST→USPS40.6742.6941.2755.93
USPS→MNIST50.6152.0451.3256.37
Table 2. Accuracy (%) of fairness transfer learning tasks 1-shot per class on Office dataset.
Table 2. Accuracy (%) of fairness transfer learning tasks 1-shot per class on Office dataset.
MethodOffice
A→DA→WD→AD→WW→AW→DAvg.
CLF [26]27.528.740.965.241.162.044.2
MME [31]21.512.223.160.914.062.432.3
CDAN [32]11.26.29.154.810.441.622.2
CAN [33]25.326.423.969.421.267.338.9
CDS [29]33.335.252.059.046.557.447.2
Ours36.643.547.175.048.261.452.0
Table 3. Accuracy (%) of fairness transfer learning tasks 3-shot per class on Office dataset.
Table 3. Accuracy (%) of fairness transfer learning tasks 3-shot per class on Office dataset.
MethodOffice
A→DA→WD→AD→WW→AW→DAvg.
CLF [26]49.246.355.385.553.886.162.7
MME [31]51.054.660.289.752.391.466.5
CDAN [32]43.750.165.191.657.089.866.2
CAN [33]48.645.341.278.239.382.355.8
CDS [29]57.058.667.686.065.781.369.3
Ours69.773.265.286.165.080.173.2
Table 4. Accuracy (%) of fairness transfer learning tasks 3% labeled source samples per class on OfficeHome dataset.
Table 4. Accuracy (%) of fairness transfer learning tasks 3% labeled source samples per class on OfficeHome dataset.
MethodOfficeHome
Ar Ar Ar Cl Cl Cl Pr Pr Pr Rw Rw Rw Avg.
Cl Pr Rw Ar Pr Rw Ar Cl Rw Ar Cl Pr
CLF [26]24.438.343.126.434.733.727.526.542.641.229.052.335.0
MME [31]4.515.425.028.734.137.025.625.444.939.329.052.030.1
CDAN [32]5.08.411.820.626.127.526.627.040.338.725.544.925.2
CAN [33]17.130.533.222.534.536.018.519.441.328.718.643.228.6
CDS [29]33.541.141.945.946.049.344.737.851.051.635.753.844.4
Ours32.237.242.239.341.539.450.744.962.961.348.460.646.8
Table 5. Accuracy (%) of fairness transfer learning tasks 6% labeled source samples per class on OfficeHome dataset.
Table 5. Accuracy (%) of fairness transfer learning tasks 6% labeled source samples per class on OfficeHome dataset.
MethodOfficeHome
Ar Ar Ar Cl Cl Cl Pr Pr Pr Rw Rw Rw Avg.
Cl Pr Rw Ar Pr Rw Ar Cl Rw Ar Cl Pr
CLF [26]28.745.751.231.939.844.137.630.854.649.936.061.842.7
MME [31]27.643.249.541.146.649.543.730.561.354.937.366.846.0
CDAN [32]26.233.744.534.842.944.742.936.059.354.940.163.643.6
CAN [33]20.434.744.729.040.438.633.321.153.436.819.158.035.8
CDS [29]38.851.754.853.253.357.053.444.265.263.745.368.654.1
Ours43.150.662.352.954.261.055.545.669.966.753.773.757.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiang, M.; Ling, Z.; Liu, Q.; Zhang, Y. Optimal Transport-Embedded Neural Network for Fairness Transfer Problem. Electronics 2023, 12, 4481. https://doi.org/10.3390/electronics12214481

AMA Style

Xiang M, Ling Z, Liu Q, Zhang Y. Optimal Transport-Embedded Neural Network for Fairness Transfer Problem. Electronics. 2023; 12(21):4481. https://doi.org/10.3390/electronics12214481

Chicago/Turabian Style

Xiang, Muchao, Zaixun Ling, Qine Liu, and Yaoxuan Zhang. 2023. "Optimal Transport-Embedded Neural Network for Fairness Transfer Problem" Electronics 12, no. 21: 4481. https://doi.org/10.3390/electronics12214481

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop