Next Article in Journal
Remark on Intuitionistic Fuzzy Temporal Modal Topological Structures
Previous Article in Journal
An Extension of the Fréchet Distribution and Applications
Previous Article in Special Issue
A Discrete Characterization of the Solvability of Equilibrium Problems and Its Application to Game Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Federated Learning Incentive Mechanism with Supervised Fuzzy Shapley Value

1
School of Mathematics and Statistics, Guizhou University, Guiyang 550025, China
2
State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, China
3
Guizhou Big Data Academy, Guizhou University, Guiyang 550025, China
4
School of Data Science and Information Engineering, Guizhou Minzu University, Guiyang 550025, China
5
Information School, Guizhou University of Finance and Economics, Guiyang 550025, China
*
Author to whom correspondence should be addressed.
Axioms 2024, 13(4), 254; https://doi.org/10.3390/axioms13040254
Submission received: 2 March 2024 / Revised: 4 April 2024 / Accepted: 9 April 2024 / Published: 11 April 2024
(This article belongs to the Special Issue Advances in Logic and Game Theory)

Abstract

:
The distributed training of federated machine learning, referred to as federated learning (FL), is discussed in models by multiple participants using local data without compromising data privacy and violating laws. In this paper, we consider the training of federated machine models with uncertain participation attitudes and uncertain benefits of each federated participant, and to encourage all participants to train the desired FL models, we design a fuzzy Shapley value incentive mechanism with supervision. In this incentive mechanism, if the supervision of the supervised mechanism detects that the payoffs of a federated participant reach a value that satisfies the Pareto optimality condition, the federated participant receives a distribution of federated payoffs. The results of numerical experiments demonstrate that the mechanism successfully achieves a fair and Pareto optimal distribution of payoffs. The contradiction between fairness and Pareto-efficient optimization is solved by introducing a supervised mechanism.

1. Introduction

The swift progression of Artificial Intelligence (AI) has caused a considerable surge in the generation of large and diverse datasets. These datasets encompass valuable information and sensitive privacy details from various domains. Notably, data islands and privacy concerns pose significant challenges in harnessing the full potential of these datasets. To address the challenges associated with data islands and data privacy, a distributed machine learning technique known as federated learning (FL) [1,2,3] was introduced by Google in 2016.
Cicceri et al. [4] put forth a proposal for the utilization of FL in healthcare, which is called “DILoCC” and built on the concept of FL and adopts the Distributed Incremental Learning (DIL) methodology to leverage the cooperation between sensing devices to build advanced ICT systems with predictive capabilities, which are capable of saving lives and avoiding economic losses. Furthermore, in realistic heterogeneous data distribution scenarios, the performance of FL applied on non-independent identically distributed (Non-IID) data degrades. Therefore, Zhang et al. [5] proposed federated continual learning (FCL) and Bidirectional Compression and Error Compensation (BCEC) algorithms to enhance the performance of non-IID data as well as to reduce communication overheads by introducing knowledge from other local models. FL enables model training to be performed locally at each node, ensuring that sensitive data remains secure and confidential without being transmitted or leaked. This approach not only maximizes the utilization of data for preferring the safeguarding of delicate data privacy, model training is given precedence [6].
Establishing a robust incentive mechanism is crucial in the FL system to encourage data owners to contribute their data and enhance model accuracy. Addressing free-riding behavior among participants, which can significantly impact the accuracy of the federated model, is a key research topic. Therefore, the development of an effective incentive mechanism holds great value in the FL system. Currently, most of the research work [7,8,9,10] assumes the completion of federated model training in the context of determining the attitudes of federated participants’ participation [11,12,13]. However, in actual model training, due to a variety of reasons, such as limited computational resources, weak economic base, limited energy, etc., it can lead to the participants’ participation attitude being uncertain and not being fully committed to the model training, i.e., the participation attitude is ambiguous.
Building upon the aforementioned points, our objective is to devise an incentive structure that guarantees fair allocation of rewards amidst uncertainty regarding the attitudes of federated participants. Consequently, we construct fuzzy Shapley-valued incentives with a supervisory mechanism under fuzzy conditions, which takes into account both fairness and Pareto efficiency considerations. However, this mechanism will increase the complexity, were the mechanism to be applied to a genuine FL situation, how would one acquire and process information about the attitudes of participants towards participation and how to design and implement the supervision mechanism. Summarizing the paper’s main contributions, they can be expressed thus:
(1)
Amidst the ambiguity and doubt of the participants’ involvement outlooks in FL, this paper proposes a fuzzy Shapley value method that can more accurately assess the degree of participants’ contribution and payoff distribution;
(2)
The clash between equity and Pareto effectiveness must be deliberated, as equity necessitates a fair distribution of payoffs among participants, while Pareto efficiency seeks that no participant’s payoff could increase without jeopardizing the payoffs of other participants;
(3)
This paper guarantees fairness and optimization of Pareto efficiency with consistency and introduces a supervisory mechanism that monitors and adjusts the behavior of participants. To guarantee that participants are treated justly, the allotment of advantages is essential in FL and to maximize the overall payoffs.
The paper is organized in the following way: An overview of the related work in the field is presented in Section 2. Preliminaries, providing necessary background information, are presented in Section 3. The FL incentive mechanism and its design are outlined in Section 4.Verification of the incentive mechanism proposal by means of numerical experiments is carried out in Section 5. A comprehensive discussion of the findings are provided in Section 6. The conclusions of this paper are described in Section 7. Proving the Theorem is given in Appendix A.

2. Related Work

The current landscape of FL incentive mechanism research comprises several prominent theoretical approaches. These include the application of the Stackelberg game [14], contract theory’s tenets [15], auctioning techniques [16], and the incorporation of the Shapley value [17]. This paper provides a comprehensive review of the research literature focusing on the design of incentive mechanisms for FL using the Shapley value. In [18], to evaluate the individual contributions of participants who own the data in training the FL model, a novel Shapley value was introduced, which is based on the contribution metric. In [19], a novel formulation for the Shapley value in FL was suggested, which eliminates the need for extra communication costs, incorporates the value of FL data, and offers incentives to the participants. In their work [20], Wang et al. employed the Shapley value to accurately assess the contribution of each participant in FL. To incentivize data owners to actively participate in federated model training and contribute their data, Liu et al. proposed a blockchain-based peer-to-peer payment system for FL in [21]. In this study [21], a blockchain-based peer-to-peer payment system was introduced for FL with the aim of providing incentives to data owners and promoting their active participation in model training. The main objective of this system is to provide a fair and practical mechanism for allocating payoffs, utilizing the principles of the Shapley value. The paper [22] presented an incentive mechanism derived from the Shapley value, aiming to promote greater participation and data sharing among participants by offering them a fair compensation. Liu et al. proposed the bootstrap-truncated gradient Shapley approach in their work [23], which focuses on the fair valuation of participants’ contributions in FL. This strategy entails reconstructing the FL model by utilizing gradient updates for computing the Shapley value. In the work [24], Nagalapattiet et al. presented a collaborative game framework in which participants exchange gradients and calculate Shapley values to identify individuals with pertinent data. The work [24] presented a cooperative game framework that facilitates the sharing of gradients among participants and utilizes Shapley values to identify individuals with relevant data. Acknowledging the disparities in computing the Shapley value for FL, Fan et al. [25] introduced an innovative comprehensive mechanism for calculating the FL Shapley value, aiming to promote fairness. Additionally, in order to mitigate the communication overhead involved in computing the Shapley value in FL, Fan et al. [26] introduced a federated Shapley value mechanism to evaluate participants’ contributions.
In [27], the authors conducted an investigation into different factors that impact FL. Additionally, they proposed an FL incentive mechanism that leverages an improved Shapley value method. In [28], an incentive design called heterogeneous client selection (IHCS) was proposed to enhance performance and mitigate security risks in FL, an approach that involves assigning a recognition value to each client using the Shapley value and is subsequently utilized to aggregate the probability of participation level. In [29], an incentive mechanism is introduced that combines the Shapley value and Pareto efficiency optimization. This approach entails incorporating a third-party entity to oversee the distribution of federated payoffs. If the payoffs can attain Pareto optimality, they are allocated utilizing the Shapley value methodology.

3. Preliminaries

3.1. Federated Learning Framework

FL is a decentralized machine learning framework designed to address data fragmentation and promote cooperation within the field of AI. The primary goal of FL is to facilitate the training of models without the need for participants’ data to leave their local environments. According to adopt this approach, collaborative model development is achieved while also ensuring the preservation of data privacy, security, and legal compliance. By breaking down barriers and allowing for decentralized training, FL promotes the sharing of knowledge and insights while preserving the confidentiality of individual data sources [1,6].
Let N = { 1 , 2 , , n } represent a set of n data participants who participate in the FL process denoted as { F 1 , F 2 , F 3 , , F n } . Each participant possesses their local dataset { D 1 , D 2 , , D n } . The shared model required for FL is denoted as M F E D , while M S U M represents the traditional machine learning model. The model accuracy of M F E D and M S U M is denoted as V F E D and V S U M , respectively. We can state that there exists a non-negative number δ 0 such that
| V F E D V S U M | < δ ,
which represents the δ -accuracy loss of the FL algorithm [6].
Figure 1 illustrates the FL framework’s schematic diagram, and training is included can be summarized as follows:
Step 1:
Individual federated participants retrieve the initial global model from the aggregation server.
Step 2:
Each participant trains their local model using the received initial global model.
Step 3:
Upon completing the local model training, participants upload the updated model and associated parameters to the aggregation server.
Step 4:
The aggregation server consolidates the models and parameters uploaded by each participant for the subsequent round of updates.
The widely adopted aggregation approach is the Federated Averaging (FedAvg) algorithm [30], and Steps 2–3 are iteratively performed until the local model reaches convergence.

3.2. Cooperative Games

Consider a cooperative game denoted as G ( N , v ) , which satisfies the following conditions [31]:
v ( S 1 ) + v ( S 2 ) v ( S 1 S 2 ) ,
S 1 S 2 = , v ( ) = 0 .
Here, N denotes a finite group of participants, S 1 , S 2 2 N , and v : 2 N R as a game-characteristic function, 2 N denotes the set of all subsets of N, v ( S ) as the payoff function for participants and v ( N ) as the coalition payoff. Additionally, φ i ( v ) represents the payoff of participant i in v ( N ) and satisfies two specific conditions:
v ( N ) = φ i ( v ) and v ( i ) φ i ( v ) , i N , i = 1 , 2 , , n ,
v ( S ) i S φ i ( v ) , S N , S .
Within this context, Equations (4) and (5) are widely recognized as the criteria for individual rationality and coalition rationality, respectively.

3.3. Shapley Value

The Shapley value was the originator of Shapley in cooperative game theory [31], addresses the problem of fair allocation of payoffs in cooperative settings, which defined as follows:
φ i ( v ) = i S w ( | S | ) v ( S ) v ( S i ) ,
w ( | S | ) = ( n | S | ) ! ( | S | 1 ) ! n ! .
Let S be a subset of participants, with i being equal to i = 1 , 2 , , n , and S N , the value | S | represents the number of subset S. The weight coefficient w ( | S | ) is associated with the size of | S | . The profit of subset S is denoted as v ( S ) , while v ( S ) v ( S i ) quantifies the participant i s contribution to coalition S was marginal. Additionally, v ( S i ) represents the payoff of participants other than i within subset | S | .

3.4. Fuzzy Coalition of FL

Let N = { 1 , 2 , , n } denote the set consisting of coalition participants in FL, and any subset of the coalition formed by the participants be S, using the fuzzy number L χ = χ 1 , χ 2 , , χ n [ 0 , 1 ] N to denote the degree of participation of a participant i in a subset of the coalition S, i.e., S ( i ) : L ( N ) [ 0 , 1 ] N , where S ( i ) = 0 denotes that the participant i does not participate in the coalition subset S at all, and S ( i ) = 1 denotes that the participant i does participate in the coalition subset S, then S = S 1 , S 2 , , S n is said to be a fuzzy coalition in FL.
Fuzzy coalitions can help solve the problem of non-cooperativeness in FL, where participants may have selfish motives and tend to keep their data and models. Establishing a fuzzy coalition can lead to the establishment of trust and cooperation among the participants to reach the training goal together. Further, under the fuzzy coalition, the stability and efficiency of the FL system are ensured by designing incentive and punishment mechanisms to motivate the cooperation of the participants.

3.5. Choquet Integral

Let f : X [ 0 , + ) , μ is a fuzzy measure defined on X, then the Choquet integral [32,33] of f with respect to μ is defined as follows:
X f ( χ ) g ( · ) = 0 g F α d α
When X is a finite set and f : X [ 0 , 1 ] , the formula for the Choquet integral is redefined as follows:
( c ) f d μ = i = 1 n f χ i f χ i 1 μ A i
where F α = { χ f ( χ ) α , χ X } , A i = χ i , χ i + 1 , χ n , α is any number belonging to the group between (0,1], g ( · ) is the fuzzy density function. In general, assume that 0 = f χ 0 f χ 1 f χ 2 f χ n 1 . If this relationship is not satisfied, it needs to be readjusted to satisfy the relationship.
In addition, the importance of Choquet integral in FL is reflected in its ability to deal with uncertainty and ambiguity information, integrate and weigh the data, and be able to solve the problem of fair benefit distribution in FL in the case of ambiguity of participants’ attitudes, etc., which can greatly help to improve the accuracy and robustness of federated models.

3.6. Pareto Optimality

Considering π ˜ = ( π ˜ 1 , π ˜ 2 , , π ˜ N ) : x R N as a target vector. For any two solutions x i and x i * in x, then Pareto optimality [34] is considered and the following condition holds:
π ˜ i ( x 1 * , , x i * , , x N * ) π ˜ i ( x 1 , , x i , , x N ) , i N .
This concept aligns with the notion of Pareto optimality, where an allocation is deemed Pareto optimal if there is no alternative allocation that can improve the well-being of one participant without simultaneously worsening the well-being of another participant [35].

3.7. Nash Equilibrium

The participant i’s payoff function is π ˜ i ( x ) , where x = x 1 , x 2 , , x N T R N represents the vector of actions taken by all participants, and x i R n represents the action specifically taken by participant i. A Nash equilibrium is a state in which no participant can improve By unilaterally altering their behavior, they increased their income. More specifically, an action profile x i * ( i = 1 , , n ) is defined as a Nash equilibrium [36,37], if
π ˜ i ( x 1 * , , x i * , , x N * ) π ˜ i ( x 1 * , , x i , , x N * ) , i N .

4. FL System Incentive Mechanism

4.1. The FL Incentive Model

In establishing the FL incentive model, we base our analysis on the following assumptions:
(1)
Economic Participation: All participants in the FL framework are capable of making financial contributions towards the FL process.
(2)
Participant Satisfaction: The final distribution of payoffs is designed to satisfy all participants.
(3)
Trustworthiness and Integrity: It is assumed that all participants in the FL system are entirely trustworthy and exhibit no instances of cheating or dishonest behavior.
(4)
Multi-Party Agreement: To ensure seamless execution of the FL strategy, it is essential to establish a multi-party agreement.
Drawing upon the principles of FL, we put forth a FL incentive model, illustrated in Figure 2. The FL incentive mechanism can be delineated through a sequence of essential steps as follows:
Step 1:
Considering n participants in federated model training, each participant possesses its own local dataset, denoted as D i ;
Step 2:
Each participant downloads the initialized model from the aggregation server, they independently train the model using their respective local dataset D i , and upload their trained model m i to the federated aggregation server:
Step 3:
To acquire a fresh global model, the federated aggregation server assumes the responsibility of gathering the model parameters m i , and subsequently employs the federated aggregation algorithm to consolidate these parameters.
Step 4:
The fuzzy Shapley values method is employed to assess the individual contributions of each participant to the global model. This approach allows for the quantification of individual contributions, considering the uncertainty and fuzziness inherent in the participants’ involvement.
Step 5:
The supervising organization assesses the attainment of Pareto optimality for each participant’s payoff. If Pareto optimality is achieved, the fuzzy Shapley value formula is utilized to allocate the payoff. Conversely, the supervising organization imposes penalties on the participant, resulting in the forfeiture of a predetermined fine.
Step 6:
Rewards are allocated to participants who achieve Pareto optimality, as determined by the payoff allocation formula.

4.2. Federated Cost Utility Function

The federated participant P i needs to take up computational resources such as memory and CPU when performing local model training, and the performance of the participant set P training model w G * is defined as F = { F 1 , F 2 , F 3 , , F n } , where F i denotes the participant P i ’s local data model training at the CPU clock frequency. In addition, the device of participant P i has a baseline memory θ i with an average consumption ratio of β i t [ 0 , 1 ] , i.e., the actual memory consumption of β i t θ i , in the tth round of local iterations. Assuming the same sample size of input data for each local model iteration and based on the processor quadratic energy model, the computational cost function [13] of the participant P i in the local model iteration is represented by the following equation:
c com P i = α ζ e i γ i f i 2 + ( 1 α ) β i t θ i
Then, define the federated cost utility function of the participant P i P after the rth round of model aggregation as
c i r ( c com P i ) = ln c com P i
where α denotes the tuning factor used to adjust the computational cost of the participant P i , ζ denotes the performance of P i ’s hardware device used to compute the effective capacitance parameter of the chipset [38], e i comes to denote the number of CPU cycles required to process a batch of data, and γ i denotes the size of data batch samples required for each round of local iteration.

4.3. Model Quality Utility Function

In this paper, the utility assessment of federated participants is defined as two metrics, model quality and freshness [39], and in the classification prediction problem, since the cross-entropy loss function [40,41] outperforms the mean-square error in terms of effectiveness, the cross-entropy loss function H f y i , z i = i = 1 n z i log f y i , the labeled data information y i , z i  can be processed efficiently to improve the accuracy and reliability of the data. To quantify the freshness of the model, the concept of trusted time is introduced, and to ensure the verifiability of the timestamps, the federation participants need to request a timer T i k w i k  from the content container [42] in the Trusted Hardware SGX, after completing the iteration of the local model w i k . The freshness F w i k  of the model w i k can be defined as F w i k = lg T i k w i k . Based on this, the quality assessment function Q w i t for model w i t is defined as
Q w i t = μ 0 F w i k μ 1 + H f x i , z i ,
where the model utility parameters μ 0 , μ 1 0 are set according to the loss function, the neural network structure, and the data distribution [43]. To assess the utility of the participants, adaptive aggregation was used, where the difference between the quality of the model of participant P i after the rth round of aggregation, Q w G r P i , and the quality of the model of those who did not take part in the aggregation, Q w G r P P i , was taken as the utility function [44]. Based on the strongly convex function property of the model quality Q w i t and the boundary Q w i 0  of the initialized global model, define the utility function of the participant P i P boosted after the rth round of model aggregation as follows:
v i r ( Q P i ) = ρ i r Q w G r P i Q w G r P P i ,
where ρ i r then denotes the payoff reward that federated participant P i receives in the rth round of model aggregation, and if v i ( r ) < 0 , it means that P i needs to pay a price compensation for obtaining a higher-quality model parameter to prevent federated participants from engaging in free-riding behavior.

4.4. Federated Optimization Function

The FL optimization function is known from Formulas (13) and (15):
arg max c i r ( c com P i ) i = 1 n [ v i r ( Q P i ) c i r ( c com P i ) ] , c i r ( c com P i ) v i r ( Q P i ) .

4.5. The Fuzzy Shapley Value

According to the fuzzy coalition of FL in Section 3.4, Choquet integral in Section 3.5 and Formulas (8) and (9). Given a fuzzy coalition S R ( N ) , let H ( S ) = { S ( i ) 0 < S ( i ) 1 , i N } , q ( S ) be the number of elements in H ( S ) , and the elements in H ( S ) be sorted into f 1 f 2 f q ( S ) in an ascending sequence. The payoff function v ˜ of the fuzzy coalitional cooperation game ( N , v ˜ , χ ) can be expressed as a Choquet integral:
v ˜ ( S ) = S d v ˜ = l = 1 q ( S ) v ( [ S ] f l ) ( f l f l 1 )
Then the fuzzy Shapley value φ ˜ i ( v ˜ ) of the fuzzy coalitional cooperation game ( N , v ˜ ) can be expressed as follows:
φ ˜ i ( v ˜ ) = l = 1 q ( S ) φ i ( v ( [ S ] f l ) ) ( f l f l 1 ) , i [ S ] f l N
where, f 0 = 0 , l = 1 , 2 , , q ( S ) , [ S ] f l = { i N S ( i ) f l } denotes the payoff function for a clear coalition of all participants with participation level S ( i ) f l , v ( [ S ] f l ) denotes the payoff function of a clear coalition of all participants with participation level S ( i ) f l , φ i ( v ( [ S ] f l ) ) denotes the payoff function of a clear coalition of all participants with participation level S ( i ) f l , and φ i ( v ( [ S ] f l ) ) denotes a clear coalition of all participants with participation level S ( i ) f l of the classical Shapley value, v ( [ S ] f l ) calculated from Formula (15). As with the classical Shapley value, the fuzzy Shapley value satisfies the following theorem:
Theorem 1
(Validity).  i N φ ˜ i ( v ˜ ) = v ˜ ( N ) .
Theorem 2
(Symmetry). If participants i , j N have v ˜ ( S { i } ) = v ˜ ( S { j } ) for any fuzzy coalition S P ( N { i , j } ) , then φ ˜ i ( v ˜ ) = φ ˜ j ( v ˜ ) .
Theorem 3
(Additivity). For any two fuzzy cooperative games < N , v ˜ 1 > and < N , v ˜ 2 > , if there exists a fuzzy cooperative game < N , v ˜ 1 + v ˜ 2 > , for any fuzzy coalition S P ( N ) , it will always satisfy ( v ˜ 1 + v ˜ 2 ) ( S ) = v ˜ 1 ( S ) + v ˜ 2 ( S ) , then φ ˜ i ( v ˜ 1 + v ˜ 2 ) = φ ˜ i ( v ˜ 1 ) + φ ˜ i ( v ˜ 2 ) .
Theorem 4
(Dumbness). If for all fuzzy coalitions S containing participant i, if v ˜ ( S ) = v ˜ ( S { i } ) is satisfied, then φ ˜ i ( v ˜ ) = 0 .
Proof. 
The proofs of Theorems 1–4 are given in Appendix A.1, Appendix A.2, Appendix A.3 and Appendix A.4 of the Appendix A. □

4.6. The Conflict of Fairness and Pareto Optimality

Consider a scenario involving n participants, where each participant i has a fuzzy coalition investment denoted as x i . The investment of participant i satisfies x i ( 0 , ) , i = 1 , 2 , , n . The fuzzy coalition investments of all participants can be represented as an n-dimensional vector x = ( x 1 , x 2 , , x n ) . Furthermore, the fuzzy coalition cost investment for participant i is denoted as c ˜ i ( x i ) , which calculated from Formula (13). The function c ˜ i ( x i ) is assumed to be differentiable, convex, and strictly monotone increasing. It satisfies the following conditions: v ˜ i ( x ) x i > 0 , 2 v ˜ i ( x ) x i 2 > 0 , c ˜ i ( 0 ) = 0 . The federated return v ˜ ( x 1 , x 2 , , x n ) calculated from Formulas (8), (9) and (15), which determined by the federated investment decisions made by the n participants. The function is strictly monotonically increasing, differentiable, and concave. It also satisfies the specified conditions: v ˜ i ( x ) x i > 0 , 2 v ˜ i ( x ) x i 2 < 0 v ˜ ( 0 , 0 , , 0 ) = 0 . The allocation of federated payoffs to the n participants follows a specific formula, as described by Equations (17) to (18).
While the fuzzy Shapley value is capable of quantifying the contribution of each participant, it fails to consider the inherent trade-off between fairness and Pareto optimality. Specifically, it overlooks the requirement of achieving Pareto optimization before to the allocation of payoffs. Consequently, it does not effectively address the issue of free-riding behavior among federated participants. Consequently, we put forward the following theorem:
Theorem 5.
While the fuzzy Shapley value method effectively promotes fairness in the distribution of payoffs after FL, it does not sufficiently tackle the problem of optimizing incentives for participants’ investments prior to FL. In other words, it does not achieve Pareto efficiency optimization prior to the commencement of FL.
Proof. 
The proof of this theorem can be found in Appendix A.5 of Appendix A. □

4.7. Introducing Supervisory Organization

The Establishment of Supervised Mechanism

4.7.1. The Implementation of a Supervisory System

The introduction of supervisors to mitigate free riding in the FL process by requiring federated participants to pay a fee to the supervisor proposed in [45]. Holmstrom [46] further emphasized the use of incentive mechanisms to address free riding, with the supervisor playing a crucial role in breaking equilibrium and creating incentives.
Assuming that the supervisor is cognizant of the fact that the federated payoff is such that it exceeds or equals the Pareto optimal payoff, they can distribute this payoff to the participants using the fuzzy Shapley value Formulas (17)–(18). However, if the federated payoffs fall short of the Pareto optimal value, the federated participants will be obligated to pay a specified fee k ˜ i in the following manner:
r ˜ i ( x ) = φ ˜ i ( v ˜ ( x ) ) , φ ˜ i ( v ˜ ( x ) ) k ˜ i , if v ˜ ( x ) v ˜ ( x * ) if v ˜ ( x ) < v ˜ ( x * )
where we let x * = ( x 1 * , x 2 * , , x n * ) be the federated investment vector that satisfies Formula (A4).

4.7.2. Penalty Conditions

Theorem 6.
In the case where the federated investment mechanism x * = ( x 1 * , x 2 * , , x n * ) meets the criteria of Pareto efficiency and the penalty k ˜ i to serve as a Nash equilibrium, it must meet the following conditions:
(1)
If the individual investment x i of participant i is less than the Pareto optimal federated investment x i * , i.e., e.g., x i < x i * , and v ˜ ( x ) is a monotonically increasing function (i.e., v ˜ ( x i , x n i * ) < v ˜ ( x i * , x n i * ) ) , participant i is fined. The remaining payoff after deducting the fine is denoted as r ˜ i [ v ˜ ( x i , x * n i ) ] = φ ˜ i ( v ˜ ( x ) ) k ˜ i , and the profit of participant i can be expressed as follows:
π ˜ i [ v ˜ ( x i , x n i * ) ] = φ ˜ i ( v ˜ ( x ) ) k ˜ i c ˜ i ( x i ) , i = 1 , 2 , , n .
(2)
If the individual investment x i of participant i is equal to the Pareto optimal federated investment x i * , i.e.,  x i = x i * , then v ˜ ( x ) = v ˜ ( x * ) . The remaining payoff after the penalty is r ˜ i [ v ˜ ( x i * , x n i * ) ] = φ ˜ i ( v ˜ ( x * ) ) , and the profit of participant i can be expressed as follows:
π ˜ i [ v ˜ ( x i * , x n i * ) ] = φ ˜ i ( v ˜ ( x * ) c ˜ i ( x i * ) , i = 1 , 2 , , n .
In this context, x n i * = ( x 1 * , x i 1 * , x i + 1 * , , x n * ) denotes the Pareto optimal federated investment vector consisting of n i participants.
Proof. 
The proof of this theorem can be found in Appendix A.6 of Appendix A. □

5. Illustrative Examples and Simulations

Within this section, we validate the reasonableness of the fuzzy Shapley value method through a numerical arithmetic example and achieve the effectiveness of this method through numerical simulations and experimental comparisons.

5.1. Illustrative Examples

Assuming that there are three federated participants jointly training the FL model, the payment function denoted as v ( x 1 , x 2 , x 3 ) = 10 x 1 + 8 x 2 + 4 x 3 + x 1 x 2 , exhibits a strictly increasing linear behavior in the deterministic federation scenario. Furthermore, the cost functions c ˜ ( x 1 ) = 1 2 x 1 2 c ˜ ( x 2 ) = x 2 2 , and  c ˜ ( x 3 ) = 1 4 x 3 2 are strictly monotonically increasing and convex. Now, due to the limitations of the energy, ability and resources of the federated participants, the levels of these three federated participants are 40 % , 60 % and 80 % , i.e., consider the distribution of benefits of each federated participant under the fuzzy coalition s = ( 2 5 , 3 5 , 4 5 ) .
Since v ( 0 ) = 0 , v ( x 1 ) = 10 x 1 , v ( x 2 ) = 8 x 2 , v ( x 3 ) = 4 x 3 , v ( x 1 , x 2 ) = x 1 x 2 , v ( x 1 , x 3 ) = v ( x 2 , x 3 ) = 0 and v ( x 1 , x 2 , x 3 ) = 10 x 1 + 8 x 2 + 4 x 3 + x 1 x 2 , and according to Equations (6) and (7), the deterministic case computes the federated participant 1, 2, and 3 in the deterministic coalition with the fuzzy Shapley values, and then based on the derived revenue sharing ratio, the revenue sharing values of each federated participant under different cooperation strategies can be derived, as shown in Table 1.
Due to the fuzzy participation attitudes of the three federated participants, the elements in Q ( S ) are ordered in increasing sequence as f 0 = 0 , f 1 = 2 5 , f 2 = 3 5 , f 3 = 4 5 , and according to the Formula (17), the fuzzy coalition under the federation payoff is
v ˜ ( x 1 , x 2 , x 3 ) = v ( x 1 , x 2 , x 3 ) ( f 1 f 0 ) + v ( x 2 , x 3 ) ( f 2 f 1 ) + v ( x 3 ) ( f 3 f 2 ) = 4 x 1 + 16 5 x 2 + 12 5 x 3 + 2 5 x 1 x 2 ,
Similarly, v ˜ ( 0 ) = 0 , v ˜ ( x 1 ) = 4 x 1 v ˜ ( x 2 ) = 24 5 x 2 v ˜ ( x 3 ) = 16 5 x 3 v ˜ ( x 1 , x 2 ) = 8 5 x 2 + 2 5 x 1 x 2 v ˜ ( x 1 , x 3 ) = 8 5 x 3 v ˜ ( x 2 , x 3 ) = 4 5 x 3 .
When the federated investments are x * = ( x 1 * , x 2 * , x 3 * ) , the fuzzy case maximizes the federated profit
max R ˜ = v ˜ ( x 1 , x 2 , x 3 ) i = 1 3 c ˜ i ( x i ) = 4 x 1 + 16 5 x 2 + 12 5 x 3 + 2 5 x 1 x 2 1 2 x 1 2 x 2 2 1 4 x 3 2
while satisfying Pareto optimality. This is achieved by ensuring that the first-order condition is met:
4 x 1 * + 1 5 x 2 * = 0 16 5 + 1 5 x 1 * 2 x 2 * = 0 12 5 1 2 x 3 * = 0
The Pareto optimal federated investments are determined as follows: x 1 * = 5.04 , x 2 * = 2.61 , and x 3 * = 4.80 . The corresponding federal payment is v ˜ ( x * ) = 45.29 , and the maximum profit under the given circumstances is max R ˜ * = 20.02 in the worst-case scenario. Using Formula (18), the fuzzy case distributes benefits among federated participants as follows:
φ ˜ 1 ( v ˜ ( x ) ) = 8 3 x 1 + 8 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 φ ˜ 2 ( v ˜ ( x ) ) = 2 3 x 1 + 32 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 φ ˜ 3 ( v ˜ ( x ) ) = 2 3 x 1 + 8 15 x 2 + 28 15 x 3
Hence, the profit function for the three federated participants can be expressed as
π ˜ 1 ( v ˜ ( x ) ) = 8 3 x 1 + 8 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 1 2 x 1 2 π ˜ 2 ( v ˜ ( x ) ) = 2 3 x 1 + 32 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 x 2 2 π ˜ 3 ( v ˜ ( x ) ) = 2 3 x 1 + 8 15 x 2 + 28 15 x 3 1 4 x 3 2
To attain Nash equilibrium in the federation, participants have the autonomy to determine their investments and aim to maximize their benefits. As a result, the first-order condition for Nash equilibrium is expressed as follows:
8 3 x 1 + 1 5 x 2 = 0 32 15 + 1 5 x 1 2 x 2 = 0 28 15 1 2 x 3 = 0 .
Based on our analysis, we identify that the federated investments that meet the Nash equilibrium conditions are x 1 = 2.94 , x 2 = 1.36 , and x 3 = 3.73 . The corresponding federated profit is v ˜ ( x ) = 26.66 , and the maximum amount of benefit achievable is max R ˜ = 17.01 .
By comparing the Pareto efficiency value and the Nash equilibrium value presented in Table 2, it becomes evident that while the fuzzy Shapley value method ensures fairness in the aftermath, it lacks in optimizing incentives beforehand. In the Nash equilibrium, the individual investments are lower than the Pareto optimal level, resulting in suboptimal federated profits.
For future research, we suggest incorporating a supervisory authority. In the event that the supervisory authority acknowledges that the federated payoff is greater than or equal to the Pareto optimal payoff of 45.29, it will distribute the payoff among the federated participants using the fuzzy Shapley value method. However, if the federated payoff falls below the Pareto optimal payoff of 45.29, the participants receive a payoff denoted by τ ˜ i . When the participants’ investments reach the Pareto optimal values, specifically x 1 * = 5.04 , x 2 * = 2.61 and x 3 * = 4.80 , then
φ ˜ 1 ( v ˜ ( x * ) ) = 8 3 x 1 * + 8 15 x 2 * + 4 15 x 3 * + 1 5 x 1 * x 2 * = 18.74 φ ˜ 2 ( v ˜ ( x * ) ) = 2 3 x 1 * + 32 15 x 2 * + 4 15 x 3 * + 1 5 x 1 * x 2 * = 12.84 φ ˜ 3 ( v ˜ ( x * ) ) = 2 3 x 1 * + 8 15 x 2 * + 28 15 x 3 * = 13.71 π ˜ 1 ( v ˜ ( x * ) ) = 8 3 x 1 * + 8 15 x 2 * + 4 15 x 3 * + 1 5 x 1 * x 2 * 1 2 x 1 * 2 = 6.04 π ˜ 2 ( v ˜ ( x * ) ) = 2 3 x 1 * + 32 15 x 2 * + 4 15 x 3 * + 1 5 x 1 * x 2 * x 2 * 2 = 6.03 π ˜ 3 ( v ˜ ( x * ) ) = 2 3 x 1 * + 8 15 x 2 * + 28 15 x 3 * 1 4 x 3 * 2 = 7.95 .
As a result, the permissible intervals for τ ˜ 1 , τ ˜ 2 , and τ ˜ 3 are 0 τ ˜ 1 6.04 , 0 τ ˜ 2 6.03 , and 0 τ ˜ 3 7.95 correspondingly. The utilities φ ˜ 1 ( v ˜ ( x ) ) , φ ˜ 2 ( v ˜ ( x ) ) , and φ ˜ 3 ( v ˜ ( x ) ) for participants 1, 2, and 3 can be represented as follows:
φ ˜ 1 ( v ˜ ( x ) ) = 8 3 x 1 + 8 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 if v ˜ ( x ) 45.29 τ ˜ 1 , if v ˜ ( x ) < 45.29
φ ˜ 2 ( v ˜ ( x ) ) = 2 3 x 1 + 32 15 x 2 + 4 15 x 3 + 1 5 x 1 x 2 , if v ˜ ( x ) 45.29 τ ˜ 2 , if v ˜ ( x ) < 45.29
φ ˜ 3 ( v ˜ ( x ) ) = 2 3 x 1 + 8 15 x 2 + 28 15 x 3 , if v ˜ ( x ) 45.29 τ ˜ 3 , if v ˜ ( x ) < 45.29 .
In our subsequent examination, we will illustrate that the Nash equilibrium, serving as the monitoring mechanism, meets the Pareto efficiency criterion as indicated by x * = ( x 1 * , x 2 * , x 3 * ) = ( 5.04 , 2.61 , 4.80 ) .
(1)
Examining the value of participant 1,when the value of x 2 * = 2.61 and x 3 * = 4.80 , then
φ ˜ 1 ( v ˜ ( x ) ) = 3.54 x 1 + 2.67 , τ ˜ 1 , if x 5.04 if x < 5.04
For participant 1, if participant 1 invests x 1 5.04 (with 0 τ ˜ 1 6.04 ), who profit function can be expressed as π ˜ 1 ( x ) = 3.54 x 1 + 2.67 1 2 x 1 2 . It is observed that π ˜ 1 ( x 1 ) x 1 = 3.54 x 1 < 0 , indicating that the revenue function π ˜ 1 ( x ) for participant 1 consistently diminishes across the range [ 5.04 , ) . Therefore, with an investment of x 1 = 5.04 , participant 1 can attain the utmost earnings of π ˜ 1 ( x ) = 6.04 . However, should participant 1 allocate an investment of x 1 < 5.04 , the return is τ ˜ 1 , while the earnings equate to π ˜ 1 ( x ) = τ ˜ 1 1 2 x 1 2 . In this case, we have π ˜ 1 ( x ) = τ ˜ 1 1 2 x 1 2 6.04 . Thus, with x 2 * = 2.61 and x 3 * = 4.8 , participant 1 can attain the maximum earning π ˜ 1 ( x ) = 6.04 through an investment of x 1 * = 5.04 .
(2)
Examining the value of participant 2 when the values of x 1 = 5.04 and x 3 = 4.80 , then
φ ˜ 2 ( v ˜ ( x ) ) = 3.14 x 2 + 4.64 , τ ˜ 2 , if x 2.61 if x < 2.61
For participant 2, when the investment is x 2 2.61 (with 0 τ ˜ 2 6.03 ), the profit function can be expressed as π ˜ 2 ( x ) = 3.14 x 2 + 4.64 x 2 2 . We observe that π ˜ 2 ( x 2 ) x 2 = 3.14 2 x 2 < 0 , indicating that the revenue function π ˜ 2 ( x ) for participant 2 consistently diminishes across the range [ 2.61 , ) . Therefore, by allocating x 2 = 2.61 in investment, participant 2 can attain the highest earnings of π ˜ 2 ( x ) = 6.03 . However, should participant 2 allocate an investment of x 1 < 2.61 , the return is τ ˜ 2 , while the earnings equate to π ˜ 2 ( x ) = τ ˜ 2 x 2 2 . In this case, we have π ˜ 2 ( x ) = τ ˜ 2 x 2 2 6.03 . Thus, with x 1 * = 5.04 and x 3 * = 4.8 , participant 2 can achieve the utmost earnings of π ˜ 2 ( x ) = 6.03 through an investment of x 2 = 2.61 .
(3)
Examining the value of participant 3 when the values of x 1 * = 5.04 and x 2 * = 2.61 , then
φ ˜ 3 ( v ˜ ( x ) ) = 1.87 x 3 + 4.75 , τ ˜ 3 , if x 4.80 if x < 4.80
For participant 3, if the investment of participant 3 is x 3 4.80 (with 0 τ ˜ 3 7.95 ), the profit function can be expressed as π ˜ 3 ( x ) = 1.87 x 3 + 4.75 1 4 x 3 2 . It is observed that π ˜ 3 ( x 3 ) x 3 = 1.87 1 2 x 3 < 0 , indicating that the revenue function  π ˜ 3 ( x ) for participant 3 consistently diminishes across the range [ 4.80 , ) . Therefore, with an investment of x 3 = 4.80 , participant 3 can attain the utmost earnings of  π ˜ 3 ( x ) = 7.95 . However, should participant 3 allocate an investment of x 3 < 4.80 , the return is τ ˜ 3 , while the earnings equate to π ˜ 3 ( x ) = τ ˜ 3 1 4 x 3 2 . In this case, we have π ˜ 3 ( x ) = τ ˜ 3 1 4 x 3 2 4.80 . Thus, with x 1 * = 5.04 and x 2 * = 2.61 , participant 3 can achieve the peak earnings of π ˜ 3 ( x ) = 7.95 through an investment of x 3 = 4.80 .

5.2. Illustrative Simulations

The results of the experimental simulation are shown in Figure 3 as follows: In Figure 3a, we can observe that the federated participants’ investments are Pareto optimal, however, the investments that satisfy the Nash equilibrium are not Pareto optimal.
In Figure 3b, we can see that the participants who satisfy the Pareto optimal investments receive higher payoffs than the participants who satisfy the Nash equilibrium investments, while Figure 3c shows that federated participants receive higher maximum profits in the case of satisfying Pareto optimal investments than in the case of satisfying Nash equilibrium investments.
In Table 3, refs. [9,19,27,28] do not take ambiguity and Pareto optimality into account, although they do take fairness into account. In particular, ref. [27] only considers the fairness of the distribution of gains under participant attitude certainty and does not discuss in depth the issue of Pareto optimality in FL, which maximizes the gains for all participants. Meanwhile, ref. [29] considers the fairness and Pareto optimality of benefit distribution in the case of certainty of participants’ attitudes. Therefore, the method in this paper is more consistent with the practical application scenario, and it can solve the fairness and Pareto optimal consistency of gain distribution under uncertainty of participants’ attitudes, which can motivate more participants to join in the FL.
With these experimental simulation results, we can conclude that for FL, the investment strategy without the introduction of the monitoring mechanism does not reach the Nash equilibrium, and the investment strategy with the addition of the monitoring mechanism reaches the Pareto optimality.

6. Discussions

Constructing a fuzzy Shapley value technique that can manage the ambiguity in the uncertainty of participation attitudes in FL is achievable, and the extent of participants’ contributions and benefits can be assessed more accurately and through the introduction of a supervised mechanism. The method, through its mechanism, guarantees that the distribution of benefits is equitable and Pareto-efficient, which enhances the accuracy of the distribution of benefits.
However, the incentive mechanism also has some disadvantages: it involves several concepts and techniques such as fuzzy sets, fuzzy probabilities and supervision mechanisms, which increases the complexity of the mechanism, and the application of the mechanism to a real FL scenario needs to consider how to obtain and process information about the attitudes of participants towards participation and how to design and implement the supervision mechanism. In addition, there may be a degree of subjectivity in developing rules and standards for the supervised mechanism.
Therefore, the method has some advantages in dealing with the uncertainty of participants’ participation attitudes and achieving fairness and Pareto efficiency consistency in FL. At the same time, it also brings disadvantages to FL in its application, such as complexity, difficulty of implementation, and subjectivity of rule-making.

7. Conclusions

In the model training of FL, we considered the case of uncertain and ambiguous attitudes toward federated participant engagement, i.e., participants are not necessarily fully engaged in FL. To promote engagement among all participants in federated model training, this paper devises an incentive system utilizing the fuzzy Shapley value approach alongside supervised functions, which resolves the tension between equity and effectiveness in the distribution of payoffs within the FL. By conducting numerical computation validation and contrasting research findings, this paper constructs a fuzzy Shapley value method with supervised functions that helps to improve the effectiveness of FL, ensures the fairness and reasonableness of the participants in the distribution of payoffs.
Consequently, using this incentive mechanism guarantees the fairness of the distribution of benefits and the Pareto optimal consistency of the federated participants in the case of uncertainty and ambiguity in their attitude to participation, which offers theoretical insights for practical implementations.

Author Contributions

Conceptualization, X.Y. and S.X.; methodology, X.Y., S.X. and C.P.; software, X.Y. and H.L.; validation, X.Y., S.X., C.P. and Y.W.; writing—original draft preparation, X.Y.; writing review and editing, X.Y., H.D. and Y.W.; visualization, X.Y., C.P. and W.T.; supervision, X.Y., S.X. and C.P.; project administration, Y.W., H.D. and H.L.; funding acquisition, Y.W. and H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the National Natural Science Foundation of China (No. [62272124], [71961003], [62002081]), the National Key Research and Development Program of China (No. 2022YFB2701401), the Guizhou Science Contract Plat Talent (No. [2020]5017), the Open Fund of Key Laboratory of Advanced Manufacturing Technology, Ministry of Education (No. GZUAMT2021KF[01]), the China Postdoctoral Science Foundation (No. 2020M673584XB), the Natural Science Researching Program of D.o.E. of Guizhou (No. Qian Edu. & Tech. [2023]065).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data sets used and/or analyzed during the current study are available from the corresponding author on reasonable request.

Acknowledgments

The authors are grateful to the referees for their careful reading of the manuscript and valuable comments. The authors thank the help from the editor too.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

NThe set of n FL participants
DThe set of FL participants’ local dataset
MThe model trained jointly by all FL participants
M F E D The FL sharing model
M S U M The traditional machine learning model
V F E D The model accuracy of M F E D
M S U M The model accuracy of M S U M
SThe alliance subset of different participants
v ( . ) The characteristic function in the determining case
v ˜ ( . ) The characteristic function in the fuzzy case
v ( S ) The participant’s payoff through the alliance S in the determining case
v ˜ ( S ) The participant’s payoff through the alliance S in the fuzzy case
v ( N ) The overall federated payoff in the determining case
v ˜ ( N ) The overall federated payoff in the fuzzy case
φ i ( v ) The participant i’s payoff in the determining case
φ ˜ i ( v ˜ ) The participant i’s payoff in the fuzzy case
xThe set of feasible actions that can be taken
π ˜ ( x ) The federated participant’s payoff function in the fuzzy case
c ˜ ( x ) The coalition cost investment of participant in the fuzzy case
k ˜ i To reach Pareto optimality in the fuzzy situation, the supervisor must meet the penalty requirement
r ˜ i ( x ) The supervisor obtained fines in the fuzzy case
R ˜ The federated participant’s profit in the fuzzy case
c i r ( c com ( P i ) ) The computational cost function
c i r ( c com ( P i ) ) The federated cost utility function
Q ( w i t ) The quality assessment function
v i r ( Q ( P i ) ) The federated utility function

Appendix A

Appendix A.1. Proof of Theorem 1

Proof. 
The validity of the traditional Shapley value implies that
i = 1 n φ ˜ i ( v ˜ ) = i = 1 n l = 1 q ( S ) φ ˜ i ( v ( [ S ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) i = 1 n φ ˜ i ( v ( [ S ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) v ( [ N ] f l ) ( f l f l 1 ) = v ˜ ( N )
Hence, the validity of fuzzy Shapley value is established. □

Appendix A.2. Proof of Theorem 2

Proof. 
From the symmetry of classical Shapley values it follows that
v ˜ ( S { i } ) = l = 1 q ( S ) v ( [ S { i } ] f l ) ( f l f l 1 ) , v ˜ ( S { j } ) = l = 1 q ( S ) v ( [ S { j } ] f l ) ( f l f l 1 ) , φ ˜ i ( v ˜ ) = l = 1 q ( S ) φ ˜ i ( v ( [ S { i } ] f l ) ) ( f l f l 1 ) , φ ˜ j ( v ˜ ) = l = 1 q ( S ) φ ˜ j ( v ( [ S { j } ] f l ) ) ( f l f l 1 ) ,
due to the fact that v ˜ ( S { i } ) = v ˜ ( S { j } ) , then
l = 1 q ( S ) φ ˜ i ( v ( [ S { i } ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) φ ˜ j ( v ( [ S { j } ] f l ) ) ( f l f l 1 )
and thus φ ˜ i ( v ˜ ) = φ ˜ j ( v ˜ ) . Hence, the symmetry of fuzzy Shapley value is proved. □

Appendix A.3. Proof of Theorem 3

Proof. 
The validity of the traditional Shapley value implies that
v ˜ 1 + v ˜ 2 ( S ) = l = 1 q ( S ) ( v 1 ( [ S ] f l ) + v 2 ( [ S ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) v 1 ( [ S ] f l ) ( f l f l 1 ) + l = 1 q ( S ) v 2 ( [ S ] f l ) ( f l f l 1 ) = v ˜ 1 ( S ) + v ˜ 2 ( S )
due to the fact that
φ ˜ i ( v ˜ 1 + v ˜ 2 ) ( S ) = l = 1 q ( S ) φ ˜ i ( v 1 ( [ S ] f l ) + v 2 ( [ S ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) φ ˜ i ( v 1 ( [ S ] f l ) ) ( f l f l 1 ) + l = 1 q ( S ) φ ˜ i ( v 2 ( [ S ] f l ) ) ( f l f l 1 ) = φ ˜ i v ˜ 1 + φ ˜ i v ˜ 2
and thus φ ˜ i v ˜ 1 + v ˜ 2 = φ ˜ i v ˜ 1 + φ ˜ i v ˜ 2 , hence the additivity of fuzzy Shapley value is proved. □

Appendix A.4. Proof of Theorem 4

Proof. 
From the dummy of classical Shapley value it follows that
v ˜ ( S ) = l = 1 q ( S ) v ( [ S ] f l ) ( f l f l 1 ) ,
v ˜ ( S { i } ) = l = 1 q ( S ) v ( [ S { i } ] f l ) ( f l f l 1 )
due to the fact that v ˜ ( S ) = v ˜ ( S { i } ) , hence
l = 1 q ( S ) v ( [ S ] f l ) ( f l f l 1 ) = l = 1 q ( S ) v ( [ S { i } ] f l ) ( f l f l 1 )
and thus
φ ˜ i ( v ˜ ) = l = 1 q ( S ) φ ˜ i ( v ( [ S ] f l ) ) ( f l f l 1 ) = l = 1 q ( S ) φ ˜ i ( v ( [ S { i } ] f l ) ( f l f l 1 ) = l = 1 q ( S ) 0 ( f l f l 1 ) = 0
Consequently, the dummy of fuzzy Shapley value is thus demonstrated. □

Appendix A.5. Proof of Theorem 5

Proof. 
With the legitimacy of the fuzzy Shapley value established, we can derive that
i = 1 n φ ˜ i ( v ˜ ( x ) ) = v ˜ ( x ) , v ˜ ( x ) .
By differentiating Formula (A1) with respect to x, we find that
i = 1 n φ ˜ i ( v ˜ ( x ) ) = 1 .
Here, φ ˜ i ( v ˜ ( x ) ) = φ ˜ i / v ˜ ( x ) represents the partial derivative of φ ˜ i with respect to v ˜ ( x ) . In the context of the Nash equilibrium, each federated participant’s investment is denoted as x i , and their profit is given by π ˜ i ( v ˜ ( x ) ) = φ ˜ i ( v ˜ ( x ) ) c ˜ i ( x i ) . The profit maximization for participant i occurs at:
max π ˜ i ( v ˜ ( x ) ) = φ ˜ i ( v ˜ ( x ) ) c ˜ i ( x i ) , i = 1 , 2 , , n
The Nash equilibrium’s first-order condition is thus
φ ˜ i ( v ˜ ( x ) ) = c ˜ i ( x ) , i = 1 , 2 , , n .
where φ ˜ i ( v ˜ ( x ) ) = φ ˜ i / x , x i = x ˜ / x i , c ˜ i ( x ) = c ˜ i ( x ) / x i . To achieve maximum federated profits, federated investments must satisfy Pareto efficiency:
x * = arg max x v ˜ ( x ) i = 1 n c ˜ i x i .
The Pareto optimality’s first-order requirement is
x i = c ˜ i ( x ) , i = 1 , 2 , , n .
By considering Formulas (A4) and (A6), we observe that the Nash equilibrium corresponds to Pareto optimality, requiring the fulfillment of the following conditions:
φ ˜ i ( v ˜ ( x ) ) = 1 , i = 1 , 2 , , n
Yet, this is inconsistent with meeting the criterion of the fuzzy Shapley value requirement i = 1 n φ ˜ i ( v ˜ ( x ) ) = 1 . □

Appendix A.6. Proof of Theorem 6

Proof. 
For the discussion to be meaningful, let us suppose that when the federated participants attain Pareto efficiency, the allocated federal revenues according to the fuzzy Shapley value method for each participant i exceed their investment costs, i.e., v ˜ ( x i * ) > c ˜ i ( x i * ) . If x * = ( x 1 * , x 2 * , , x n * ) represents the Nash equilibrium of this mechanism, the following condition should hold: π ˜ i [ v ˜ ( x i * , x n i * ) ] π ˜ i [ v ˜ ( x i , x n i * ) ] and
φ ˜ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) φ ˜ i ( v ˜ ( x ) ) k ˜ i c ˜ i ( x i ) k ˜ i [ φ ˜ i ( v ˜ ( x ) ) c ˜ i ( x i ) ] [ φ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] .
In the FL process, the coalition’s invisible and non-unique investment cost c ˜ i ( x i ) makes Formula (A8) unsuitable for formulating penalties. However, it is important to note that c ˜ i ( x i ) 0 , leading to φ ˜ i ( v ˜ ( x ) ) [ φ ˜ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] [ φ ˜ i ( v ˜ ( x ) ) c ˜ i ( x i ) ] [ φ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] . Hence, Equation (A8) holds true solely under the condition that k ˜ i φ ˜ i ( v ˜ ( x ) ) [ φ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] , guaranteeing the realization of Pareto efficiency. Therefore, the penalty requirement for the overseer to achieve Pareto efficiency is
k ˜ i φ ˜ i ( v ˜ ( x ) ) [ φ ˜ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] .
To enhance participants’ enthusiasm, it is vital to ensure that penalties are not overly steep and comply with the principles of restricted involvement and restricted liability. Assuming that all participants in FL have limited liability, the penalty amount should not exceed the participant’s payoff. Hence, if a participant’s payoff is zero, no penalty is necessary. Consequently, based on Formula (19), we derive φ ˜ i ( v ˜ ( x ) ) k ˜ i 0 , leading to the conclusion that
k ˜ i φ ˜ i ( v ˜ ( x ) ) .
From Formulas (A9) and (A10), we establish the following expression:
φ ˜ i ( v ˜ ( x ) ) [ φ ˜ i ( v ˜ ( x * ) ) c ˜ i ( x i * ) ] k ˜ i φ ˜ i ( v ˜ ( x ) ) .
If we denote the net gain of participant i following the penalty as δ ˜ i , then δ ˜ i = φ ˜ i ( v ˜ ( x ) ) k ˜ i . By utilizing Formula (A11), we obtain the following expression:
0 δ ˜ i φ i ( v ˜ ( x * ) ) c ˜ i ( x i * )
Building upon the earlier analysis and taking into account the constraint of restricted liability, the optimization of the Pareto value is accomplished. If the overseer’s penalty conforms to Equation (A11), the optimal mechanism can be delineated as:
r ˜ i ( x ) = φ ˜ i ( v ˜ ( x ) ) , δ ˜ i , if v ˜ ( x ) v ˜ ( x * ) if v ˜ ( x ) < v ( x * )
In this scenario, the value of δ ˜ i must comply with Equation (A13), where i = 1 , 2 , , n . We will now proceed to explain the mechanism under two different penalty value conditions as specified in Formula (A13).
If k ˜ i = φ ˜ i ( v ˜ ( x ) ) , it indicates that the overseer recognizes that the federated reward equals or exceeds the Pareto-efficient reward. Under such circumstances, the overseer will distribute the federated reward among all participants according to the equation of fuzzy Shapley value. Should the federated reward not reach the Pareto-efficient level, the entire will be affected by the overseer. The equation is as follows:
r ˜ i ( x ) = φ ˜ i ( v ˜ ( x ) ) , 0 , if v ˜ ( x ) v ˜ ( x * ) if v ˜ ( x ) < v ( x * ) .
Furthermore, we will illustrate that the federated investment x * = ( x 1 * , x 2 * , , x n * ) leading to Pareto efficiency constitutes the Nash equilibrium of this mechanism. If we assume that participant i invests x i < x * in the FL process, as the other participants allocate x n i * , owing to the monotonic nature of the function v ˜ ( x ) , it follows that v ˜ ( x i , x n i * ) < v ˜ ( x i * , x n i * ) . Moreover, r ˜ i ( x i , x n i * ) = 0 , and participant i’s earnings amount to π ˜ i ( x i ) = c ˜ i ( x i ) 0 . Consequently, a rational participant i would not invest x i < x * . Given that participant i invests x i x * in the FL process, and considering the monotonicity of the function v ˜ ( x ) , the profit of participant i can be expressed as π ˜ i ( x i ) = φ ˜ i ( x i ) c ˜ i ( x i ) > 0 . Consequently, a rational participant i would choose to invest x i x * . □

References

  1. Konečnỳ, J.; McMahan, H.B.; Ramage, D.; Richtárik, P. Federated optimization: Distributed machine learning for on-device intelligence. arXiv 2016, arXiv:1610.02527. [Google Scholar]
  2. McMahan, H.B.; Moore, E.; Ramage, D.; y Arcas, B.A. Federated learning of deep networks using model averaging. arXiv 2016, arXiv:1602.05629. [Google Scholar]
  3. Konečnỳ, J.; McMahan, H.B.; Yu, F.X.; Richtárik, P.; Suresh, A.T.; Bacon, D. Federated learning: Strategies for improving communication efficiency. arXiv 2016, arXiv:1610.05492. [Google Scholar]
  4. Cicceri, G.; Tricomi, G.; Benomar, Z.; Longo, F.; Puliafito, A.; Merlino, G. Dilocc: An approach for distributed incremental learning across the computing continuum. In Proceedings of the 2021 IEEE International Conference on Smart Computing (SMARTCOMP), Irvine, CA, USA, 23–27 August 2021; pp. 113–120. [Google Scholar]
  5. Zhang, Z.; Zhang, Y.; Guo, D.; Zhao, S.; Zhu, X. Communication-efficient federated continual learning for distributed learning system with non-iid data. Sci. China Inf. Sci. 2023, 66, 122102. [Google Scholar] [CrossRef]
  6. Yang, Q.; Liu, Y.; Chen, T.; Tong, Y. Federated machine learning: Concept and applications. ACM Trans. Intell. Syst. Technol. (Tist) 2019, 10, 1–19. [Google Scholar] [CrossRef]
  7. Kairouz, P.; McMahan, H.B.; Avent, B.; Bellet, A.; Bennis, M.; Bhagoji, A.N.; Bonawitz, K.; Charles, Z.; Cormode, G.; Cummings, R.; et al. Advances and open problems in federated learning. Found. Trends Mach. Learn. 2021, 14, 1–210. [Google Scholar] [CrossRef]
  8. Sarikaya, Y.; Ercetin, O. Motivating workers in federated learning: A stackelberg game perspective. IEEE Netw. Lett. 2019, 2, 23–27. [Google Scholar] [CrossRef]
  9. Yu, H.; Liu, Z.; Liu, Y.; Chen, T.; Cong, M.; Weng, X.; Niyato, D.; Yang, Q. A fairness-aware incentive scheme for federated learning. In Proceedings of the Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society, New York, NY, USA, 7–9 February 2020; Volume 2020, pp. 393–399. [Google Scholar]
  10. Li, T.; Sanjabi, M.; Beirami, A.; Smith, V. Fair resource allocation in federated learning. arXiv 2019, arXiv:1905.10497. [Google Scholar]
  11. Zhan, Y.; Li, P.; Qu, Z.; Zeng, D.; Guo, S. A learning-based incentive mechanism for federated learning. IEEE Internet Things J. 2020, 7, 6360–6368. [Google Scholar] [CrossRef]
  12. Kang, J.; Xiong, Z.; Niyato, D.; Xie, S.; Zhang, J. Incentive mechanism for reliable federated learning: A joint optimization approach to combining reputation and contract theory. IEEE Internet Things J. 2019, 6, 10700–10714. [Google Scholar] [CrossRef]
  13. Tran, N.H.; Bao, W.; Zomaya, A.; Nguyen, M.N.; Hong, C.S. Federated learning over wireless networks: Optimization model design and analysis. In Proceedings of the IEEE INFOCOM 2019-IEEE conference on computer communications, Paris, France, 29 April–2 May 2019; pp. 1387–1395. [Google Scholar]
  14. Kim, H.; Park, J.; Bennis, M.; Kim, S.L. Blockchained on-device federated learning. IEEE Commun. Lett. 2019, 24, 1279–1283. [Google Scholar] [CrossRef]
  15. Feng, S.; Niyato, D.; Wang, P.; Kim, D.I.; Liang, Y.C. Joint service pricing and cooperative relay communication for federated learning. In Proceedings of the 2019 International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Atlanta, GA, USA, 14–17 July 2019; pp. 815–820. [Google Scholar]
  16. Yang, X. An exterior point method for computing points that satisfy second-order necessary conditions for a C1, 1 optimization problem. J. Math. Anal. Appl. 1994, 187, 118–133. [Google Scholar] [CrossRef]
  17. Wang, Z.; Hu, Q.; Li, R.; Xu, M.; Xiong, Z. Incentive mechanism design for joint resource allocation in blockchain-based federated learning. IEEE Trans. Parallel Distrib. Syst. 2023, 34, 1536–1547. [Google Scholar] [CrossRef]
  18. Song, T.; Tong, Y.; Wei, S. Profit allocation for federated learning. In Proceedings of the 2019 IEEE International Conference on Big Data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 2577–2586. [Google Scholar]
  19. Wang, T.; Rausch, J.; Zhang, C.; Jia, R.; Song, D. A principled approach to data valuation for federated learning. In Federated Learning: Privacy and Incentive; Springer: Berlin/Heidelberg, Germany, 2020; pp. 153–167. [Google Scholar]
  20. Wang, G.; Dang, C.X.; Zhou, Z. Measure contribution of participants in federated learning. In Proceedings of the 2019 IEEE international conference on big data (Big Data), Los Angeles, CA, USA, 9–12 December 2019; pp. 2597–2604. [Google Scholar]
  21. Liu, Y.; Ai, Z.; Sun, S.; Zhang, S.; Liu, Z.; Yu, H. Fedcoin: A peer-to-peer payment system for federated learning. In Federated Learning: Privacy and Incentive; Springer: Berlin/Heidelberg, Germany, 2020; pp. 125–138. [Google Scholar]
  22. Zeng, R.; Zeng, C.; Wang, X.; Li, B.; Chu, X. A comprehensive survey of incentive mechanism for federated learning. arXiv 2021, arXiv:2106.15406. [Google Scholar]
  23. Liu, Z.; Chen, Y.; Yu, H.; Liu, Y.; Cui, L. Gtg-shapley: Efficient and accurate participant contribution evaluation in federated learning. ACM Trans. Intell. Syst. Technol. (Tist) 2022, 13, 1–21. [Google Scholar] [CrossRef]
  24. Nagalapatti, L.; Narayanam, R. Game of gradients: Mitigating irrelevant clients in federated learning. Proc. AAAI Conf. Artif. Intell. 2021, 35, 9046–9054. [Google Scholar] [CrossRef]
  25. Fan, Z.; Fang, H.; Zhou, Z.; Pei, J.; Friedlander, M.P.; Liu, C.; Zhang, Y. Improving fairness for data valuation in horizontal federated learning. In Proceedings of the 2022 IEEE 38th International Conference on Data Engineering (ICDE), Kuala Lumpur, Malaysia, 9–12 May 2022; pp. 2440–2453. [Google Scholar]
  26. Fan, Z.; Fang, H.; Zhou, Z.; Pei, J.; Friedlander, M.P.; Zhang, Y. Fair and efficient contribution valuation for vertical federated learning. arXiv 2022, arXiv:2201.02658. [Google Scholar]
  27. Yang, X.; Tan, W.; Peng, C.; Xiang, S.; Niu, K. Federated Learning Incentive Mechanism Design via Enhanced Shapley Value Method. Wirel. Commun. Mob. Comput. 2022, 2022, 9690657. [Google Scholar] [CrossRef]
  28. Pene, P.; Liao, W.; Yu, W. Incentive Design for Heterogeneous Client Selection: A Robust Federated Learning Approach. IEEE Internet Things J. 2023, 11, 5939–5950. [Google Scholar] [CrossRef]
  29. Yang, X.; Xiang, S.; Peng, C.; Tan, W.; Li, Z.; Wu, N.; Zhou, Y. Federated Learning Incentive Mechanism Design via Shapley Value and Pareto Optimality. Axioms 2023, 12, 636. [Google Scholar] [CrossRef]
  30. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; y Arcas, B.A. Communication-efficient learning of deep networks from decentralized data. In Proceedings of the Artificial Intelligence and Statistics. PMLR, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  31. Shapley, L.S. A value for n-person games. Contrib. Theory Games 1953, 2, 307–317. [Google Scholar]
  32. Murofushi, T.; Sugeno, M. An interpretation of fuzzy measures and the Choquet integral as an integral with respect to a fuzzy measure. Fuzzy Sets Syst. 1989, 29, 201–227. [Google Scholar] [CrossRef]
  33. Mesiar, R. Fuzzy measures and integrals. Fuzzy Sets Syst. 2005, 156, 365–370. [Google Scholar] [CrossRef]
  34. Zhou, Z.H.; Yu, Y.; Qian, C. Evolutionary Learning: Advances in Theories and Algorithms; Springer: Berlin/Heidelberg, Germany, 2019. [Google Scholar]
  35. Pardalos, P.M.; Migdalas, A.; Pitsoulis, L. Pareto Optimality, Game Theory and Equilibria; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008; Volume 7. [Google Scholar]
  36. Nas, J. Non-cooperative games. Ann. Math. 1951, 54, 286–295. [Google Scholar] [CrossRef]
  37. Ye, M.; Hu, G. Distributed Nash Equilibrium Seeking by a Consensus Based Approach. IEEE Trans. Autom. Control 2017, 62, 4811–4818. [Google Scholar] [CrossRef]
  38. Kang, J.; Xiong, Z.; Niyato, D.; Yu, H.; Liang, Y.C.; Kim, D.I. Incentive design for efficient federated learning in mobile networks: A contract theory approach. In Proceedings of the 2019 IEEE VTS Asia Pacific Wireless Communications Symposium (APWCS), Singapore, 28–30 August 2019; pp. 1–5. [Google Scholar]
  39. Pandey, S.R.; Tran, N.H.; Bennis, M.; Tun, Y.K.; Manzoor, A.; Hong, C.S. A crowdsourcing framework for on-device federated learning. IEEE Trans. Wirel. Commun. 2019, 19, 3241–3256. [Google Scholar] [CrossRef]
  40. Deng, J.; Guo, J.; Zafeiriou, S. Arcface: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  41. De Boer, P.T.; Kroese, D.P.; Mannor, S.; Rubinstein, R.Y. A tutorial on the cross-entropy method. Ann. Oper. Res. 2005, 134, 19–67. [Google Scholar] [CrossRef]
  42. Jauernig, P.; Sadeghi, A.R.; Stapf, E. Trusted execution environments: Properties, applications, and challenges. IEEE Secur. Priv. Mag. 2020, 8, 56–60. [Google Scholar] [CrossRef]
  43. Wang, H.; Kaplan, Z.; Niu, D.; Li, B. Optimizing federated learning on non-iid data with reinforcement learning. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020; pp. 1698–1707. [Google Scholar]
  44. Karimireddy, S.P.; Kale, S.; Mohri, M.; Reddi, S.; Stich, S.; Suresh, A.T. Scaffold: Stochastic controlled averaging for federated learning. Int. Conf. Mach. Learn. 2019, 119, 5132–5143. [Google Scholar]
  45. Alchian, A.A.; Demsetz, H. Production, information costs, and economic organization. Am. Econ. Rev. 1972, 62, 777–795. [Google Scholar]
  46. Holmstrom, B. Moral hazard in teams. Bell J. Econ. 1982, 13, 324–340. [Google Scholar] [CrossRef]
Figure 1. Federated learning framework.
Figure 1. Federated learning framework.
Axioms 13 00254 g001
Figure 2. Federated learning incentive model.
Figure 2. Federated learning incentive model.
Axioms 13 00254 g002
Figure 3. Comparison of investments and payoffs.
Figure 3. Comparison of investments and payoffs.
Axioms 13 00254 g003
Table 1. Calculated value of benefits to federated participants under the defined coalition φ i ( v ( [ S ] f l ) ) .
Table 1. Calculated value of benefits to federated participants under the defined coalition φ i ( v ( [ S ] f l ) ) .
Benefits under Different Coalitions
[ S ] f l { 1 } { 2 } { 3 } { 1 , 2 } { 1 , 3 } { 2 , 3 } { 1 , 2 , 3 }
1 10 x 1 00 40 x 1 2 x 2 + 8 x 1 x 2 2 + 4 x 1 x 2 x 3 + 3 x 1 2 x 2 2 50 x 1 + 40 x 2 + 8 x 3 + 6 x 1 x 2 00 20 3 x 1 + 4 3 x 2 + 2 3 x 3 + 1 2 x 1 x 2
20 8 x 2 0 10 x 1 2 x 2 + 32 x 1 x 2 2 + 4 x 1 x 2 x 3 + 3 x 1 2 x 2 2 50 x 1 + 40 x 2 + 8 x 3 + 6 x 1 x 2 00 5 3 x 1 + 16 3 x 2 + 2 3 x 3 + 1 2 x 1 x 2
300 4 x 3 000 5 3 x 1 + 4 3 x 2 + 8 3 x 3
Table 2. Federated investment and profit comparison.
Table 2. Federated investment and profit comparison.
Invest and Profit ComparisonInvest x 1 Invest x 2 Invest x 3 Federated PayoffMaximum Profit
Pareto optimality5.042.614.8045.2920.02
Nash equilibrium2.941.363.7326.6617.01
Table 3. Comparison of experimental results.
Table 3. Comparison of experimental results.
Research MethodsFuzzinessFairnessPareto Optimality
[9,19,27,28]
[29]
This paper
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, X.; Xiang, S.; Peng, C.; Tan, W.; Wang, Y.; Liu, H.; Ding, H. Federated Learning Incentive Mechanism with Supervised Fuzzy Shapley Value. Axioms 2024, 13, 254. https://doi.org/10.3390/axioms13040254

AMA Style

Yang X, Xiang S, Peng C, Tan W, Wang Y, Liu H, Ding H. Federated Learning Incentive Mechanism with Supervised Fuzzy Shapley Value. Axioms. 2024; 13(4):254. https://doi.org/10.3390/axioms13040254

Chicago/Turabian Style

Yang, Xun, Shuwen Xiang, Changgen Peng, Weijie Tan, Yue Wang, Hai Liu, and Hongfa Ding. 2024. "Federated Learning Incentive Mechanism with Supervised Fuzzy Shapley Value" Axioms 13, no. 4: 254. https://doi.org/10.3390/axioms13040254

APA Style

Yang, X., Xiang, S., Peng, C., Tan, W., Wang, Y., Liu, H., & Ding, H. (2024). Federated Learning Incentive Mechanism with Supervised Fuzzy Shapley Value. Axioms, 13(4), 254. https://doi.org/10.3390/axioms13040254

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop