Next Article in Journal
Optimized Trajectory Tracking for Robot Manipulators with Uncertain Dynamics: A Composite Position Predictive Control Approach
Next Article in Special Issue
Joint Overlapping Event Extraction Model via Role Pre-Judgment with Trigger and Context Embeddings
Previous Article in Journal
Pixel-Level Degradation for Text Image Super-Resolution and Recognition
Previous Article in Special Issue
Applying Image Analysis to Build a Lightweight System for Blind Obstacles Detecting of Intelligent Wheelchairs
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Resolving Agent Conflicts Using Enhanced Uncertainty Modeling Tools for Intelligent Decision Making

1
School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
2
Key Laboratory of Computational Intelligence and Chinese Information Processing of Ministry of Education, Shanxi University, Taiyuan 030006, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Electronics 2023, 12(21), 4547; https://doi.org/10.3390/electronics12214547
Submission received: 12 October 2023 / Revised: 26 October 2023 / Accepted: 3 November 2023 / Published: 5 November 2023
(This article belongs to the Special Issue Advances in Intelligent Data Analysis and Its Applications, Volume II)

Abstract

:
Conflict analysis in intelligent decision making has received increasing attention in recent years. However, few researchers have analyzed conflicts by considering trustworthiness from the perspective of common agreement and common opposition. Since L-fuzzy three-way concept lattice is able to describe both the attributes that objects commonly possess and the attributes that objects commonly do not possess, this paper introduces an L-fuzzy three-way concept lattice to capture the issues on which agents commonly agree and the issues which they commonly oppose, and proposes a hybrid conflict analysis model. In order to resolve conflicts identified by the proposed model, we formulate the problem as a knapsack problem and propose a method for selecting the optimal attitude change strategy. This strategy takes into account the associated costs and aims to provide the decision maker with the most favorable decision in terms of resolving conflicts and reaching consensus. To validate the effectiveness and feasibility of the proposed model, a case study is conducted, providing evidence of the model’s efficacy and viability in resolving conflicts.

1. Introduction

Conflict analysis [1,2,3] aims to analyze complex conflict situations with appropriate models by studying the conflict relationships between individuals or groups on issues, identifying the internal causes of conflict and providing some guidance for conflict resolution in intelligent decision making such as labor negotiation [4], diplomatic relations [5], and urban planning [6].
Many scholars have proposed various models for conflict analysis from different perspectives. Pawlak [5] first considered the uncertainty of agents’ attitudes toward issues and divided the agents into three groups (i.e., coalition, neutrality, and conflict). Yao [7] extended the Pawlak model [5] by examining three levels of conflict and proposed three-way conflict analysis. Lang et al. [8] further improved the Yao model [7] by employing an alliance measure and a conflict measure. In addition, considering uncertainty and data complexity in actual conflict situations, Lang et al. [9] used Pythagorean fuzzy sets to describe conflict situations and proposed a Bayesian minimal risk theory based conflict analysis method. Li et al. [10] proposed a conflict analysis model to cope with trapezoidal fuzzy numbers in agents’ attitudes toward issues. Yang et al. [11] investigated a three-way conflict analysis to deal with diverse rating types in situation tables. Suo et al. [12] studied a three-way conflict analysis model to deal with incomplete three-valued situation tables. Furthermore, since psychological factors and risk attitudes of agents may affect the results of conflict analysis, Wang et al. [13] proposed a compound risk preference model for three-way decision based on different types of risk preferences.
Zhi et al. [14] introduced a three-way concept for conflict analysis [15]. Three-way concept analysis was proposed in [15] by combining three-way decision theory [16,17,18,19,20] and formal concept analysis [21,22], with the ability of describing the properties that objects possess in common and those that they do not possess in common [23,24]. Zhi et al. [14] analyzed the conflict relationships between agents and identified the binary relationships between sets of agents and sets of issues by using three-way concepts, and then analyzed the causes of conflict. The above studies were conducted in the case that agents are completely trustworthy. In some situations (for example, in network); however, agents may be untrustworthy. To this end, Zhi et al. [25] proposed a multilevel conflict analysis that analyzed and resolved uncertainty in agents’ trustworthiness and uncertainty in agents’ attitudes toward issues. However, in [25], when describing the consistency of agents using fuzzy concepts, only the common agreement consistency is considered and the common opposition consistency is ignored. This may lead to a misjudgment of conflict in some cases. For example, when agents oppose an issue, according to [25] the agents are inconsistent on this issue, but in fact, agents are consistent.
On the other hand, when agents have disputes, it is necessary to find appropriate conflict resolution strategies to promote cooperation among them. To this end, Sun et al. [26] constructed a probabilistic rough set model and provided an effective method to find the feasible consensus strategy to facilitate the resolution of conflict situations. To select an effective feasible strategy, Xu et al. [4] formulated the criteria for selecting feasible strategies based on consistency measurement of cliques. Based on game-theoretic rough sets, Bashir et al. [1] designed a novel conflict resolution model by constructing a game among all agents, computing the payoff of different strategies, and classifying them according to the equilibrium rules. From the perspective of multi-criteria decision analysis, Du et al. [2] introduced three kinds of relations among agents into multi-criteria large-scale group decision making in linguistic context, obtained the coalition of decision models and the weights of criteria and finally proposed a conflict coordination and feedback mechanism to solve conflicts.
Most of the existing solution strategies resolve conflicts by selecting an optimal subset of issues most agents agree on. However, if conflicts have not yet reached a serious level, some compromises can be made through third-party mediation to promote cooperation between the agents. For example, Iran and Saudi Arabia have resumed diplomatic relations after China-mediated talks. Therefore, this paper considers changing the attitudes of agents to make them reach a consensus. Since such changes will bring a certain cost, it is necessary to measure the costs to determine an optimal change. In addition, the cost may also change as trust degree is introduced and changed. Consequently, conflict resolution strategies that only consider costs, without considering trust, are often ineffective.
In order to solve the above problems, this paper introduces an L-fuzzy three-way concept lattice (L-3WCL) [27], which is mainly used in knowledge representation [28] and fuzzy three-way concept lattices reduction [29], to conflict analysis and resolves the conflict using the dynamic programming method of the knapsack problem [30,31,32]. As a result, a hybrid conflict analysis model is developed. The model first employs L-fuzzy three-way concept to capture the issues on which agents commonly agree and the issues which they commonly oppose, and then measures the relative inconsistency of a set of agents. By relative inconsistency, we identify the state of a set of agents and categorize the issues into different types, which may help us find the causes of conflict. To facilitate cooperation between agents, we act as a third-party mediator, seeking to compromise between the agents at minimal cost to reach a consensus. We model this problem as a knapsack problem, which is a combinatorial optimization problem that can be solved using dynamic programming method. Finally, we propose an optimal attitude change strategy based on dynamic programming and solve conflicts with minimum cost. Furthermore, we verify the effectiveness of this strategy in intelligent decision-making instances such as business decision making.
Section 2 will briefly review the models in [5,25]. Section 3 analyzes the shortcomings of the fuzzy-concept-lattice-based conflict analysis model and proposes an L-3WCL-based conflict analysis model. Section 4 develops an optimal attitude change strategy and illustrates the effectiveness of the strategy with a case study. Finally, Section 5 concludes the paper with an outlook.

2. Related Works

This section briefly reviews some basic conflict analysis models, including the Pawlak’s conflict analysis model and the fuzzy-concept-lattice-based conflict analysis model.

2.1. Pawlak’s Conflict Analysis Model

Pawlak’s conflict analysis model is built on conflict analysis information system (CAIS).
Definition 1 
([5]). A CAIS is a triple K = ( U , V , R ) , where U = { x 1 , x 2 , , x n } is a non-empty finite set of agents, V = { a 1 , a 2 , a m } is a non-empty finite set of issues, and R : U × V { 1 , 0 , 1 } is defined by
R ( x , a ) = 1 , x is positive towards a 0 , x is neutral towards a 1 , x is negative towards a
where x U and a V .
In order to determine the relationship between two agents, Pawlak the defined auxiliary function [5].
Definition 2 
([5]). Let K = ( U , V , R ) be a CAIS and a V . Define the auxiliary function ϕ a with regard to a for x i , x j U as
ϕ a ( x i , x j ) = 1 , if R ( x i , a ) R ( x j , a ) = 1 or x i = x j 0 , if R ( x i , a ) R ( x j , a ) = 0 and x i x j 1 , if R ( x i , a ) R ( x j , a ) = 1 .
By Definition 2, ϕ a ( x i , x j ) = 1 means that both agents x i and x j have the same attitude towards issue a; ϕ a ( x i , x j ) = 1 means that agents x i and x j have different attitudes towards a; ϕ a ( x i , x j ) = 0 means that at least one agent has a neutral attitude towards a.
Based on the auxiliary function, the distance between agents can be measured.
Definition 3 
([5]). Let K = ( U , V , R ) be a CAIS and B V . Define the distance function with regard to B for x i , x j U as
ρ B ( x i , x j ) = 1 | B | a B 1 ϕ a ( x i , x j ) 2 .
According to the distance function, Pawlak divided the relationships between agents into three groups [5].
Definition 4 
([5]). Let K = ( U , V , R ) be a CAIS and B V . For x i , x j U ,
1 
If ρ B ( x i , x j ) > 0.5 , then x i , x j are called in a conflict state;
2 
If ρ B ( x i , x j ) = 0.5 , then x i , x j are called in a neutral state;
3 
If ρ B ( x i , x j ) < 0.5 , then x i , x j are called in an alliance state.

2.2. Fuzzy-Concept-Lattice-Based Conflict Analysis Model

Zhi et al. [25] employed fuzzy concepts to characterize the uncertainty of agents and proposed a fuzzy-concept-lattice-based conflict analysis model.
Definition 5 
([33]). A complete residuated lattice (CRL) is a seven-tuple ( L , , , , , 0 , 1 ) with
  • ( L , , , 0 , 1 ) is a complete lattice;
  • ( L , , 1 ) is a commutative monoid;
  • ⊗ and → are adjoint, i.e., for x, y, z L ,
    x y z x y z .
Definition 6 
([34]). Let ( L , , , , , 0 , 1 ) be a CRL. For a L , define ¬ : L L as ¬ a = a 0 and call ¬ a negation operation of L.
CRL has the following properties.
Proposition 1 
([34,35]). If L is a CRL and ¬ is a negation operation of L, then the following conclusions hold
1 
x y ( ( x y ) = 1 ) ;
2 
( x x ) = ( x 1 ) = ( 0 x ) = 1 , ( 1 x ) = x ;
3 
¬ 1 = 0 , ¬ 0 = 1 .
CRL generalizes various types of truth structures. For example, the Łukasiewicz implication operator is:
x y = ( ( 1 x + y ) 1 )
where x, y [ 0 , 1 ] . For a universe U and a CRL L, S : U L is called an L-fuzzy set on U. The set of all the fuzzy sets on U is denoted by L U .
Definition 7 
([36,37]). A fuzzy formal context is a triple K = ( U , V , R ) , where U is the set of objects, V is the set of attributes, and R L U × V is a fuzzy relation between U and V.
Based on fuzzy formal context, fuzzy concepts can then be defined.
Definition 8 
([36,37]). Let K = ( U , V , R ) be a fuzzy formal context. For A L U and B L V , : L U L V and : L V L U can be defined as
A ( y ) = x U ( A ( x ) R ( x , y ) ) , y V
B ( x ) = y V ( B ( y ) R ( x , y ) ) , x U
If A = B and B L V , then ( A , B ) is called a fuzzy concept in K, where A and B are called the extent and intent of ( A , B ) , respectively. For all the fuzzy concepts in K, the order defined by ( A 1 , B 1 ) ( A 2 , B 2 ) A 1 A 2 B 2 B 1 forms a complete lattice F ( K ) , called the fuzzy concept lattice of K.
Definition 9 
([25]). Let K = ( U , V , R ) be a CAIS. The inconsistency measure for X L U is defined as
m ( X ) = 1 | V | y V ( 1 X ( y ) )
Given two thresholds t 1 and t 2 such that 0 t 1 t 2 1 , for X L U
1 
If m ( X ) > t 2 , then X is called in a conflict state;
2 
If t 1 m ( X ) t 2 , then X is called in a neutral state;
3 
If m ( X ) < t 1 , then X is called in an alliance state.
If a set X of agents is in a conflict state, then we call X a conflict set of agents; if X is in a neutral state, then we call X a neutral set of agents; and if X is in an alliance state, then we call X an alliance set of agents.
In this paper, we refer to the conflict analysis model proposed in [25] as the Zhi model.

3. L -3WCL Based Conflict Analysis Model

The Zhi model can effectively analyze conflict situations with the uncertainty of agents in trustworthiness, but it may lead to misjudgment of conflicts. Example 1 illustrates the cause of misjudgment.
Example 1. 
Let K = ( U , V , R ) be a CAIS, where U = { x 1 , x 2 } , V = { y } and R ( x 1 , y ) = R ( x 2 , y ) = 0 .
For K, since the attitudes of agents x 1 and x 2 towards issue y are both negative, they reach a consensus on y; therefore, the inconsistency should be 0. However, according to the Zhi model, when using the Łukasiewicz implication, for X = { 1 / x 1 , 1 / x 2 } L U , according to Definitions 8 and 9, X ( y ) = 0 and m ( X ) = 1 . Thus, the inconsistency of agents x 1 and x 2 on y is 1, indicating that the attitudes of the two agents are inconsistent. Similar problems also exist in the Zhi model when using other commonly used implication operators.
From Example 1, we can see that the Zhi model misjudges the conflict. This is due to the fact that the Zhi model only considers the common agreement of agents as consistency and the common opposition of agents as inconsistency. Since L-3WCL [27] is able to characterize both the attributes that objects commonly possess and the attributes that they commonly do not possess, we will utilize L-3WCL to describe the issues that agents commonly agree on and commonly oppose, and propose a hybrid conflict analysis method based on L-3WCL.
Definition 10 
([27]). For a fuzzy formal context K = ( U , V , R ) , define T : L U L V × L V and T : L V × L V L U as:
X T ( y ) = ( X + ( y ) , X ( y ) )
( Y + , Y ) T ( x ) = y V ( Y + ( y ) R ( x , y ) ) y V ( Y ( y ) ¬ R ( x , y ) ) ,
where X L U , y V , ( Y + , Y ) L V × L V , x U , and
X + ( y ) = x U ( X ( x ) R ( x , y ) )
X ( y ) = x U ( X ( x ) ¬ R ( x , y ) ) .
When X T = ( Y + , Y ) and ( Y + , Y ) T = X , ( X , ( Y + , Y ) ) is called an object-derived L-fuzzy three-way concept and X and ( Y + , Y ) are called the extent and intent of ( X , ( Y + , Y ) ) , respectively. All the fuzzy concepts of K form a complete lattice A T W L ( K ) , called the L-3WCL of K, defined as
( X 1 , ( Y 1 + , Y 1 ) ) ( X 2 , ( Y 2 + , Y 2 ) ) X 1 X 2 ( Y 2 + , Y 2 ) ( Y 1 + , Y 1 ) .
A CAIS can be regarded as a fuzzy formal context, where R ( x , y ) represents the attitude of the agent x towards the issue y, and X ( x ) represents the trust degree of x. By Definition 10, X + ( y ) represents the consistency degree of X in agreeing with the issue y; X ( y ) represents the consistency degree of X in opposing y. Since the larger the value of R ( x , y ) the higher the degree of agreement, when X ( x ) remains constant, X + ( y ) increases with the increase of R ( x , y ) ; in other words, X + ( y ) is the agreement consistency. Similarly, since the larger the value of ¬ R ( x , y ) the higher the degree of opposition, when X ( x ) remains constant, X ( y ) increases with the increase of ¬ R ( x , y ) , and therefore X + ( y ) is the opposition consistency.
Similarly, in Definition 10, Y + ( y ) denotes the agreement consistency with the issue y and Y ( y ) denotes the opposition consistency with y. If Y + ( y ) = X + ( y ) and Y ( y ) = X ( y ) , i.e., Y + ( y ) is the agreement consistency of X with y and Y ( y ) is the opposition consistency of X with y, ( Y + , Y ) T ( x ) returns a new set ( Y + , Y ) T L U such that X ( x ) ( Y + , Y ) T ( x ) = X T T ( x ) for any x U (see [27]), implying that one can increase the trust degree of x while keeping the same agreement consistency and opposition consistency because X T T T ( x ) = X T ( x ) (see [27]).
In our case, we require L to be {0, 0.5, 1}. In particular, we have the following conclusions.
  • Both X + ( y ) and X ( y ) are also in {0, 0.5, 1}. If X + ( y ) = 1 , then the set X of agents completely reach a consensus in agreeing with the issue y; if X + ( y ) = 0.5 , then X partially reach a consensus in agreeing with y; and if X + ( y ) = 0 , then X does not reach a consensus in agreeing with y. Similar analysis applies to X ( y ) .
  • The value of ( Y + , Y ) T ( x ) also falls in { 0 , 0.5 , 1 } . In this case, if ( Y + , Y ) T ( x ) = 1 , then the agent x is fully trusted. If ( Y + , Y ) T ( x ) = 0.5 , then x is partially trusted; if ( Y + , Y ) T ( x ) = 0 , then x is not trusted.
Example 2. 
Since the Gödel implication operator is not suitable for building fuzzy logic systems [38], we will choose the Łukasiewicz implication operator for illustration. In this case, for X L U and y V , we have
X + ( y ) = x U ( X ( x ) R ( x , y ) ) = x U ( ( 1 X ( x ) + R ( x , y ) ) 1 )
X ( y ) = x U ( X ( x ) ¬ R ( x , y ) ) = x U ( ( 1 X ( x ) + ¬ R ( x , y ) ) 1 ) .
When all the agents x U are trustworthy, i.e., X ( x ) = 1 , we have X + ( y ) = x U R ( x , y ) and X ( y ) = x U ( ¬ R ( x , y ) ) . This is reasonable because when all the agents are trustworthy, the consistency of X will completely depend on their attitudes.
If there is an agent x U with X ( x ) = 0 , then we have X ( x ) R ( x , y ) = X ( x ) ¬ R ( x , y ) = 1 , i.e., the values of X + ( y ) and X ( y ) are not affected by R ( x , y ) . This is also reasonable because if the agent x is not trustworthy, their attitude can be ignored.
Based on the agreement consistency and opposition consistency, the comprehensive consistency X ( y ) can be defined.
Definition 11. 
Let K = ( U , V , R ) be a CAIS and X L U . The comprehensive consistency X ( y ) of X on the issue y V is defined as
X ( y ) = X + ( y ) X ( y ) .
According to Definition 11, we can see that
  • If X + ( y ) = 1 and X ( y ) = 1 , then the set X of agents reaches a complete consensus both on agreeing with y and on opposing y. In fact, when using the Łukasiewicz implication operator, for X L U and y V , if X ( x ) = 0 for all the agents x U , then we have X + ( y ) = 1 and X ( y ) = 1 , i.e., when the attitudes of all the agents are ignorable, one can conclude X + ( y ) = 1 and X ( y ) = 1 . Similarly, if for all the agents x U , X ( x ) = 0.5 implies R ( x , y ) = 0.5 , then we have X + ( y ) = 1 and X ( y ) = 1 , i.e., when the trustworthiness and the attitudes of agents are uncertain, one can also conclude X + ( y ) = 1 and X ( y ) = 1 . In both the cases, it can be considered either that X is unanimously agreeing on y or that X is unanimously opposing on y, and y is an alliance issue for X.
  • If X + ( y ) = 1 and X ( y ) 0.5 , then X reaches a complete consensus on agreeing with y and does not reach a complete consensus on opposing y. In this case, the agents within X have achieved a unanimous agreement on y. Similarly, if X + ( y ) 0.5 and X ( y ) = 1 , then X reaches a complete consensus on opposing y but does not reach a complete consensus on agreeing with y. It can be considered that the agents in X have achieved a unanimous opposition on y. In both the cases, X achieves a consensus, i.e., X ( y ) = 1 , and y is an alliance issue for X.
  • If X + ( y ) = 0 and X ( y ) = 0 , then there must exist at least one pair of agents with opposite attitudes. Therefore, its comprehensive consistency is 0. In fact, when using the Łukasiewicz implication operator, if X + ( y ) = 0 , then there must exist x 1 U such that X ( x 1 ) R ( x 1 , y ) = 0 , which implies R ( x 1 , y ) = 0 . Similarly, if X ( y ) = 0 , then there must exist x 2 U such that x 2 U , which implies ¬ R ( x 2 , y ) = 0 , yielding R ( x 2 , y ) = 1 . Hence, the attitudes of agents x 1 and x 2 towards y are opposite, indicating that X does not reach a consensus, i.e., X ( y ) = 0 . At this point, y is a conflict issue of X.
  • If X + ( y ) = 0.5 and X ( y ) = 0 , then X reaches a partial consensus on agreeing with y but does not reach a consensus on opposing y. In this case, X only partially agrees on y. Similarly, if X ( y ) = 0.5 and X + ( y ) = 0.5 , then the agents reach a partial consensus on opposing y but does not reach a consensus on agreeing with y. Thus, X only partially agrees on y. If X + ( y ) = 0.5 and X ( y ) = 0.5 , X partially reaches a consensus on agreeing with and opposing y. This indicates that X reaches a partial consensus on y. In all the three cases, X only reaches a partial consensus on y, i.e., X ( y ) = 0.5 , and y is a neutral issue of X.
In Example 1, because the Zhi model does not consider the opposition consistency, a discrepancy with the actual situation occurs when analyzing the consistency of agents. Definition 11 considers both the agreement and opposition consistency, leading to a more reasonable result than the Zhi model. For example, for the issue y in Example 1, two agents x 1 and x 2 have the agreement consistency X + ( y ) = 0 , the opposition consistency X ( y ) = 1 , and the comprehensive consistency X ( y ) = 1 , which is in line with the actual situation.
The relative inconsistency of a set of agents can be determined by the comprehensive consistency over all the issues.
Definition 12. 
Let K = ( U , V , R ) be a CAIS and X L U . Define the relative inconsistency of X as
m ( X ) = 1 | V | y V ( 1 X ( y ) ) .
Let t 1 and t 2 be two thresholds such that 0 t 1 t 2 1 . For X L U , if m ( X ) < t 1 , then X is called an alliance set of agents; if t 1 m ( X ) t 2 , then X is called a neutral set of agents; if m ( X ) > t 2 , then X is called a conflict set of agents.
The difference between the relative inconsistency (Equation (17)) and the inconsistency in the Zhi model (Equation (8)) is illustrated by Example 3.
Example 3. 
Table 1 shows a CAIS K = ( U , V , R ) , where U = { x 1 , x 2 } and V = { y 1 , y 2 , y 3 , y 4 , y 5 , y 6 } .
If X = { 1 / x 1 , 1 / x 2 } L U , then according to the Zhi model, we have X ( y 1 ) = 0 , X ( y 2 ) = 0 , X ( y 3 ) = 0.5 , X ( y 4 ) = 1 , X ( y 5 ) = 0.5 , and X ( y 6 ) = 0 . By Equation (8) we can obtain m ( X ) = 0.67 . According to the proposed model, we have X ( y 1 ) = 1 , X ( y 2 ) = 0.5 , X ( y 3 ) = 0.5 , X ( y 4 ) = 1 , X ( y 5 ) = 0.5 , and X ( y 6 ) = 0 , yielding m ( X ) = 0.42 < 0.67 = m ( X ) . From the calculation process, it can be seen that the two models are consistent on the set { y 3 , y 4 , y 5 , y 6 } , but on the set { y 1 , y 2 } , the Zhi model considers that the set is completely inconsistent on y 1 and y 2 (i.e., X ( y 1 ) = 0 and X ( y 2 ) = 0 ), while the proposed model considers that the set is completely consistent on y 1 (i.e., X ( y 1 ) = 1 ) and partially consistent on y 2 (i.e., X ( y 2 ) = 0.5 ). Intuitively, the attitude of X on y 1 is consistent and therefore the calculation result of the proposed model is reasonable; for y 2 , if the consistency of the attitudes of X towards this issue is considered to be 0, the same result can be obtained in the case of R ( x 1 , y 2 ) = 0 and R ( x 2 , y 2 ) = 1 . In other words, the result of the Zhi model on y 2 cannot distinguish the two cases: 1. R ( x 1 , y 2 ) = 0 and R ( x 2 , y 2 ) = 0.5 ; 2. R ( x 1 , y 2 ) = 0 and R ( x 2 , y 2 ) = 1 .
When considering the subset V 1 = { y 1 , y 2 , y 3 , y 4 , y 5 } , by Equation (8), we can obtain m ( X ) = 0.6 and by Equation (17), we can obtain m ( X ) = 0.3 . Thus, when the issue y 6 is not taken into consideration, m ( X ) will decrease from 0.67 to 0.6 , while m ( X ) will decrease from 0.42 to 0.3 . Obviously, the decrease of m ( X ) is greater than that of m ( X ) . This is because the Zhi model considers that X has conflicts not only on issue y 6 , but also on issues y 1 and y 2 , and thus removing issue y 6 does not eliminate the conflicts in X. The proposed model considers that X has the highest level of conflict on y 6 and thus removing y 6 results in a greater decrease in the inconsistency of X.
From Example 3, we can derive m ( X ) m ( X ) , a general relationship between Equations (8) and (17).
Theorem 1. 
Let K = ( U , V , R ) be a CAIS and X L U . Then, we have m ( X ) m ( X ) .
Proof of Theorem 1. 
For any y V , we have X ( y ) = X + ( y ) X ( y ) X + ( y ) = X ( y ) and thus the conclusion holds by Definitions 11 and 12.    □
Theorem 1 is obvious, because the proposed model captures both the agreement consistency and the opposition consistency of agents towards issues, while the Zhi model only captures the agreement consistency of agents towards issues. Thus, compared to the Zhi model, the proposed model reduces the inconsistencies of the sets of agents.
Theorem 2. 
Let K = ( U , V , R ) be a CAIS, and X, X L U . If X X , then we have m ( X ) m ( X ) .
Proof of Theorem 2. 
If X X , for any u U , then we have X ( u ) X ( u ) , and by the definition of X + ( y ) and X ( y ) , when R ( u , y ) is constant, if X ( u ) increases, then both X + ( y ) and X ( y ) will decrease. Thus, we obtain X + ( y ) X + ( y ) and X ( y ) X ( y ) . By Definitions 11 and 12, the conclusion holds.    □
Theorem 2 shows that the more agents there are, the higher the probability of inconsistency. In fact, suppose that X, X 1 L U satisfy X 1 ( u 1 ) X ( u 1 ) , X 1 ( u 2 ) = X ( u 2 ) , X 1 ( u 3 ) = X ( u 3 ) , …, X 1 ( u | U | ) = X ( u | U | ) . One can easily obtain X 1 ( y ) X ( y ) for any y and thus m ( X 1 ) m ( X ) . In other words, when the trust degree of agent u 1 increases from X ( u 1 ) to X 1 ( u 1 ) , since X 1 ( u 1 ) R ( u 1 , y ) X ( u 1 ) R ( u 1 , y ) , the importance of this agent’s opinion in X increases, leading to a decrease in the agreement consistency X + ( y ) or the opposition consistency X ( y ) , which in turn leads to an increase in the inconsistency of X. Similarly, suppose that X 2 L U satisfies X 2 ( u 1 ) = X 1 ( u 1 ) , X 2 ( u 2 ) X 1 ( u 2 ) , X 2 ( u 3 ) = X 1 ( u 3 ) , …, X 2 ( u | U | ) = X 1 ( u | U | ) . One can easily obtain X 2 ( y ) X 1 ( y ) for any y and thus m ( X 2 ) m ( X 1 ) , i.e., the increase of the trust degree in agent u 2 leads to the increase of the inconsistency in X. Continue the process and suppose that X | U | = X L U satisfies X | U | ( u 1 ) = X | U | 1 ( u 1 ) , X | U | ( u 2 ) = X | U | 1 ( u 2 ) , X | U | ( u 3 ) = X | U | 1 ( u 3 ) , …, X | U | ( u | U | ) X | U | 1 ( u | U | ) . Then, one can easily obtain X | U | ( y ) X | U | 1 ( y ) and thus m ( X | U | ) m ( X | U | 1 ) . In summary, we have
m ( X ) m ( X | U | 1 ) m ( X | U | 2 ) m ( X 1 ) m ( X ) .
By Definition 12, one can determine the state of a set of agents. If the set is allied, one may be more concerned about whether the set is unanimously agreed on or opposed some issues; if it is neutral, then the set is partially agreed and one may be more concerned about the issues that the agents are partially agreed on and the issues that they oppose; if it is conflicted, then the set does not reach a consensus on some issues and one may have to analyze the reasons of the conflicts. For this purpose, one should further analyze issues according to their states.
Definition 13. 
Let K = ( U , V , R ) be a CAIS. For X L U and y V ,
1 
If X + ( y ) = 1 and X ( y ) 1 , then y is called a unanimous agreement issue of X. The unanimous agreement set of issues of X is defined as
α + ( X ) = { y | X + ( y ) = 1 X ( y ) 1 , y V } ;
2 
If X + ( y ) 1 and X ( y ) = 1 , then y is called a unanimous opposition issue of X. The unanimous opposition set of issues of X is defined as
α ( X ) = { y | X + ( y ) 1 X ( y ) = 1 , y V } ;
3 
If X + ( y ) = 1 and X ( y ) = 1 , then y is called a unanimous issue of X. The unanimous set of issues of X is defined as
α ( X ) = { y | X + ( y ) = 1 X ( y ) = 1 , y V } ;
4 
If X + ( y ) = 0.5 and X ( y ) = 0 , then y is called an agreement-neutral issue of X. The agreement-neutral set of issues of X is defined as
γ + ( X ) = { y | X + ( y ) = 0.5 X ( y ) = 0 , y V } ;
5 
If X + ( y ) = 0 and X ( y ) = 0.5 , then y is called an opposition-neutral issue of X. The opposition-neutral set of issues of X is defined as
γ ( X ) = { y | X + ( y ) = 0 X ( y ) = 0.5 , y V } ;
6 
If X + ( y ) = 0.5 and X ( y ) = 0.5 , then y is called a completely neutral issue of X. The completely neutral set of issues of X is defined as
γ ( X ) = { y | X + ( y ) = 0.5 X ( y ) = 0.5 , y V } ;
7 
If X + ( y ) = 0 and X ( y ) = 0 , then y is called a conflict issue of X. The conflict set of issues of X is defined as
β ( X ) = { y | X + ( y ) = 0 X ( y ) = 0 , y V } .
By Definition 13, the attitude of a set of agents on an issue can be derived. For X L U and y V , if y α + ( X ) , then X agrees unanimously on y; if y α ( X ) , then X opposes unanimously y; if y α ( X ) , as mentioned before, then either X agrees unanimously on y or X opposes unanimously y. Since X is unanimous on y when y α + ( X ) α ( X ) α ( X ) , these issues are collectively referred to as alliance issues. If y γ + ( X ) , then X is neutral on agreeing on y; if y γ ( X ) , then X is neutral on opposing y; if y γ ( X ) , then X is neutral on either agreeing on or opposing y. Since X is neutral on y when y γ + ( X ) γ ( X ) γ ( X ) , these issues are collectively referred to as neutral issues. If y β ( X ) , then X is in conflict on y, and these issues will be called conflict issues. Neutral and conflict issues are collectively referred to as non-coalition issues.
Theorem 3. 
Let K = ( U , V , R ) be a CAIS, and X, X L U . If X X , then
1 
α ( X ) α ( X )
2 
β ( X ) β ( X )
3 
α + ( X ) α + ( X ) α ( X )
4 
α ( X ) α ( X ) α ( X )
5 
γ + ( X ) γ + ( X ) β ( X )
6 
γ ( X ) γ ( X ) β ( X )
7 
γ ( X ) γ ( X ) γ + ( X ) γ ( X ) β ( X )
8 
γ ( X ) γ ( X ) α + ( X ) α ( X ) α ( X ) .
Proof of Theorem 3. 
(1) If y α ( X ) , then we have X + ( y ) = 1 and X ( y ) = 1 . Since X X , we obtain X + ( y ) X + ( y ) = 1 and X ( y ) X ( y ) = 1 , i.e., X + ( y ) = 1 and X ( y ) = 1 , and hence y α ( X ) .
(2) If y β ( X ) , then we have X + ( y ) = 0 and X ( y ) = 0 . Since X X , we obtain X + ( y ) X + ( y ) = 0 and X ( y ) X ( y ) = 0 , i.e., X + ( y ) = 0 and X ( y ) = 0 , and hence y β ( X ) .
(3) If y α + ( X ) , then we have X + ( y ) = 1 and X ( y ) 1 . Since X X , we obtain X + ( y ) X + ( y ) = 1 and X ( y ) X ( y ) 1 , i.e., X + ( y ) = 1 and X ( y ) 0 . If X + ( y ) = 1 and X ( y ) 0.5 , then we have y α + ( X ) ; if X + ( y ) = 1 and X ( y ) = 1 , then we have y α ( X ) .
The proofs of (4), (5), (6), (7), and (8) are similar.    □
Theorem 3 (1) states that the more agents there are, the fewer issues they reach a consensus on. Theorem 3 (2) states that the more agents there are, the more conflict issues there are. Theorem 3 (3) and Theorem 3 (4) state that the unanimous agreement issues or unanimous opposition issues may become the unanimous issues when there are fewer agents. Theorem 3 (5) and Theorem 3 (6) state that the agreement-neutral issues or the opposition-neutral issues may become conflict issues when there are more agents. Theorem 3 (7) states that the completely neutral issues may become the agreement-neutral issues or the opposition-neutral issues or the conflict issues when there are more agents. Theorem 3 (8) states that the completely neutral issues may become the unanimous agreement issues or the unanimous opposition issues or the unanimous issues when there are fewer agents.
Example 4. 
Table 2 shows a CAIS K = ( U , V , R ) , where U = { x 1 , x 2 } and V = { y 1 , y 2 , y 3 } .
For X = { 0.5 / x 1 } and X = { 0.5 / x 1 , 0.5 / x 2 } , according to Definition 10, we can obtain X + ( y 1 ) = 1 , X ( y 1 ) = 1 , X + ( y 2 ) = 1 , X ( y 2 ) = 0.5 , X + ( y 3 ) = 1 , and X ( y 3 ) = 1 ; therefore, α ( X ) = { y 1 , y 3 } , α + ( X ) = { y 2 } . Similarly, we have X + ( y 1 ) = 1 , X ( y 1 ) = 0.5 , X + ( y 2 ) = 0.5 , X ( y 2 ) = 0.5 , X + ( y 3 ) = 1 , and X ( y 3 ) = 1 ; therefore, α + ( X ) = { y 1 } , α ( X ) = { y 3 } and γ ( X ) = { y 2 } . When the number of agents increases, i.e., X becomes X , according to Theorem 3 (1), the unanimous issues become fewer, from { y 1 , y 3 } to { y 3 } . When the number of agents decreases, i.e., X becomes X, according to Theorem 3 (3), the unanimous agreement issue y 1 become the unanimous issue; according to Theorem 3 (8), the completely neutral issues y 2 become the unanimous agreement issues. In addition, it can be found that although X X , α + ( X ) has no inclusion relation with α + ( X ) .
According to the above discussion, for a given set X of agents, the state of X and its attitude towards each issue can be determined. When the trust degrees of agents changes frequently; however, the workload of performing conflict analysis may be huge. It is easy to conclude that for a set of agents containing n agents, the number of sets of agents under | L | is | L | n ; even when L = { 0 , 0.5 , 1 } , the number of sets of agents will be 3 n . In fact, the basic nature of L-3WCL states that sets of agents with different trust degrees may produce the same conflict analysis results. Specially, for the sets X 1 and X 2 of agents with different trust degrees, if X 1 + = X 2 + and X 1 = X 2 , then their conflict analysis results are exactly the same. In other words, the number of conflict analysis results is equal to the number of L-fuzzy three-way concepts, much smaller than | L | n . Therefore, in order to reduce the workload, we can establish L-3WCL to calculate all possible conflict analysis results, and when trust degrees change, we only need to find the result corresponding to X. In addition, by analyzing L-3WCL, we can also observe the change of conflict analysis results when trust degree changes.
In A T W L ( K ) , the extent of a concept represent a set of agents with different trust degrees, and the intent represent the agreement degree and opposition degree. For any set X of agents, the corresponding fuzzy three-way concept ( X T T , X T T T ) can be found in A T W L ( K ) , and the relative inconsistency m ( X T T ) and sets of issues α + ( X T T ) , α ( X T T ) , α ( X T T ) , γ + ( X T T ) , γ ( X T T ) , γ ( X T T ) , and β ( X T T ) can be derived by Definitions 12 and 13.
For three concepts ( X 1 , X 1 T ) , ( X 2 , X 2 T ) and ( X 3 , X 3 T ) in A T W L ( K ) , if ( X 3 , X 3 T ) ( X 1 , X 1 T ) ( X 2 , X 2 T ) , then X 3 X 1 X 2 . According to Theorems 1 and 2, if the trust degrees of X 1 raises to X 2 , the conflict issues in X 1 must be the conflict issues in X 2 , the opposition-neutral issues must be the opposition-neutral issues or the conflict issues in X 2 , the agreement-neutral issues must be the agreement-neutral issues or the conflict issues in X 2 , the complete neutral issues must be the complete neutral issues, the opposition-neutral issues, and the agreement-neutral issues or the conflict issues in X 2 . If the trust degrees of X 1 decreases to X 3 , then the unanimous issues in X 1 must be the unanimous issues in X 3 , the unanimous agreement issues must be the unanimous agreement issues or unanimous issues in X 3 , the unanimous opposition issues must be the unanimous opposition issues or the unanimous issues in X 3 , the completely neutral issues must be the completely neutral issues, the unanimous agreement issues, the unanimous opposition issues, or the unanimous issues in X 3 .
In the following, we illustrate the L-3WCL based conflict analysis model in Example 5.
Example 5. 
Consider the situation in which some agents are discussing some strategies for the development of an enterprise. Since the circumstances and the interests of agents may be different, there may be some conflicts among agents. Assume that three agents and four development strategies constitute a CAIS K = ( U , V , R ) , where U = { x 1 , x 2 , x 3 } and V = { y 1 , y 2 , y 3 , y 4 } with y 1 representing talent introduction, y 2 representing project management, y 3 representing marketing and sales, and y 4 representing technology development. The attitude of each agent towards each issue is shown in Table 3. For example, agent x 1 has a positive attitude towards y 1 , agent x 2 has a negative attitude towards y 1 , and agent x 3 has a neutral attitude towards y 1 .
Since agents have different discourse powers, we adopt three levels of analysis, i.e., agent has very large discourse power, agent has relatively small discourse power, and agent has no discourse power. For example, when the discourse power of agent x 1 is relatively small, the discourse power of agent x 2 is very large, and agent x 3 has no discourse power, we represent it as { 0.5 / x 1 , 1 / x 2 , 0 / x 3 } . When Łukasiewicz implication operator is used, A T W L ( K ) of K is shown in Figure 1.
The concepts in Figure 1 and the corresponding conflict analysis results are shown in Table 4.
In Table 4, for the concepts with at least two agents in the extents, the conflict analysis results obtained are the consistency of different agents; for the concepts with only one agent in the extents, the conflict analysis results obtained are the consistency of the agent for different issues; for the concept having no agents in the extent, the conflict analysis results are not meaningful.
The conflict analysis result of a set of agents can be used to infer the conflict analysis results of other sets of agents. For example, the conflict issue of X 4 = { 0.5 / x 1 , 1 / x 2 , 1 / x 3 } is y 4 ; according to Theorem 3, we can infer that issue y 4 must also be a conflict issue of X 1 = { 1 / x 1 , 1 / x 2 , 1 / x 3 } ; the opposition-neutral issues y 1 and y 3 of X 4 must be the opposition-neutral issues or the conflict issues of X 1 ; and the agreement-neutral issue y 2 of X 4 must be an agreement-neutral issue or a conflict issue of X 1 .
The extents of the concepts in Table 4 contain only some sets in L U and other sets of agents can be described by the corresponding concepts in A T W L ( K ) , as shown in Table 5.
Combining Table 4 and Table 5, all the conflict analysis results of all sets of agents can be calculated. For example, in Table 5, we can see that for the set of agents X = { 0.5 / x 1 , 1 / x 3 } , i.e., x 1 has a relatively large discourse power, x 3 has a very large discourse power, and x 2 has no discourse power, the corresponding concept can be obtained as ( { 0.5 / x 1 , 0.5 / x 2 , 1 / x 3 } , ( { 0.5 / y 1 , 0.5 / y 2 } , { 0.5 / y 1 , 1 / y 3 , 0.5 / y 4 } ) ) , and α + ( X ) = , α ( X ) = { y 3 } , α ( X ) = , γ + ( X ) = { y 2 } , γ ( X ) = { y 4 } , γ ( X ) = { y 1 } , and β ( X ) = . This indicates that the agents in X reach a complete consensus on opposing the marketing and sales issue y 3 , a partial consensus on agreeing on the project management issue y 2 and opposing the technology development issue y 4 , and a partial consensus on the talent introduction issue y 1 . Since m ( X ) = 0.375 , the conflict between agents is small.
When applying the Zhi model to Table 3, the conflict analysis results can be obtained, as shown in Table 6. The comparison with the proposed model is also shown in Table 6.
According to Table 6, among the 17 results obtained by the Zhi model, 16 results are different from those obtained by the proposed model, indicating that there is a 94% probability that the Zhi model is problematic. The proposed model reduces the inconsistency in 16 results compared with the Zhi model, as shown in Figure 2. According to Figure 2 and Table 6, it is clear that the inconsistency decreases as the number of agents decreases, which verifies the correctness of Theorem 2. It can be observed that the difference between the two models occurs when the opposition consistency X ( y ) of X for an issue y is greater than the agreement consistency X + ( y ) . This is because the consistency in the Zhi model is the agreement consistency, and the consistency in the proposed model is the maximum of the opposition consistency and the agreement consistency. By examining the disparity distribution of two inconsistencies which is shown in Figure 3, we can observe that the disparities between the two inconsistencies are primarily situated in the interval [ 0.125 , 0.375 ) . These variations in value are relatively minor. On the interval [ 0.375 , 0.5 ) , there are two results with a large disparity between the two inconsistencies, which is mainly caused by the Zhi method misclassifying an alliance issue as a conflict issue. In particular, the disparity between the two inconsistencies in the 14th result is 0.5, the largest disparity among all the results. This is because the Zhi model treats the two issues y 3 and y 4 , which are unanimously opposed by X, as conflict issues, and the proposed model considers the two issues as the unanimous opposition issues. This means that if the Zhi model determines that X is a non-coalition, the proposed model must consider that X is an alliance. Thus, the disparity between the two inconsistencies directly affects the disparity in the states of X. Similarly, in the 15th-16th results, the inconsistency of the Zhi model indicates that the sets of agents are divergent and neutral on several issues, and no consensus is reached. The inconsistencies of the proposed model are all 0 and the sets of agents reach a consensus on all the issues.

4. Conflict Resolution Based on the Proposed Model

In Section 3, we examined the alliance issues, the neutrality issues, and the conflict issues of a given set of agents. For a set of agents in alliance, minimal divergence exists in the set, enabling the agents to easily reach a consensus. A non-alliance set of agents may experience substantial divergence in issues, presenting challenges for consensus-building among the agents. To address these conflicts, it becomes crucial to develop strategies that eliminate the divergence between agents and mitigate the relative inconsistency, thereby transforming the non-alliance set of agents into an alliance set.
For X L U , it follows from Definition 12 that the relative inconsistency m ( X ) is determined by the comprehensive consistency X ( y ) of all the issues y, and thus m ( X ) can be decreased only by increasing X ( y ) . According to Definition 13, if y α + ( X ) α ( X ) α ( X ) , then we have X ( y ) = 1 , so the comprehensive consistency X ( y ) of an alliance issue cannot be increased; if y γ + ( X ) γ ( X ) γ ( X ) , then we have X ( y ) = 0.5 , and the comprehensive consistency X ( y ) can be increased from 0.5 to 1, i.e., a neutral issue can be transformed into an alliance issue, thus decreasing the relative inconsistency m ( X ) ; if y β ( X ) , then we have X ( y ) = 0 , and the comprehensive consistency X ( y ) can be increased from 0 to 0.5 or 1, which transforms a conflict issue into a neutral issue or an alliance issue, thus decreasing the relative inconsistency m ( X ) . Therefore, the relative inconsistency can be effectively decreased only by increasing the comprehensive consistency of non-alliance issues. In the following, we consider how to increase the comprehensive consistency of the issues in γ + ( X ) , γ ( X ) , γ ( X ) , and β ( X ) .
According to Definition 11, the comprehensive consistency X ( y ) is determined by the agreement consistency X + ( y ) and the opposition consistency X ( y ) , and from Definition 10, it follows that X + ( y ) and X ( y ) are jointly determined by the trust degree X ( x ) and the attitude R ( x , y ) . To increase X ( y ) , it is necessary to increase X + ( y ) or X ( y ) . Depending on the nature of implication operator, X + ( y ) and X ( y ) can be increased by decreasing X ( x ) ; however, in most conflict situations, X ( x ) cannot be easily changed, e.g., when an enterprise makes a development plan, the discourse powers of agents in different positions are generally fixed. Therefore, in order to increase X + ( y ) or X ( y ) , it is necessary to change agents’ attitudes towards issues.
Let K = ( U , V , R ) be a CAIS and X L U . In order to increase X ( y ) , X + ( y ) can be increased by changing the attitudes of agents. If X + ( y ) = 0 , then in order to increase X + ( y ) to 0.5, according to the definition of X + ( y ) , the set of agents whose attitudes should be changed is
X 0 + ( y ) = { x | X ( x ) R ( x , y ) = 0 }
For x X 0 + ( y ) , we have X ( x ) 0 ; otherwise, we have X ( x ) R ( x , y ) = 0 R ( x , y ) = 1 . Therefore, we denote
X ˙ 1 + ( y ) = { x | X ( x ) R ( x , y ) = 0 , X ( x ) = 1 }
X ˙ 0.5 + ( y ) = { x | X ( x ) R ( x , y ) = 0 , X ( x ) = 0.5 }
and then X 0 + ( y ) = X ˙ 1 + ( y ) X ˙ 0.5 + ( y ) .
If X + ( y ) = 0 , then in order to increase X + ( y ) to 1, the set of agents whose attitudes should be changed is
X 0 , 0.5 + ( y ) = X 0 + ( y ) X 0.5 + ( y ) = X ˙ 1 + ( y ) X ˙ 0.5 + ( y ) X 0.5 + ( y )
where X 0.5 + ( y ) = { x | X ( x ) R ( x , y ) = 0.5 } .
If X + ( y ) = 0.5 , then in order to increase X + ( y ) to 1, the set of agents whose attitudes should be changed is X 0.5 + ( y ) .
Similarly, in order to increase X ( y ) , it is also feasible to increase X ( y ) by changing the attitudes of the agents. If X ( y ) = 0 , then in order to increase X ( y ) to 0.5, according to the definition of X ( y ) , the set of agents whose attitude should be changed is
X 0 ( y ) = { x | X ( x ) ¬ R ( x , y ) = 0 }
For x X 0 ( y ) , we have R ( x , y ) 0 ; otherwise, we have X ( x ) ¬ 0 = X ( x ) 1 = 1 . Therefore, we denote
X ˙ 0.5 ( y ) = { x | X ( x ) ¬ R ( x , y ) = 0 , R ( x , y ) = 0.5 }
X ˙ 1 ( y ) = { x | X ( x ) ¬ R ( x , y ) = 0 , R ( x , y ) = 1 }
and then X 0 ( y ) = X ˙ 0.5 ( y ) X ˙ 1 ( y ) .
If X ( y ) = 0 , then in order to increase X ( y ) to 1, the set of agents whose attitudes should be changed is
X 0 , 0.5 ( y ) = X 0 ( y ) X 0.5 ( y ) = X ˙ 0.5 ( y ) X ˙ 1 ( y ) X 0.5 ( y )
where X 0.5 ( y ) = { x | X ( x ) ¬ R ( x , y ) = 0.5 } .
If X ( y ) = 0.5 , then in order to increase X ( y ) to 1, the set of agents whose attitudes should be changed is X 0.5 ( y ) .
The following conclusions hold.
Theorem 4. 
Let K = ( U , V , R ) be a CAIS, x U and X L U .
1 
Suppose y γ + ( X ) .
(a) 
If | X 0.5 + ( y ) | | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · | X 0.5 + ( y ) | . In this case, the attitude of x needs to be increased to 1 for X ( x ) = 1 and R ( x , y ) = 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 0 .
(b) 
If | X 0.5 + ( y ) | > | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) . In this case, if ¬ 0.5 = 0 , then the attitude of x needs to be reduced to 0 for X ( x ) 0.5 and R ( x , y ) 0.5 ; if ¬ 0.5 = 0.5 , then the attitude of x needs to be reduced to 0 for X ( x ) = 1 and R ( x , y ) 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 1 .
2 
Suppose y γ ( X ) .
(a) 
If | X 0.5 ( y ) | 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · | X 0.5 ( y ) | . In this case, the attitude of x needs to be reduced to 0 for X ( x ) = 1 and R ( x , y ) = 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 1 .
(b) 
If | X 0.5 ( y ) | > 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · ( 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | ) . In this case, the attitude of x needs to be increased to 1 for X ( x ) = 1 and R ( x , y ) 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 0 .
3 
Suppose y γ ( X ) .
(a) 
If | X 0.5 + ( y ) | | X 0.5 ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · | X 0.5 + ( y ) | . In this case, the attitude of x needs to be increased to 1 for X ( x ) = 1 and R ( x , y ) = 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 0 .
(b) 
If | X 0.5 + ( y ) | > | X 0.5 ( y ) | , then X ( y ) can be increased from 0.5 to 1 at a cost of at least 0.5 · | X 0.5 ( y ) | . In this case, the attitude of x needs to be reduced to 0 for X ( x ) = 1 and R ( x , y ) = 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 1 .
4 
Suppose y β ( X ) .
(a) 
If | X ˙ 1 + ( y ) | | X 0 ( y ) | , then X ( y ) can be increased from 0 to 0.5 at a cost of at least 0.5 · | X ˙ 1 + ( y ) | . In this case, the attitude of x needs to be increased to 0.5 for X ( x ) = 1 and R ( x , y ) = 0 .
(b) 
If | X ˙ 1 + ( y ) | > | X 0 ( y ) | , then X ( y ) can be increased from 0 to 0.5 at a cost of at least 0.5 · | X 0 ( y ) | . In this case, the attitude of x needs to be reduced to 0.5 for X ( x ) = 1 and R ( x , y ) = 1 .
(c) 
If 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | , then X ( y ) can be increased from 0 to 1 at a cost of at least 0.5 · ( 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | ) . In this case, the attitude of x needs to be increased to 1 for X ( x ) = 1 and R ( x , y ) 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 0 .
(d) 
If 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | > | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | , then X ( y ) can be increased from 0 to 1 at a cost of at least 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) . In this case, if ¬ 0.5 = 0 , then the attitude of x needs to be reduced to 0 for X ( x ) 0.5 and R ( x , y ) 0.5 ; if ¬ 0.5 = 0.5 , then the attitude of x needs to be reduced to 0 for X ( x ) = 1 and R ( x , y ) 0.5 and to 0.5 for X ( x ) = 0.5 and R ( x , y ) = 1 .
Proof of Theorem 4. 
1. For y γ + ( X ) , in order to increase X ( y ) from 0.5 to 1, we can increase X + ( y ) from 0.5 to 1, or increase X ( y ) from 0 to 1.
In order to increase X + ( y ) from 0.5 to 1, according to the definition of X + ( y ) , the values of X ( x ) R ( x , y ) of all the agents in X 0.5 + ( y ) must be increased from 0.5 to 1. If x X 0.5 + ( y ) , then consider the following three cases.
(1) If X ( x ) = 1 , then we have 1 R ( x , y ) = 0.5 , and by Proposition 1 (2), we have R ( x , y ) = 0.5 . In order to increase the value of X ( x ) R ( x , y ) from 0.5 to 1, we need to change R ( x , y ) to R 1 + ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) a = 1 } . Since 1 a = 1 , we have a = 1 and R 1 + ( x , y ) = 1 . Thus, the cost is R 1 + ( x , y ) R ( x , y ) = 0.5 .
(2) If X ( x ) = 0.5 , then we have 0.5 R ( x , y ) = 0.5 and thus R ( x , y ) = 0 ; otherwise, by Proposition 1 (1), if 0.5 R ( x , y ) , then we have 0.5 R ( x , y ) = 1 . In order to increase the value of X ( x ) R ( x , y ) from 0.5 to 1, we need to change R ( x , y ) to R 1 + ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) a = 1 } . Since 0.5 a = 1 , we have a = 0.5 or a = 1 and thus R 1 + ( x , y ) = 0.5 . Therefore, the cost is R 1 + ( x , y ) R ( x , y ) = 0.5 .
(3) If X ( x ) = 0 , then by Proposition 1 (1), R ( x , y ) does not exist, such that 0 R ( x , y ) = 0.5 .
Therefore, the total cost is x X 0.5 + ( y ) ( R 1 + ( x , y ) R ( x , y ) ) = 0.5 · | X 0.5 + ( y ) | .
In order to increase X ( y ) from 0 to 1, according to the definition of X ( y ) , the values of X ( x ) ¬ R ( x , y ) of all the agents in X 0 , 0.5 ( y ) must be increased to 1. Since X 0 , 0.5 ( y ) = X ˙ 0.5 ( y ) X ˙ 1 ( y ) X 0.5 ( y ) and X ˙ 0.5 ( y ) , X ˙ 1 ( y ) and X 0.5 ( y ) are disjoint, we need to consider the three sets, respectively.
If x X ˙ 0.5 ( y ) , then consider the following three cases.
(1) If X ( x ) = 1 , then, since R ( x , y ) = 0.5 , we have 1 ¬ R ( x , y ) = ¬ R ( x , y ) = ¬ 0.5 = 0 . In order to increase X ( x ) R ( x , y ) of all the agents in X ˙ 0.5 ( y ) from 0 to 1, we need to change R ( x , y ) to R 1 ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) ¬ a = 1 } . Since 1 ¬ a = ¬ a = 1 , we have a = 0 and thus R 1 ( x , y ) = 0 . The cost is R ( x , y ) R 1 ( x , y ) = 0.5 .
(2) If X ( x ) = 0.5 , then because R ( x , y ) = 0.5 , we have 0.5 ¬ R ( x , y ) = 0 . Since 0.5 0.5 = 0.5 1 = 1 , we have ¬ 0.5 = 0 . At this time, we need to change R ( x , y ) to R 1 ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) ¬ a = 1 } , so we have ¬ a = 0.5 or ¬ a = 1 . If ¬ a = 0.5 , then according to Proposition 1 (3), we have a = 0.5 and thus ¬ 0.5 = 0.5 , which is contradictory to ¬ 0.5 = 0 . Thus, we have R 1 ( x , y ) = 0 , and the cost is R ( x , y ) R 1 ( x , y ) = 0.5 .
(3) If X ( x ) = 0 , then R ( x , y ) does not exist, such that 0 ¬ R ( x , y ) = 0.5 .
Therefore, the total cost is x X ˙ 0.5 ( y ) ( R ( x , y ) R 1 ( x , y ) ) = 0.5 · | X ˙ 0.5 ( y ) | .
If x X ˙ 1 ( y ) , then consider the following three cases.
(1) If X ( x ) = 1 , then in order to increase X ( x ) ¬ R ( x , y ) from 0 to 1, we need to change R ( x , y ) to R 1 ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) ¬ a = 1 } . Since 1 ¬ a = ¬ a = 1 , we have a = 0 and thus R 1 ( x , y ) = 0 . The cost is R ( x , y ) R 1 ( x , y ) = 1 .
(2) If X ( x ) = 0.5 , then since R ( x , y ) = 1 , we have 0.5 ¬ R ( x , y ) = 0.5 ¬ 1 = 0.5 0 = ¬ 0.5 = 0 . Since 0.5 ¬ a = 1 , we have ¬ a = 0.5 or ¬ a = 1 . If ¬ a = 0.5 , according to Proposition 1 (3), we can conclude a = 0.5 and thus ¬ 0.5 = 0.5 , which is contradictory to ¬ 0.5 = 0 . Therefore, we have ¬ a = 1 , i.e., a = 0 . In this case, R 1 ( x , y ) = 0 and the cost is R ( x , y ) R 1 ( x , y ) = 1 .
(3) If X ( x ) = 0 , then R ( x , y ) does not exist, such that 0 ¬ R ( x , y ) = 0.5 .
Therefore, the total cost is x X ˙ 1 ( y ) ( R ( x , y ) R 1 ( x , y ) ) = x X ˙ 1 ( y ) 1 = | X ˙ 1 ( y ) | .
If x X 0.5 ( y ) , then consider the following three cases.
(1) If X ( x ) = 1 , we have 1 ¬ R ( x , y ) = ¬ R ( x , y ) = 0.5 , and only when R ( x , y ) = 0.5 , ¬ R ( x , y ) = 0.5 holds. In order to increase the value of X ( x ) ¬ R ( x , y ) from 0.5 to 1, we need to change R ( x , y ) to R 1 ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) ¬ a = 1 } . Since 1 ¬ a = 1 , we have ¬ a = 1 and thus a = 0 . Thus, we have R 1 ( x , y ) = 0 , and the cost is R ( x , y ) R 1 ( x , y ) = 0.5 .
(2) If X ( x ) = 0.5 , then we have 0.5 ¬ R ( x , y ) = 0.5 and thus ¬ R ( x , y ) = 0 , yielding ¬ 0.5 = 0.5 0 = 0.5 . Since ¬ R ( x , y ) = 0 , we have R ( x , y ) = 1 or R ( x , y ) = 0.5 . If R ( x , y ) = 0.5 , then we have ¬ 0.5 = 0 , which is contradictory to ¬ 0.5 = 0.5 , and thus R ( x , y ) = 1 . In order to increase the value of X ( x ) ¬ R ( x , y ) from 0.5 to 1, we need to change R ( x , y ) to R 1 ( x , y ) = { a { 0 , 0.5 , 1 } | X ( x ) ¬ a = 1 } . Since 0.5 ¬ a = 1 , we have ¬ a = 0.5 or ¬ a = 1 , and thus a = 0.5 or a = 0 . At this time, we have R 1 ( x , y ) = 0.5 , and the cost is R ( x , y ) R 1 ( x , y ) = 0.5 .
(3) If X ( x ) = 0 , then R ( x , y ) does not exist, such that 0 ¬ R ( x , y ) = 0.5 .
Therefore, the total cost is x X 0.5 ( y ) ( R ( x , y ) R 1 ( x , y ) ) = 0.5 · | X 0.5 ( y ) | .
In summary, in order to increase X ( y ) from 0.5 to 1, we can increase X + ( y ) to 1 at the cost of 0.5 · | X 0.5 + ( y ) | , or increase X ( y ) to 1 at the cost of 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) . If | X 0.5 + ( y ) | | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | , then X + ( y ) can be increased to 1 at a cost of at least 0.5 · | X 0.5 + ( y ) | . Otherwise, X ( y ) can be increased to 1 at a cost of at least 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) .
Thus, if X + ( y ) is increased to 1, then we need to increase the attitude of x with X ( x ) = 1 and R ( x , y ) = 0.5 to 1, as well as increasing the attitude of x with X ( x ) = 0.5 and R ( x , y ) = 0 to 0.5. If X ( y ) is increased to 1, and if ¬ 0.5 = 0 , then we need to reduce the attitude of x with X ( x ) 0.5 and R ( x , y ) 0.5 to 0. If ¬ 0.5 = 0.5 , then the attitude of x with X ( x ) = 1 and R ( x , y ) 0.5 should be reduced to 0, and the attitude of x with X ( x ) = 0.5 and R ( x , y ) = 1 should be reduced to 0.5.
Other results can be proved similarly.    □
By Definition 13, if m ( X ) > t 1 , then X is a neutral or conflict set of agents. In order to transform X into an alliance set, it is necessary to reduce m ( X ) to t 1 , with the difference of m ( X ) t 1 .
In order to reduce m ( X ) , it is necessary to increase X ( y ) . If X ( y ) = 0 , then X ( y ) can be increased either to 0.5 or to 1; if X ( y ) = 0.5 , then X ( y ) can be increased to 1.
Now, denote the number of issues such that X ( y ) = 0 as n 0 and the number of issues such that X ( y ) = 0.5 as n 0.5 . According to Definition 12, we have
m ( X ) = y V ( 1 X ( y ) ) | V | = n 0 + 0.5 n 0.5 | V |
Denote by n 0 0.5 the number of issues whose X ( y ) need to be increased from 0 to 0.5; denote by n 0 1 the number of issues whose X ( y ) need to be increased from 0 to 1; denote by n 0.5 1 the number of issues whose X ( y ) need to be increased from 0.5 to 1. Thus, the relative inconsistency will become
m ( X ) = 0.5 n 0 0.5 + n 0 n 0 0.5 n 0 1 + 0.5 ( n 0.5 n 0.5 1 ) | V |
and the difference is
m ( X ) m ( X ) = Δ m ( X ) = 0.5 n 0 0.5 + 0.5 n 0.5 1 + n 0 1 | V |
We require Δ m ( X ) m ( X ) t 1 , i.e., n 0 0.5 , n 0 1 and n 0.5 1 satisfy Equation (37)
0.5 n 0 0.5 + 0.5 n 0.5 1 + n 0 1 ( m ( X ) t 1 ) | V |
Obviously, there may be different values of n 0 0.5 , n 0.5 1 and n 0 1 that satisfy Equation (37) and can transform X into an alliance set. Since the cost of increasing X ( y ) varies for different issues y, in real life, people tend to seek the least costly way to make decisions. Therefore, it is necessary to consider which issues to change so that the cost of transforming X into an alliance set is minimal.
For an issue y, there may be more than one way to increase X ( y ) . For y γ + ( X ) γ ( X ) γ ( X ) , X ( y ) can only be increased from 0.5 to 1 with a cost of 0.5; for y β ( X ) , X ( y ) can be increased from 0 to 0.5 or from 0 to 1 with a cost of 0.5 or 1. Therefore, there are at most two methods to increase X ( y ) for issue y, but at most one method can be chosen.
The problem can be formulated as follows. For a set of s issues, r i { 1 , 2 } denotes the number of increasing methods for the i-th issue; the variation of the k-th increasing method for the i-th issue is w i k , and the corresponding cost is c i k . The objective is to choose an optimal solution M = { p 1 1 , . . . , p 1 r 1 , p 2 1 , , p 2 r 2 , , p s 1 , , p s r s } , such that the sum of the costs i = 1 s k = 1 r i p i k c i k is minimized while the sum of variations i = 1 s k = 1 r i p i k w i k is not less than W = ( m ( X ) t 1 ) | V | , where p i k { 0 , 1 } indicates whether the optimal solution chooses the k-th increasing method for the i-th issue, i.e.,
min i = 1 s k = 1 r i p i k c i k
s . t . i = 1 s k = 1 r i p i k w i k W
It can be observed that the problem is actually a grouped knapsack problem, which is a special kind of knapsack problem. There are several solutions to solve the knapsack problem, such as branch-and-bound method [39], dynamic programming method [40], and approximation algorithm [41]. Since the dynamic programming method can solve the optimal solution quickly for smaller knapsack problems, we adopt the dynamic programming method to select the least costly set of issues. The method of solving the optimal attitude change strategy based on the dynamic programming method is shown in Algorithm 1.
Algorithm 1 The optimal attitude change strategy based on dynamic programming method.
Require: A CAIS K = ( U , V , R ) , threshold t 1 , a set X L U of agents
Ensure: An optimal attitude change strategy
1:
Compute γ + ( X ) , γ ( X ) , γ ( X ) , and β ( X ) according to Definition 13
2:
for each  y  in  γ + ( X )   do
3:
       Compute | X 0.5 + ( y ) | and | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) |
4:
      r [ y ] = 1         ▹ r [ y ] denotes the number of increasing methods of y
5:
      w [ y ] [ 1 ] = 0.5    ▹ w [ y ] [ 1 ] denotes the variation of the first increasing method of y
6:
      c [ y ] [ 1 ] = min ( 0.5 · | X 0.5 + ( y ) | , 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) )     ▹ c [ y ] [ 1 ] denotes the cost of the first increasing method of y
7:
end for
8:
for each  y  in  γ ( X )   do
9:
      Compute | X 0.5 ( y ) | and 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) |
10:
      r [ y ] = 1
11:
      w [ y ] [ 1 ] = 0.5
12:
      c [ y ] [ 1 ] = min ( 0.5 · | X 0.5 ( y ) | , 0.5 · ( 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | ) )
13:
end for
14:
for each  y  in  γ ( X )   do
15:
      Compute | X 0.5 + ( y ) | and | X 0.5 ( y ) |
16:
      r [ y ] = 1
17:
      w [ y ] [ 1 ] = 0.5
18:
      c [ y ] [ 1 ] = min ( 0.5 · | X 0.5 + ( y ) | , 0.5 · | X 0.5 ( y ) | )
19:
end for
20:
for each  y  in  β ( X )   do
21:
      Compute | X ˙ 1 + ( y ) | , | X 0 ( y ) | , 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | and | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) |
22:
      r [ y ] = 2             ▹ the number of increasing methods is 2
23:
      w [ y ] [ 1 ] = 0.5
24:
      c [ y ] [ 1 ] = min ( 0.5 · | X ˙ 1 + ( y ) | , 0.5 · | X 0 ( y ) | )
25:
      w [ y ] [ 2 ] = 1 w [ y ] [ 2 ] denotes the variation of the second increasing method of y
26:
      c [ y ] [ 2 ] = min ( 0.5 · ( 2 | X ˙ 1 + ( y ) | + | X ˙ 0.5 + ( y ) | + | X 0.5 + ( y ) | ) , 0.5 · ( | X ˙ 0.5 ( y ) | + 2 | X ˙ 1 ( y ) | + | X 0.5 ( y ) | ) )          ▹ c [ y ] [ 2 ] denotes the cost of the second increasing method of y
27:
end for
28:
Compute m ( X ) according to Definition 12
29:
Δ = ( m ( X ) t 1 ) | V |    ▹ Δ is the amount of variations needed to transform X into an alliance set
30:
Utilize the dynamic programming method in [30] or [31,32] to solve the knapsack problem and obtain the optimal solution M = { p 1 1 , , p 1 r 1 , p 2 1 , , p 2 r 2 , , p s 1 , , p s r s }
31:
According to M, determine the agents that need to change their attitudes, and add the agents and the issues to the corresponding sets
32:
return R 0 0.5 , R 0.5 1 , R 0 1 , R 0.5 0 , R 1 0.5 , R 1 0
Steps 2–27 of Algorithm 1 first determine the number of increasing methods r [ y ] , the variation w [ y ] [ k ] , and the cost c [ y ] [ k ] of the k-th increasing method for issue y according to Theorem 4. Then, Algorithm 1 solves the optimal attitude change strategy by dynamic programming method in Step 30 and obtains an optimal solution M = { p 1 1 , , p 1 r 1 , p 2 1 , , p 2 r 2 , , p s 1 , , p s r s } . Finally, in Step 31, according to the optimal solution M and Theorem 4, Algorithm 1 identifies the agents that are required to change their attitudes towards some specific issues, and adds the pairs of agent and issue to the corresponding sets: R 0 0.5 , R 0.5 1 , R 0 1 , R 0.5 0 , R 1 0.5 , and R 1 0 .
Example 6 verifies the validity of Algorithm 1.
Example 6. 
For the CAIS in Example 5, given the threshold t 1 = 0.35 , we have X = { 1 / x 1 , 0.5 / x 2 , 1 / x 3 } . According to Table 3, we have m ( X ) = 0.625 , γ + ( X ) = { y 1 } , γ ( X ) = { y 3 , y 4 } , and β ( X ) = { y 2 } . Since m ( X ) = 0.625 > 0.35 , X should be transformed into an alliance set. According to Algorithm 1, X ( y 1 ) of issue y 1 can be increased from 0.5 to 1 with the variation w [ y 1 ] [ 1 ] = 0.5 and the cost c [ y 1 ] [ 1 ] = 0.5 ; X ( y 3 ) of issue y 3 can be increased from 0.5 to 1 with the variation w [ y 3 ] [ 1 ] = 0.5 and the cost c [ y 3 ] [ 1 ] = 0.5 ; X ( y 4 ) of issue y 4 can be increased from 0.5 to 1 with the variation w [ y 4 ] [ 1 ] = 0.5 and the cost c [ y 4 ] [ 1 ] = 0.5 ; X ( y 2 ) of issue y 2 can be increased from 0 to 0.5 with the variation w [ y 2 ] [ 1 ] = 0.5 and the cost c [ y 2 ] [ 1 ] = 0.5 , or increased from 0 to 1 with the variation w [ y 2 ] [ 2 ] = 1 and the cost c [ y 2 ] [ 2 ] = 1 . According to the dynamic programming method, only p y 2 2 = 1 and p y 4 1 = 1 can be obtained in M.
Since y 2 β ( X ) , we can compute 2 | X ˙ 1 + ( y 1 ) | + | X ˙ 0.5 + ( y 1 ) | + | X 0.5 + ( y 1 ) | = 2 = | X ˙ 0.5 ( y 1 ) | + 2 | X ˙ 1 ( y 1 ) | + | X 0.5 ( y 1 ) | . Furthermore, since R ( x 1 , y 2 ) = 0 and X ( x 2 ) = 1 , we have R 0 1 = { ( x 1 , y 2 ) } . Since y 4 γ ( X ) , we can compute | X ˙ 0.5 ( y 4 ) | = 1 < 4 = 2 | X ˙ 1 + ( y 4 ) | + | X ˙ 0.5 + ( y 4 ) | + | X 0.5 + ( y 4 ) | . Since R ( x 2 , y 4 ) = 1 and X ( x ) = 0.5 , we have R 1 0.5 = { ( x 2 , y 4 ) } .
Therefore, the optimal attitude change strategy is R 0 1 = { ( x 1 , y 2 ) } , R 1 0.5 = { ( x 2 , y 4 ) } , i.e., changing the attitude of x 1 towards y 2 from opposing to agreeing and changing the attitude of x 2 towards y 4 from agreeing to neutral. After changing attitudes, we have m ( X ) = 0.25 < 0.35 , α + ( X ) = { y 2 } , α ( X ) = { y 4 } , γ + ( X ) = { y 1 } and γ ( X ) = { y 3 } , i.e., X becomes an alliance set with the minimal cost 2. The changes in conflict analysis results caused by attitude changes is shown as Figure 4.

5. Conclusions

In this paper, a hybrid conflict analysis model is proposed by combining conflict analysis and L-3WCL. The model employs L-fuzzy three-way concepts to capture the issues on which agents commonly agree and the issues which they commonly oppose, and identifies the state of a set of agents. In addition, this paper seeks minimal cost to reach a consensus of agents by developing an optimal attitude change strategy.
In fact, the existing studies [9,10,11] used different fuzzy numbers to describe agents’ uncertainty about issues in conflict analysis situations, and these studies were conducted in the case that agents are completely trustworthy. The study [25] further considered to what extent the agent can be trusted. However, [25] used fuzzy concepts to describe agents’ consistency by considering only the common agreement consistency and ignoring the common opposition consistency. In this study, we use L-3WCL to describe both agreement consistency and disagreeing consistency, which can help us analyze conflict situations more comprehensively.
On the other hand, the existing conflict analysis solution methods [1,4,6,26] used as a solution strategy the selection of the optimal subsets of issues most agents agree on. Our method introduces a third-party assumption to mediate between agents in search of an optimal attitude change strategy, and is better suited to deal with situations where the conflict has not yet reached a serious level. Thus, our study provides a new perspective on conflict resolution.
However, there are some problems that need to be explored and improved. For example, in the proposed model, both the agreement consistency and the opposition consistency may reach the value 1, i.e., the set of agents may agree on and oppose an issue unanimously. Although we have explained this situation as a special case, this may be unreasonable. In fact, by Proposition 1 (1), this is a result of introducing CRL to characterize the problem of conflict analysis. In addition, the proposed model employs CRL to characterize both attitudes and trust degrees to { 0 , 0.5 , 1 } . Since attitude values and trust values have different meanings and operators, for example, it seems unreasonable that the negative attitude 0 is less than the positive attitude 1, but it seems reasonable that the trust degree 0 is less than the trust degree 1. Furthermore, the different meanings of their values means that different operators are needed. This may lead to unreasonable results and therefore needs to be further explored.

Author Contributions

Conceptualization, Y.Z. and Z.J.; methodology, Y.Z., Z.J. and D.L.; software, Y.Z.; validation, Y.Z. and Z.J.; formal analysis, Y.Z., Z.J. and D.L.; investigation, Z.J.; resources, Y.Z. and D.L.; data curation, Y.Z. and Z.J.; writing—original draft preparation, Y.Z. and Z.J.; writing—review and editing, Y.Z., Z.J. and D.L.; visualization, Y.Z. and Z.J.; supervision, Y.Z. and D.L.; project administration, Y.Z.; funding acquisition, Y.Z. and D.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant numbers 61972238 and 62072294).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zia, B.; Saima, M.; Malik, M.G.A. Conflict resolution using game theory and rough sets. Int. J. Intell. Syst. 2020, 36, 237–259. [Google Scholar]
  2. Du, J.; Liu, S.; Liu, Y.; Tao, L. Multi-criteria large-scale group decision-making in linguistic contexts: A perspective of conflict analysis and resolution. Group Decis. Negot. 2022, 32, 177–207. [Google Scholar] [CrossRef]
  3. Pawlak, Z. On conflicts. Int. J. Man–Mach. Stud. 1984, 21, 127–134. [Google Scholar] [CrossRef]
  4. Xu, F.; Cai, M.; Song, H.; Dai, J. The selection of feasible strategies based on consistency measurement of cliques. Inf. Sci. 2022, 583, 33–55. [Google Scholar] [CrossRef]
  5. Pawlak, Z. An inquiry into anatomy of conflicts. Inf. Sci. 1998, 109, 65–78. [Google Scholar] [CrossRef]
  6. Du, J.; Liu, S.; Liu, Y.; Yi, J. A novel approach to three-way conflict analysis and resolution with pythagorean fuzzy information. Inf. Sci. 2022, 584, 65–88. [Google Scholar] [CrossRef]
  7. Yao, Y. Three-way conflict analysis: Reformulations and extensions of the pawlak model. Knowl. Based Syst. 2019, 180, 26–37. [Google Scholar] [CrossRef]
  8. Lang, G.; Yao, Y. New measures of alliance and conflict for three-way conflict analysis. Int. J. Approx. Reason. 2021, 132, 49–69. [Google Scholar] [CrossRef]
  9. Lang, G.; Miao, D.; Fujita, H. Three-way group conflict analysis based on pythagorean fuzzy set theory. IEEE Trans. Fuzzy Syst. 2019, 28, 447–461. [Google Scholar] [CrossRef]
  10. Li, X.; Yang, Y.; Yi, H.; Yu, Q. Conflict analysis based on three-way decision for trapezoidal fuzzy information systems. Int. J. Mach. Learn. Cybern. 2022, 13, 929–945. [Google Scholar] [CrossRef]
  11. Yang, H.; Wang, Y.; Guo, Z. Three-way conflict analysis based on hybrid situation tables. Inf. Sci. 2023, 628, 522–541. [Google Scholar] [CrossRef]
  12. Suo, L.; Yang, H. Three-way conflict analysis based on incomplete situation tables: A tentative study. Int. J. Approx. Reason. 2022, 145, 51–74. [Google Scholar] [CrossRef]
  13. Wang, T.; Huang, B.; Li, H.; Liu, D.; Yu, H. Three-way decision for probabilistic linguistic conflict analysis via compounded risk preference. Inf. Sci. 2023, 631, 65–90. [Google Scholar] [CrossRef]
  14. Zhi, H.; Qi, J.; Qian, T.; Ren, R. Conflict analysis under one-vote veto based on approximate three-way concept lattice. Inf. Sci. 2019, 516, 316–330. [Google Scholar] [CrossRef]
  15. Qi, J.; Qian, T.; Wei, L. The connections between three-way and classical concept lattices. Knowl. Based Syst. 2016, 91, 143–151. [Google Scholar] [CrossRef]
  16. Yao, Y. An outline of a theory of three-way decisions. In Rough Sets and Current Trends in Computing: Proceedings of the 8th International Conference, RSCTC 2012, Chengdu, China, 17–20 August 2012; Springer: Berlin/Heidelberg, Germany, 2012; Volume 7413, pp. 1–17. [Google Scholar]
  17. Yao, Y. Three-way decision and granular computing. Int. J. Approx. Reason. 2018, 103, 107–123. [Google Scholar] [CrossRef]
  18. Yang, Z.; Ren, J.; Zhang, Z.; Sun, Y.; Zhang, C.; Wang, M.; Wang, L. A New Three-Way Incremental Naive Bayes Classifier. Electronics 2023, 12, 1730. [Google Scholar] [CrossRef]
  19. Wu, J.; Xu, J.; Lin, D.; Tu, M. Optical Flow Filtering-Based Micro-Expression Recognition Method. Electronics 2020, 9, 2056. [Google Scholar] [CrossRef]
  20. Zhang, C.; Zhang, M.; Yang, G.; Xue, T.; Zhang, Z.; Liu, L.; Wang, L.; Hou, W.; Chen, Z. Three-Way Selection Random Forest Optimization Model for Anomaly Traffic Detection. Electronics 2023, 12, 1788. [Google Scholar] [CrossRef]
  21. Wille, R. Restructuring lattice theory: An approach based on hierarchies of concepts. In Ordered Sets (Banff, Alta., 1981); D. Reidel Publishing Company: Dordrecht, The Netherlands, 1982; pp. 445–470. [Google Scholar]
  22. Wille, R.; Ganter, B. Fuzzy concept lattice. In Formal Concept Analysis: Mathematical Foundations; Springer: Berlin, Germany, 1999; pp. 183–217. [Google Scholar]
  23. Wei, L.; Liu, L.; Qi, J.; Qian, T. Rules acquisition of formal decision contexts based on three-way concept lattices. Inf. Sci. 2020, 516, 529–544. [Google Scholar] [CrossRef]
  24. Zhai, Y.; Qi, J.; Li, D.; Zhang, C.; Xu, W. Three-way decision and granular computing. Int. J. Approx. Reason. 2022, 146, 157–173. [Google Scholar] [CrossRef]
  25. Zhi, H.; Li, J.; Li, Y. Multilevel conflict analysis based on fuzzy formal contexts. IEEE Trans. Fuzzy Syst. 2022, 30, 5128–5142. [Google Scholar] [CrossRef]
  26. Sun, B.; Chen, X.; Zhang, L.; Ma, W. Three-way decision making approach to conflict analysis and resolution using probabilistic rough set over two universes. Inf. Sci. 2020, 507, 809–822. [Google Scholar] [CrossRef]
  27. He, X.; Wei, L.; She, Y. L-fuzzy concept analysis for three-way decisions: Basic definitions and fuzzy inference mechanisms. Int. J. Mach. Learn. Cybern. 2018, 9, 1857–1867. [Google Scholar] [CrossRef]
  28. Wang, Z.; Qi, J.; Shi, C.; Ren, R.; Wei, L. Multiview granular data analytics based on three-way concept analysis. Appl. Intell. 2023, 53, 14645–14667. [Google Scholar] [CrossRef]
  29. Zhang, C.; Li, J.; Lin, Y. Matrix-based reduction approach for one-sided fuzzy three-way concept lattices. J. Intell. Fuzzy Syst. 2021, 40, 11393–11410. [Google Scholar] [CrossRef]
  30. Ezugwu, A.E.; Pillay, V.; Hirasen, D.; Sivanarain, K.; Govender, M. A comparative study of meta-heuristic optimization algorithms for 0 -1 knapsack problem: Some initial results. IEEE Access 2019, 7, 43979–44001. [Google Scholar] [CrossRef]
  31. Sin, S.T.T. The parallel processing approach to the dynamic programming algorithm of knapsack problem. In Proceedings of the IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (ElConRus), St. Petersburg, Russia, 26–29 January 2021; pp. 2252–2256. [Google Scholar]
  32. Clautiaux, F.; Detienne, B.; Guillot, G. An iterative dynamic programming approach for the temporal knapsack problem. Eur. J. Oper. Res. 2021, 293, 442–456. [Google Scholar] [CrossRef]
  33. Ward, M.; Dilworth, R.P. Residuated Lattices. In The Dilworth Theorems; Kenneth, P.B., Freese, R., Joseph, P.S.K., Eds.; Birkhäuser: Boston, MA, USA, 1990; pp. 317–336. [Google Scholar]
  34. Pei, D. The characterization of residuated lattices and regular residuated lattices. Acta Math. Sin. 2002, 45, 271–278. [Google Scholar]
  35. Bělohlávek, R. Some properties of residuated lattices. Czechoslov. Math. J. 2003, 53, 161–171. [Google Scholar] [CrossRef]
  36. Bělohlávek, R. Algorithms for fuzzy concept lattices. Acta Math. Sin. 2002, 45, 271–278. [Google Scholar]
  37. Bělohlávek, R.; De Baets, B.; Outrata, J.; Vychodil, V. Computing the lattice of all fixpoints of a fuzzy closure operator. IEEE Trans. Fuzzy Syst. 2010, 18, 546–557. [Google Scholar] [CrossRef]
  38. Wang, W.; Wang, G.J. Gödel implication operator is not suitable for establishing fuzzy propositional logic. Fuzzy Sets Syst. 2005, 19, 14–18. [Google Scholar]
  39. Stefano, C.; Fabio, F.; Pablo, S. A new combinatorial branch-and-bound algorithm for the knapsack problem with conflicts. Eur. J. Oper. Res. 2021, 289, 435–455. [Google Scholar]
  40. Djeumou, F.F. A lifted-space dynamic programming algorithm for the quadratic knapsack problem. Discret. Appl. Math. 2023, 335, 52–68. [Google Scholar] [CrossRef]
  41. Boeckmann, J.; Thielen, C.; Pferschy, U. Approximating single- and multi-objective nonlinear sum and product knapsack problems. Discret. Optim. 2023, 48, 100771. [Google Scholar] [CrossRef]
Figure 1. A T W L ( K ) .
Figure 1. A T W L ( K ) .
Electronics 12 04547 g001
Figure 2. The differences in inconsistency between two models.
Figure 2. The differences in inconsistency between two models.
Electronics 12 04547 g002
Figure 3. The disparity distribution of two inconsistencies.
Figure 3. The disparity distribution of two inconsistencies.
Electronics 12 04547 g003
Figure 4. Changes in conflict analysis results caused by attitude changes.
Figure 4. Changes in conflict analysis results caused by attitude changes.
Electronics 12 04547 g004
Table 1. A CAIS for Example 3.
Table 1. A CAIS for Example 3.
y 1 y 2 y 3 y 4 y 5 y 6
x 1 000.5111
x 2 00.50.510.50
Table 2. A CAIS for Example 4.
Table 2. A CAIS for Example 4.
y 1 y 2 y 3
x 1 0.510.5
x 2 100.5
Table 3. A CAIS for Example 5.
Table 3. A CAIS for Example 5.
y 1 y 2 y 3 y 4
x 1 100.50
x 2 00.50.51
x 3 0.5100
Table 4. Conflict analysis results.
Table 4. Conflict analysis results.
( X , ( X + , X ) ) α + ( X ) α ( X ) α ( X ) γ + ( X ) γ ( X ) γ ( X ) β ( X ) m ( X )
C 1 : ( { 1 x 1 , 1 x 2 , 1 x 3 } , ( , { 0.5 y 3 } ) ) { y 3 } { y 1 , y 2 , y 4 } 0.875
C 2 : ( { 1 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 } ) ) { y 2 } { y 3 } { y 1 , y 4 } 0.75
C 3 : ( { 1 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 } , { 0.5 y 3 , 0.5 y 4 } ) ) { y 1 } { y 3 , y 4 } { y 2 } 0.625
C 4 : ( { 0.5 x 1 , 1 x 2 , 1 x 3 } , ( { 0.5 y 2 } , { 0.5 y 1 , 0.5 y 3 } ) ) { y 2 } { y 1 , y 3 } { y 4 } 0.625
C 5 : ( { 1 x 1 , 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } ) ) { y 1 } { y 2 , y 4 } { y 3 } 0.5
C 6 : ( { 0.5 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 } ) ) { y 4 } { y 1 } { y 2 , y 3 } 0.5
C 7 : ( { 0.5 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 , 0.5 y 2 } , { 0.5 y 1 , 1 y 3 , 0.5 y 4 } ) ) { y 3 } { y 2 } { y 4 } { y 1 } 0.375
C 8 : ( { 1 x 1 , 0.5 x 3 } , ( { 1 y 1 , 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 , 1 y 4 } ) ) { y 1 } { y 4 } { y 2 } { y 3 } 0.25
C 9 : ( { 0.5 x 1 , 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) ) { y 3 } { y 1 , y 2 , y 4 } 0.375
C 10 : ( { 1 x 1 , 0.5 x 2 } , ( { 0.5 y 1 , 0.5 y 3 } , { 1 y 2 , 0.5 y 3 , 0.5 y 4 } ) ) { y 2 } { y 1 } { y 4 } { y 3 } 0.375
C 11 : ( { 1 x 2 , 0.5 x 3 } , ( { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 1 y 1 , 0.5 y 2 , 0.5 y 3 } ) ) { y 1 } { y 4 } { y 2 , y 3 } 0.375
C 12 : ( { 1 x 1 } , ( { 1 y 1 , 0.5 y 3 } , { 1 y 2 , 0.5 y 3 , 1 y 4 } ) ) { y 1 } { y 2 , y 4 } { y 3 } 0.125
C 13 : ( { 0.5 x 1 , 0.5 x 3 } , ( { 1 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 1 y 4 } ) ) { y 1 } { y 3 , y 4 } { y 2 } 0.125
C 14 : ( { 0.5 x 1 , 0.5 x 2 } , ( { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } , { 0.5 y 1 , 1 y 2 , 1 y 3 , 0.5 y 4 } ) ) { y 2 } { y 3 } { y 1 , y 4 } 0.25
C 15 : ( { 1 x 2 } , ( { 0.5 y 2 , 0.5 y 3 , 1 y 4 } , { 1 y 1 , 0.5 y 2 , 0.5 y 3 } ) ) { y 4 } { y 1 } { y 2 , y 3 } 0.25
C 16 : ( { 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 1 y 2 , 0.5 y 3 , 0.5 y 4 } , { 1 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) ) { y 2 } { y 1 , y 3 } { y 4 } 0.125
C 17 : ( { 1 x 3 } , ( { 0.5 y 1 , 1 y 2 } , { 0.5 y 1 , 1 y 3 , 1 y 4 } ) ) { y 2 } { y 3 , y 4 } { y 1 } 0.125
C 18 : ( { 0.5 x 1 } , ( { 1 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } , { 0.5 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) ) { y 1 } { y 2 , y 4 } { y 3 } 0
C 19 : ( { 0.5 x 2 } , ( { 0.5 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } , { 1 y 1 , 1 y 2 , 1 y 3 , 0.5 y 4 } ) ) { y 4 } { y 1 } { y 2 , y 3 } 0
C 20 : ( { 0.5 x 3 } , ( { 1 y 1 , 1 y 2 , 0.5 y 3 , 0.5 y 4 } , { 1 y 1 , 0.5 y 2 , 1 y 3 , 1 y 4 } ) ) { y 2 } { y 3 , y 4 } { y 1 } 0
C 21 : ( , ( { 1 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } , { 1 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) ) { y 1 , y 2 , y 3 , y 4 } 0
Table 5. Sets of agents and the corresponding concepts.
Table 5. Sets of agents and the corresponding concepts.
The Set X of AgentsThe Corresponding Concepts ( X T T , X T T T )
{ 1 x 1 , 1 x 2 } C 2 : ( { 1 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 } ) )
{ 1 x 1 , 1 x 3 } C 3 : ( { 1 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 } , { 0.5 y 3 , 0.5 y 4 } ) )
{ 1 x 2 , 1 x 3 } C 4 : ( { 0.5 x 1 , 1 x 2 , 1 x 3 } , ( { 0.5 y 2 } , { 0.5 y 1 , 0.5 y 3 } ) )
{ 0.5 x 1 , 1 x 2 } C 6 : ( { 0.5 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 } ) )
{ 0.5 x 2 , 1 x 3 } C 7 : ( { 0.5 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 , 0.5 y 2 } , { 0.5 y 1 , 1 y 3 , 0.5 y 4 } ) )
{ 0.5 x 1 , 1 x 3 } C 7 : ( { 0.5 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 , 0.5 y 2 } , { 0.5 y 1 , 1 y 3 , 0.5 y 4 } ) )
Table 6. Comparison results.
Table 6. Comparison results.
NO.Results of Zhi Model m ( X ) Results of the Proposed Model m ( X )
1 ( { 1 x 1 , 1 x 2 , 1 x 3 } , ) 1 ( { 1 x 1 , 1 x 2 , 1 x 3 } , ( , { 0.5 y 3 } ) ) 0.875
2 ( { 1 x 1 , 1 x 2 , 0.5 x 3 } , { 0.5 y 3 } ) 0.875 ( { 1 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 } ) ) 0.75
3 ( { 1 x 1 , 0.5 x 2 , 1 x 3 } , { 0.5 y 1 } ) 0.875 ( { 1 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 } , { 0.5 y 3 , 0.5 y 4 } ) ) 0.625
4 ( { 0.5 x 1 , 1 x 2 , 1 x 3 } , { 0.5 y 2 } ) 0.875 ( { 0.5 x 1 , 1 x 2 , 1 x 3 } , ( { 0.5 y 2 } , { 0.5 y 1 , 0.5 y 3 } ) ) 0.625
5 ( { 1 x 1 , 0.5 x 2 , 0.5 x 3 } , { 0.5 y 1 , 0.5 y 3 } ) 0.75 ( { 1 x 1 , 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } ) ) 0.5
6 ( { 0.5 x 1 , 1 x 2 , 0.5 x 3 } , { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } ) 0.625 ( { 0.5 x 1 , 1 x 2 , 0.5 x 3 } , ( { 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 } ) ) 0.5
7 ( { 0.5 x 1 , 0.5 x 2 , 1 x 3 } , { 0.5 y 1 , 0.5 y 2 } ) 0.75 ( { 0.5 x 1 , 0.5 x 2 , 1 x 3 } , ( { 0.5 y 1 , 0.5 y 2 } , { 0.5 y 1 , 1 y 3 , 0.5 y 4 } ) ) 0.375
8 ( { 1 x 1 , 0.5 x 3 } , { 1 y 1 , 0.5 y 3 } ) 0.625 ( { 1 x 1 , 0.5 x 3 } , ( { 1 y 1 , 0.5 y 3 } , { 0.5 y 2 , 0.5 y 3 , 1 y 4 } ) ) 0.25
9 ( { 0.5 x 1 , 0.5 x 2 , 0.5 x 3 } , { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } ) 0.5 ( { 0.5 x 1 , 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) ) 0.375
10 ( { 0.5 x 1 , 0.5 x 3 } , { 1 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } ) 0.375 ( { 0.5 x 1 , 0.5 x 3 } , ( { 1 y 1 , 0.5 y 2 , 0.5 y 3 , 0.5 y 4 } , { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 1 y 4 } ) ) 0.125
11 ( { 0.5 x 1 , 0.5 x 2 } , { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) 0.375 ( { 0.5 x 1 , 0.5 x 2 } , ( { 0.5 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } , { 0.5 y 1 , 1 y 2 , 1 y 3 , 0.5 y 4 } ) ) 0.25
12 ( { 1 x 2 } , { 0.5 y 2 , 0.5 y 3 , 1 y 4 } ) 0.5 ( { 1 x 2 } , ( { 0.5 y 2 , 0.5 y 3 , 1 y 4 } , { 1 y 1 , 0.5 y 2 , 0.5 y 3 } ) ) 0.25
13 ( { 0.5 x 2 , 0.5 x 3 } , { 0.5 y 1 , 1 y 2 , 0.5 y 3 , 0.5 y 4 } ) 0.375 ( { 0.5 x 2 , 0.5 x 3 } , ( { 0.5 y 1 , 1 y 2 , 0.5 y 3 , 0.5 y 4 } , { 1 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) ) 0.125
14 ( { 1 x 3 } , { 0.5 y 1 , 1 y 2 } ) 0.625 ( { 1 x 3 } , ( { 0.5 y 1 , 1 y 2 } , { 0.5 y 1 , 1 y 3 , 1 y 4 } ) ) 0.125
15 ( { 0.5 x 1 } , { 1 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } ) 0.25 ( { 0.5 x 1 } , ( { 1 y 1 , 0.5 y 2 , 1 y 3 , 0.5 y 4 } , { 0.5 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) ) 0
16 ( { 0.5 x 2 } , { 0.5 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) 0.125 ( { 0.5 x 2 } , ( { 0.5 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } , { 1 y 1 , 1 y 2 , 1 y 3 , 0.5 y 4 } ) ) 0
17 ( , { 1 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) 0 ( , ( { 1 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } , { 1 y 1 , 1 y 2 , 1 y 3 , 1 y 4 } ) ) 0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhai, Y.; Jia, Z.; Li, D. Resolving Agent Conflicts Using Enhanced Uncertainty Modeling Tools for Intelligent Decision Making. Electronics 2023, 12, 4547. https://doi.org/10.3390/electronics12214547

AMA Style

Zhai Y, Jia Z, Li D. Resolving Agent Conflicts Using Enhanced Uncertainty Modeling Tools for Intelligent Decision Making. Electronics. 2023; 12(21):4547. https://doi.org/10.3390/electronics12214547

Chicago/Turabian Style

Zhai, Yanhui, Zihan Jia, and Deyu Li. 2023. "Resolving Agent Conflicts Using Enhanced Uncertainty Modeling Tools for Intelligent Decision Making" Electronics 12, no. 21: 4547. https://doi.org/10.3390/electronics12214547

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop