You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

17 August 2022

Hierarchical Analysis Process for Belief Management in Internet of Drones

,
and
1
Transportation Research Institute (IMOB), Hasselt University, 3500 Hasselt, Belgium
2
Computer Sciences Department, German University of Technology in Oman (GUtech), Athaibah, Muscat 130, Oman
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Applied Data Science and Intelligence

Abstract

Group awareness is playing a major role in the efficiency of mission planning and decision-making processes, particularly those involving spatially distributed collaborative entities. The performance of this concept has remarkably increased with the advent of the Internet of Things (IoT). Indeed, a myriad of innovative devices are being extensively deployed to collaboratively recognize and track events, objects, and activities of interest. A wide range of IoT-based approaches have focused on representing and managing shared information through formal operators for group awareness. However, despite their proven results, these approaches are still refrained by the inaccuracy of information being shared between the collaborating distributed entities. In order to address this issue, we propose in this paper a new belief-management-based model for a collaborative Internet of Drones (IoD). The proposed model allows drones to decide the most appropriate operators to apply in order to manage the uncertainty of perceived or received information in different situations. This model uses Hierarchical Analysis Process (AHP) with Subjective Logic (SL) to represent and combine opinions of different sources. We focus on purely collaborative drone networks where the group awareness will also be provided as service to collaborating entities.

1. Introduction

Smart cities applications are built open distributed sensors and equipment like sensors embedded in infrastructure, vehicles, UAVs, etc. These devices could collect useful data which can be integrated and analyzed to infer meaningful information and improve the quality of life [1]. In novel smart cities applications, intelligence could be implemented in devices which allows to assign a level of autonomy. Autonomy opens important opportunities for collaboration and cooperation among these devices, such as cluster-wise cooperative automated trucks [2] and collaborative smart drones [3]. Using novel processing technologies, traffic components could collaborate to automatically detect traffic accidents [4] and road hazards. The collaboration could be useful not only to decrease task completion time or to coordinate the tasks but also to improve group awareness and increase productivity and efficiency 3]. In our previous work [5], we modeled collaborative join task planning in multi-UAV application where the actions of the drones are guided by a firefly algorithm using a reward and cost ratio. Perceived and received information in agent technology could be modeled as beliefs. Smart cities applications include different aspect of uncertainty. Co-existing agents could receive and perceive conflicting and uncertain information. With the scarcity of possible available resources, decision making cannot only be guided by the reward–cost ratio but should also take in consideration resource value and uncertainty of information. A variety of uncertainty management approaches have been proposed in literature. Different beliefs fusion operators could be used to fuse received and perceived beliefs. Whereas no single belief fusion operator is suitable in every situation [6]. The suitable belief operator depends on the situation to model, the conflict between uncertain beliefs, and the reliability attributed to source of information.
In this research work, we aim to propose a belief management approach to model group beliefs under uncertainty in a drone collaboration for Intelligent Transportation System (ITS) application. We base this model on a probabilistic belief function to represent uncertainty. We propose a structure of individual agent belief base to model group and individual beliefs with different levels of uncertainty. To fuse perceived and received information, we use different belief fusion operators. The selection of the fusion operator is based on a Hierarchical Analysis Process (HAP). Moreover, to provide an in-demand belief-sharing service with a different level of uncertainty, we decompose agent belief bases in three repositories. In the application scenario, we aim to allow traffic agents (vehicles, drones, etc.) to select the most suitable fusion operators and combine received and perceived information with uncertainty, manage beliefs in the belief base, and provide group beliefs as services. In the following section, we present a short literature review of the main existing uncertainty management methods, belief fusions, and their combination with multi-criteria decision making (MCDM) methods. In the second section, we present the proposed agent dynamic belief management model. In the third and last section, we suggest an application of the proposed model in smart cities and the ITS domain.

3. Hierarchical Analysis Process for Intelligent Collaborative Belief Management

In this work, we propose to model the beliefs of the collaborative group in a decentralized manner. In the environment, different collaborative teams could be created on demand based on the need as explained in our previous work [5]. Each team has a team leader agent, which initiates the collaboration process, and member agents, which accept or refuse to join the collaboration. Once an agent accepts to join a group, he became a member. The team members will execute sub-tasks of the global mission. The leader agent will serve as a group belief repository for the team, which allows the team members to subscribe to receive reported information. The beliefs contained in the team leader are represented in four repositories:
  • Temporary Beliefs: This beliefs repository contains the beliefs received from the team members. The team members will report collected information or events to the team leader. Once the team leader receives this information, he will store this information in the Temporary Beliefs repository.
  • Individual Beliefs: This beliefs repository contains the beliefs perceived by the agent.
  • Promoted Beliefs: This beliefs repository contains the beliefs frequently reported by different team members.
  • Shared Beliefs: This belief repository contains the beliefs in which all the team members belief on it. These beliefs could be considered as an alternative model to model common beliefs presented in [28].
Figure 2 represents the belief base of the team leader agent. The collaborative team members have the same model of belief management, except that they do not store temporary beliefs from the team members of the collaborative group to reduce communication. Based on their subscription to the geographical area of interest on a selected repository, the collaborative team members will receive new reported information in the specified beliefs repository. This model of belief management allows collaborative team members to benefit from being a part of the group and increase awareness. At the same time, it evaluates the uncertainty of the information and allows member agents to subscribe to specific repositories. The reported information could also be used by the collaborative agent to guide his decision-making process. The collaborative agent has a degree of freedom to choose between selfish or collaborative behaviour [5]. However, the members will not benefit from the group awareness if they decide to disband the team.
Figure 2. Representation of beliefs management module in the leader agent.
To allow all members of the collaborative team to benefit from this beliefs model, create a team awareness, and reduce the communication messages, we allow each team member to subscribe to a region of interest. Each agent interested in receiving the information reported about a region will subscribe to the region of interest in the leader agent. The leader agent will also include a module to handle the subscription of the agent members. The team members could subscribe to temporal, promoted and shared beliefs. All the team members which will execute a task in the region of interest will publish the reported information to the leader agent, which will store it in the temporary belief repository. Based on the certainty and the frequency of the reported information. The team leader can then decide to store the belief in the promoted or shared repository or keep it in the temporary beliefs. Figure 3 represents the architecture of collaborative team of agents.
Figure 3. Decentralized management of team beliefs.
To manage the uncertainty of perceived and received information, the agents will use subjective logic to create opinions about the perceived environment. The leader agent will receive reported information and use an HAP module to identify the operator that should be used to fuse the perceived and received opinions. The proposed module uses the following criteria to decide the operator that will be used:
  • Time: The time difference between the perception of the first opinion and the time of the perception of the second opinion;
  • Location: The location of the agent reporting the opinion. The location of the agent will be used to calculate the distance between the agent reporting the event and the team leader;
  • Source: The agent sending the opinion. The source will be used to find the trust in the source. The evaluation of the trust is out of the scope of this work;
  • Risk: The risk of the perceived and received information. The risk could depend on the nature of the region where the event is perceived or the type of the perceived event or hazard;
  • Uncertainty: The uncertainty of the perceived and received opinions;
  • Conflict: The conflict between the perceived and received opinions.
The HAP module will take the five SL operators as alternatives. We define the weight of each criterion before the execution of the drone’s mission. The weights could be defined from the literature, from domain experts or derived from a training phase. These weights could be refined after the execution of the mission. During the mission, for each received information, the leader drone agent will evaluate the criteria (time, location, source, risk, uncertainty, and conflict) using the received and perceived information to decide the operator or strategy that should be used to combine the opinions. The selected operator will be used to fuse two opinions and produce a new opinion, as presented in Section 2.2. The selection of the operator will affect the uncertainty of the resulting opinion. Since the leader opinion will be shared with the team member’s drone, the selected operator will also affect the state of the belief bases of the drone members and their actions.

4. Implementation in the Context of Intelligent Transportation Systems

We are focusing in this paper on the collaborative operations of an Internet of Drones in the context of ITS.

4.1. Formulation

To model uncertainty, we use Equation (1) from subjective logic. We model the environment as a grid of cells. Several events could be randomly created and diffused in the environment. The events could also disappear after a certain time. In the environment, we model drones as Belief–Desire–Intension (BDI) agents. Each drone will have parametrized characteristics, such as the capacity of the battery, the range of the field of view, and the charging time. The agent drone will be able to perceive only the cell in his field of view. Once an event is perceived, the agent will calculate the opinion based on the base rate, the belief mass, the disbelief mass, and the uncertainty and send the opinion to the agent leader.
In the belief module a hierarchical structure model will be built to decide the best fusion operator, as shown in Figure 4. First, a Pair-Wise Comparison Matrix Criteria weights (Matrix 1) is represented to identify the importance of each criterion compared to the others. In our case, we choose to construct a judgement matrix with the (1)–(9) scale method:
V = { V t i m e ,   V l o c a t i o n   , V r i s k ,   V t r u s t ,   V u n c e r a i n t y , V c o n f l i c t } A = [ a 1 , 1 a 1 , j a i , 1 a i , j ]   ( Matrix   1 )
Figure 4. HAP for the selection of belief fusion strategy.
To evaluate the consistency index of the proposed matrix, we process the following: steps.
  • We calculate the judgement matrix normalized by column:
b i , j = a i j / a i j
2.
The normalized matrix is summed by row:
c i = j = 1 n b i j
3.
c i is normalized and weights are obtained
w i 2 = c i / c i
4.
Find the maximum eigenvalue corresponding to weight;
w 2 :   λ max = 1 n i ( ( A W ( 2 ) ) i w i ( 2 ) )
5.
Consistency testing: We calculate the degree of inconsistency or Consistency Index (CI) of the matrix A to make sure that the rankings given by different decision makers and used as inputs to the AHP application are consistent:
CI = λ max n n 1
6.
We finally calculate the Consistency Ratio (CR). The ratio of Consistency Index (CI) and the Random Consistency Index (RCI). The CI measure the degree of inconsistency. The larger the inconsistency between comparisons, the larger the consistency index. The comparisons should have a much lower consistency index than what would be produced by random entries. The RCI is the mean CI for random entries. The RCI is defined for different sizes of the matrices [49]. For our case the size of the matrix is six, which means the RCI = 1.24:
CR = CI/RCI
Saaty [54] states that an acceptable consistency ratio should be less than 0.1, yet a ratio of less than 0.2 is considered acceptable. To calculate the benefits of each strategy, we define the utility values for each strategy. The value of each utility change for each strategy. For example, the belief constraint fusion operator is suitable when opinions are totally conflicting or totally uncertain. So, the conflict and uncertainty are the most beneficial criteria. For Average belief fusion, time criterion is the most important than uncertainty. Following Table 1, for cumulative belief fusion, the value t_cum is greater than t_avg, t_const, t_wg, and t_comp.
Table 1. Assignment of criteria values for different alternative strategies.
The total of the utility for each operator (example: Average Belief Fusion) is calculated using the Equation (9) and the benefit of each option is calculated using the Equation (10):
u a v g = 1 ( t a v g + d a v g + c a v g + u a v g ) + ( r a v g + t r a v g )
B o p = [ w 1 2 w n 2 ] × [ u 1 u n ]
For each operator, the benefits will be evaluated and the operator with the maximum benefits will be selected from the candidates. After the selection of the operator, the leader agent will fuse the two opinions and produce a new opinion. This opinion will be stored in the temporal belief. Based on the received information, this belief could afterward be moved to a promoted belief or the shared belief repository.

4.2. Simulations and Results

To show the utility of the proposed belief management module, we here present an example of the application in ITSs where the environment contains vehicles, road infrastructure, roadside units, drones, etc. For simplicity, the environment will be divided in a grid-based decomposition. Each drone is assigned to a region as presented in Figure 5. In each of the region one or many roadside units will be placed. The vehicles will report traffic events to the roadside unit. The roadside units will report the information to the closest drone, which reports the event to the team leader drone. The leader drone will use HAP to combine the received information and the perceived information and generate new belief. Then, it will use a decision module to decide the repository to store the belief. Once the drone identifies many traffic events that should be monitored which exceed his capabilities, the drone will request collaboration from drones within his communication range and share sub-tasks with the team members. Once the drones accept the collaboration, they will be able to select the best subtasks according to their states and capabilities. The drone team members will leave their regions and move to the region of the team leader to maintain adherence to the collaboration commitment.
Figure 5. Simplified Architecture of belief management module application in ITS.
The team leader drone will allow the drones which are members of the group to subscribe to the three repositories and share the stored information with interested members of the group to update the collaborative plan and take in consideration the new traffic events in the team member regions.
To simulate the described application, we used Agent-Based Modeling (ABM) to model the autonomy and the proactiveness of the drone and prototype the collaboration. We used BDI agents with GAMA simulator. We modeled the environment as a grid cell, and we defined two options for the distribution of the events. The first option is to generate a random number of event sources, as shown in the Figure 6b, or a single event with a random distribution (Figure 6a). We set three drones in the simulation, one leader drone and two drone members, and one charging station (yellow cube). The team formation of the collaborative network is out of the scope of this paper. In [5], we proposed an approach to form the team network. For each drone, we define a set of characteristics such as battery capacity, range of the field of the view, speed, etc. Each drone can perceive the events in his field of view. Once an event is perceived, the drone will calculate his opinions, and send it as a FIPA message to the leader. The leader will use the HAP module to select the operator that should be used to combine the different received and perceived beliefs.
Figure 6. Simulation setting with (a) Single origin of event; (b) Multiple origin of event.
We represented the view of the state of the world from the opinions of each agent (using his individual belief). The member drones will initially only be able to identify the perceived cells. Combining the received and perceived opinions, the drone leader will revise his temporal belief repository. The leader drone will publish the combined opinions to the team members. To show the difference between each fusion operator and proposed belief management module, we represent the state of the environment (Figure 7a) and the state of the individual belief of each agent (Figure 7c–e) and the temporal belief of the leader agent (Figure 7b). The intensity of the colour in the cells represents the value of uncertainty (higher intensity indicate a higher value of uncertainty).
Figure 7. Simulated Environment and state of individual belief of each drone agent and the temporal belief of team leader agent: (a) Environment; (c,d) Individual belief of drone members; (e) Individual belief of leader drone; (b) Temporal belief of drone leader.
In the following, we track the changes of the mean uncertainty in the temporal belief of the drone leader based on the opinions received and the meta-data of the opinions (reception-time and location and the source of the information and risk) using the three operators of subjective logic (Cumulative fusion, Averaging fusion, and Weighted fusion). The cumulative fusion operator assumes that the amount of independent evidence increases and the uncertainty decreases by including more and more sources. As depicted in Figure 8, at the beginning of the simulation, the team leader operates alone without the collaboration of the member’s drone. After the diffusion of the event and the participation of the other drone. The mean value in the temporary belief base of the leader drone decreases by involving more sources. The value continues to decrease even with an important time difference between the last perceived/received state of the cell and the received opinion about the same cell.
Figure 8. (a) Variation in the mean uncertainty over time in the temporal belief of the drone leader using only cumulative fusion operator; (b) Difference between the perceived and received opinions.
The average belief fusion assumes that including more sources does not mean that more evidence is supporting the conclusion. Assume that agents A and B observe the same outcomes of the same process over the same time period, so their opinions are necessarily dependent. However, their perceptions might be different (e.g., because their cognitive capabilities are different). The average operator is suitable when no prior knowledge is given for the reliability of each source. So, the opinions of the sources are considered equally reliable. As depicted in Figure 9, the mean uncertainty in the temporal beliefs of the leader drone first decrease then increase when time difference increases.
Figure 9. Variation in the (a) Mean uncertainty over time in the temporal belief of the drone leader using only average fusion operator; (b) Difference between the perceived and received opinions.
The weighted belief operator is suitable when the opinions should not have the same importance for deriving the resulting opinion. For example, if the agent receives an event from another agent that he trusts, even if he has a conflict with his current belief, he will adopt the opinion of the second agent. In this case, trust is more important than conflict. Another example is the risk of the event. If an agent receives information about an event with high risk (example: fire in industrial facility), more weight will be assigned to this opinion, even it has a low uncertainty to mitigate undesirable consequences. The reception of an event with a high risk should guide the agent to choose a weighted belief fusion operator in which the new opinion will follow the agent with less uncertainty. In the simulation, we evaluated the risk based on the distribution of the event in the neighbor cells of each cell. The value of risk will be in the range [0, 1].
Using a cumulative fusion operator, the mean uncertainty increased due to the participation of other drone members in the collaborative awareness (Figure 8). The usage of this operator will be beneficial for events reported with small time interval. When the time interval increases, the use of the average fusion operator will be more beneficial (The uncertainty increases when the time interval increases (Figure 9). However, the average fusion operator assumes that opinions have the same importance, which is not the case if one of the reported events has an important risk value, whereas the weighted belief fusion operator considers the risk value. Thus, the uncertainty decreases when the risk increases (Figure 10). In the realized simulation, the AHP module allows the agent to alternate between operators based on the distance difference, time difference, risk difference, and trust difference. As shown in Figure 10, using the AHP module, the agent adopted the cumulative fusion operator at the beginning of the simulation. Afterward, the leader switched to the average fusion operator. Receiving event messages with a high-risk difference, the leader adopted the weighted belief fusion operator, which decreased the uncertainty in the temporal belief of the leader drone. When the risk reduced, the agent switched back to the average belief operator.
Figure 10. Variation in the: (a) Mean uncertainty over time in the temporal belief of the drone leader using only weighted fusion operator; (b) Time difference between the perceived and received opinions (c) Risk difference between the perceived and received opinions.
Using the belief management module, we allow the team leader agent to select the best operators and receive the benefit of each operator in the correct situation. For example, as presented in Figure 11. At the beginning, the leader drone used a cumulative belief fusion operator then switched to an average belief fusion operator. When the risk value increased, the drone used the weighted belief fusion operator. In the simulation, the leader will select the belief operator (strategy) that maximize the benefit referring to the Equation (8), in which he will evaluate the time difference, distance, risk, and trust.
Figure 11. Variation in the (a) Mean uncertainty over time in the temporal belief of the drone leader using AHP module; (b) Variation of selected belief fusion operator in the leader drone (0: cumulative, 1: averaging, 2: weighted); (c) Risk difference between the perceived and received opinions.
After the drone selects the strategy to adopt, the agent will combine the received belief with existing beliefs. If the belief did not exist before, the leader agent will adopt it and add it to his temporal belief. The temporal beliefs will be cleared each time interval. The new belief will move to one of the leader drone belief repositories. For the selection of the belief repository, we base it on the frequency of the reported information to decide the transition from one belief repository to the other. The transition from a belief repository to another will be modelled as more sophisticated decision-making module in a future works. We model beliefs repositories as repositories that will be used by the agent to store the combined belief; this repository will be revised after the insertion of the new beliefs to insure the consistency of the belief repository.

5. Conclusions

One of the important aspects of the collaboration is the management of group knowledge and awareness. The smart cities and ITS applications are integrating more and more data from a variety of sources or IoT devices with different uncertainty. Now IoT devices can perform more processing onboard, such as the identification of events, vehicles, hazards. These sensors and equipment have heterogonous capabilities and can generate information with different uncertainties. Combining the opinions of group members provides a collaborative awareness. In some situations, increasing the number of sources of the information leads to a lower uncertainty. However, this is not always correct, since events could be dependents. The use of average uncertainty gives an equal importance to each opinion, which is also invalid in some situations. Many approaches have been suggested to manage the uncertainties. Combining MCDM with evidence theory has also been proposed in some research works. However, most of these proposed approaches use a single fusion operator and do not focus on the selection of the operator but on the selection of the opinion. In our work, we proposed an AHP module to decide the suitable fusion operator to use as a strategy to combine uncertain opinions, and we apply this model in an autonomous collaborative drone network. The selection of a correct fusion operator will guide the drone to make better decisions about his actions. In addition, the afforded leader drone belief repositories will provide the agents in the group the opportunity to receive continuous updates about the region with different levels of reliability. The created repositories could also be provided as services to other drone networks. We aim to extend the proposed belief management module empowering the leader agent by a decision mechanism to decide the transition of the belief from a repository to another and to provide the collected data to other inter-group collaboration. In addition, we aim to improve the stability of the proposed model using theories such as OODA (Observation, Orientation, Decision, Action) ring [55] or Game-Theoretic Utility Tree, which are suitable for adversarial environments [56].

Author Contributions

Conceptualization, H.G., N.J. and A.U.-H.Y.; Methodology, H.G., N.J. and A.U.-H.Y.; Software, H.G.; Validation, N.J. and A.U.-H.Y.; Investigation, H.G., N.J. and A.U.-H.Y.; Writing—Original Draft Preparation, H.G., N.J. and A.U.-H.Y.; Writing—Review & Editing, H.G., N.J. and A.U.-H.Y.; Visualization H.G., N.J. and A.U.-H.Y.; Supervision: N.J. and A.U.-H.Y.; Project Administration: N.J. and A.U.-H.Y.; Funding Acquisition: N.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jo, T.H.; Ma, J.H.; Cha, S.H. Elderly Perception on the Internet of Things-Based Integrated Smart-Home System. Sensors 2021, 21, 1284. [Google Scholar] [CrossRef]
  2. Zhang, P.; Zhu, H.; Zhou, Y. Modeling cooperative driving strategies of automated vehicles considering trucks’ behavior. Phys. A Stat. Mech. Its Appl. 2022, 585, 126386. [Google Scholar] [CrossRef]
  3. Alsamhi, S.H.; Ma, O.; Ansari, M.S.; Almalki, F.A. Survey on Collaborative Smart Drones and Internet of Things for Improving Smartness of Smart Cities. IEEE Access 2019, 7, 128125–128152. [Google Scholar] [CrossRef]
  4. Tian, D.; Zhang, C.; Duan, X.; Wang, X. An Automatic Car Accident Detection Method Based on Cooperative Vehicle Infrastructure Systems. IEEE Access 2019, 7, 127453–127463. [Google Scholar] [CrossRef]
  5. Gharrad, H.; Jabeur, N.; Yasar, A.U.H.; Galland, S.; Mbarki, M. A five-step drone collaborative planning approach for the management of distributed spatial events and vehicle notification using multi-agent systems and firefly algorithms. Comput. Netw. 2021, 198, 108282. [Google Scholar] [CrossRef]
  6. Jøsang, A. Subjective Logic; Springer: Berlin/Heidelberg, Germany, 2009; Volume 171. [Google Scholar]
  7. Dong, S.; Ma, M.; Feng, L. A smart city simulation platform with uncertainty. Assoc. Comput. Mach. 2021, 1, 229–230. [Google Scholar] [CrossRef]
  8. Jöckel, L.; Iese, F.; Kläs, M.; Iese, F. Model-Based Engineering of Collaborative Embedded Systems: Extensions of the Spes Methodology; Springer Nature: Berlin/Heidelberg, Germany, 2021. [Google Scholar] [CrossRef]
  9. Li, Y.; Chen, J.; Feng, L. Dealing with uncertainty: A survey of theories and practices. IEEE Trans. Knowl. Data Eng. 2013, 25, 2463–2482. [Google Scholar] [CrossRef]
  10. Machot, F.A. A Robust Event Detection under Uncertainty in Video/Audio Surveillance Systems, Alpen-Adria-Universitat Klagenfurt. Faculty of Technical Sciences Austria 2013. Available online: https://www.researchgate.net/profile/Fadi-Al-Machot/publication/303544127_A_Robust_Event_Detection_under_Uncertainty_in_VideoAudio_Surveillance_Systems/links/57478ba108ae2301b0b80694/A-Robust-Event-Detection-under-Uncertainty-in-Video-Audio-Surveillance-Systems.pdf (accessed on 27 June 2022).
  11. Benferhat, S.; Leray, P.; Tabia, K. Belief Graphical Models for Uncertainty Representation and Reasoning; A Guided Tour of Artificial Intelligence Research; Springer: Cham, Switzerland, 2020; pp. 209–246. [Google Scholar] [CrossRef]
  12. Weller, A. Methods for Inference in Graphical Models. Doctor of Philosophy in the Graduate School of Arts and Sciences. Ph.D. Thesis, Columbia University, New York, NY, USA, 2014. [Google Scholar]
  13. Doherty, P.; Dunin-Kȩplicz, B.; Szałas, A. Dynamics of approximate information fusion. In International Conference on Rough Sets and Intelligent Systems Paradigms; Springer: Berlin/Heidelberg, Germany, 2007; Volume 4585, pp. 668–677. [Google Scholar] [CrossRef]
  14. Deng, X.; Jiang, W. Exploring the combination rules of D numbers from a perspective of conflict redistribution. In Proceedings of the 2017 IEEE 20th International Conference on Information Fusion, Xi’an, China, 10–13 July 2017. [Google Scholar] [CrossRef]
  15. Almond, R.G. Graphical Belief Modeling; CRC Press: Boca Raton, FL, USA, 1995. [Google Scholar]
  16. Ferson, S.; Sentz, K. Epistemic Uncertainty in Agent-Based Modeling. In Proceedings of the 7th International Workshop on Reliable Engineering Computing, Bochum, Germany, 15–17 June 2016; pp. 65–82. [Google Scholar]
  17. Kaplan, D. On the Quantification of Model Uncertainty: A Bayesian Perspective. Psychometrika 2021, 86, 215–238. [Google Scholar] [CrossRef]
  18. Sahlin, U.; Helle, I.; Perepolkin, D. This Is What We Don’ t Know’: Treating Epistemic Uncertainty in Bayesian Networks for Risk Assessment. Integr. Environ. Assess. Manag. 2021, 17, 221–232. [Google Scholar] [CrossRef]
  19. Ni, S.; Lei, Y.; Tang, Y. Improved base belief function-based conflict data fusion approach considering belief entropy in the evidence theory. Entropy 2020, 22, 801. [Google Scholar] [CrossRef]
  20. Tang, Y.; Zhou, D.; Xu, S.; He, Z. A weighted belief entropy-based uncertainty measure for multi-sensor data fusion. Sensors 2017, 17, 928. [Google Scholar] [CrossRef] [PubMed]
  21. Wang, D.; Gao, J.; Wei, D. A new belief entropy based on deng entropy. Entropy 2019, 21, 987. [Google Scholar] [CrossRef]
  22. Zhou, D.; Tang, Y.; Jiang, W. A modified belief entropy in Dempster-Shafer framework. PLoS ONE 2017, 12, e0176832. [Google Scholar] [CrossRef]
  23. Xiao, F. A Multiple-Criteria Decision-Making Method Based on D Numbers and Belief Entropy. Int. J. Fuzzy Syst 2019, 21, 1144–1153. [Google Scholar] [CrossRef]
  24. Liu, W.; Williams, M. A Framework for Multi Agent Belief Revision. Springer Studia Log. Int. J. Symb. Logic. 2016, 67, 291–312. [Google Scholar]
  25. Fagin, R.; Hapern, J.Y.; Moses, Y.; Vardi, M. Reasining about Knowledge; The MIT Press: London, UK, 1995. [Google Scholar]
  26. Andreas, H. Belief Revision. In Dynamic Tractable Reasoning; Springer: Cham, Switzerland, 2020; pp. 49–65. [Google Scholar]
  27. Liu, J.; Tang, Y. Conflict data fusion in a multi-agent system premised on the base basic probability assignment and evidence distance. Entropy 2021, 23, 820. [Google Scholar] [CrossRef]
  28. Jiang, W.; Zhan, J. A modified combination rule in generalized evidence theory. Appl. Intell. 2017, 46, 630–640. [Google Scholar] [CrossRef]
  29. Tacnet, J.-M.; Batton-Hubert, M.; Dezert, J. A two-step fusion process for multi-criteria decision applied to natural hazards in mountains. In Proceedings of the Workshop on the Theory of Belief Functions, Brest, France, 1–2 April 2010. [Google Scholar] [CrossRef]
  30. Dunin-Keplicz, B.; Verbrugge, R. Teamwork in Multi-Agent Systems; John Wiley & Sons: Hoboken, NJ, USA, 2011; Volume 22. [Google Scholar]
  31. Loke, S.W. Intelligent Information Agents; Springer: Berlin/Heidelberg, Germany, 2001; Volume 2112. [Google Scholar] [CrossRef]
  32. Singh, R.; Sonenberg, L.; Miller, T. Communication and Shared Mental Models for Teams Performing Interdependent Tasks; Springer Coordination, Organizations, Institutions, and Norms in Agent Systems XII: Singapore, 2017; Volume 10315, pp. 81–97. [Google Scholar] [CrossRef]
  33. Clark, H.H.; Marshall, C.R. Definite knowledge and mutual knowledge. In Elements of Discourse Understanding; Cambridge University Press: Cambridge, UK, 1981; pp. 10–63. [Google Scholar]
  34. Niler, A.A.; Mesmer-Magnus, J.R.; Larson, L.E.; Plummer, G.; DeChurch, L.A.; Contractor, N.S. Conditioning team cognition: A meta-analysis. Organ. Psychol. Rev. 2021, 11, 144–174. [Google Scholar] [CrossRef]
  35. Ronal, R.S. A Shared Mental Model; School of Computing and Information Systems, The University of Melbourne: Melbourne, Australia, 2018. [Google Scholar]
  36. Scheutz, M.; DeLoach, S.A.; Adams, J.A. A Framework for Developing and Using Shared Mental Models in Human-Agent Teams. J. Cogn. Eng. Decis. Mak. 2017, 11, 203–224. [Google Scholar] [CrossRef]
  37. Gervits, F.; Thurston, D.; Thielstrom, R.; Fong, T.; Pham, Q.; Scheutz, M. Toward genuine robot teammates: Improving human-robot team performance using robot shared mental models. In Proceedings of the International Joint Conference on Autonomous Agents and Multiagent Systems, London, UK, 3–7 May 2021; pp. 429–437. [Google Scholar]
  38. Sarkadi, S.; Panisson, A.R.; Bordini, R.H.; McBurney, P. Towards an Approach for Modelling Uncertain Theory of Mind in Multi-Agent Systems. In Proceedings of the 6th International Conference on Agreement Technologies, Bergen, Norway, 6–7 December 2018. [Google Scholar]
  39. Wang, Y.; Li, H.; Qian, L. Belief Propagation and Quickest Detection-Based Cooperative Spectrum Sensing in Heterogeneous and Dynamic Environments. IEEE Trans. Wirel. Commun. 2017, 16, 7446–7459. [Google Scholar] [CrossRef]
  40. Li, Y.; Wang, Y.; Guan, X. Belief propagation based multi-AUV cooperative localization in anchor-free environments. In Proceedings of the 4th Underwater Communications and Networking Conference, Lerici, Italy, 28–30 August 2018; pp. 1–5. [Google Scholar] [CrossRef]
  41. Fan, X.; Shi, Z. Loopy belief propagation algorithm in distributed wireless cooperative spectrum sensing. In Proceedings of the 2011 3rd International Conference on Communications and Mobile Computing, Qingdao, China, 18–20 April 2011; pp. 282–285. [Google Scholar] [CrossRef]
  42. Kim, H.; Choi, S.W.; Kim, S. Connectivity information-aided belief propagation for cooperative localization. IEEE Wirel. Commun. Lett. 2018, 7, 1010–1013. [Google Scholar] [CrossRef]
  43. Chen, H.; Xian-Bo, W.; Liu, J.; Wang, J.; Ye, W. Collaborative Multiple UAVs Navigation with GPS/INS/UWB Jammers Using Sigma Point Belief Propagation. IEEE Access 2020, 8, 193695–193707. [Google Scholar] [CrossRef]
  44. Jung, S.; Yang, P.; Quek, T.Q.S.; Kim, J.H. Belief Propagation based Scheduling for Energy Efficient Multi-drone Monitoring System. In Proceedings of the 2020 International Conference on Information and Communication Technology Convergence, Jeju Island, Korea, 21–23 October 2020; pp. 261–263. [Google Scholar] [CrossRef]
  45. Li, J.; Yang, G.; Cai, Q.; Niu, H.; Li, J. Cooperative navigation for UAVs in GNSS-denied area based on optimized belief propagation. Meas. J. Int. Meas. Confed. 2022, 192, 110797. [Google Scholar] [CrossRef]
  46. Zhilenkov, A.A.; Epifantsev, I.R. The Use of Convolution Artificial Neural Networks for Drones Autonomous Trajectory Planning. In Proceedings of the IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus), 29 January–1 February 2018; pp. 1044–1047. [Google Scholar]
  47. Salh, A.; Audah, L.; Alhartomi, M.A.; Kim, K.S.; Alsamhi, S.H.; Almalki, F.A.; Abdullah, Q.; Saif, A.; Algethami, H. Smart Packet Transmission Scheduling in Cognitive IoT Systems: DDQN Based Approach. IEEE Access 2022, 10, 50023–50036. [Google Scholar] [CrossRef]
  48. Salh, A.; Audah, L.; Kim, K.S.; Alsamhi, S.H.; Alhartomi, M.A.; Almalki, F.A.; Algethami, H. Refiner GAN Algorithmically Enabled Deep-RL for Guaranteed Traffic Packets in Real-Time URLLC B5G Communication Systems. IEEE Access 2022, 10, 50662–50676. [Google Scholar] [CrossRef]
  49. Triantaphyllou, E. Multi-Criteria Decision-Making Methods: A Comparative Study; Kluwer Academic Publishers: Dordrecht, The Netherlands, 2000. [Google Scholar]
  50. Altieri, M.G.; Dell’Orco, M.; Marinelli, M.; Sinesi, S. Evidence (Dempster—Shafer) Theory-Based evaluation of different Transport Modes under Uncertainty.: Theoretical basis and first findings. Transp. Res. Procedia 2017, 27, 508–515. [Google Scholar] [CrossRef]
  51. Tacnet, J.M.; Dezert, J. New Belief Function Based Methods for Multi-Criteria Decision-Making; Euro Working Group—Decision Support Systems Workshop: Paris, France, 2011; Volume 121, p. 14. [Google Scholar] [CrossRef]
  52. Fei, L.; Xia, J.; Feng, Y.; Liu, L. An ELECTRE-Based Multiple Criteria Decision-Making Method for Supplier Selection Using Dempster-Shafer Theory. IEEE Access 2019, 7, 84701–84716. [Google Scholar] [CrossRef]
  53. Ganguly, K. Integration of analytic hierarchy process and Dempster-Shafer theory for supplier performance measurement considering risk. Int. J. Product. Perform. Manag. 2014, 63, 85–102. [Google Scholar] [CrossRef]
  54. Wedley, W.C. Consistency Prediction for Incomplete AHP Matrices. Mathl. Comput. Model. 1993, 17, 151–161. [Google Scholar] [CrossRef]
  55. Lu, X.; Ma, H.; Wang, Z. Analysis of OODA Loop based on Adversarial for Complex Game Environments. March 2022. Available online: http://arxiv.org/abs/2203 (accessed on 29 June 2022).
  56. Yang, Q.; Parasuraman, R. A General Cooperative Multi-Agent Hierarchical Decision Architecture in Adversarial Environments. arXiv 2020, arXiv:2009.00288v2. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.