Next Article in Journal
Numerical Simulation of Hot Air Anti-Icing Characteristics for Intake Components of Aeronautical Engine
Previous Article in Journal
Deep-Reinforcement-Learning-Enhanced Kriging Modeling Method with Limit State Dominant Sampling for Aeroengine Structural Reliability Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Survey on Personalized Conflict Resolution Approaches in Air Traffic Control

1
Institute of Flight Guidance, German Aerospace Center (DLR), Lilienthalplatz 7, 38108 Braunschweig, Germany
2
Air Traffic Management Research Institute (ATMRI), Nanyang Technological University, Singapore 637460, Singapore
*
Author to whom correspondence should be addressed.
Aerospace 2025, 12(9), 751; https://doi.org/10.3390/aerospace12090751
Submission received: 1 July 2025 / Revised: 31 July 2025 / Accepted: 20 August 2025 / Published: 22 August 2025
(This article belongs to the Section Air Traffic and Transportation)

Abstract

The global shortage of air traffic controllers (ATCOs) has led to significant challenges. One of them is the high workload of ATCOs, often resulting in flight delays. This makes it essential to develop solutions that reduce ATCOs’ workload in order to increase capacity. One promising approach is the integration of decision-support systems. A typical task for which these systems are used for is the resolution of aircraft conflicts in the upper airspaces. A key challenge in implementing these support systems is to ensure a high acceptance and adoption rate of the proposed advisories. One potential solution to this problem is to personalize the advisories, aligning them with individual ATCOs’ preferences and conflict resolution strategies. As this approach offers many promising research directions, this literature review aims to provide a comprehensive overview of existing research in this domain and highlight potential opportunities and open challenges. Overall, 16 papers are discussed in detail to examine the diversity of conflict resolution strategies among ATCOs, the impact of personalization on the acceptance rate of advisories, the technical feasibility of implementing personalization, and the balance between personalized advisories and operational efficiency. Additionally, this paper highlights the opportunities such personalization presents, along with the unresolved challenges that should be addressed in future research.

1. Introduction

Air traffic and, correspondingly, the workload of air traffic controllers (ATCOs) is expected to increase in the future [1,2], while there is an ongoing ATCO shortage in many parts of the world [3,4]. Accordingly, solutions should be found that support ATCOs in managing current and future demands. According to Mertes et al. [5], already in 1969, the Air Traffic Control Advisory Committee (ATCAC) of the US Department of Transportation (DOT) predicted a significant growth in air traffic. This lead them to propose recommendations for future systems that emphasized higher levels of automation. Their goal was to achieve highly automated systems to be operational between 1995 and 2020 [5]. As noted by Jenney and Ratner [6], this high level of automation was intended to automate approximately 75–80% of tasks. In such a system, human operators were envisioned to work collaboratively with automation in different roles [6]. The ATCAC’s vision for increased automation aligns with the predictions of Hunt and Zellweger in 1987 [7]. They projected that by 1995, many planning and control tasks traditionally performed by ATCOs would be handled by computers. Beyond 1995, computers were expected to autonomously manage routine operations, such as generating and transmitting clearances directly to aircraft. In such automated systems, the role of ATCOs would shift to supervising the system and addressing exceptions [7].
However, this prediction did not come true and ATCOs managing the upper airspace still have a wide variety of tasks. According to Kallus et al. [8], these can be summarized as follows: (1) maintaining situation awareness, (2) developing and receiving a sector control plan/conflict avoidance, (3) making decisions on control actions, (4) solving conflicts, (5) providing tactical air traffic management (ATM), and (6) other complementary tasks. All these tasks contribute to ensuring safe operations. While most of the mentioned tasks are less time-critical and focus more on pre-planning and maintaining situation awareness, “solving conflicts” stands out as a high priority task that is also time-critical. A conflict occurs when two or more aircraft are projected to violate minimum separation standards, as shown in Figure 1. Because such conflicts demand timely and well-considered action, en-route ATCOs must continuously monitor traffic, anticipate potential conflicts, and implement, if necessary, resolutions to maintain flight safety. Therefore, conflict advisory tools play a crucial role in supporting en-route ATCOs by suggesting possible resolutions in the future. Accordingly, a notable early development during the 1980s was the application of symbolic artificial intelligence (AI) in the form of rule-based expert systems to automate conflict resolution [9]. This demonstrated the potential of AI to automate tasks of ATCOs and, therefore, potentially reduce the workload of ATCOs. This set the stage for research into more sophisticated AI technologies to support ATCOs.
With the rapid advancements in AI, particularly in the domain of deep learning during the 2010s and 2020s, its potential applications in aviation and air traffic control (ATC) have garnered significant attention, leading administrative aviation organizations to publish different AI roadmaps [10,11]. Accordingly, current research investigates the decision support and automation for tasks such as conflict resolution using AI [12]. A significant challenge researchers are facing is the interaction between AI and human ATCOs. An important aspect of this interaction is the trustworthiness of AI, which goes hand in hand with the ATCO’s acceptance of the system [11]. Previous research has shown that ATCOs are often reluctant to relinquish control, which could result in low acceptance of highly automated systems [13]. Even for low-level automation in the form of recommender systems, a low acceptance could lead ATCOs to reject advisories suggested by the system frequently, not using the full potential of these tools.
Accordingly, current research is investigating different approaches to increase trust or acceptance of these systems. A common approach in many domains, including aviation, is to increase transparency or explainability. This approach was demonstrated by Würfel et al. [14], who developed a pilot advisory system with explainability components to meet the trust requirements of AI by the EASA [11], or Ternus [15], who investigated the influence of transparency on user trust in such a pilot advisory system. Similar concepts are also implemented in the ATC domain. For example, Tyburzy et al. proposed a human–machine interaction display for ATC [16] to improve transparency and Westin et al. [17] investigated the effects of more transparent conflict resolution advisories.
Another approach aims to align system-generated advisories with human decisionmaking patterns, making them more intuitive and thus easier for air traffic ATCOs to understand and, therefore, accept [18]. This direction holds significant potential, particularly in the aviation domain and, more specifically, in ATC. As previously discussed, one of the core responsibilities of ATCOs is resolving conflicts between aircraft [8]. While Westin et al. explore the broader implications of personalization in human–automation interaction, they explicitly highlight air traffic conflict resolution as a relevant use case for personalized systems [18]. Typically, ATCOs can resolve conflicts using lateral, vertical, or speed-based maneuvers. Lateral resolutions involve assigning new headings or directing aircraft to alternative waypoints, while vertical resolutions are achieved by assigning different flight levels. Additionally, controllers may resolve conflicts by adjusting aircraft speed. To address a conflict, an ATCO may issue one or multiple instructions to one or several aircraft. This makes it a complex task with many different strategies that ATCOs can follow to find a suitable solution. Accordingly, it is an ideal problem for implementig and testing different approaches to personalized advisories and testing the effects of these approaches.
Therefore, this literature review explores current works that contribute to this concept in the domain of air traffic conflict resolution and provides an overview of the opportunities and open challenges of this research direction.
First, papers that aim to identify different strategies that ATCOs use to resolve conflicts are investigated. After that, the technologies available to model these strategies and develop personalized systems capable of generating advisories tailored to either individual users or groups are examined. Then, the paper provides an overview of approaches that aim to combine and balance the benefits of personalized advisories with advisories that strive for efficiency. Finally, papers that examine the influence of personalization on the acceptance rate of conflict resolution advisories are reviewed. Accordingly, the research questions to be answered by this literature review are as follows:
  • RQ 1: What strategies do ATCOs use to resolve conflicts and how can these strategies be identified?
  • RQ 2: What methods can be used to train a personalized model?
  • RQ 3: What approaches can be used to balance personalized advisories with advisories that strive for efficiency?
  • RQ 4: Does the personalization of conflict resolution advisories (group-based or individual) improve their acceptance rates among ATCOs?
The remainder of this paper is structured as follows: First, in Section 2, the effects and benefits of personalized systems are discussed. Then, in Section 3 the method of the conducted literature review is described. After that, current publication trends are analyzed in Section 4. This is followed by the core of the paper, presenting the results from the conducted review in Section 5. After that, the opportunities resulting from the current research are explained in Section 6 and the challenges in Section 7. Finally, the results of the conducted review are concluded in Section 8 and an outlook is provided in Section 9.

2. Personalization in Artificial Intelligence

The use of personalization has proven to be an effective approach for increasing user acceptance and system efficiency across a range of domains. In web-based recommender systems, for instance, the goal is to tailor content to the needs and preferences of individual users. These systems typically analyze user behavior to detect patterns in interests and preferences. Methods such as clustering are used to uncover underlying structures in the data, enabling the generation of targeted and relevant suggestions [19]. Personalized recommendations have been shown to enhance user engagement and satisfaction [20].
Techniques such as behavioral cloning are also applied in more complex and interactive domains, such as autonomous driving. In these cases, systems are trained on demonstrations from human drivers to learn behaviors like steering or lane following [21,22]. While the primary goal in this context is often functional performance rather than personalization, the system may nonetheless inherit individual driving styles as a side effect. This form of imitation learning typically requires relatively little training data to achieve reasonable performance [21]. However, the resulting personalization can also be a limitation: The human driver may not always demonstrate ideal behavior or cover rare but critical scenarios, such as how to recover from mistakes [21].
In other scenarios, system behavior is shaped through human feedback: Models are guided to act in ways that align with human expectations and are discouraged when deviating from them. This helps prevent undesired or unpredictable behavior and supports applications in control tasks. It also extends to domains like recommender systems and generative models, where systems are adapted to better reflect the preferences and goals of individual users [23].
The observed benefits in other domains using personalized systems suggest that similar approaches could be valuable in ATC. Given the central role of conflict resolution in the work of ATCOs, it is promising to investigate how the personalization of system advisories could increase the acceptance of proposed solutions. Accordingly, this paper reviews and analyzes existing research that explores personalization in conflict resolution to support ATCOs and discusses the potential benefits, challenges, and implications of such approaches.

3. Method

The core of this work is to analyze the current literature in relation to personalized conflict resolution approaches and to analyze the effects of these approaches on the acceptance rates of the proposed solutions. For this reason, a structured literature review was conducted.

3.1. Search Strategy

Google Scholar was selected as the primary database for this literature review due to its broad coverage of the academic literature. Since Google Scholar provides access to a wide range of sources, it requires a thorough filtering process to ensure relevance and quality. Therefore, a combination of relevant search terms was used with a combination of ’AND’ and ’OR’ operators resulting in the following search query: allintitle: (“air traffic” OR controllers OR ATC OR ATCO) (conflict OR conformal OR personalized OR Supervised OR data-driven OR “instructional pattern” OR strategy OR maneuver). In order to gather the most up-to-date advancements in the field, the search was limited to works published between 2015 and 2024. This time frame was chosen to capture recent developments as well as valuable contributions from prior years. On 6 December 2024, this search resulted in a total of 439 papers found.

3.2. Exclusion Criteria

A paper was excluded from this review if it met any of the following conditions:
  • The paper is a literature review. Review papers usually summarize already conducted research to provide researchers an overview about a specific topic, but the goal of this study is to focuses on original research.
  • The study addresses conflict resolution approaches that did not account for the preferences or actual behavior of ATCOs.
  • A full-text of the paper is not accessible.
  • The work is a thesis or dissertation, as these are typically not peer-reviewed or reference already published research, in the case of a cumulative dissertation.

3.3. Paper Selection and Analysis Process

The paper selection process followed a step-by-step approach, whereby the initial dataset of 439 papers was systematically refined, according to the defined exclusion criteria, as illustrated in Figure 2. A first round of filtering was conducted by screening the titles of the papers for relevance, reducing the selection to 71 papers. Following this, the selection underwent a cleaning process. This step involved removing duplicates, theses, inaccessible papers, literature reviews and older studies incorrectly identified by Google Scholar within the specified time frame. Next, the abstracts of the remaining 56 papers were screened, further narrowing the selection to 38 papers. Finally, a full-text review of these papers was conducted to evaluate their relevance, leaving a total of 16 papers that were included in this literature review.

4. Publication Trends

There has been consistent publication and growing interest in personalized conflict resolution approaches in ATC, as can be seen in Figure 3. An increase in publications can be observed, particularly in the past five years, reflecting the expanding relevance of the topic. This trend may be driven in part by the rise in automation research within the ATC domain. As automation becomes more integrated into aviation systems, the need for human-centric approaches has gained greater significance, as EASA has emphasized that automation in aviation should prioritize human-centric design to ensure safety and efficiency [11,24].

5. Thematic Analysis of Human-Centered Conflict Resolution

To further explore this landscape of research, the selected 16 papers were grouped into thematic categories based on their primary focus and contributions. An overview of these categories can be seen in Figure 4 and an overview of the papers in Table 1. This classification aims to highlight the primary focus and key contributions of each study. Given the interdisciplinary nature of the field, some papers were relevant to multiple categories and were, therefore, included in more than one grouping. The topics identified are “Finding Strategies to Resolve Conflicts”, “Implementing Personalized Conflict Resolution”, “Balancing Different Conflict Resolution Strategies”, and, lastly, “Influence of Personalization on the Acceptance Rates of Advisories”.

5.1. Finding Strategies to Resolve Conflicts

Recent research provides insights into the decision-making strategies of ATCOs, revealing individual differences, inconsistencies, and preferences in conflict resolution methods. To implement personalized conflict resolution advisories, these findings have to be taken into account. Zakaria et al. [31] classified ATCO strategies into three categories: altitude changes, heading changes, and a mixed approach. Since speed changes are rarely used to resolve conflicts, they are not considered as a viable strategy [31]. While some studies, such as Westin et al. [25], found that heading changes were the most frequently used actions to resolve conflicts, others, such as Palma et al. [29] and Zakaria et al. [31], reported a preference for altitude changes, with heading adjustments and speed changes less favored. Trapsilawati et al. [30] conducted a study with undergraduate students trained for only three hours in ATC procedures. Their findings showed a strong preference for altitude changes (72%), followed by speed changes (19%) and heading changes (9%), diverging from professional ATCOs’ behaviors. This demonstrates that conducting experiments with students in this context is often not meaningful, given that their approaches differ greatly from that of professional ATCOs. Furthermore, the findings of Palma et al. [29] suggest that strategy selection is often driven by three main considerations: the characteristics of the conflict, a desire to improve customer service (e.g., providing more direct routes), and sector-specific constraints.
Beyond strategy selection, research highlights differences in how consistently ATCOs apply their preferred approaches. Guleria et al. analyzed decision-making patterns across multiple ATCOs, initially studying two individuals [27] and later expanding their research to eight [28]. Their results showed that ATCOs differ not only in their strategic preferences, but also in their adherence to them. Some ATCOs demonstrated high consistency in their maneuver choices, while others showed greater flexibility. The follow-up study [28] suggested that experience may play a role in consistency, with more experienced ATCOs showing greater consistency with a particular strategy. This has implications for incorporating ATCO behavior into automation models, as inconsistencies could affect the reliability of personalized recommendations. The challenge of inconsistency is a recurring theme across multiple studies. Westin et al. [25] also found that more experienced ATCOs seem to be more consistent. To address this problem, Regtuit et al. [26] proposed an approach to distinguishing between a ATCO’s primary strategy and occasional deviations. Their study introduced an approach using k-means clustering with average silhouette width to differentiate between strategic preferences and inconsistencies. By applying this method to a dataset where an ATCO was instructed to behave inconsistently across three predefined strategies, the algorithm successfully identified the strategies used, demonstrating the potential for advanced analytical tools in understanding decision-making [26].
These studies highlight the complexity of ATCO decision-making in conflict resolution, revealing that ATCOs adopt a variety of strategies that can differ not only between different ATCOs but also within the same ATCO over time. These strategies are influenced by individual decision-making tendencies, experience levels, and contextual factors such as regional conventions, rules, and airspace structures. The research presents mixed results regarding the most common conflict resolution strategies. Some studies emphasize heading changes, while others highlight altitude adjustments. This variability in strategy choice reflects the potential influence of regional differences, but it is also compounded by the inconsistency with which ATCOs may adhere to particular strategies. In addressing RQ 1—What strategies do ATCOs use to resolve conflicts and how can these strategies be identified?—the findings suggest that ATCOs’ conflict resolution strategies are diverse and dynamic, shaped by both personal and contextual factors. To identify these strategies, most studies rely on recording ATCOs’ actions during conflict resolution and analyzing various metrics to describe the strategies employed. Clustering techniques can also be utilized to uncover the underlying strategies, helping to reduce random variations and providing a clearer picture of the primary strategies used by individual ATCOs. This approach not only helps in identifying the main strategies but also in assessing the consistency with which an ATCO adheres to a particular strategy.

5.2. Implementing Personalized Conflict Resolution

Several studies have explored the use of machine learning to create systems that are able to provide personalized conflict resolution advisories. Typically, these works either use reinforcement learning (RL) or supervised learning (SL). RL is a popular solution to develop conflict resolution advisory systems [12]. The process to resolve a conflict can be described by a Markov Decision Process (MDP). A MDP is a mathematical framework used for modeling decision-making in a specific environment. As described by Naeem et al. [40], it can be defined as a tuple: M = ( S , A , P , R ) where S is a set of states, A is a set of actions, P : S × A × S [ 0 , 1 ] describes a transition probability, and R : S × A R describes a reward function. Problems that can be described by an MDP can typically be tackled with RL. The goal is to find an optimal policy—a strategy for choosing actions—that maximizes cumulative rewards over time. Accordingly, by solving an MDP, it is possible to determine the best sequence of actions to take in order to achieve success in a given environment [40]. This process is visualized in Figure 5.
When implementing an RL approach, one of the primary challenges is to define a reward function that leads the system to learn the desired behavior. As described in Section 5.1, Regtuit et al. [26] used clustering algorithms to find and isolate conflict resolution strategies of ATCOs. Based on these strategies, a reward function is constructed that compares the agent’s actions to an estimated trace based on the strategy used by the ATCO. Accordingly, an important factor that influences the received reward is the alignment of the agent’s actions with the strategy of the ATCO. Regtuit et al. demonstrated the effectiveness of their approach using the Q-learning algorithm, a fundamental RL method that enables an agent to learn optimal decision-making through trial and error by interacting with its environment and updating its strategy based on received rewards. Q-learning assigns and updates values to state-action pairs, allowing the agent to learn which actions yield the highest expected future rewards over time [42]. Similarly, Tran et al. [32,33] applied RL using the Deep Deterministic Policy Gradient (DDPG), a more advanced method compared to Q-learning. While Q-learning discretely maps state-action values to learn optimal policies, the DDPG is an actor–critic approach designed for continuous action spaces. In this approach, the reward function is structured to reward resolved conflicts and penalize violations of minimum separation standards. Additionally, since the DDPG is an actor–critic method, it uses two neural networks: the actor network, which determines the agent’s actions, and the critic network (value network), which evaluates how well those actions align with optimal decision-making [43]. In this case, the value network assesses the similarity between the agent’s predictions and actual ATCO behavior, ensuring the trained agent aligns with real-world operational strategies [32,33].
Compared to RL, a more straightforward way to develop personalized systems is SL, which trains models on examples. Consequently, several recent studies have explored SL approaches to train personalized models for conflict resolution. Van Rooijen et al. [34] investigated the feasibility of personalized automation using convolutional neural networks (CNNs) on data from 12 novice ATCOs. The study trained both personalized and general models, finding that personalized models outperformed general models in predicting ATCO decision-making, confirming that ATCOs follow distinct strategies. Moreover, a correlation was observed between ATCO consistency and model performance, suggesting that more consistent ATCOs are better suited for personalized automation. This underscores the potential for personalized machine learning models in ATC, provided that ATCOs maintain a certain level of decision consistency. A similar pattern emerged in studies by Guleria et al. [27,28], where random forests, XGBoost, and support vector machines (SVMs) were used in a multi-step prediction pipeline. Their approach sequentially predicted conflict resolution factors such as the aircraft chosen for maneuvering, the maneuver initiation time, heading angles, etc. Initially, the framework was tested on data from two ATCOs, demonstrating that the models could successfully replicate individual ATCOs’ decision patterns. The follow-up study expanded the dataset to eight experienced ATCOs, further confirming that their approach is suitable to create personalized models. Bastas et al. [35] approached the problem using variational autoencoders (VAEs) to model when and how ATCOs resolve conflicts. In their initial study, the authors trained a VAE, finding that it outperformed a basic encoder-based model. The study primarily focused on predicting ATCO reaction timing, rather than specific resolution maneuvers. In their follow-up research [36], random forests, gradient tree boosting, and neural networks were used to predict how ATCOs resolve conflicts. While the models for classification tasks performed well overall, the study noted difficulties in accurately predicting values for the vectoring, likely due to a small number of training samples.
A key limitation across multiple studies is the exclusion of flight level changes as a conflict resolution strategy, as seen in both Bastas et al. [35,36] and Guleria et al. [27,28]. This contradicts findings from [29] and [31] that for many ATCOs, flight-level adjustments are the preferred solution to resolve conflicts.
Accordingly, RQ 2—What methods can be used to train a personalized model?—can be answered as follows: Typically, a personalized model can be trained using different approaches. SL is as a common one, where data are collected from ATCOs and used to train models based on these examples. Another approach involves RL, in which ATCO data can either help determine strategies that form the basis of a reward function, guiding the training process, or be used to compare the actions of an AI agent with the actual behavior of ATCOs. This comparison allows the model’s performance to be assessed and improved over time. Both SL and RL provide effective methods for training personalized models, with each offering unique benefits.

5.3. Balancing Different Conflict Resolution Strategies

Recent studies have explored approaches to balance the preferences of ATCOs with optimal solutions. Westin et al. [17] discussed a hybrid learning approach to balancing personalized and optimal models. They proposed combining SL (to achieve conformance) with RL to explore and identify optimal solutions. While the study did not implement this system, it emphasized that a combined framework could allow AI to replicate ATCO preferences while improving upon them where necessary, maintaining a balance between high acceptance rates and efficiency.
Another study, based on the work of Guleria et al. [28], investigated how ATCOs respond to different conflict resolution advisories [37]. Using an SL model, three types of resolution advisories were generated: (1) an conformal advisory maximizing alignment with the ATCO’s typical behavior, (2) an optimal advisory that minimizes travel time for greater efficiency, and (3) a balanced advisory, which is a trade-off between conformance and optimality. To find a balanced advisory, the paper proposes an approach that assigns an optimality and a conformance score to multiple generated advisories that lie between the conformal and optimal solution. Based on a balancing factor, the scores are weighted and summed up. The advisory with the highest overall score is selected as the balanced advisory. Results showed that the conformal advisory was the most favored, closely followed by the balanced advisory, while the purely optimal advisory was less accepted. This suggests that ATCOs prefer solutions that align with their existing strategies. Both approaches would be a valid answer to RQ 3—What approaches can be used to balance personalized advisories with advisories that strive for efficiency?

5.4. Influence of Personalization on the Acceptance Rates of Advisories

Acceptance of AI is fundamental to ensure its adoption and effective use. The acceptance of AI is shaped by multiple factors including technological readiness, perceived usefulness and risk, ease of use, and impacts on organizational and societal levels. A study by Rane et al. [44] highlights that the technological attributes of AI, such as performance and reliability, alongside ease of integration into existing systems, critically influence user acceptance. Trust, transparency, and fairness are also key components, while ethical concerns, privacy issues, and potential job displacement present notable challenges to acceptance [44]. In the context of conflict resolution advisories, acceptance is often measured by the ratio of advisories accepted to those declined by the ATCO. Previous studies have shown that the alignment between conflict resolution advisories and ATCOs’ own resolution strategies influences the acceptance rate of such advisories. Hilburn et al. [45] and Westin et al. [46] indicate that ATCOs are more likely to accept advisories that mirror their own prior solutions, especially in highly complex scenarios. Personalized advisories not only had higher acceptance rates and agreement ratings but also resulted in faster response times compared to non-conformal advisories, suggesting that the alignment of AI with familiar strategies can foster trust and reduce cognitive load in critical settings [46].
Westin et al. [38] examined the role of personalization in AI-generated conflict resolution advisories. In their experiment, resolution advisories were either conformal, aligning with the ATCO’s own conflict resolution style, or non-conformal, based on the decision patterns of a different ATCO. Despite not being explicitly informed about the conformance manipulation, ATCOs accepted 77.8% of conformal advisories compared to 66.7% of non-conformal ones, highlighting a preference for suggestions that matched their own behavior; however, the results were not statistically significant. The study also explored the role of transparency, finding that while richer visual interfaces helped ATCOs better understand conflict resolution options, transparency did not significantly influence the acceptance rates of advisories. Building on this, Westin et al. [17] explored the relationship between transparency, conformance, and optimality within the SEASAR project MAHALO (https://mahalo01.lr.tudelft.nl/, accessed on 25 February 2025). They found that personalization appeared to increase the acceptance and transparency to nudge ATCOs toward more optimal solutions while maintaining trust. However, the study also found that combining transparency with conformal solutions can have unintended negative effects. Because a strategic conformal advisory can be a sub-optimal solution, transparency can highlight its inefficiency, leading ATCOs to reject it despite its familiarity [17]. This suggests that transparent AI should not merely emphasize optimality but also account for user preferences and biases in decision-making.
In another paper, by Westin et al. [39], the behavior of 34 ATCOs was examined in different conflict scenarios. They used different approaches to generate individually personalized, group-based, and optimal solutions. An individually personalized model is trained solely on the recorded data of one specific person meant to use that specific model, while the group-based model uses data recorded from all participants to train the model. The results indicated that the effects of individual personalization and group-based models were not consistent across all scenarios, potentially due to small sample sizes or inherent ATCO inconsistencies in conflict resolution. The study identified that ATCOs with stable, predictable strategies were better suited for personalized models, while variability in decision-making negatively affected the effectiveness of both group-based and individually personalized models. Moreover, the study found that ATCOs’ preferences for separation distances played a key role in their acceptance rate of AI-generated advisories. Moreover, the study found that ATCOs’ acceptance rate of optimal advisories also depends on how closely the optimal advisory’s closest point of approach (CPA) matches their personal preferences. Those whose preferences were further from the optimal CPA showed lower acceptance rates and made more adjustments with increased transparency [39]. In a study by Guleria et al. [28], two ATCOs, ATCO A and ATCO B, were compared. These two ATCOs exhibited different levels of consistency in their conflict resolution strategies. ATCO A demonstrated highly consistent decision-making and rarely deviated from their preferred resolution strategies, accepting another ATCO’s solution only when it was visually identical to their own preferred approach. In contrast, ATCO B, who had less radar operations experience, was more open to alternative solutions and occasionally adopted ATCO A’s strategies when they were perceived as more efficient. This suggests that experienced ATCOs may place greater importance on conformance, whereas less experienced ATCOs may be more adaptable to new strategies. According to Guleria et al. [28], these findings raise an important research question: Does conformance become more relevant only for experienced ATCOs who have already established consistent decision-making strategies? If so, this could have implications on how personalized advisory systems should be tailored based on ATCO experience levels [28]. Since Guleria et al. [28] tested this only with two ATCOs, it would be interesting to analyze this on a larger scale.
Overall, the literature suggests that personalization, whether at a group or individual level, plays an important role in improving acceptance rates of conflict resolution advisories. This answers RQ 4: Does the personalization of conflict resolution advisories (group-based or individual) improve their acceptance rates among ATCOs? Individually personalized models tend to demonstrate the highest acceptance rates, but they require substantial data from each ATCO, which presents challenges for the actual implementation. Furthermore, their effectiveness is contingent upon the consistency of the ATCO’s strategies, which may vary. Group-based models, while less demanding in terms of individual data collection, pose the risk of incorporating conflicting strategies within the group, potentially reducing conformance and, consequently, acceptance rates.

6. Opportunities

There are multiple opportunities that arise from personalized conflict resolution advisories and the concepts used to implement these in general. These opportunities will be further examined in the following sections.

6.1. Workload Reduction

Current estimates show that automation and decision support systems for ATC could reduce the workload of ATCOs by more than 40% [47]. However, systems that require ATCOs to actively decide whether to accept or reject a suggestion may introduce additional cognitive load. This is especially true if the majority or all of the system’s suggestions are rejected by the ATCO, as the ATCO must assess each recommended conflict resolution and, in the case of rejection, resolve the conflict manually. To mitigate this issue, achieving a high acceptance rate of advisories is crucial. As described in Section 5.4, conformal advisory systems can enhance this acceptance rate [45,46]. By aligning system recommendations more closely with the ATCO’s strategies, these systems could not only improve acceptance rates but also reduce the number of manual interventions required. Furthermore, conformal advisories could make it easier for ATCOs to comprehend and evaluate system-generated suggestions, reducing the overall time to resolve a conflict [46].

6.2. Increasing Situation Awareness with Generative AI

One of the key aspects to effectively resolve conflicts is maintaining a high level of situation awareness. Current advisory systems support ATCOs by providing conflict alerts and resolution suggestions, but they often lack a deep understanding of the controller’s specific strategy and intent. Generative AI could bridge this gap by dynamically adapting to the ATCO’s mental model, thereby enhancing shared situation awareness. A key approach to achieving this is through interactive learning between the AI system and the user. By leveraging Large Language Models (LLMs), the AI could engage in a dialogue with ATCOs, gradually refining its understanding of their preferred conflict resolution strategies. This exchange of mental models allows the system to align its advisories more closely with the ATCO’s objectives in a given situation. Instead of merely suggesting generic conflict resolutions, the AI can anticipate the ATCO’s decision-making style and offer context-aware recommendations that complement their strategy. This shared mental model between AI and ATCOs could foster enhanced decision support.

6.3. Co-Evolution of ATCO-AI Systems

Previous research has shown that humans can adapt to the behavior of AI systems, especially when these systems make better decisions compared to the human [48]. Similarly, users can become accustomed to the linguistic or procedural patterns of AI systems, such as the language use of LLMs [49]. This opens up a bidirectional learning opportunity: While AI systems can learn from human behavior to generate personalized advisories, humans can also improve their decision-making through repeated interaction with the system. This creates the potential to dynamically balance system-generated suggestions that are tailored to human preferences with optimal conflict resolutions. As mentioned in Section 5.3, Guleria et al. [37] provide the foundation for this by optimizing personalized solutions, trying to find a balance between optimized and personalized advisories. Such a strategy could maximize the acceptance rates of the generated advisories while simultaneously fostering continuous improvement in human decision-making. Through this interplay, higher efficiency in ATC conflict resolution could be achieved in the future.
Dynamically balancing the level of personalization and optimization would also addresses the paradox highlighted by Westin et al. [18], who point out that operators may only develop trust in a system after using it, yet they may be reluctant to use a system they do not already trust. This makes an initially high personalization a crucial factor in the successful deployment of these systems. This is underscored by effects observed in other domains, where user satisfaction tends to increase linearly with the amount of time the system is used [50]. This means that the ATCOs might accept the system’s suggestions more frequently the longer they use the system. Accordingly, dynamically changing the level of personalization might be promising, as in early states, the system could be highly personalized to increase acceptance and and system usage, while slowly moving towards more optimized solutions, increasing the efficiency of the overall system while allowing the ATCO to slowly adapt to the more optimized advisories [18].

7. Open Challenges

There are various challenges involved in implementing personalized conflict resolution systems, with three stakeholder groups being particularly affected. These are the challenges faced by developers, those encountered by certifying bodies, and those experienced by ATCOs.

7.1. Challenges for Developers

Developers face the challenge of translating current research into practical, real-world applications. At the same time, they must remain mindful of the responsibility and potential impact that personalized systems can have on ATCOs. These considerations result in the challenges outlined in the following chapters.
Most existing approaches rely on artificially created, isolated scenarios. However, implementing such a system in a real-world environment would require a thorough analysis of ATCOs’ strategies under actual operational conditions and is still an open challenge. One potential method for developing a personalized conflict resolution system using previously recorded traffic data is through a group-based approach. This would first necessitate identifying the strategies employed by ATCOs in real environments, which could be achieved using methods similar to those outlined in Section 5.1. Techniques such as clustering could be applied to datasets like the one created by Gaume et al. [51] to distinguish between different approaches taken by ATCOs and allow the identification of underlying strategies. Once these strategies are established, the behavior of an ATCO interacting with the system could be classified accordingly. This classification would allow for tailoring the systems advisories to align with the preferences and strategies of the specific ATCO.
Another key challenge for developers is finding the right balance between personalized advisories and optimized advisories. Relying solely on personalized advisories may limit their efficiency to the individual performance of the ATCO, whereas optimized advisories may not align with the ATCO’s expectations or decision-making style and result in low acceptance rates. Future research should investigate how to dynamically integrate both approaches in a way that enhances overall efficiency without compromising acceptance rates. Guleria et al. [37] highlight the potential of combining optimized advisories and personalization; however, identifying the optimal trade-off between system performance and the personalization of advisories remains an open challenge.
Another challenge, often overlooked by existing works, is that conflicts are dynamic situations that evolve over time. As a result, any solution, personalized or optimal, should also adapt over time. However, current methods looking into personalized conflict resolution typically provide static solutions.
Furthermore, because ATC is a safety-critical domain, systems must be thoroughly validated and there must be some sort of redundancy or control system to ensure that any proposed action can be safely executed. The validation of AI, especially deep learning, is still an ongoing research topic [52]. This is especially a problem if the system should continue to learn to further adapt to the preferences of an ATCO, because this would require monitoring the system continuously to ensure that no undesired behavior is learned. Accordingly, a continuous verification and validation process would be necessary. In addition, ensuring the safety of the predictions made by deep learning-based systems is also an important research direction. There are many approaches that could be applied and tested for personalized conflict resolution [53].
Moreover, personalization requires substantial amounts of data from each ATCO using the system. Collecting these data can be both time-consuming and challenging. For example, Guleria et al. [28] recorded over 570 conflict scenarios from two different ATCOs to train two personalized models. However, these models were limited to predicting heading changes. Training more complex models that also allow flight level or speed changes would require even larger datasets. Consequently, personalizing the conflict resolution advisories for each individual ATCO would require enormous effort. Therefore, developers need to explore ways to reduce the amount of data needed. One approach could be to train a model on a group of ATCOs, similar to the group-based models proposed by Westin et al. [39], and then fine-tune the model for individual controllers. What makes this even more difficult is that the developers need to be aware that ATCOs not always work consistently, as discussed in Section 5.1. It is, therefore, a major challenge not only to record sufficient data, but also to filter and decide which of the recorded data is meaningful and which should perhaps be excluded based on the consistency of the ATCOs.
Lastly, the developers have to consider that a decision support system, especially one that is designed to generate high acceptance rates, could lead to issues like over-trust [54]. Accordingly, developers have to ensure that their systems are not too “persuasive”, leading to bad or even dangerous decisions. This is a problem that has not been widely considered for decision support systems for ATC tasks yet.

7.2. Challenges for ATCOs

Several human factors challenges emerge when ATCOs collaborate with AI systems, which extend even further to personalized conflict resolution systems.
First of all, there is a need to ensure that ATCOs do not extensively rely on decision support systems in the future. Previous research has shown that people tend to rely too much on recommendations by AI in different situations [54,55], which can lead to knowledge complacency [56]. Accordingly, ATCOs need to learn about the limitations of AI-based recommendation systems and be aware of the risks in using them. They need to critically question advisories and learn to calibrate their trust to a sufficient level and avoid complacency. Therefore, they must be capable of detecting AI errors and verifying the system’s outputs, which depends on an adequate understanding of the system’s reasoning and constraints [57]. Furthermore, it has been found, in the past, that pilots’ cognitive skills can be negatively impacted when relying on a high degree of automation in the cockpit [58]. This problem might also apply to ATCOs when using highly automated systems. To prevent this, ATCOs would potentially need to regularly take part in additional trainings to keep their skill level and stay prepared for situations where they need to act without system support. Alternatively, the systems would have to be designed in a way to avoid this problem in the first place. Generally, achieving shared situation awareness and building shared mental models remains a central challenge in the collaboration of humans with autonomous systems [59], like the proposed personalized conflict resolution systems.

7.3. Certification and Adoption Challenges

There are also open challenges for the certification and adoption of personalized conflict resolution systems. These concern, e.g., training procedures and the certification of systems that may change dynamically depending on the user.
To enable the certification and safe adoption of personalized conflict resolution systems in ATC, a clearly defined and consistently applied concept for human–AI collaboration is essential. This includes specifying how roles are distributed and how communication takes place with such systems. As mentioned in the previous section, effective collaboration depends on both human and AI sharing a common mental model of the situation. This can be supported by such systems that transparently communicate their own intentions, goals, and reasoning processes, as proposed by Chen et al. [59]. Hereby, the alignment of the mental models may be further enhanced through personalization, allowing the AI to adapt its behavior to the specific strategies of operators.
Nevertheless, as noted by Westin et al. [25], balancing personalization with standardization remains a critical challenge. To certify a system, it is necessary that this system follows certain regulations and standards. Introduction personalization would make it more difficult to ensure the safety of the system, making the process challenging for certification agencies. Accordingly, further research is needed to explore how personalization can be implemented without compromising operational safety and procedural uniformity.
Another challenge is that new training procedures would be needed to train first-time users and to avoid problems like over-trust and knowledge complacency. This is a challenge on the administrative side in that these new training procedures would need to be developed, and on the operational side in that time would be needed for ATCOs to learn and adapt to these systems.

8. Conclusions

For this survey, papers published between 2015 and 2024 were analyzed, focusing on personalized conflict resolution approaches in air traffic control (ATC). The findings indicate that tailoring conflict resolution advisories to align with the preferences and strategies of air traffic controllers (ATCOs) tends to increase the acceptance rate of these advisories. Furthermore, this review shows that ATCOs employ a wide range of strategies and differ in the consistency with which they apply them. In addition, approaches that can be used to implement personalized conflict resolution, typically, supervised learning (SL) or reinforcement learning (RL), have been discussed within the paper. Additionally, the methods that exist to combine and balance the advantages of personalized and optimality-seeking conflict resolution advisories were examined. Overall, the results of this paper suggest that both individual and group-based personalization in conflict resolution advisory systems represent a promising research direction. Such systems have significant potential to enhance ATCOs’ acceptance of decision support tools, which could contribute to reducing their workload in the future. As can be seen in Section 6 and Section 7, there are many opportunities and open challenges which highlight the potential of future research in this direction. These opportunities include reduced ATCO workload through higher advisory acceptance rates, improved decision-making as both artificial intelligence (AI) based systems and humans learn from each other over time, and enhanced situation awareness through generative AI. Nevertheless, implementing personalized conflict resolution systems also involves challenges for developers, certifying bodies, and ATCOs, including the need to model and adapt to real-time changes in controller behavior, ensure the safety and reliability of AI-driven decisions, navigate complex certification requirements, and design training procedures that prevent over-reliance on automation. Addressing these challenges opens up numerous promising avenues for future research.

9. Outlook

Current research efforts are focused on automating various ATC tasks to develop digital assistants that can serve ATCOs as collaborative team partners [60]. The concepts discussed in this literature review could be extended to other functionalities of digital assistants, enabling more personalized support across a wider range of ATC tasks and responsibilities. Accordingly, an important direction for further research is to examine how different tools impact ATCO workload, identifying which tools benefit from personalization and which are better suited to be standardized. This would allow for a dynamic adjustment of personalization levels, ultimately leading to more efficient human–machine collaboration.
Additionally, the survey highlights that personalization can support users even in safety-critical domains such as ATC. Personalized systems have been shown to increase the acceptance of system advisories, as discussed in Section 5.4, which is a significant advantage. This suggests that further exploring such approaches in similarly high-stakes environments could help address current challenges more effectively and support the adaption of these techniques.

Author Contributions

Conceptualization, J.R. and Y.G.; methodology, J.R. and S.T.; formal analysis, J.R.; investigation, J.R.; resources, J.R. and Y.G.; writing—original draft preparation, J.R. and S.T.; writing—review and editing, S.T. and Y.G.; visualization, J.R.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Eurocontrol. Eurocontrol Forecast Update 2024–2030. 2024. Available online: https://www.eurocontrol.int/publication/eurocontrol-forecast-2024-2030 (accessed on 24 January 2025).
  2. Eurocontrol. Network Manager Annual Report 2023. 2024. Available online: https://www.eurocontrol.int/publication/network-manager-annual-report-2023 (accessed on 22 January 2025).
  3. Eurocontrol. Eurocontrol Think Paper #19—ATC Mobility and Capacity Shortfalls. 2022. Available online: https://www.eurocontrol.int/sites/default/files/2022-12/eurocontro-think-paper-19-atc-mobility-capacity.pdf (accessed on 22 January 2025).
  4. Jay, T. US Air Traffic Controller Shortage Stalls Aviation and Air Cargo Growth—globaltrademag.com. 2024. Available online: https://www.globaltrademag.com/us-air-traffic-controller-shortage-stalls-aviation-and-air-cargo-growth/ (accessed on 22 January 2025).
  5. Mertes, F.; Jenney, L. Automation Applications in an Advanced Air Traffic Management System: Volume 1. Summary; Technical Report; United States Department of Transportation. Office of the Secretary: Washington, DC, USA, 1974. [Google Scholar]
  6. Jenney, L.; Ratner, R. Man as Manager of Automated Resources in an Advanced Air Traffic System. J. Aircr. 1975, 12, 936–942. [Google Scholar] [CrossRef]
  7. Hunt, V.R.; Zellweger, A. The FAA’s advanced automation system: Strategies for future air traffic control systems. Computer 1987, 20, 19–32. [Google Scholar] [CrossRef]
  8. Kallus, K.W.; Van Damme, D.; Dittmann, A. Integrated Task and Job Analysis of Air Traffic Controllers-Phase 2-Task Analysis of Enroute Controllers; European Organisation for the Safety of Air Navigation: Brussels, Belgium, 1999. [Google Scholar]
  9. Spencer, D. Applying artificial intellignece techniques to air traffic control automation. Linc. Lab. J. 1989, 2, 537–554. [Google Scholar]
  10. FAA. Roadmap for Artificial Intelligence Safety Assurance. 2024. Available online: https://www.faa.gov/aircraft/air_cert/step/disciplines/artificial_intelligence (accessed on 27 November 2024).
  11. EASA. EASA Artificial Intelligence Roadmap 2.0—A human-centric approach to AI in Aviation. EASA. 2023. Available online: https://www.easa.europa.eu/en/document-library/general-publications/easa-artificial-intelligence-roadmap-20 (accessed on 27 November 2024).
  12. Wang, Z.; Pan, W.; Li, H.; Wang, X.; Zuo, Q. Review of deep reinforcement learning approaches for conflict resolution in air traffic control. Aerospace 2022, 9, 294. [Google Scholar] [CrossRef]
  13. Langford, K.; Kille, T.; Lee, S.Y.; Zhang, Y.; Bates, P.R. “In automation we trust”—Australian air traffic controller perspectives of increasing automation in air traffic management. Transp. Policy 2022, 125, 352–362. [Google Scholar] [CrossRef]
  14. Würfel, J.; Papenfuß, A.; Wies, M. Operationalizing AI Explainability Using Interpretability Cues in the Cockpit: Insights from User-Centered Development of the Intelligent Pilot Advisory System (IPAS). In Proceedings of the Artificial Intelligence in HCI, Washington, DC, USA, 29 June–4 July 2024; Degen, H., Ntoa, S., Eds.; Springer: Cham, Switzerland, 2024; pp. 297–315. [Google Scholar] [CrossRef]
  15. Ternus, S. User-Centered Design of an Intelligent Pilot Advisory System for Non-Emergency Scenarios: A Case Study. Master’s Thesis, Hochschule Trier, Trier, Germany, 2024. [Google Scholar]
  16. Tyburzy, L.; Jameel, M.; Hunger, R.; Böhm, J. Empowering Human-AI Collaboration in Air Traffic Control through Smart Interaction Design. In Proceedings of the 2024 AIAA DATC/IEEE 43rd Digital Avionics Systems Conference (DASC), San Diego, CA, USA, 29 September–3 October 2024; pp. 1–9. [Google Scholar] [CrossRef]
  17. Westin, C.; Hilburn, B.; Borst, C.; Van Kampen, E.J.; Bång, M. Building transparent and personalized AI support in air traffic control. In Proceedings of the 2020 AIAA/IEEE 39th Digital Avionics Systems Conference (DASC), Online, 11–16 October 2020; pp. 1–8. [Google Scholar] [CrossRef]
  18. Westin, C.; Borst, C.; Hilburn, B. Strategic conformance: Overcoming acceptance issues of decision aiding automation? IEEE Trans. Hum.-Mach. Syst. 2015, 46, 41–52. [Google Scholar] [CrossRef]
  19. Bobadilla, J.; Ortega, F.; Hernando, A.; Gutiérrez, A. Recommender systems survey. Knowl.-Based Syst. 2013, 46, 109–132. [Google Scholar] [CrossRef]
  20. Gupta, A.K. Real-World Evaluation: Hybrid Recommender System and User Engagement. In Proceedings of the 2024 International Conference on Advancements in Power, Communication and Intelligent Systems (APCI), Kannur, India, 21–22 June 2024; pp. 1–6. [Google Scholar]
  21. Bojarski, M.; Del Testa, D.; Dworakowski, D.; Firner, B.; Flepp, B.; Goyal, P.; Jackel, L.D.; Monfort, M.; Muller, U.; Zhang, J.; et al. End to end learning for self-driving cars. arXiv 2016, arXiv:1604.07316. [Google Scholar] [CrossRef]
  22. Farag, W.; Saleh, Z. Behavior Cloning for Autonomous Driving using Convolutional Neural Networks. In Proceedings of the 2018 International Conference on Innovation and Intelligence for Informatics, Computing, and Technologies (3ICT), Sakhier, Bahrain, 18–19 November 2018; pp. 1–7. [Google Scholar] [CrossRef]
  23. Kaufmann, T.; Weng, P.; Bengs, V.; Hüllermeier, E. A survey of reinforcement learning from human feedback. arXiv 2024, arXiv:2312.14925. [Google Scholar] [CrossRef]
  24. EASA. EASA Artificial Intelligence Roadmap 1.0—A Human-Centric Approach to AI in Aviation|EASA—easa.europa.eu. 2020. Available online: https://www.easa.europa.eu/en/document-library/general-publications/easa-artificial-intelligence-roadmap-10 (accessed on 4 October 2024).
  25. Westin, C.; Hilburn, B.; Borst, C. Air Traffic Controller decision-making consistency and consensus in conflict solution performance. In Proceedings of the CEAS 2015: 5th CEAS Air and Space Conference “Challenges in European Aerospace”, Delft, The Netherlands, 7–11 September 2015; Delft University of Technology: Delft, The Netherlands, 2015. [Google Scholar]
  26. Regtuit, R.; Borst, C.; Van Kampen, E.J. Building strategic conformal automation for air traffic control using machine learning. In Proceedings of the 2018 AIAA Information Systems-AIAA Infotech@ Aerospace, Kissimmee, FL, USA, 8–12 January 2018; p. 74. [Google Scholar] [CrossRef]
  27. Guleria, Y.; Tran, P.; Pham, D.T.; Durand, N.; Alam, S. A machine learning framework for predicting ATC conflict resolution strategies for conformal automation. In Proceedings of the 11th SESAR Innovation Days, Online, 7–9 December 2021. [Google Scholar]
  28. Guleria, Y.; Pham, D.T.; Alam, S.; Tran, P.N.; Durand, N. Towards conformal automation in air traffic control: Learning conflict resolution strategies through behavior cloning. Adv. Eng. Inform. 2024, 59, 102273. [Google Scholar] [CrossRef]
  29. Palma Fraga, R.; Kang, Z.; Crutchfield, J.M.; Mandal, S. Visual search and conflict mitigation strategies used by expert en route air traffic controllers. Aerospace 2021, 8, 170. [Google Scholar] [CrossRef]
  30. Trapsilawati, F.; Wickens, C.D.; Herliansyah, M.K.; Sari, M.P.F.; Tissamodie, G. Why do controllers choose the conflict resolution maneuvers that they do? Int. J. Aerosp. Psychol. 2022, 32, 24–38. [Google Scholar] [CrossRef]
  31. Zakaria, Z.; Lye, S.W.; Endy, S. Unearthing ATCO Decision Instructional Pattern from Simulator Data. ATMRI Conference Papers. 2023. Available online: https://hdl.handle.net/10356/173504 (accessed on 21 June 2025).
  32. Tran, N.P.; Pham, D.T.; Goh, S.K.; Alam, S.; Duong, V. An intelligent interactive conflict solver incorporating air traffic controllers’ preferences using reinforcement learning. In Proceedings of the 2019 Integrated Communications, Navigation and Surveillance Conference (ICNS), Herndon, VI, USA, 9–11 April 2019; pp. 1–8. [Google Scholar] [CrossRef]
  33. Tran, P.N.; Pham, D.T.; Goh, S.K.; Alam, S.; Duong, V. An interactive conflict solver for learning air traffic conflict resolutions. J. Aerosp. Inf. Syst. 2020, 17, 271–277. [Google Scholar] [CrossRef]
  34. van Rooijen, S.; Ellerbroek, J.; Borst, C.; van Kampen, E. Conformal automation for air traffic control using convolutional neural networks. In Proceedings of the 13th USA/Europe Air Traffic Management Research and Development Seminar 2019, ATM 2019, Vienna, Austria, 17–21 June 2019. [Google Scholar]
  35. Bastas, A.; Vouros, G. Data-driven prediction of Air Traffic Controllers reactions to resolving conflicts. Inf. Sci. 2022, 613, 763–785. [Google Scholar] [CrossRef]
  36. Bastas, A.; Vouros, G.A. Data-Driven Modeling of Air Traffic Controllers’ Policy to Resolve Conflicts. Aerospace 2023, 10, 557. [Google Scholar] [CrossRef]
  37. Guleria, Y.; Pham, D.T.; Alam, S. Advancing beyond conformal conflict resolution in air traffic control: Balancing efficiency and conformance. In Proceedings of the 11th International Conference on Research in Air Transportation (ICRAT), Singapore, 1–4 July 2024. [Google Scholar]
  38. Westin, C.; Borst, C.; Hilburn, B. Automation transparency and personalized decision support: Air traffic controller interaction with a resolution advisory system. IFAC-PapersOnLine 2016, 49, 201–206. [Google Scholar] [CrossRef]
  39. Westin, C.; Borst, C.; Kampen, E.; Nunes, T.M.; Boonsong, S.; Hilburn, B.; Cocchioni, M.; Bonelli, S. Personalized and transparent AI support for ATC conflict detection and resolution: An empirical study. In Proceedings of the 12th SESAR Innovation Days, Budapest, Hungary, 5–8 December 2022; pp. 5–8. [Google Scholar]
  40. Naeem, M.; Rizvi, S.T.H.; Coronato, A. A Gentle Introduction to Reinforcement Learning and its Application in Different Fields. IEEE Access 2020, 8, 209320–209344. [Google Scholar] [CrossRef]
  41. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 2018; Volume 2. [Google Scholar]
  42. Wickens, C.D. Automation in air traffic control: The human performance issues. In Automation Technology and Human Performance: Current Research and Trends; Scerbo, M.W., Mouloua, M., Eds.; Lawrence Erlbaum Associates: Mahwah, NJ, USA, 1999; pp. 2–10. [Google Scholar]
  43. Lillicrap, T. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar] [CrossRef]
  44. Rane, N.L.; Choudhary, S.P.; Rane, J. Acceptance of artificial intelligence: Key factors, challenges, and implementation strategies. J. Appl. Artif. Intell. 2024, 5, 1–33. [Google Scholar] [CrossRef]
  45. Hilburn, B.; Westin, C.; Borst, C. Will controllers accept a machine that thinks like they think? The role of strategic conformance in decision aiding automation. Air Traffic Control Q. 2014, 22, 115–136. [Google Scholar] [CrossRef]
  46. Westin, C.A.; Borst, C.; Hilburn, B. Mismatches between automation and human strategies: An investigation into future air traffic management decision aiding. In Proceedings of the 17th International Symposium on Aviation Psychology, Dayton, OH, USA, 6–9 May 2013; p. 530. [Google Scholar]
  47. Schier-Morgenthal, S.; Böhm, J.; Meier, J.; Renkhoff, J.; Wüstenbecker, N. DLRK2024-Weg von der Traditionellen Arbeitsposition-Das Potential von Digitalen Lotsen für die Flugsicherung; DLRK: Hamburg, Germany, 2024. [Google Scholar]
  48. Shin, M.; Kim, J.; Kim, M. Measuring human adaptation to AI in decision making: Application to evaluate changes after AlphaGo. arXiv 2020, arXiv:2012.15035. [Google Scholar] [CrossRef]
  49. Kobak, D.; González-Márquez, R.; Horvát, E.Á.; Lause, J. Delving into ChatGPT usage in academic writing through excess vocabulary. arXiv 2024, arXiv:2406.07016. [Google Scholar] [CrossRef]
  50. Mullany, M.J. The use of analyst-user cognitive style differentials to predict aspects of user satisfaction with information systems. Ph.D. Thesis, AUT University, Auckland, New Zeeland, 2006. [Google Scholar]
  51. Gaume, K.; Olive, X.; Gianazza, D.; Alligier, R.; Durand, N. A Catalogue of Deconfliction Actions Extracted from Historical ADS-B Data. TU Delft OPEN Publ. 2024, 1. [Google Scholar] [CrossRef]
  52. Renkhoff, J.; Feng, K.; Meier-Doernberg, M.; Velasquez, A.; Song, H.H. A Survey on Verification and Validation, Testing and Evaluations of Neurosymbolic Artificial Intelligence. IEEE Trans. Artif. Intell. 2024, 5, 3765–3779. [Google Scholar] [CrossRef]
  53. Gu, S.; Yang, L.; Du, Y.; Chen, G.; Walter, F.; Wang, J.; Knoll, A. A review of safe reinforcement learning: Methods, theory and applications. arXiv 2022, arXiv:2205.10330. [Google Scholar] [CrossRef]
  54. Klingbeil, A.; Grützner, C.; Schreck, P. Trust and reliance on AI—An experimental study on the extent and costs of overreliance on AI. Comput. Hum. Behav. 2024, 160, 108352. [Google Scholar] [CrossRef]
  55. Shekar, S.; Pataranutaporn, P.; Sarabu, C.; Cecchi, G.A.; Maes, P. People over trust AI-generated medical responses and view them to be as valid as doctors, despite low accuracy. arXiv 2024, arXiv:2408.15266. [Google Scholar] [CrossRef]
  56. Rodriguez, S.S.; O’Donovan, J.; Schaffer, J.A.; Höllerer, T. Knowledge Complacency and Decision Support Systems. In Proceedings of the 2019 IEEE Conference on Cognitive and Computational Aspects of Situation Management (CogSIMA), Las Vegas, NV, USA, 8–11 April 2019; pp. 43–51. [Google Scholar] [CrossRef]
  57. Kirwan, B. Human factors requirements for human-ai teaming in aviation. Future Transp. 2025, 5, 42. [Google Scholar] [CrossRef]
  58. Casner, S.M.; Geven, R.W.; Recker, M.P.; Schooler, J.W. The Retention of Manual Flying Skills in the Automated Cockpit. Hum. Factors 2014, 56, 1506–1516. [Google Scholar] [CrossRef]
  59. Chen, J.Y.C.; Lakhmani, S.G.; Stowers, K.; Selkowitz, A.R.; Wright, J.L.; Barnes, M. Situation awareness-based agent transparency and human-autonomy teaming effectiveness. Theor. Issues Ergon. Sci. 2018, 19, 259–282. [Google Scholar] [CrossRef]
  60. Jameel, M.; Tyburzy, L.; Gerdes, I.; Pick, A.; Hunger, R.; Christoffels, L. Enabling Digital Air Traffic Controller Assistant through Human-Autonomy Teaming Design. In Proceedings of the 2023 IEEE/AIAA 42nd Digital Avionics Systems Conference (DASC), Barcelona, Spain, 1–5 October 2023; pp. 1–9. [Google Scholar] [CrossRef]
Figure 1. A radar display showing a conflict in a simulation software. The black lines indicate the projected trajectories of the aircraft. The system highlights areas in which conflicts are predicted to occur in red and attaches an alert to the affected aircraft. The darker gray area is the sector that is currently managed by the ATCO. Accordingly, the aircraft represented by the black diamond is within the sector, while the ones in blue lie outside the sector. Additionally, two navigational aids (NAVAIDs) can be seen on the map indicated by gray crosses. The aircraft indicated with black diamonds are currently assumed by the ATCO while the aircraft colored in blue are currently not assumed by the ATCO.
Figure 1. A radar display showing a conflict in a simulation software. The black lines indicate the projected trajectories of the aircraft. The system highlights areas in which conflicts are predicted to occur in red and attaches an alert to the affected aircraft. The darker gray area is the sector that is currently managed by the ATCO. Accordingly, the aircraft represented by the black diamond is within the sector, while the ones in blue lie outside the sector. Additionally, two navigational aids (NAVAIDs) can be seen on the map indicated by gray crosses. The aircraft indicated with black diamonds are currently assumed by the ATCO while the aircraft colored in blue are currently not assumed by the ATCO.
Aerospace 12 00751 g001
Figure 2. Paper selection and analysis process.
Figure 2. Paper selection and analysis process.
Aerospace 12 00751 g002
Figure 3. Number of publications from 2015 to 2024.
Figure 3. Number of publications from 2015 to 2024.
Aerospace 12 00751 g003
Figure 4. Thematic clustering of the selected literature. One paper can belong to multiple categories.
Figure 4. Thematic clustering of the selected literature. One paper can belong to multiple categories.
Aerospace 12 00751 g004
Figure 5. The concept of reinforcement learning. Figure based on Sutton and Barto [41].
Figure 5. The concept of reinforcement learning. Figure based on Sutton and Barto [41].
Aerospace 12 00751 g005
Table 1. Overview of papers analyzed within this literature review.
Table 1. Overview of papers analyzed within this literature review.
CategoryPapers
Finding Strategies to Resolve Conflicts[25,26,27,28,29,30,31]
Implementing Personalized Conflict Resolution[26,27,28,32,33,34,35,36]
Balancing Different Conflict Resolution Strategies[17,37]
Influence of Personalization on the Acceptance Rates of Advisories[17,28,38,39]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Renkhoff, J.; Ternus, S.; Guleria, Y. A Survey on Personalized Conflict Resolution Approaches in Air Traffic Control. Aerospace 2025, 12, 751. https://doi.org/10.3390/aerospace12090751

AMA Style

Renkhoff J, Ternus S, Guleria Y. A Survey on Personalized Conflict Resolution Approaches in Air Traffic Control. Aerospace. 2025; 12(9):751. https://doi.org/10.3390/aerospace12090751

Chicago/Turabian Style

Renkhoff, Justus, Sarah Ternus, and Yash Guleria. 2025. "A Survey on Personalized Conflict Resolution Approaches in Air Traffic Control" Aerospace 12, no. 9: 751. https://doi.org/10.3390/aerospace12090751

APA Style

Renkhoff, J., Ternus, S., & Guleria, Y. (2025). A Survey on Personalized Conflict Resolution Approaches in Air Traffic Control. Aerospace, 12(9), 751. https://doi.org/10.3390/aerospace12090751

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop