Next Article in Journal
Research on Trajectory Prediction Algorithm Based on Unmanned Aerial Vehicles Behavioral Intentions
Previous Article in Journal
A New Perspective on Urban Mobility Through Large-Scale Drone Experiments for Smarter, Sustainable Cities
Previous Article in Special Issue
A Robust Aerodynamic Design Optimization Methodology for UAV Airfoils Based on Stochastic Surrogate Model and PPO-Clip Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on the Construction and Resource Optimization of a UAV Command Information System Based on Large Language Models

1
Air Defense and Antimissile School, Air Force Engineering University, Xi’an 710043, China
2
People’s Liberation Army Unit 32705, Xi’an 710000, China
3
Hangzhou International Innovation Institute, Beihang University, Beijing 100191, China
4
National University of Defense Technology, Wuhan 430019, China
*
Author to whom correspondence should be addressed.
Drones 2025, 9(9), 639; https://doi.org/10.3390/drones9090639
Submission received: 28 July 2025 / Revised: 23 August 2025 / Accepted: 4 September 2025 / Published: 12 September 2025

Abstract

Highlights

What are the main findings?
  • A three-layer “cloud-edge-terminal” UAV command information system has been developed, deploying LLMs of different scales (cloud 175B, edge 70B, terminal 7B) to achieve optimal resource allocation and support elastic expansion from 10 to over 1000 UAVs. An improved Grey Wolf Optimization Algorithm (ILGWO) integrating Lévy flight, adaptive weighting, elite learning, and chaotic initialization strategies has been proposed to enhance system performance.
  • The system achieves remarkable improvements compared to traditional methods: 34.2% reduction in task latency, 29.6% optimization in energy consumption, and 31.8% improvement in resource utilization. In urban rescue scenarios, response latency decreased by 44.7% and coordination efficiency increased by 39.5%, while LLM integration enhanced decision-making accuracy across multiple dimensions—task priority adjustment (76.3% to 94.7%), dynamic resource allocation (68.9% to 91.2%), and anomaly handling (71.5% to 93.8%).
What is the implication of the main finding?
  • System Performance and Generalization Capabilities. Integrating large language models into unmanned aerial vehicle command information systems enables autonomous decision-making and intelligent planning without relying on preset rules, effectively addressing the shortcomings of existing information systems in handling dynamic scenarios and complex environments. The system demonstrates superior generalization ability with 82.5% performance retention in new scenarios, significantly outperforming the reinforcement learning method’s 56.7%, enabling rapid deployment in rescue, emergency response, inspection, and reconnaissance scenarios.
  • Technical Specifications and Practical Implementation. The proposed system architecture and optimization algorithm provide a scalable and robust solution for large-scale UAV swarm operations. With fault tolerance capabilities maintaining an 88.4% task completion rate even under 20% node failure conditions and a real-time inference delay of only 156 ms for edge-deployed 7B models, the system meets practical requirements for emergency rescue, environmental monitoring, and intelligent surveillance applications. This work establishes a theoretical and practical foundation for integrating cutting-edge AI technologies into unmanned aerial vehicle systems, promoting the development of intelligent, adaptive, and resilient autonomous systems for critical mission scenarios.

Abstract

As UAVs are increasingly deployed in complex scenarios such as disaster monitoring, emergency rescue, and power-line inspection, traditional command and control systems face severe challenges in intelligent decision-making, resource allocation, and elastic scalability. To address these issues, we first propose a distributed UAV command and control system based on large language models of the LLaMA2 family. The system adopts a “cloud–edge–terminal” architecture, using 5G as the backbone network and the Internet of Things as a supplement, with edge computing serving as the computing platform. LLMs of various parameter scales are deployed on demand at different hierarchical levels to support both training and inference, enabling intelligent decision-making and optimal resource allocation. Second, we establish a multidimensional system model that integrates computation, communication, and energy consumption, providing a theoretical analysis of network dynamics, resource constraints, and task heterogeneity. Furthermore, we develop an improved Grey Wolf Optimizer (ILGWO) that incorporates adaptive weights, an elite learning strategy, and Lévy flights to solve the multi-objective optimization problem posed by the system. Experimental results show that the proposed system improves task latency, energy efficiency, and resource utilization by 34.2%, 29.6%, and 31.8%, respectively, compared with conventional methods. Real-world field tests demonstrate that, in urban rescue scenarios, the system reduces response latency by 44.7% and increases coordination efficiency by 39.5%. This work offers a reference for the optimized design and practical deployment of UAV command and control systems in complex environments.

1. Introduction

In recent years, rapid technological advancements have enabled unmanned aerial vehicles (UAVs) to be widely employed in emergency rescue, power-line inspection, disaster monitoring, and intelligence reconnaissance. Nevertheless, as application scenarios and mission requirements expand, existing UAV command and control (C2) systems confront significant challenges.
First, they exhibit insufficient intelligent decision-making capabilities. Most current systems rely on preset rules and heuristic algorithms, which struggle to cope efficiently with dynamic and complex environments. In multi-UAV cooperative missions, heavy human intervention is required, and capabilities such as intelligent situational awareness, dynamic resource adjustment, and task allocation are lacking.
Second, resource allocation efficiency is limited. Computing resources, communication channels, and energy budgets are configured independently without holistic optimization, resulting in a pronounced gap between resource utilization and system performance. This shortcoming is especially critical in emergency rescue operations, resource-constrained environments, and regions with deficient infrastructure, where efficient resource orchestration is indispensable.
Third, scalability is inadequate. Traditional centralized architectures scale poorly, making it difficult to support large-scale UAV swarms executing heterogeneous missions. As system size increases, overload risks emerge, leading to exponential growth in computational complexity, extended communication latency, and amplified fault impacts.
Consequently, scholars have made various efforts to address these challenges. Rule-engine-based approaches [1] offer rapid decision-making and straightforward implementation yet suffer from poor adaptability to unfamiliar scenarios. Reinforcement-learning-based methods [2,3,4], including DQN, PPO, and DPO, learn optimal policies via trial-and-error interactions with the environment; however, they entail high training costs, low sample efficiency, and limited generalization. Traditional machine learning techniques perform well in specific contexts but are constrained by limited adaptability, insufficient generalization, and dependence on manual feature engineering.
Recently, large language models (LLMs) have emerged as a novel AI paradigm, demonstrating remarkable intelligence and being regarded as a viable route toward artificial general intelligence [5,6,7,8]. Pre-trained on massive unsupervised corpora, LLMs acquire high-dimensional representations of general knowledge. Leveraging in-context learning (ICL) and few-shot prompting, they enable UAV systems to adapt to diverse task scenarios through succinct natural-language interactions and to accomplish situational awareness, task allocation, and command decision-making effectively in complex missions.
Motivated by these insights, this paper proposes a cloud–edge–terminal UAV command and control system empowered by LLMs. The principal contributions are as follows:
(1)
Architectural innovation: A three-tier distributed architecture integrated with large language models is designed, optimizing the entire pipeline from data acquisition to intelligent decision-making.
(2)
Modeling innovation: A multidimensional system-effectiveness optimization model is established that explicitly considers network dynamics, task heterogeneity, and resource constraints.
(3)
Algorithmic innovation: An improved Lévy-flight-enhanced Grey Wolf Optimizer (ILGWO) incorporating multiple enhancement strategies is devised to solve complex multi-objective optimization problems effectively.
(4)
Evaluation innovation: Comprehensive performance assessments and practical validations are conducted via simulations and real-world scenarios, demonstrating the system’s efficacy and utility.

2. Related Work

2.1. Unmanned Command and Control Systems

Command and control (C2) systems are defined differently across industry, telecommunications, and defense, yet they share a common essence: a system that supports personnel, equipment, or other systems in executing assigned tasks by issuing directives and maintaining control over assets [9]. Zhong et al. [10] constructed a two-stage optimization model for unmanned weapon systems that maximizes mission-resilience metrics, thereby improving the coordination between central and edge nodes after mission interruptions. Johansson et al. [11] advocate integrating UAV command and control into a single, cohesive system rather than treating them as separate functions, arguing that such unification yields synergistic advantages. Radovanović et al. [12] evaluate the feasibility of incorporating UAV mission execution into modern air-warfare C2 information systems and illustrate recent worldwide armed-conflict use cases. Wang et al. [13] propose a reinforcement-learning-based task-offloading method for post-disaster rescue scenarios.
Nevertheless, existing studies predominantly concentrate either on task offloading and communication-resource optimization on the UAV-task side or in vertical-domain applications. There remains a dearth of dedicated command-and-information systems tailored for UAVs. Moreover, most solutions rely on simple preset rules or heuristic algorithms, rendering them ill-suited for highly dynamic, complex environments where multiple heterogeneous tasks must be processed in parallel.

2.2. Large Language Model Compression Techniques

Large-scale models entail enormous parameter counts, data volumes, and computational demands. Although they exhibit near-general artificial intelligence capabilities, their heavy reliance on computational resources severely constrains their deployment scenarios. Model compression has emerged as an effective remedy. Techniques such as quantization and pruning drastically reduce model size while preserving predictive accuracy. Quantization maps high-precision floating-point weights to low-bit-width representations. LLM-QAT [14] introduces a data-free distillation scheme that leverages data synthesized by the pre-trained model itself, thereby minimizing dependence on external datasets during quantization-aware training and enhancing model extensibility.
Pruning identifies and eliminates redundant parameters to achieve lightweighting. LLM-Pruner [15] proposes a task-agnostic structured pruning method that constructs a dependence graph using gradient information. By leveraging first-order statistics and an approximated Hessian matrix, it evaluates parameter importance and interdependencies to selectively remove non-critical structures. Collectively, these techniques lay the groundwork for deploying large language models in resource-constrained environments.

2.3. Intelligent Decision-Making Approaches

Intelligent decision-making is one of the core capabilities of unmanned aerial vehicle command and control systems, and it is a concentrated reflection of the overall intelligence level of the system. The existing intelligent decision-making methods mainly include reinforcement learning, rule-based engines, game theory methods, and traditional machine learning methods.
Reinforcement-learning-related methods include the Deep Q-Network (DQN), Near-End Policy Optimization (PPO), and Actor Critic (SAC). This type of method learns the optimal strategy through trial-and-error interaction with the environment, demonstrating higher accuracy and faster speed in intelligent decision-making than human operators. However, there is a significant dependence on the number of training samples and reward function, and the model’s generalization ability is limited. The method based on a rule engine constructs a rule base based on domain knowledge, which has the advantages of fast decision-making speed and clear and accurate logic. However, it relies heavily on rules and has poor adaptability to application scenarios. The game theory method models the UAV decision-making problem as a differential game or multi-agent game and seeks optimization through a Nash equilibrium solution. It is naturally suitable for adversarial scenarios, but modeling dynamic and high-capacity scenarios is difficult, and optimization is challenging. Traditional machine learning methods mainly include support vector machines and random forests, which perform well in specific scenarios but have a high dependence on artificial features and scene characteristics.
The LLM method introduced in this article is a method that is convenient to use, has strong generalization ability, and performs well in intelligence. The LLM is mainly based on the transformer architecture, which learns general knowledge representations through pre-training on massive unsupervised data, empowering unmanned aerial vehicle systems to reduce the threshold of human–computer interaction, associate a wide range of relevant knowledge, and accurately solve intelligence situations. We compared the characteristics of several methods in Table 1 to provide quick screening for researchers in this field.
In the above table, n is the input dimension, d is the feature dimension, S and A are the state and action space sizes, R is the number of rules, and L is the sequence length.

3. System Architecture Design

This chapter analyzes the system architecture proposed in this article. Section 3.1 provides a schematic diagram of the overall system framework and introduces the functions of each level. Section 3.2 summarizes and evaluates the key technologies covered by the system, helping readers quickly grasp the prerequisite knowledge in the relevant fields. Section 3.3 provides a schematic diagram of the system workflow, visualizing the system’s operational process. Section 3.4 and Section 3.5, respectively, discuss the scalability and reliability of the system.

3.1. Overall Architecture Overview

Figure 1 depicts the proposed LLM-based UAV command and control system, structured as a three-tier “cloud–edge–terminal” architecture. The design uses 5G as its backbone network and edge computing servers as its computational substrate [16]. By co-locating Multi-access Edge Computing (MEC) nodes with the 5G base-station CU (Central Unit), the architecture pushes both computing power and intelligent services to the mission edge, ensuring an unimpeded pipeline from data acquisition to intelligent decision-making.
The central layer of this architecture establishes a system center cloud by deploying high-performance general-purpose servers, base stations, and data centers for cloud deployment of UAVs and related business services, as well as the full LLM (LLaMA-175B). Cloud resources are abundant, suitable for handling large-scale and highly complex tasks. Through methods such as fiber optic cable network, 5G mobile communication, and satellite communication, the system maintains high-speed interconnection with other layers by routing and jumping through the aggregation layer, providing global situational awareness, intelligent decision-making, device control, and task distribution services for the system.
The edge layer is located between the center layer and the edge layer. As a middleware for instruction parsing, data distribution, and task offloading, it integrates the MEC system with the 5G mobile communication system and deploys them in the same base station through the 5G base-station control unit (CU) and the edge computing node to build the edge cloud for processing delay-sensitive and medium-sized computing tasks.
The terminal layer includes unmanned aerial vehicle platforms, ground control stations, vehicle-mounted base-station nodes, and sensor devices. UAVs carry different types of payloads, such as electro-optical pods, radar systems, and communication relays, to carry out the corresponding tasks. Terminal devices have limited computing power and energy storage, so they mainly undertake tasks such as data collection, data transmission, and instruction execution.

3.2. Key Technology Architecture

3.2.1. Distributed Large Language Model Deployment Architecture

We adopt a layered deployment strategy for the system architecture designed in this article. We deploy a 175B full-scale model in the central layer to handle complex inference and global optimization tasks; deploy a 70B large language model at the edge layer to handle general complex tasks and latency sensitive tasks, saving the latency overhead of task feedback; deploy the 7B lightweight large language model together with the base-station CU at the terminal layer, and provide simple data-parsing tasks at the task site. The LLM scale selected for the three-layer network not only adapts to the existing edge computing servers but also adapts to the upper limit of hardware computing power of intelligent terminal devices.

3.2.2. Fifth-Generation (5G) Network Slicing and QoS Assurance

Leveraging 5G network-slicing technology [17], the system can logically share a single physical infrastructure among heterogeneous tasks. Depending on whether the UAV is engaged in data collection, reconnaissance, swarm control, or data dissemination, the appropriate slice—tailored in terms of latency, bandwidth, and compute—is selected on demand. A QoS framework further performs dynamic bandwidth allocation, task-queue optimization, and adaptive modulation and coding. By continuously matching resource allocation to task priority and type, the system guarantees end-to-end service quality.

3.2.3. Edge-Intelligence Computing Framework

Edge computing is realized through a containerized architecture that delivers elastic compute resources and scheduling. Built on Kubernetes [18], the platform automatically scales pod resources in response to real-time system load, ensuring efficient utilization of constrained hardware. A task scheduler allocates workloads according to task requirements, resource availability, and network conditions. By consulting a pre-built task-feature library and a demand-prediction model, the framework forecasts the performance of different scheduling policies. Reinforcement-style algorithms then learn the optimal policy, enhancing the system’s adaptability to evolving task environments.

3.2.4. LLM Empowers Intelligent Decision Analysis

In the previous section on intelligent decision-making methods, we briefly introduced a comparison between large language models and other intelligent decision-making methods. This section will analyze the applicability of LLMs in empowering intelligent decision-making for unmanned aerial vehicle command and control systems from a requirements perspective. There are three main requirements for UAVs to perform diverse tasks.
One is the real-time requirement. Unmanned aerial vehicle control has extreme requirements for system decision delay, and in complex scenarios, obstacle avoidance and emergency response usually require millisecond-level response delays. The rule engine method not only has an efficient reasoning process but also has fast decision-making ability. But it is difficult to effectively cope with complex task scenarios. Compressing the LLM while maintaining model performance can effectively reduce deployment thresholds and exhibit acceptable inference latency.
The second is the demand for environmental adaptability. Emergency rescue scenarios involving unmanned aerial vehicles are usually dynamic, complex, and uncertain. Reinforcement learning methods perform well in limited space training, but they have difficulty dealing with unfamiliar scenarios and interference factors outside the data distribution. And large language models rely on in-context learning and few-shot capabilities, which allow them to effectively cope with out-of-distribution scenarios.
The third requirement is multitask collaboration. With the application of UAVs in industries such as manufacturing agriculture, and power, their tasks are becoming more diverse. Tasks such as disaster reconnaissance, agricultural sowing, and power inspection require training models based on corresponding scenarios. An LLM, relying on its strong generalization ability, can quickly adapt to different scenarios and switch task scenarios through prompt engineering and model training.
In addition, during the selection phase of the system technology roadmap, we conducted extensive simulation experiments to test similar methods and compare the performance of different intelligent decision-making methods in unmanned aerial vehicle systems, as shown in Table 2.
The training cost of the LLMs in the table mainly includes the fine-tuning cost for unmanned aerial vehicle task scenarios and does not include pre-training costs. As this article uses an open-source LLM (LLaMA2) as the base model, most of the work expenses are in the post-training stage. In summary, this article has decided to use an LLM as the intelligent decision-making core, 5G MEC as the computing communication platform, and distributed deployment of large-scale models of different scales to support unmanned aerial vehicles in carrying out diverse tasks.

3.3. System Workflow

The system is designed for emergency response scenarios such as earthquakes, mudslides, and floods. Field intelligence is collected by UAVs, unmanned ground vehicles, and other sensors at the terminal tier. At this tier, the collected data are first preprocessed or feature-extracted by either a sensor-aggregation node or the lightweight LLM co-located with the base station. Based on the task’s type, requirements, and complexity, an execution policy is then selected: the task is offloaded to the lowest tier that still meets its minimal resource and latency constraints. After the central or edge tier has completed processing, the results are distributed back to the terminal endpoints via the 5G downlink. The overall workflow is illustrated in Figure 2.

3.4. System Scalability Design

In order to enhance the flexibility of the system in different scenarios, this article fully considers the scalability requirements during the system design phase, ensuring that the system can adapt to different application scenarios while meeting constantly changing application requirements. Therefore, we comprehensively consider four dimensions: horizontal expansion, vertical expansion, functional module expansion, and cross-domain collaborative expansion.
The first is the ability to expand horizontally. The central layer adopts a distributed cluster configuration, while the cloud computing center adopts containerized deployment and microservice architecture, supporting agile scaling of computing and storage nodes. The edge computing server of the edge layer adopts co-station deployment with the 5G base station and supports node plug and play, and the edge nodes are interconnected through standardized interfaces to build a flexible and elastic edge cloud. The terminal layer supports device access for different network standards, with devices equipped with 5G baseband and RF chips for on-demand access. Heterogeneous terminals access the 5G backbone network through protocol converters or IoT aggregation nodes.
The second is vertical scalability. In terms of computing resources, nodes at all levels support hardware upgrades and expansion. The central layer can increase GPU clusters to enhance computing power. The edge layer can expand computing resources by increasing computing power or upgrading processors and memory. Terminal devices can upgrade payloads or expand sensor computing and corresponding functions. In terms of storage resources, a distributed storage architecture is adopted, which can support linear expansion of storage resources for nodes at all levels as needed.
The third is the expansion of functional modules, mainly considering the expansion of intelligent engine models and other business functions. Large language models can be upgraded online and deployed locally, and vertical-domain models can be deployed according to task requirements, such as image segmentation, image recognition, and remote sensing image interpretation. The system deployment model management system can efficiently manage version iteration and rollback backtracking. The extension of other business functions supports plug-in deployment, enabling on-demand loading and agile deployment without the need for downtime upgrades.
The fourth is cross-domain collaborative expansion. On the one hand, the system supports distributed deployment, achieving regional linkage, efficient collaboration, and resource sharing. On the other hand, the system supports industry expansion, and the system architecture design fully considers the common needs of different scenarios while considering universality. Therefore, it can efficiently adapt to industrial applications such as power inspection, disaster monitoring, and logistics distribution.

3.5. System Reliability Design

UAVs require high reliability in critical scenarios such as emergency rescue, line inspection, and communication support. To this end, we enhance system reliability through three aspects of design: network security, redundancy design, and fault resolution.
The first is network security protection. On the one hand, security measures are designed for the access layer, network layer, and application layer. At the access layer, identity authentication mechanisms are set up through hardware certificates and dynamic keys, and access control whitelists and behavior patterns are dynamically controlled. In addition, end-to-end AES-256 encryption is used to ensure data transmission security. At the network layer, logical isolation is achieved through VLANs and software-defined networks, while a real-time network traffic monitoring mechanism is set up to monitor abnormal communication modes. An API gateway is built at the application layer for agreeing to API access control and traffic restrictions, and data anonymization and role-based granular permission control mechanisms are set up at the edge layer to ensure network layer security. On the other hand, intelligent intrusion detection systems are deployed at the edge and center layers, with built-in intrusion detection algorithms that perform abnormal behavior detection, threat intelligence fusion, and automatic response mechanisms.
The second is redundant design. On the one hand, design redundancy mechanisms are established for communication links. Multiple independent communication links are established, with 5G as the backbone network for the main link and satellite or shortwave communication as the backup link. On the other hand, redundancy design is implemented for computing resources, deploying N + 2 service instances at the central layer and deploying the same instance as a backup in data centers in different geographical regions to achieve automatic transfer in case of failure. The edge layer adopts an adjacent node backup scheme, and tasks are automatically offloaded to the optimal node based on optimization algorithms according to resource usage. In addition, data storage also adopts redundant settings, maintaining three copies of critical data backup, backing up data in data centers (DCs) in different regions, and keeping all three modified synchronously.
The third is the fault resolution mechanism. A multidimensional fault detection and resolution mechanism is established. One is the fault detection system, which regularly monitors the health status of hardware devices, monitors real-time indicators such as CPU, memory, and network, and uses large language models for predictive maintenance based on historical data. It conducts regular usability checks on software, analyzes abnormal logs, and monitors performance baseline deviations. Regular connectivity testing is conducted to check for network failures, monitor real-time latency changes, and measure packet loss rates. The second is automatic fault transfer. A multi-level fault transfer strategy is designed, with service-level transfer using automatic restart, instance switching, and downgrade service strategies. The node level adopts task migration, data synchronization, and resource reallocation strategies; system-level transfer adopts strategies such as backup system activation, data recovery, and business continuity support. The third is the service recovery strategy. Measures such as fault isolation, service reconstruction, and state recovery are adopted to support rapid recovery from system failures, ensuring data consistency through transaction rollback, data verification, and incremental synchronization.

4. System Modeling

This section establishes corresponding mathematical models for the designed system architecture and workflow, providing support for the subsequent introduction of LLMs and algorithm empowerment. Section 4.1 describes the network model; Section 4.2 and Section 4.3 establish corresponding models from the perspectives of communication and computation; Section 4.4 models the inference and accuracy of the LLM; Section 4.5 establishes the system energy consumption model, including communication, computation, and flight energy consumption; Section 4.6 and Section 4.7 model the system performance indicators and robustness; and Section 4.8 describes the optimization problem, creating conditions for subsequent algorithm introduction.

4.1. Network Model

Consider a three-layer architecture system consisting of N UAV terminals, M edge nodes, and 1 cloud center. Let N = { 1 , 2 , , N } be the set of UAVs, M = { 1 , 2 , , M } the set of edge nodes, and e the set of tasks to be processed.
Each task k K has the following attributes: D k represents input data size (bits), C k represents computational complexity (CPU cycles), T k m a x represents maximum tolerable latency (seconds), ω k represents task priority weight, and ϵ k represents result accuracy requirement.
The basic attributes of UAV i include the following: f i l o c a l represents local CPU frequency (Hz), P i m a x represents maximum transmit power (watts), E i b a t t e r y represents battery capacity (joules), v i represents flight speed (m/s), and x i ,   y i ,   z i represents 3D coordinate position.
Since the large language models deployed at different levels of cloud, edge, and end are responsible for solving different types of tasks, the system uses a dynamic scheduling mechanism to determine the specific level of offloading based on the estimated amount of computation, response requests, and queuing sequence, combined with the advanced rules of edge computing and cloud computing centers, as well as the requirements of the task scenario for delay, network conditions, resource availability, and so on. Table 3 lists common task types and example offloading levels.

4.2. Communication Model

4.2.1. Air-to-Ground Channel Model

Communication between the UAV and the terrestrial base station is modeled using an Air-to-Ground (A2G) channel [19]. Both Line-of-Sight (LoS) and Non-Line-of-Sight (NLoS) propagation paths are considered, and the channel gain is expressed as
h i , j = η LoS · d i , j α LoS · ξ i , j , if   LoS η NLoS · d i , j α NLoS · ξ i , j , if   NLoS
where η LoS and η NLoS denote the path-loss constants for the LoS and NLoS links, respectively; α LoS and α NLoS are the corresponding path-loss exponents; and ξ represents the shadow-fading component, which follows a log-normal distribution.
The LoS probability is a function of the UAV’s altitude and its horizontal distance from the base station and is expressed as
P LoS ( h i , d i , j h o r ) = 1 1 + a · exp ( b ( θ i , j a ) )
where θ i , j = arctan ( h i / d i , j h o r ) is the elevation angle, and a and b are environment-dependent parameters.

4.2.2. Data Transmission Rate

Based on Shannon’s theorem, the uplink transmission rate from UAV i to edge node j is given by
R i , j u p = B i , j log 2 1 + P i , j · h i , j σ 2 + k I i , j P k , j · h k , j
where B i , j denotes the bandwidth allocated to link i , j , P i , j is the transmit power of UAV i on this link, σ 2 represents the noise power, and I i , j is the set of UAVs that create interference toward link i , j .
Accounting for the OFDMA technology in 5G networks, the bandwidth allocation must satisfy
i = 1 N B i , j B j t o t a l ,       j M
B i , j m i n B i , j B i , j m a x , i N , j M

4.2.3. Transmission Delay Model

The delay for task k transmitted from UAV i to edge node j comprises propagation delay, transmission delay, and protocol overhead:
τ k , i , j t r a n s = d i , j c + D k R i , j u p + τ protocol
where c is the speed of light, and τ protocol is the protocol-processing overhead.
Owing to the UAV’s mobility, the channel state varies; thus, the average rate during transmission must account for the Doppler effect:
R i , j a v g = 0 T t r a n s R i , j ( t ) d t / T t r a n s
where R i j ( t ) is the instantaneous transmission rate at time t .

4.3. Computation Model

4.3.1. Local Computation Model

The latency for UAV i to process task k locally is
τ k , i l o c a l = C k f i l o c a l + δ k o v e r h e a d
where δ k o v e r h e a d represents system overhead, including operating-system scheduling and memory-access latency.
The energy consumption model for local computation accounts for both dynamic and static power:
E k , i l o c a l = κ · ( f i l o c a l ) 2 · C k + P i s t a t i c · τ k , i l o c a l
where κ is the effective switched capacitance, and P i s t a t i c is the static power consumption.

4.3.2. Edge Computing Model

Edge node j ’s computing resources are dynamically allocated through virtualization. Let the total computing capacity of edge node j be F j t o t a l , and let the amount of computing resources allocated to task k be f k , j e d g e .
Edge computing latency comprises the queuing delay plus the computation delay:
τ k , j e d g e = τ k , j q u e u e + C k f k , j e d g e
The queuing delay is modeled as an M/M/1 queue that takes task priorities into account:
τ k , j q u e u e = λ j μ j ( μ j λ j ) · 1 + l H k ρ l 1 ρ l
where λ j is the task arrival rate, μ j is the service rate, H k is the set of tasks with higher priority than task k , and ρ l is the traffic intensity of task l .
The allocation of computing resources at the edge node must satisfy
k K j f k , j e d g e F j t o t a l
f k , j m i n f k , j e d g e f k , j m a x
where K j is the set of tasks assigned to edge node j .

4.3.3. Cloud Computing Model

The cloud has ample computing resources; the dominant factor is network transmission delay.
τ k c l o u d = τ k , i , c l o u d t r a n s + C k f k c l o u d + τ k , c l o u d , i b a c k
Let τ k , c l o u d , i b a c k denote the round-trip delay for returning results.
Because the backhaul link from the cloud to the edge typically offers higher bandwidth, the return delay is comparatively small:
τ k , c l o u d , i b a c k = D k r e s u l t R c l o u d , i d o w n
where D k r e s u l t is the size of the result data, which is usually much smaller than the input data.

4.4. Large Language Model Inference Model

4.4.1. Hierarchical Model Deployment

Based on model complexity and computational demand, the large language model is divided into three tiers:
Cloud full model: Θ c l o u d = 175 × 10 9 parameters, supporting complex inference tasks.
Edge compressed model: Θ e d g e = 7 × 10 9 parameters, obtained via knowledge distillation.
Terminal lightweight model: Θ l o c a l = 0.7 × 10 9 parameters, dedicated to emergency decisions.

4.4.2. Inference Latency Modeling

The inference delay of an LLM is mainly composed of three factors: linear computational complexity, quadratic computational complexity, and hardware limitations [20]. The mainstream LLMs, including LLaMA used in this article, are usually based on the transformer architecture [21], and their total parameter count can be represented by the following equation:
Θ     L l a y e r s   ×   d m o d e l   ×   4 d m o d e l   +   d f f
where d m o d e l represents the model dimension, d f f represents the feedforward layer dimension, and L l a y e r s represents the number of layers. In the self-attention mechanism, the projection of query, health, and value is 3   ×   L s e q   ×   d m o d e l 2 , the attention weight is represented as L s e q 2   ×   d m o d e l , and the output projection is L s e q   ×   d m o d e l 2 . The feedforward network has two projection layers with different dimensions: the first layer is represented as L s e q   ×   d m o d e l   ×   d f f , and the second layer is represented as L s e q   ×   d f f   ×   d m o d e l . Therefore, the total computational complexity is represented as
FLOP total = L l a y e r s × 6 × L s e q × d m o d e l 2 + L s e q 2 × d m o d e l + 2 × L s e q × d m o d e l × d f f
The complexity of feedforward networks and self-attention mechanisms can be expressed as
O ( L s e q · d m o d e l · d f f )
O ( L s e q 2 · d m o d e l )
The former has a linear relationship with sequence length and parameter quantity, while the latter has a quadratic relationship with sequence length. Hardware bottlenecks are mainly limited by computational frequency and memory bandwidth and are related to model parameter quantity and hardware performance [22]. Therefore, inference delay can be modeled as
τ L L M = α · L s e q · Θ + β · L s e q 2 + γ · Θ f c o m p u t e
where L s e q is the length of the input sequence; α , β , and γ are model-related constants; and f c o m p u t e is the frequency of the computing unit. For the transformer architecture, the computational complexity of the self-attention mechanism is O ( L s e q 2 · d m o d e l ) , where d m o d e l is the model dimension.

4.4.3. Model Accuracy Modeling

The commonly used methods for compressing large language models include quantization, pruning, and knowledge distillation. Model compression is based on information theory, and the relationship between model capacity and parameter quantity follows the logarithmic law:
C a p a c i t y     l o g Θ
Therefore, when the model is compressed from the original parameter quantity Θ o r i g i n a l to Θ c o m p r e s s e d , the loss can be modeled as
L o s s i n f o = l o g Θ o r i g i n a l / Θ c o m p r e s s e d
The performance can be maintained at R e t e n t i o n r a t e   =   1     λ d i s t i l l   ×   L o s s i n f o ; therefore the final accuracy of the LLM after compression can be expressed as
A c c c o m p r e s s e d = A c c o r i g i n a l · ( 1 λ d i s t i l l · log ( Θ o r i g i n a l Θ c o m p r e s s e d ) )
where λ d i s t i l l is the knowledge distillation loss coefficient. In addition, we have compiled experimental compression data from mainstream LLMs [23,24,25,26], which are summarized in Table 4.
According to statistical data, an average of λ d i s t i l l = 0.062 ± 0.018 can be obtained, while the goodness of fit R 2   =   0.94 .

4.5. Energy Consumption Model

4.5.1. Communication Energy

The UAV’s communication energy comprises the power consumed by the RF power amplifier and baseband processing:
E c o m m = P t x η P A · T t r a n s + P c i r c u i t · T t r a n s
where P t x is the transmit power, η P A is the power-amplifier efficiency, and P c i r c u i t is the circuit power.
The power-amplifier efficiency is a function of transmit power:
η P A = η m a x · 1 exp P t x P s a t
where η m a x is the maximum efficiency, and P s a t is the saturation power.

4.5.2. Computation Energy

Computation energy depends on processor operating frequency and computational complexity.
E c o m p = κ · f 2 · C + P s t a t i c · T c o m p
For GPU computation, memory-access energy must also be considered:
E m e m o r y = N a c c e s s · E p e r _ a c c e s s
where N a c c e s s is the number of memory accesses, and E p e r _ a c c e s s is the energy per access.

4.5.3. Flight Energy

The UAV’s flight energy consumption is related to its speed, weight, and aerodynamic characteristics:
E f l i g h t = 1 2 ρ S C D v 3 · T f l i g h t + m g 2 2 ρ S · 1 v · T f l i g h t
where ρ is air density, S is the reference area, C D is the drag coefficient, m is the mass, and g is the gravitational acceleration.

4.6. System Performance Metrics

4.6.1. Latency Metric

The overall system latency is the weighted sum of all task latencies:
Ψ d e l a y = k = 1 K ω k · τ k t o t a l T k m a x
where τ k t o t a l denotes the total latency of task k , including transmission, queuing, computation, and return delays.

4.6.2. Energy Metric

The total system energy consumption comprises the communication, computation, and flight energy of all UAVs:
Ψ e n e r g y = i = 1 N E i t o t a l E i b a t t e r y
where E i t o t a l = E i c o m m + E i c o m p + E i f l i g h t is the corresponding energy weighting factor.

4.6.3. Reliability Metric

System reliability accounts for communication interruption, computation failure, and model accuracy degradation:
Ψ r e l i a b i l i t y = k = 1 K ω k · ( 1 P k s u c c e s s )
where the probability of successful task completion is
P k s u c c e s s = P k c o m m · P k c o m p · P k m o d e l

4.7. System Robustness Modeling

System robustness is an important indicator of the ability of unmanned aerial vehicles to perform tasks under complex conditions. This section defines and models system robustness.

4.7.1. Definition of Robustness Indicators

The robustness of the system is quantified through the following four dimensions:
Availability   indicator :   A s y s = M T B F M T B F + M T T R
where M T B F is the average time between failures, and M T T R is the average time to repair.
Performance   degradation   degree :   D p e r f = 1 P f a u l t P n o r m a l
where P f a u l t represents the system performance under fault conditions, and P n o r m a l represents the system performance under normal conditions.
Fault   tolerance :   F t o l e r a n c e   = N s u c c e s s N t o t a l
where N s u c c e s s is the number of tasks successfully completed under fault conditions, and N t o t a l is the total number of tasks.
Recovery   ability :   R r e c o v e r y = e λ t r e c o v e r y
where λ is the recovery rate parameter, and t r e c o v e r y is the system recovery time.

4.7.2. Fault Model

(1)
Communication link failure model
Considering the randomness of link failures, the available probability of link a is modeled as
P l i n k ( i , j ) = e λ l i n k · d i , j · ( 1 I i n t e r f e r e n c e ) v
where λ l i n k is the link failure rate, d i , j is the distance between nodes, and I i n t e r f e r e n c e is the interference indicator function.
(2)
Calculate node fault model
The failure probability of edge node j is related to its load:
P f a i l ( j ) = 1 e α · U j β
where U j = k K j f k , j e d g e F j t o t a l is the resource utilization rate of node j , and α and β are fault model parameters.
(3)
Environmental interference model
The impact of environmental interference on system performance is modeled as follows:
I e n v ( t ) = I b a s e + m = 1 M A m sin ( 2 π f m t + ϕ m )
where I b a s e is the basic interference level, and A m , f m , and ϕ m are the amplitude, frequency, and phase of the m interference source, respectively.

4.7.3. Cascade Fault Model

Considering the interdependence between nodes in the system, a single fault may trigger cascading effects:
P c a s c a d e ( t ) = P i n i t i a l · k = 1 K ( 1 + β k · P d e p e n d e n c y k ( t ) )
where P i n i t i a l is the initial failure probability, β k is the cascading factor, and P d e p e n d e n c y k ( t ) is the k -th level dependent failure probability.

4.7.4. System Elasticity Model

System resilience reflects the ability of a system to recover from a faulty state to a normal state:
Resilience = 0 T Q ( t ) d t / T
where Q ( t ) is the system performance quality function at time t :
Q ( t ) = 1 , n o r m a l Q d e g r a d e d ( t ) , d e g r a d e d 0 , f a u l t
Performance quality in degraded state:
Q d e g r a d e d ( t ) = Q m i n + ( 1 Q m i n ) · e γ ( t t f a u l t )
where Q m i n is the minimum performance level, γ is the recovery rate, and t f a u l t is the time of failure occurrence.

4.8. Optimization Problem Description

Integrating latency, energy, and reliability, the system optimization objective is
min Φ = λ 1 Ψ d e l a y + λ 2 Ψ e n e r g y + λ 3 Ψ r e l i a b i l i t y
Decision variables:
x k , d { 0 , 1 } : deployment decision for task k ( d { l o c a l , e d g e , c l o u d } ); f k , j e d g e : computing resources assigned by edge node j to task k ; B i , j : bandwidth allocation on link i , j ; P i , j : power allocation on link i , j .
Constraints are as follows:
Deployment constraint: d x k , d = 1 , k ; resource constraint: k K j f k , j e d g e F j t o t a l , j i B i , j B j t o t a l , j j P i , j P i m a x , i ; QoS constraint: τ k t o t a l T k m a x , k E i t o t a l E i b a t t e r y , i .

5. Improved Grey Wolf Optimizer

5.1. Basic Grey Wolf Optimizer

The Grey Wolf Optimization Algorithm [27] simulates the hierarchical concept and hunting behavior of wolf packs. The algorithm divides the wolf pack into four levels: α (optimal solution), β (suboptimal solution), δ (third optimal solution), and ω (remaining solutions). The hunting process includes three stages: search, encirclement, and attack.

5.1.1. Encircling Prey

Wolves encircle the prey by adjusting their positions; the position update equation is
D = | C · X p ( t ) X ( t ) |
X ( t + 1 ) = X p ( t ) A · D
where X p is the prey’s position, X is the wolf’s current position, and A and C are coefficient vectors.

5.1.2. Hunting Mechanism

Since the prey’s exact location is unknown, the algorithm assumes that the α , β , and δ wolves possess better information about its position; the remaining wolves update their own positions based on these three leaders:
D α = | C 1 · X α X | D β = | C 2 · X β X | D δ = | C 3 · X δ X |
X 1 = X α A 1 · D α X 2 = X β A 2 · D β X 3 = X δ A 3 · D δ
X ( t + 1 ) = X 1 + X 2 + X 3 3

5.1.3. Attacking and Searching for Prey

Parameter a decreases linearly from 2 to 0 to balance exploration and exploitation:
A = 2 a · r 1 a C = 2 · r 2
where r 1 and r 2 are random vectors in the range [0, 1].

5.2. Algorithmic Enhancement Strategy

5.2.1. Chaotic-Map Initialization

Traditional random initialization often yields uneven population distributions and degrades algorithmic performance. This paper adopts the Tent chaotic map [28] to generate the initial population:
x n + 1 = x n μ , 0 x n < μ 1 x n 1 μ , μ x n 1
where μ = 0.7 is the control parameter. The chaotic sequence is then mapped to the search space as
X i ( 0 ) = X m i n + ( X m a x X m i n ) · x i

5.2.2. Adaptive Weight Strategy

To balance global exploration and local exploitation, an adaptive weight adjustment mechanism is designed:
w ( t ) = w m a x ( w m a x w m i n ) · t T m a x 2
where w m a x   =   0.9 ,   w m i n   =   0.1 ,   T m a x is the maximum iteration count.
The weight parameters are dynamically tuned according to the algorithm’s convergence state:
w a d a p t i v e = w ( t ) · 1 + σ f i t n e s s | μ f i t n e s s |
where σ f i t n e s s and μ f i t n e s s are the standard deviation and mean of the current population fitness, respectively.

5.2.3. Lévy Flight Search Strategy

Lévy flight [29] exhibits long-range correlation and can enhance the global search capability of the algorithm. The probability density function of the Lévy distribution is
L ( s ) ~ s 1 β , 0 < β 2
Lévy flight step lengths are generated using the Mantegna algorithm:
s t e p = u | v | 1 / β
where u ~ N ( 0 , σ u 2 ) , v ~ N ( 0 , σ v 2 ) , and
σ u = Γ ( 1 + β ) sin ( π β / 2 ) Γ ( ( 1 + β ) / 2 ) β 2 ( β 1 ) / 2 1 / β
σ v = 1 during position updating. Lévy flight is incorporated as follows:
X n e w = X o l d + α · L e v y ( β ) ( X b e s t X o l d )
where α is a step-size control parameter, and denotes element-wise multiplication.

5.2.4. Elite Learning Strategy

To accelerate convergence and prevent the loss of high-quality solutions, an elite learning mechanism is designed. The population is sorted by fitness, and the top 20% of individuals are selected as the elite group. The position update formula for elite individuals is
X e l i t e ( t + 1 ) = X e l i t e ( t ) + ϕ · ( X α X e l i t e ( t ) ) + ψ · r a n d ( 1 , 1 )
where ϕ is the learning factor, and ψ is the perturbation intensity.

5.2.5. Dynamic Boundary Handling

When a search solution exceeds the boundary, a dynamic boundary-handling strategy is adopted:
X i n e w = X m i n + r a n d · ( X m a x X m i n ) , X i < X m i n X m a x r a n d · ( X m a x X m i n ) , X i > X m a x X i , o t h e r w i s e

5.3. The Framework of the Improved Algorithm

The overall framework of the Improved Grey Wolf Optimizer (ILGWO) (Algorithm 1) is presented below.
Algorithm 1. ILGWO
ILGWO pseudocode
Input: Population size N , maximum number of iterations T m a x , problem dimension D .
Output: Optimal solution X b e s t .
1. Initialization phase
 Generate initial population using Tent chaotic mapping.
 Calculate the fitness value of each individual.
 Identify X α , X β , X δ .
2. while  t   <   T m a x  do
3. for each wolf  X i  do
 Calculate adaptive weight w a d a p t i v e .
 Update coefficients A and C .
 Update the current wolf position based on the α ,   β ,   δ wolf positions.
 Apply Lévy flight search strategy.
 Perform dynamic boundary processing.
4. end for
5. Elite Learning Update
 Sort populations by fitness.
 Perform learning updates on elite individuals.
6. Update the leader
 Recalculate fitness value.
 Update X α , X β , X δ .
7.  t   =   t   +   1
8. end while
9. return  X α

5.4. Algorithm Complexity Analysis

The time complexity of the ILGWO algorithm mainly consists of the following parts:
initialization :   O ( N · D )
Fitness calculation: O ( N · f e v a l ) , where f e v a l is the complexity of evaluating the fitness function.
Location   update :   O ( N · D )
Sorting   operation :   O ( N log N )
The   overall   time   complexity   is   O ( T m a x · N · ( D + f e v a l + log N ) )
Space complexity is mainly used to store population individuals and temporary variables, known as O ( N · D ) .
Compared with the basic GWO algorithm, the increase in time complexity of ILGWO mainly comes from Lévy flight calculations and elite learning operations, but the complexity of these operations is still at a linear level and will not significantly affect the efficiency of the algorithm.

5.5. Algorithm Scalability Analysis

5.5.1. Parallel Computing Expansion

If the edge nodes are cluster servers, the ILGWO algorithm naturally supports parallel evaluation of individual populations. The fitness calculation of each individual is independent of other individuals and can be executed in parallel on multi-core CPUs or GPUs, achieving near-linear acceleration effects. In addition, the algorithm can be extended to a distributed version, distributing the population across multiple nodes. Through scheduled information exchange and migration operations, distributed collaborative optimization is achieved, which is suitable for large-scale optimization problems.

5.5.2. Dynamic Scale Adjustment

Algorithms can dynamically adjust population size based on task complexity and available resources, efficiently adapting to various task scenarios. In addition, for large-scale problems, algorithms can adopt a hierarchical optimization strategy, decomposing complex problems into subproblems, optimizing them separately, and then coordinating globally.

5.5.3. Multi-Objective Expansion

The algorithm can be extended to a multi-objective optimization version while optimizing multiple objectives, such as latency, energy consumption, and reliability, generating a Pareto-optimal solution set for decision selection. In addition, supporting more complex constraint conditions, including dynamic constraints, probabilistic constraints, etc., enhances the adaptability of the algorithm in practical applications.

6. Experimental Design and Result Analysis

6.1. Experimental Environment and Parameter Settings

6.1.1. Hardware Platform

The experiment was conducted on the following platform: server: 2 × Intel Xeon Gold 5218R CPU (3.2 GHz), 1 T DDR4 memory, 4 × NVIDIA A100 GPU. Edge node: Intel i9-11900K CPU (3.5 GHz), 64 GB DDR4 memory, NVIDIA RTX 3090 GPU. UAV simulation: based on AirSim platform, equipped with ARM Cortex-A78 core and 8 GB memory.

6.1.2. Network Environment Configuration

Fifth-generation (5G) network: Center frequency of 3.5 GHz, bandwidth of 100 MHz, transmission power of 43 dBm. WiFi backhaul: IEEE 802.11ax standard 6 GHz frequency band, bandwidth of 160 MHz. Satellite link: Ka band, uplink 30–31 GHz, downlink 20–21 GHz, bandwidth of 500 MHz [30].

6.1.3. Algorithm Parameter Settings

The ILGWO algorithm parameter settings are as follows:
Population size: N   =   50 ; maximum iteration number: T max =   500 ; Lévy flight parameter: β = 1.5 ; adaptive weights: w max   =   0.9 , w min   =   0.1 ; elite ratio: 20%.
Comparative algorithms include the following:
Basic Grey Wolf Optimization Algorithm (GWO), Particle Swarm Optimization Algorithm (PSO), Genetic Algorithm (GA), Differential Evolution Algorithm (DE), and Artificial Bee Colony Algorithm (ABC).

6.1.4. Task Load Design

Five typical tasks were defined, as shown in Table 5.

6.2. Benchmark Function Testing

6.2.1. Test Function Set

Table 6 shows some benchmark functions for CEC2017, from which we selected ten representative functions for testing.

6.2.2. Performance Evaluation Indicators

Algorithm performance was evaluated using the following metrics:
Mean fitness value (Mean): The average result of 30 independent runs.
Standard deviation (Std): Reflects the stability of the algorithm.
Best value: The best result among 30 runs.
Convergent Algebra (Conv): The number of iterations required to achieve threshold accuracy.
Success rate (SR): The proportion of runs that achieve the preset accuracy.

6.2.3. Benchmark Test Results

We show the results of running the relevant algorithms on the F1, F3, and F5 test functions for 30 rounds in Table 7. Obviously, the ILGWO algorithm has certain advantages.

6.2.4. Convergence Performance Analysis

Figure 3 selected five representative functions from the function test set, namely, F1 single peak, F3 multipeak, F5 multipeak (highly multimodal), F7 composite multipeak, and F10 with strong deception, and tested them separately, generating convergence curves. In terms of convergence speed, ILGWO converges quickly in the first 100 iterations, with a convergence speed about 35% faster than GWO and about 60% faster than PSO. In terms of convergence accuracy, ILGWO can converge to higher precision, reaching machine accuracy level on the F1 function. In terms of stability, the convergence curve of ILGWO is smoother and less volatile, indicating that the algorithm has good stability. In terms of global search capability, ILGWO is better able to jump out of local optima on the multimodal function F3, demonstrating the effectiveness of Lévy flight and elite learning strategies.
From the benchmark test results, it can be seen that the ILGWO algorithm is significantly better than the compared algorithms in terms of convergence accuracy, stability, and success rate. However, comparing the greatly improved ILGWO with the original algorithms, such as GWO, PSO, GA, etc., lacks a certain degree of fairness. In order to comprehensively evaluate the performance of the ILGWO algorithm, we conducted comparative experiments with improved metaheuristic algorithms proposed in recent years. Table 8 shows the comparative experimental results between ILGWO and several advanced improved algorithms in recent years. Including Improved Grey Wolf Optimization Algorithm (IGWO) [31], which combines elite reverse learning and Lévy flight strategy; Adaptive Particle Swarm Optimization (APSO) algorithm [32], which adopts dynamic parameter adjustment and diversity maintenance mechanism; Hybrid Genetic Algorithm (HGA) [33], which combines simulated annealing and a local search strategy; Multi-Objective Grey Wolf Algorithm (MOGWO) [34], which was specifically improved for multi-objective optimization problems; and Quantum Behavioral Particle Swarm Optimization (QPSO) [35], which is an improved PSO based on Quantum Mechanics Principles.
Figure 4 shows the convergence results of the algorithm performance comparison experiment.
Statistical significance test:
The Wilcoxon signed-rank test (α = 0.05) was used to statistically compare ILGWO with other improved algorithms, and the results showed the following:
Comparison between ILGWO and IGWS-2023: p = 0.0023 < 0.05, significant difference.
Comparison between ILGWO and APSO-2024: p = 0.0012 < 0.05, significant difference.
Comparison between ILGWO and other algorithms: p-values are all less than 0.05, indicating significant differences.

6.3. System Performance Simulation Experiment

6.3.1. Experimental Scene Setting

A simulation environment was constructed consisting of 20 UAVs, five edge nodes, and one cloud center. UAVs perform tasks within a 100 km × 100 km area, with edge nodes covering a radius of 20 km. The specific parameter configuration is shown in Table 9.

6.3.2. Task Load Change Experiment

Changes in system performance under different task numbers were tested, increasing the number of tasks from 50 to 500. Table 10 shows the system performance under different task scales.
The results indicate that as the number of tasks increases, the performance of all algorithms decreases, but the ILGWO algorithm consistently maintains optimal performance. In large-scale task scenarios, the advantages of ILGWO are especially pronounced. However, as the demand for tasks continues to rise, the system may face overload risks in certain extremely special scenarios. Therefore, we have conducted experiments on extreme scenarios and summarized their patterns as follows:
(1)
Performance degradation law
As the number of tasks increases, the performance of all algorithms shows a non-linear degradation trend. When the number of tasks exceeds 3000, the success rate of traditional algorithms drops to below 70%, while ILGWO can still maintain a success rate of over 80%.
(2)
System bottleneck identification
Under extreme loads, system bottlenecks manifest as a surge in transmission delay caused by communication bandwidth saturation, intensified competition for edge node computing resources, and exponential growth in task scheduling complexity.
(3)
Scalability strategy
In extreme scenarios, the system can increase its scalability by dynamically increasing the number of edge nodes, performing hierarchical task scheduling, and dynamically adjusting task priorities.

6.3.3. Network Condition Impact Experiment

The impact of different network conditions on system performance was tested, including changes in signal strength and interference levels.
Figure 5 shows the impact of network conditions on system performance. From the experimental results, it can be seen that when the signal strength decreases from −80 dBm to −110 dBm, the performance of all algorithms decreases, but the decrease in ILGWO is the smallest, showing better robustness. Secondly, in high-interference environments, ILGWO can maintain good performance by dynamically adjusting power allocation and offloading strategies. In addition, ILGWO can adaptively adjust optimization strategies according to changes in network conditions, reflecting the intelligence and adaptability of the algorithm.

6.3.4. Effect of Integrating Large Language Models

In order to evaluate the improvement effect of the integrated large language models on the intelligent decision-making capability of unmanned aerial vehicle command and control systems, we defined the decision quality indicators shown in Table 11.
Based on this, we tested the improvement effect of the integrated LLM on the system’s intelligent decision-making capability, as shown in Table 12.
From the experimental results, it can be seen that in emergency rescue scenarios, rule-based systems for task priority adjustment cannot take into account various changes in conditions, such as changes in the disaster situation, number of casualties, geographical location, and resources, resulting in secondary solutions. After the LLM is integrated into the system, the large language model can analyze dynamically changing external conditions based on prior knowledge and generate better priority sequences. In terms of dynamic resource allocation, most existing methods are based on load balancing, while the introduction of the LLM can combine multiple constraints, such as task requirements, remaining resources, and energy consumption limitations, to achieve efficient and accurate resource allocation. In terms of handling abnormal situations, the LLM relies on pre-trained knowledge representations to efficiently handle abnormal situations outside the distribution and has good generalization ability.
After integrating the large language model, the intelligent decision-making ability of the system significantly improved, and all decision quality indicators improved by more than 20%. Moreover, the method proposed in this article not only improves the average performance in relevant experimental scenarios but also reduces the standard deviation, indicating that stability also has certain advantages. And the confidence intervals do not overlap, indicating that this approach has significant differences from traditional approaches.

6.3.5. Comparative Experiment with Reinforcement Learning Methods

1.
Experimental setup
This section conducts comparative experiments with mainstream methods to provide specific differences between our method and reinforcement learning methods. We have selected DQN, PPO, SAC, QMIX, and LLM-7B (edge side). The experiment uses AirSim + Gym as the simulation platform, with a map of 10 km × 10 km and 10 UAVs. The obstacle density is set to 50 per square kilometer, and four subtasks are set: target search, collaborative reconnaissance, dynamic obstacle avoidance, and resource allocation. The climate conditions are set to clear, strong wind, and rain and fog, and the communication conditions are set to normal, partially interrupted, and strongly interfered with.
The hyperparameters for reinforcement learning are set as follows: the epoch number is 50,000, the number of steps per round is 1000, the learning rate is 3 × 10−4, the batch size is 256, the discount factor is 0.99, and the experience pool size is 1e6. Evaluation metrics include task success rate, convergence speed, average reward, sample efficiency, generalization performance, and computational cost.
2.
Task scenario design
Based on the characteristics of emergency rescue scenarios, we have designed four tasks, namely, dynamic target search, multi-aircraft collaborative reconnaissance, complex environment obstacle avoidance, and emergency resource scheduling, in order to approximate the actual scenario requirements as closely as possible.
3.
Analysis of experimental results
(1)
Task performance comparison
Figure 6 shows the performance comparison of different methods in four task scenarios. From the results, it can be intuitively concluded that the LLM method has certain advantages in specific scenarios or performs at average levels. The experimental results are shown in Table 13.
From the table, it can be seen that the lightweight large language model has significant advantages in target search and obstacle avoidance navigation scenarios, but it leads with a slight advantage in collaborative reconnaissance and resource scheduling. This is because we roughly selected the lightweight 7B LLM with edge-side quantization compression, and the computing and planning capabilities of the large language model in complex scenarios have not been fully realized.
(2)
Comparison of Learning Efficiency
As shown in Figure 7, the LLM method converges after only 2000 rounds in the same experimental setup. In zero-shot scenarios, the LLM can achieve a success rate of 82.4% without post-training. The specific results are shown in Table 14.
(3)
Comparison of generalization ability
To accurately reflect the generalization advantage of the LLM method, we adjusted the terrain, weather, task, and scale to test the performance retention level of different methods in new scenarios. The experimental results are shown in Table 15.
From the experimental results, it can be seen that the large language model method has significant advantages in unfamiliar scenarios.
(4)
Real-time comparison
Considering that the average delay is more sensitive to extreme values, resulting in significant discrepancies between the actual experience and experimental values of the UAV system, we introduce P95 delay and P99 delay to better capture daily delay experiences and delays in extreme scenarios. Figure 8 is a visualization of comparative experiments using different methods. The left subgraph in Figure 8 reflects the central trend and dispersion of different methods through a block diagram, and points out outliers that exceed the IQR range by 1.5 times. The subgraph on the right of Figure 8 displays the density distribution of delay values using different methods through a violin plot, and marks the real-time threshold and ideal delay (green line), as well as the P95 and P99 percentages. Overall, it reflects that the large language model method has a concentrated distribution and advantages in real-time performance. The specific experimental results are shown in Table 16.
4.
Key findings and discussion
From the experimental results, the LLM method has certain advantages in sample efficiency, generalization ability, and real-time performance. On the one hand, because the large language model learned a lot of prior knowledge during the pre-training stage, it can quickly adapt to the data distribution of new scenes in an unfamiliar environment. On the other hand, in future work, we will consider mixing the LLM with reinforcement learning methods, using the LLM to provide initial strategies to enhance early optimization of reinforcement learning and then improving the decision accuracy of the LLM in specific scenarios through online reinforcement learning. Theoretically, better results can be achieved.

6.4. Ablation Experiment

6.4.1. Analysis of Contribution of Algorithm Components

By gradually removing the key components of the ILGWO algorithm, the contribution of each component to algorithm performance can be analyzed. The average fitness in the table refers to the comprehensive average fitness value of system-level multi-objective optimization, which can be expressed as f =   λ 1   ×   Ψ d e l a y + λ 2   ×   Ψ e n e r g y   +   λ 3   ×   Ψ r e l i a b i l i t y , where Ψ d e l a y represents the standardized delay index, Ψ e n e r g y represents the standardized energy consumption index, Ψ r e l i a b i l i t y represents the standardized reliability index, and λ 1 ,   λ 2 ,   λ 3 represent weight coefficients, the sum of which is 1. We have presented the specific results of the ablation experiment in Table 17.
The results indicate that the Lévy flight strategy leads to the most significant improvement in algorithm performance, followed by adaptive weighting and elite learning mechanisms.

6.4.2. Parameter Sensitivity Analysis

Figure 9 illustrates the impact of key parameters on algorithm performance. Population size: When the population size is between 30 and 60, the algorithm performs better; being too small can affect diversity, while being too large can increase computational costs. Lévy flight parameter β: The optimal range is 1.3–1.7, and the algorithm performs best when β = 1.5. Elite ratio: The optimal range is 15–25%, and a high ratio can lead to insufficient diversity.

6.4.3. Impact of Large Language Model Configuration

The impact of different model configurations on system performance was tested. We have presented the test results of different LLMs in Table 18.
The results indicate that the 7B compressed version model achieves the best balance between performance and resource consumption.

6.5. System Robustness Experiment

6.5.1. Experimental Design

To comprehensively verify the robustness of the system, four types of fault scenario experiments were designed:
(1)
Node Failure Experiment
Fault type: Randomly select edge nodes for fault injection.
Failure rates: 5%, 10%, 15%, 20%, 25%.
Fault duration: 30 s, 60 s, 120 s, 300 s.
Evaluation indicators: Task completion rate, average latency, resource utilization rate.
(2)
Network Interruption Experiment
Interrupt types: Link interruption, bandwidth limitation, high latency.
Interruption intensity: Mild (10–30%), moderate (30–60%), severe (60–90%).
Scope of impact: Local interruption, regional interruption, and network-wide interruption.
Evaluation indicators: Communication success rate, data integrity, retransmission frequency.
(3)
Load impact test
Impact mode: Sudden load, sustained high load, periodic load.
Load intensity: 1.5×, 2×, 3×, 5× normal load.
Impact duration: 10 s, 30 s, 60 s, 180 s.
Evaluation indicators: Response time, throughput, resource overflow rate.
(4)
Environmental interference experiment
Interference sources: Electromagnetic interference, meteorological interference, obstacle obstruction.
Interference intensity: −5 dB, −10 dB, −15 dB, −20 dB, signal-to-noise ratio decrease.
Interference modes: Continuous interference, intermittent interference, random interference.
Evaluation indicators: Signal quality, bit error rate, connection stability.

6.5.2. Experimental Environment Configuration

Simulation parameter settings: Simulation time: 1800 s; number of UAVs: 30; edge nodes: 8; cloud center: 1; task load: 800 concurrent tasks; fault detection cycle: 5 s; recovery strategy: active redundancy + passive reconfiguration. We present the comparative results of task completion effectiveness under different failure rates in Table 19. In addition, we set up other differentiated experimental scenarios, and Table 20 shows the system performance under different network interruption intensities; Table 21 shows the changes in performance under different loads; Table 22 shows the communication quality under different interference strengths.

6.5.3. Experimental Results

(1)
Node fault tolerance test
Table 19. System performance under different failure rates.
Table 19. System performance under different failure rates.
Failure RateAlgorithmTask Completion Rate (%)Average Latency (ms)Resource Utilization Rate (%)Recovery Time (s)
5%ILGWO97.3178.489.212.3
GWO94.8203.783.618.7
PSO91.2245.378.125.4
GA87.6287.972.832.1
10%ILGWO94.7198.685.415.8
GWO89.3238.278.924.3
PSO84.6291.771.234.7
GA78.9345.865.343.2
15%ILGWO91.8223.581.719.4
GWO83.2278.973.231.6
PSO76.4342.164.845.9
GA68.7412.356.958.7
20%ILGWO88.4251.277.324.1
GWO76.8324.666.839.8
PSO67.9398.757.158.3
GA57.2489.447.676.5
(2)
Analysis of the Impact of Network Interruptions
Table 20. System performance under different network interruption intensities.
Table 20. System performance under different network interruption intensities.
Interruption StrengthCommunication Success Rate (%)Data Integrity (%)Average Number of RetransmissionsEnd-to-End Delay (ms)
Mild (10–30%)94.698.71.23189.4
Moderate (30–60%)87.395.22.67267.8
Severe (60–90%)76.889.64.58398.2
Network-wide interruption45.267.38.94756.3
(3)
Load impact adaptability test
Table 21. System response under sudden load.
Table 21. System response under sudden load.
Load MultiplierResponse Time (ms)Throughput (ops/s)Resource Overflow Rate (%)Task Loss Rate (%)
1.5×198.71847.32.10.8
267.41623.85.72.3
389.61289.412.46.8
567.8892.123.615.7
(4)
Environmental interference resistance test
Table 22. Communication quality under different interference intensities.
Table 22. Communication quality under different interference intensities.
Signal-to-Noise RatioDecrease Signal Quality (RSSI)Error RateConnection Stability (%)Switching Frequency (Times/min)
−5 dB−67.3 dBm1.2 × 10−496.80.7
−10 dB−72.1 dBm3.8 × 10−492.41.9
−15 dB−78.6 dBm9.6 × 10−485.74.2
−20 dB−85.2 dBm2.3 × 10−376.38.5

6.5.4. Cascade Fault Analysis

To verify the system’s ability to resist cascading faults, a cascading fault experiment was designed. Figure 10 shows the dynamic topology of cascading fault propagation. In addition, Table 23 presents the results of cascading fault propagation analysis.
Figure 11 shows the performance evolution and recovery performance comparison during cascading faults, visually demonstrating the performance advantages of the proposed algorithm ILGWO in this paper.

6.5.5. Analysis of Experimental Results

(1)
Stability advantage analysis
The experimental results show that the ILGWO algorithm has significant advantages in system robustness: (1) Fault tolerance. At a node failure rate of 20%, ILGWO can still maintain a task completion rate of 88.4%, which is 11.6 percentage points higher than GWO. (2) Quick recovery. The average fault recovery time is reduced by 35–40% compared to traditional algorithms, reflecting the algorithm’s adaptive reconfiguration capability. (3) Slow performance degradation. Under moderate interference, the performance degradation amplitude is controlled within 15%, demonstrating good anti-interference ability. (4) Cascade control. The propagation of cascading faults is effectively controlled, and the number of cascading connections is kept at a low level.
(2)
Key technological contributions
(1) Multi-layer redundancy mechanism. The cloud–edge–terminal architecture provides multiple layers of backup, and a single point of failure will not cause system crashes. (2) Dynamic load balancing. The Lévy flight search strategy enables the algorithm to quickly find alternative resource allocation solutions. (3) Intelligent fault detection. Anomaly detection based on large language models can identify potential faults in advance. (4) Adaptive reconfiguration. The elite learning mechanism ensures rapid resource reallocation after a malfunction.
(3)
Practical application significance
The robustness experiment verified the reliability of the system in actual complex environments, mainly reflected in the following aspects. (1) Applicable to critical scenarios of the task. In critical tasks such as emergency rescue, the system can continue to operate in the event of partial equipment failure. (2) Strong adaptability to the environment. It can still maintain basic functions under conditions such as electromagnetic interference and severe weather. (3) Good scalability. As the system scale expands, the robustness indicators can still be maintained within an acceptable range. (4) Low maintenance cost. The self-healing ability reduces the need for manual intervention and lowers operational costs.

6.6. Discussion of Experimental Results

6.6.1. Analysis of Algorithmic Performance Advantages

The superior performance of the ILGWO algorithm is mainly attributed to the following aspects. The first is initialization optimization: chaotic-map initialization provides a better population distribution, offering the algorithm a favorable starting point. The second is enhancement of the search strategy: the Lévy flight mechanism strengthens the algorithm’s global search capability, enabling it to escape local optima. The third is adaptive mechanisms: dynamic weight adjustment allows the algorithm to achieve a better balance between exploration and exploitation. The fourth is elite preservation: the elite learning strategy ensures the retention and evolution of high-quality solutions.

6.6.2. Analysis of System Architecture Advantages

The “cloud–edge–terminal” architecture aligns with the existing network information system, and its advantages are reflected in the following. The first is hierarchical processing: tasks of different complexities are handled at appropriate layers, improving overall efficiency. The second is intelligent decision-making: the integration of large language models significantly enhances the system’s intelligence. The third is resource optimization: dynamic resource allocation strategies improve resource utilization efficiency. The fourth is strong robustness: multi-layer backup mechanisms ensure system reliability.

6.6.3. Practical Application Value

Experimental verification shows that the designed system possesses practical application value, as it introduces the intelligent capabilities of large language models into UAV command and control information systems. First, performance is significantly improved: key metrics exhibit gains of 20–60%. Second, adaptability is strong: the system can adapt to different application scenarios and environmental conditions. Third, scalability is excellent: it supports coordinated operations of large-scale UAV swarms. Fourth, deployment is convenient: based on standardized interfaces, it can be easily integrated with existing systems.

6.6.4. Scalability Analysis

The system designed in this article has three advantages in scalability: system layer, algorithm layer, and data layer, as follows:
(1)
Scalability advantage at the system level. This is mainly reflected in the following aspects: Elastic architecture: The three-layer architecture, consisting of the cloud, edge, and end, naturally supports elastic expansion, and each layer can be independently expanded without affecting the operation of the other layers. Standardized interfaces: Using standardized communication protocols and interface specifications ensures seamless integration of heterogeneous devices and systems. Service-oriented design and microservice architecture: Functional modules can be independently developed, deployed, and expanded, improving system flexibility.
(2)
Scalability advantage at the algorithmic level. This is mainly reflected in the following aspects: Parallel processing capability: The parallel nature of the ILGWO algorithm enables it to fully utilize multi-core and distributed computing resources. Parameter adaptation: The algorithm parameters can be dynamically adjusted according to the size of the problem and resource conditions, ensuring optimization effectiveness at different scales. Decomposition strategy: The decomposition and solution of large-scale problems are supported, reducing the complexity of individual optimization tasks.
(3)
Scalability advantage at the data flow level. This is mainly reflected in the following aspects: Hierarchical processing: Data are preprocessed and filtered at different levels to reduce the processing pressure on core nodes. Cache mechanism: The multi-level caching strategy improves data access efficiency and supports large-scale concurrent access. Flow control: Intelligent traffic scheduling and load balancing ensure the stable operation of the system under high loads.

6.6.5. System Model Validation and Deep Analysis

(1)
Communication model validation
To verify the accuracy of the communication model established in Section 4.2, we conducted actual testing and simulation comparison experiments.
Validation of air–ground channel model. The testing environment includes three typical environments: urban, suburban, and rural. The altitude of the UAV is 50 m–500 m, with an interval of 50 m. The horizontal distance is 1 km–10 km, with an interval of 1 km. The test indicators are path loss, channel gain, and LoS probability. Table 24 shows the specific results of the communication model validation.
The verification results show that the average relative error of the established communication model is less than 2%, which meets the requirements of engineering applications.
(2)
Computational model validation
Edge computing delay model validation. By deploying computing tasks of different complexities on the actual MEC platform, the computing model established in Section 4.3.2 is verified. Table 25 shows the validation results of the delay model.
(3)
Energy consumption model validation
Flight energy consumption model validation. The flight energy consumption model established in Section 4.5.3 is verified using actual UAV flight data. Table 26 shows the validation results of the flight energy consumption model.
From the above system model experiments, it can be seen that the model can be further improved in the following directions in the future: the first improvement is to introduce machine learning methods to enhance the model’s adaptive ability; the second is to increase the dynamic adjustment of model parameters by environmental factors; the third is to establish a multi-level model validation system to improve the credibility of the model. These verification results indicate that the system model established in this paper has good accuracy and practicality, providing a reliable theoretical basis for the design and optimization of unmanned aerial vehicle command and control systems.

6.7. Actual Scenario Verification Experiment

In order to verify the performance of the system in real environments, we collaborated with Beihang University to deploy and test the system in simulated urban emergency rescue scenarios. The specific experimental environment deployment, test scenario design, and key performance indicator testing are provided in Appendix A.
In addition, we selected an emergency rescue drill in a practical scenario experiment to verify the performance of the system proposed in this paper in complex urban rescue. The scene of the experimental site includes factors such as building obstruction, multipath effects, and high-density interference, which have a strong degree of realistic simulation. Figure 12 shows a comparison of system performance in urban emergency rescue scenarios.
The experimental results show that among several methods, the method proposed in this paper has certain advantages in urban rescue scenarios. Capable of handling over 30 concurrent tasks simultaneously, with a response time of less than 500 ms, the system has strong parallel capability. Secondly, in an environment with building obstructions and electromagnetic interference, the system maintains communication quality through intelligent routing and power control. However, there are still some challenges in actual testing that highlight aspects that need to be optimized and improved in future work. The first goal is to increase the wind resistance and positioning accuracy of the UAV, the second is to expand the training data by including Chinese dialects and professional terminology, and the third is to optimize the system architecture to achieve larger-scale concurrency.

7. Summary

Aiming to solve the problem of the UAV command and control system’s inability to perform intelligent decision-making and resource allocation, this paper proposes a “cloud, edge, and end” command and control system based on a large language model. With 5G as the backbone network, the Internet of Things as a supplement, edge computing as the computing platform, and the large language model as the intelligent engine, this system enhances the system’s intelligent decision-making, environmental adaptation, and resource allocation abilities. The main contributions are as follows:
(1)
System architecture contribution. A three-layer distributed architecture empowered by large language models has been proposed, which achieves the optimal matching of intelligent decision-making capabilities and computing resources through the layered deployment of LLMs of different scales (cloud 175B, edge 70B, terminal 7B). The system supports elastic expansion from 10 to over 1000 UAVs, with good scalability.
(2)
Contribution to modeling theory. A multidimensional system efficiency optimization model covering communication, computing, and energy consumption has been established, which, for the first time, incorporates the inference delay and accuracy loss of the large language model into the system modeling, providing a theoretical basis for the quantitative analysis of unmanned aerial vehicle command and control systems.
(3)
Contribution of algorithmic technology. The ILGWO algorithm is proposed, which integrates Lévy flight, adaptive weighting, elite learning, and other strategies, and outperforms traditional algorithms in terms of convergence accuracy and speed. The benchmark function test shows that the algorithm’s convergence accuracy and convergence speed are improved by 35% and 60%, respectively.
(4)
Application verification contribution. The effectiveness of the system has been demonstrated through large-scale simulation experiments and practical scenario verification. Task latency, energy efficiency, and resource utilization are improved by 34.2%, 29.6%, and 31.8%, respectively. In actual emergency rescue exercises, the response time was shortened by 44.7%, and collaborative efficiency was improved by 39.5%.
In addition, the system proposed in this article has the following advantages:
(1)
Advantages of intelligent decision-making. By introducing the LLM, the accuracy of the system in task priority adjustment, dynamic resource allocation, exception handling, collaborative path planning, and other aspects exceeds 90%, which is 20–30% higher than that of traditional rule systems.
(2)
System robustness advantage. Despite a 20% node failure rate, it can still maintain a task completion rate of 88.4%, with the average fault recovery time reduced by 35–40% compared to traditional algorithms, demonstrating excellent fault tolerance.
(3)
Real-time advantage. The inference delay of the 7B model deployed on the edge side is only 156 ms, which can meet the real-time decision-making requirements of emergency scenarios.
(4)
Generalization ability advantage. In scenarios such as new terrain, new climate, new tasks, and scale adjustments, the performance retention rate reaches 82.5%, significantly better than the reinforcement learning method’s 56.7%.
However, the system still faces the following challenges in practical applications:
(1)
Calculation resource requirements. The deployment of large language models requires significant computing resources and storage space, which may pose challenges in environments with extremely limited resources.
(2)
Communication bandwidth dependence. Cloud–edge collaboration requires a stable, high-bandwidth communication link, and system performance may decrease in harsh communication conditions.
(3)
Standardization level. The compatibility between the system and existing equipment needs to be further verified, and industry-standardized interface specifications need to be established.
With the continuous maturation and improvement of related technologies, this system is expected to play an important role in more fields, providing strong support for social and economic development and the improvement of people’s lives.

Author Contributions

Conceptualization, S.H. and C.F.; methodology, S.H. and P.W.; software, Z.L.; validation, M.W., D.L. and Z.L.; formal analysis, S.H.; investigation, S.H.; resources, C.F. and Z.L.; data curation, D.L.; writing—original draft preparation, S.H.; writing—review and editing, S.H.; visualization, P.W.; supervision, P.W.; project administration, C.F.; funding acquisition, C.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by [Shaanxi Province Natural Science Basic Research Program] grant number [2025JC-YBMS-804, 2025JC-YBMS-691] and The APC was funded by [2025JC-YBMS-804].

Data Availability Statement

Due to the integration of the improved algorithm and simulation environment proposed in this article as the research object, namely, a large-scale model-based unmanned aerial vehicle command information system, the entire system and source code settings are currently in the patent application stage. Considering the standardization of the patent review workflow, we have not yet released the source code. In a later stage, when the patent review is approved, we will promptly publish the source code at https://github.com/YMLLM/UAV_sys-based-LLM (accessed on 3 September 2025).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

A.1. Experimental Environment and Deployment

Experimental location: A city complex in a certain city (with an area of approximately 2 km × 2 km). Equipment involved: UAVs: 8 DJI M300 RTKs equipped with Hesai LiDAR and electro-optical pods; edge node: three mobile edge computing vehicle platforms (Intel i9-11900K + RTX 3090); central layer: Huawei Cloud Server Cluster (8 cores, 32 GB × 4 units); communication: China Mobile’s 5G private network coverage. (DJI M300 RTKs are manufactured by Chinese company DJI; Hesai LiDAR company has offices in Shanghai, Silicon Valley, Stuttgart and other places in China)
Deployment architecture: Central layer: Deploys the LLaMA-2 175B model, responsible for global decision coordination. Edge layer: Each edge node deploys a 70B compression model to handle real-time task scheduling. Terminal: The UAV is equipped with a 0.7B–7B lightweight model as needed for emergency decision-making.

A.2. Test Scenario Design

Scenario 1: Search and rescue for building collapse.
Task description: Simulate building collapse after an earthquake and search for trapped individuals.
Test area: 2000 m2, including 3 simulated collapsed buildings.
Six UAVs participated, each responsible for different tasks (thermal imaging search, structural detection, communication relay).
Scenario 2: Large-scale fire monitoring.
Task description: Simulate a fire in a chemical industrial park, requiring real-time monitoring of the spread of the fire.
Test area: 5000 m2 open area, with multiple smoke generation points set up.
Participated in 4 UAVs for regional cruising and hotspot positioning
Scenario 3: Emergency response to communication interruption.
Task Description: Simulate ground communication interruption caused by extreme weather conditions.
Test conditions: Artificially blocking some 5G signals, simulating network instability.
Response measures: Unmanned aerial vehicle’s self-organizing network, establishing temporary communication links.

A.3. Key Performance Indicator Testing

A.3.1. Real-Time Response Capability Test

Test method: Randomly initiate 50 emergency tasks during the search and rescue process.
Test indicator: Time from receiving the task to the arrival of the UAV on site.
Comparison of results:
Traditional scheduling system: Average response time of 3.8 min.
This article’s system has an average response time of 2.1 min.
Performance improvement: 44.7%.

A.3.2. Efficiency Testing of Collaborative Work

Test scenario: Eight UAVs collaborate to search a 2000 m2 area.
Evaluation indicators: Search completion time and coverage rate.
Comparison of results:
Manual command mode: Completion time 18.5 min, coverage rate 87.3%.
This article’s system: Completion time 11.2 min, coverage rate 96.8%.
Efficiency improvement: Time reduced by 39.5%, coverage increased by 10.9%.

A.3.3. Exception Handling Capability Test

Test method: Intentionally causing equipment malfunctions and communication interruptions during the exercise.
Test case:
Case 1: UAV 1 emergency landing due to battery depletion.
Traditional system: Requires manual reassignment of tasks, takes 4.2 min.
This article’s system automatically detects and reallocates, taking 47 s.
Case 2: Communication interruption at edge node 2.
Traditional system: The task in this area is paused and waiting for manual intervention.
This article’s system automatically switches to cloud processing, with increased latency but continuation of tasks.
Case 3: GPS signal interference leads to inaccurate positioning.
Traditional system: UAVs return and wait for signal recovery.
This article’s system: Enables visual SLAM backup positioning; the task is proceeding normally.

References

  1. Zhou, W.Q.; Zhu, J.H.; Kuang, M.C.; Shi, H.; Zhou, W.; Zhu, J.; Kuang, M.; Shi, H. Multi-UAV Cooperative Swarm Algorithm in Air Combat Based on Predictive Game Tree. Sci. Sin. Technol. 2023, 53, 187–199. [Google Scholar] [CrossRef]
  2. Xiao, X.; Wang, X.; Lin, W. Joint AoI-Aware UAVs Trajectory Planning and Data Collection in UAV-Based IoT Systems: A Deep Reinforcement Learning Approach. IEEE Trans. Consum. Electron. 2024, 70, 6484–6495. [Google Scholar] [CrossRef]
  3. Xing, X.; Zhou, Z.; Li, Y.; Xiao, B.; Xun, Y. Multi-UAV Adaptive Cooperative Formation Trajectory Planning Based on an Improved MATD3 Algorithm of Deep Reinforcement Learning. IEEE Trans. Veh. Technol. 2024, 73, 12484–12499. [Google Scholar] [CrossRef]
  4. Wang, L.; Zheng, S.; Tai, S.; Liu, H.; Yue, T. UAV Air Combat Autonomous Trajectory Planning Method Based on Robust Adversarial Reinforcement Learning. Aerosp. Sci. Technol. 2024, 153, 109402. [Google Scholar] [CrossRef]
  5. Han, S.; Wang, M.; Zhang, J.; Li, D.; Duan, J. A Review of Large Language Models: Fundamental Architectures, Key Technological Evolutions, Interdisciplinary Technologies Integration, Optimization and Compression Techniques, Applications, and Challenges. Electronics 2024, 13, 5040. [Google Scholar] [CrossRef]
  6. Radford, A.; Narasimhan, K.; Salimans, T.; Sutskever, I. Improving Language Understanding by Generative Pre-Training. 2018. Available online: https://www.semanticscholar.org/paper/Improving-Language-Understanding-by-Generative-Radford-Narasimhan/cd18800a0fe0b668a1cc19f2ec95b5003d0a5035 (accessed on 3 September 2025).
  7. Touvron, H.; Lavril, T.; Izacard, G.; Martinet, X.; Lachaux, M.-A.; Lacroix, T.; Rozière, B.; Goyal, N.; Hambro, E.; Azhar, F.; et al. LLaMA: Open and Efficient Foundation Language Models. arXiv 2023, arXiv:2302.13971. [Google Scholar] [CrossRef]
  8. DeepSeek-AI; Guo, D.; Yang, D.; Zhang, H.; Song, J.; Zhang, R.; Xu, R.; Zhu, Q.; Ma, S.; Wang, P.; et al. DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning. arXiv 2025, arXiv:2501.12948. [Google Scholar]
  9. Zhong, Y.; Zhang, Y.; Zhang, J.; Wan, L. Research on Resilience Measurement and Optimization of Networked Command Information System. Reliab. Eng. Syst. Saf. 2025, 261, 111048. [Google Scholar] [CrossRef]
  10. Zhong, Y.; Li, H.; Zhuang, X. A Resilience-Driven Two-Stage Operational Chain Optimization Model for Unmanned Weapon System-of-Systems under Limited Resource Environments. Eksploat. Niezawodn. Maint. Reliab. 2024, 26, 124. [Google Scholar] [CrossRef]
  11. Johansson, B.; Tordenlid, J.; Lundberg, J.; Alfredson, J. One System to Connect Them All-a Core System for Realizing Integrated Command and Control Research. In Proceedings of the 21st International ISCRAM Conference, Münster, Germany, 25–29 May 2024. [Google Scholar] [CrossRef]
  12. Radovanović, M.; Petrovski, A.; Žnidaršič, V. The C5ISR System Integrated with Unmanned Aircraft in the Large-Scale Combat Operations. Vojen. Rozhl. 2023, 32, 98–118. [Google Scholar] [CrossRef]
  13. Wang, L.; Jiao, H. Multi-Agent Reinforcement Learning-Based Computation Offloading for Unmanned Aerial Vehicle Post-Disaster Rescue. Sensors 2024, 24, 8014. [Google Scholar] [CrossRef]
  14. Liu, Z.; Oguz, B.; Zhao, C.; Chang, E.; Stock, P.; Mehdad, Y.; Shi, Y.; Krishnamoorthi, R.; Chandra, V. LLM-QAT: Data-Free Quantization Aware Training for Large Language Models. arXiv 2023, arXiv:2305.17888. [Google Scholar]
  15. Ma, X.; Fang, G.; Wang, X. LLM-Pruner: On the Structural Pruning of Large Language Models. In Proceedings of the Advances in Neural Information Processing Systems 36 (Neurips 2023); Oh, A., Neumann, T., Globerson, A., Saenko, K., Hardt, M., Levine, S., Eds.; Neural Information Processing Systems (nips): La Jolla, CA, USA, 2023. [Google Scholar]
  16. Zhou, P.Y.; Fu, S.; Finley, B.; Li, X.; Tarkoma, S.; Kangasharju, J.; Ammar, M.; Hui, P. 5G MEC Computation Handoff for Mobile Augmented Reality. In Proceedings of the 2024 IEEE International Conference on Metaverse Computing, Networking, and Applications, Metacom 2024, Hong Kong, China, 12–14 August 2024; IEEE Computer Soc: Los Alamitos, UC, USA, 2024; pp. 129–136. [Google Scholar]
  17. Vatten, T. Investigating 5G Network Slicing Resilience through Survivability Modeling. In Proceedings of the 2023 IEEE 9th International Conference on Network Softwarization, Netsoft, Madrid, Spain, 19–23 June 2023; Bernardos, C.J., Martini, B., Rojas, E., Verdi, F.L., Zhu, Z., Oki, E., Parzyjegla, H., Eds.; IEEE: New York, NY, USA, 2023; pp. 370–373. [Google Scholar]
  18. Chen, Q.; He, Y.; Yu, G.; Xu, C.; Liu, M.; Li, Z. KBMP: Kubernetes-Orchestrated IoT Online Battery Monitoring Platform. IEEE Internet Things J. 2024, 11, 25358–25370. [Google Scholar] [CrossRef]
  19. Chen, R.; Ding, Y.; Zhang, B.; Li, S.; Liang, L. Air-to-Ground Cooperative OAM Communications. IEEE Wirel. Commun. Lett. 2024, 13, 1063–1067. [Google Scholar] [CrossRef]
  20. Hoffmann, J.; Borgeaud, S.; Mensch, A.; Buchatskaya, E.; Cai, T.; Rutherford, E.; de Las Casas, D.; Hendricks, L.A.; Welbl, J.; Clark, A.; et al. Training Compute-Optimal Large Language Models. In Proceedings of the Advances in Neural Information Processing Systems 35 (Neurips 2022), New Orleans, LA, USA, 28 November–9 December 2022; Koyejo, S., Mohamed, S., Agarwal, A., Belgrave, D., Cho, K., Oh, A., Eds.; Neural Information Processing Systems (nips): La Jolla, CA, USA, 2022. [Google Scholar]
  21. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems 30 (Nips 2017), Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Neural Information Processing Systems (nips): La Jolla, CA, USA, 2017; Volume 30. [Google Scholar]
  22. Narayanan, D.; Shoeybi, M.; Casper, J.; LeGresley, P.; Patwary, M.; Korthikanti, V.A.; Vainbrand, D.; Kashinkunti, P.; Bernauer, J.; Catanzaro, B.; et al. Efficient Large-Scale Language Model Training on GPU Clusters Using Megatron-LM. In Proceedings of the International Conference for High Performance Computing, Networking, Storage and Analysis, St. Louis, MO, USA, 14–19 November 2021. [Google Scholar]
  23. Wang, W.; Wei, F.; Dong, L.; Bao, H.; Yang, N.; Zhou, M. Minilm: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers. In Proceedings of the Advances in Neural Information Processing Systems 33, Neurips 2020, Online, 6–12 December 2020; Larochelle, H., Ranzato, M., Hadsell, R., Balcan, M.F., Lin, H., Eds.; Neural Information Processing Systems (nips): La Jolla, CA, USA, 2020; Volume 33. [Google Scholar]
  24. Jiao, X.; Yin, Y.; Shang, L.; Jiang, X.; Chen, X.; Li, L.; Wang, F.; Liu, Q. TinyBERT: Distilling BERT for Natural Language Understanding. In Proceedings of the Findings of the Association for Computational Linguistics, Emnlp 2020, Online, 16–20 November 2020; Cohn, T., He, Y., Liu, Y., Eds.; Assoc Computational Linguistics-Acl: Stroudsburg, PA, USA, 2020; pp. 4163–4174. [Google Scholar]
  25. Sanh, V.; Debut, L.; Chaumond, J.; Wolf, T. DistilBERT, a Distilled Version of BERT: Smaller, Faster, Cheaper and Lighter. arXiv 2019, arXiv:1910.01108. [Google Scholar]
  26. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. arXiv 2015, arXiv:1503.02531. [Google Scholar] [CrossRef]
  27. Qiu, Y.; Yang, X.; Chen, S. An Improved Gray Wolf Optimization Algorithm Solving to Functional Optimization and Engineering Design Problems. Sci. Rep. 2024, 14, 14190. [Google Scholar] [CrossRef]
  28. Liu, J.; Deng, Y.; Liu, Y.; Chen, L.; Hu, Z.; Wei, P.; Li, Z. A Logistic-Tent Chaotic Mapping Levenberg Marquardt Algorithm for Improving Positioning Accuracy of Grinding Robot. Sci. Rep. 2024, 14, 9649. [Google Scholar] [CrossRef]
  29. Armstrong, G.; Bogdan, K.; Grzywny, T.; Lezaj, L.; Wang, L. Yaglom Limit for Unimodal Lévy Processes. Ann. L’institut Henri Poincare (B) Probab. Stat. 2023, 59, 1688–1721. [Google Scholar] [CrossRef]
  30. Maulana, A.H.; Wellem, T. Studi Modul Wifi Dan Simulasi Jaringan Ieee 802.11ax Pada Simulator Ns-3 Untuk Pengukuran Throughput. 3. Available online: https://ejournal.uksw.edu/itexplore/article/view/11955 (accessed on 3 September 2025).
  31. Alsheikh, N.M.; Munassar, N.M. Improving Software Effort Estimation Models Using Grey Wolf Optimization Algorithm. IEEE Access 2023, 11, 143549–143579. [Google Scholar] [CrossRef]
  32. Yang, S.; Wei, B.; Deng, L.; Jin, X.; Jiang, M.; Huang, Y.; Wang, F. A Leader-Adaptive Particle Swarm Optimization with Dimensionality Reduction Strategy for Feature Selection. Swarm Evol. Comput. 2024, 91, 101743. [Google Scholar] [CrossRef]
  33. Zheng, J.; Zhong, J.; Chen, M.; He, K. A Reinforced Hybrid Genetic Algorithm for the Traveling Salesman Problem. Comput. Oper. Res. 2023, 157, 106249. [Google Scholar] [CrossRef]
  34. Torabi, A.; Yosefvand, F.; Shabanlou, S.; Rajabi, A.; Yaghoubi, B. Optimization of Integrated Operation of Surface and Groundwater Resources Using Multi-Objective Grey Wolf Optimizer (MOGWO) Algorithm. Water Resour. Manag. 2024, 38, 2079–2099. [Google Scholar] [CrossRef]
  35. Rugveth, V.S.; Khatter, K. Sensitivity Analysis on Gaussian Quantum-Behaved Particle Swarm Optimization Control Parameters. Soft Comput. 2023, 27, 8759–8774. [Google Scholar] [CrossRef]
Figure 1. Architecture of UAV command and control system based on LLM.
Figure 1. Architecture of UAV command and control system based on LLM.
Drones 09 00639 g001
Figure 2. System workflow diagram.
Figure 2. System workflow diagram.
Drones 09 00639 g002
Figure 3. Convergence curve of algorithm testing. (F1) is the comparison result between ILGWO and classical algorithms on the F1 function; (F3) is the comparison result between ILGWO and classical algorithms on the F3 function; (F5) is the comparison result between ILGWO and classical algorithms on the F5 function; (F7) is the comparison result between ILGWO and classical algorithms on the F7 function; (F10) is the comparison result between ILGWO and classical algorithms on the F10 function.
Figure 3. Convergence curve of algorithm testing. (F1) is the comparison result between ILGWO and classical algorithms on the F1 function; (F3) is the comparison result between ILGWO and classical algorithms on the F3 function; (F5) is the comparison result between ILGWO and classical algorithms on the F5 function; (F7) is the comparison result between ILGWO and classical algorithms on the F7 function; (F10) is the comparison result between ILGWO and classical algorithms on the F10 function.
Drones 09 00639 g003
Figure 4. Performance comparison experiment between ILGWO algorithm and improved algorithm.
Figure 4. Performance comparison experiment between ILGWO algorithm and improved algorithm.
Drones 09 00639 g004
Figure 5. Impact of network conditions on system performance.
Figure 5. Impact of network conditions on system performance.
Drones 09 00639 g005
Figure 6. Performance comparison of different methods in four task scenarios.
Figure 6. Performance comparison of different methods in four task scenarios.
Drones 09 00639 g006
Figure 7. Comparative experimental results of convergence performance.
Figure 7. Comparative experimental results of convergence performance.
Drones 09 00639 g007
Figure 8. Real-time comparison experiment results.
Figure 8. Real-time comparison experiment results.
Drones 09 00639 g008
Figure 9. Results of parameter sensitivity analysis.
Figure 9. Results of parameter sensitivity analysis.
Drones 09 00639 g009
Figure 10. Dynamic diagram of cascading fault propagation.
Figure 10. Dynamic diagram of cascading fault propagation.
Drones 09 00639 g010
Figure 11. Performance evolution and recovery performance comparison of cascade faults.
Figure 11. Performance evolution and recovery performance comparison of cascade faults.
Drones 09 00639 g011
Figure 12. System performance in urban emergency rescue scenarios.
Figure 12. System performance in urban emergency rescue scenarios.
Drones 09 00639 g012
Table 1. Comparative analysis of different intelligent decision-making methods.
Table 1. Comparative analysis of different intelligent decision-making methods.
Characteristic DimensionReinforcement LearningRule EngineTraditional Machine LearningLarge Language Models
Knowledge sourceEnvironment interactionExpert experienceAnnotation dataPre-training fine-tuning
Learning paradigmTrial and errorReward symbol reasoningSupervised learningSelf-monitoring
Reasoning mechanismValue and Strategy FunctionLogic reasoningFeature MappingAttention mechanism
Time complexity O n 2 O n O n l o g n O n 2 d
Space complexity O S x A
Sample efficiency Low   ( 10 6 levels)No need for samples Medium   ( 10 3 levels) High   ( 10 2 levels)
Table 2. Performance comparison of intelligent decision-making methods.
Table 2. Performance comparison of intelligent decision-making methods.
MethodDecision AccuracyResponse TimeGeneralization AbilityTraining CostReasoning Memory
DQN89.3% ± 2.1%120 ms68.5%2000 GPU-h4 GB
PPO91.2% ± 1.8%135 ms71.2%1500 GPU-h3.5 GB
Rule engine68.9% ± 3.5%23 ms45.3%00.5 GB
Random forest82.4% ± 2.3%85 ms76.8%50 CPU-h2 GB
This article’s LLM-7B91.2% ± 1.5%156 ms85.6%200 GPU-h 27 GB
This article’s LLM-3B87.4% ± 1.8%89 ms82.3%100 GPU-h12 GB
Table 3. Example of offloading UAV tasks in emergency rescue scenarios.
Table 3. Example of offloading UAV tasks in emergency rescue scenarios.
Task DomainSpecific Task Type Specific Task Type Task Offloading Hierarchy
Natural Language UnderstandingInstruction translationFly to a certain coordinatesTerminal Layer
Semantic understandingMultistep task planning in fire scenariosEdge layer
Complex reasoningEmergency response plan generation for this scenarioCentral layer
Image RecognitionBasic object detectionReal-time obstacle recognitionTerminal Layer
Scene understanding analysisDisaster situation assessmentEdge layer
Visual reasoningGlobal situation analysisCentral layer
Table 4. Experimental data of mainstream LLM compression.
Table 4. Experimental data of mainstream LLM compression.
Original ModelModel CompressionCompression RatioAccuracy Retention Rate λ d i s t i l l
GPT-3 175BGPT-3 13B13.50.870.051
BERT-LargeBERT-Base3.00.950.042
T5-11BT5-3B3.70.910.068
LLaMA-65BLLaMA-7B9.30.820.088
Table 5. Typical UAV testing tasks.
Table 5. Typical UAV testing tasks.
Task TypeImage RecognitionVideo AnalysisSensor Data FusionPath PlanningNatural Language Processing
Data size1–5 MB5–20 MB50–200 KB500 KB-2 MB10–100 KB
Computational complexity0.8–2.5 G cycle2–8 G cycle100–500 M cycle500 M–3 G cycle1–10 G cycle
Delay requirements100–500 ms0.5–2 s10–50 ms200 ms–1 s50–200 ms
Accuracy requirementsOver 95%Over 90%Over 98%Over 99%Over 85%
Table 6. Characteristics of benchmark functions.
Table 6. Characteristics of benchmark functions.
FunctionNameTypeDimensionSearch IntervalGlobal Optimal
F1SphereSingle peak30[−100, 100]0
F2RosenbrockMultimodal30[−30, 30]0
F3RastriginMultimodal30[−5.12, 5.12]0
F4GriewankMultimodal30[−600, 600]0
F5AckleyMultimodal30[−32, 32]0
F6WeierstrassMultimodal30[−0.5, 0.5]0
F7SchwefelMultimodal30[−500, 500]0
F8KatsuuraMultimodal30[−100, 100]0
F9LunacekMultimodal30[−5, 5]0
F10Shifted SphereSingle peak30[−100, 100]100
Table 7. Comparative experimental results of ILGWO and several classic algorithms.
Table 7. Comparative experimental results of ILGWO and several classic algorithms.
FunctionAlgorithmMeanStdBestConvSR (%)
F1ILGWO2.31 × 10−158.45 × 10−161.23 × 10−1878100
GWO3.67 × 10−121.23 × 10−114.56 × 10−1514295
PSO1.45 × 10−83.21 × 10−82.31 × 10−1029878
GA2.34 × 10−51.12 × 10−48.91 × 10−745645
F3ILGWO0.00240.0156015695
GWO2.453.780.018923472
PSO15.6712.343.4538723
GA45.2323.4512.784898
F5ILGWO4.12 × 10−92.34 × 10−88.88 × 10−1489100
GWO1.23 × 10−64.56 × 10−63.45 × 10−816789
PSO0.00340.01231.23 × 10−529856
GA0.2340.5670.034544512
Table 8. Comparative experimental results of ILGWO and several advanced improved algorithms.
Table 8. Comparative experimental results of ILGWO and several advanced improved algorithms.
FunctionAlgorithmMeanStdBestConvSR(%)
F1ILGWO2.31 × 10−158.45 × 10−161.23 × 10−1878100
IGWO-20234.56 × 10−141.23 × 10−132.34 × 10−169895
APSO-20241.23 × 10−123.45 × 10−125.67 × 10−1513490
HGA-20232.34 × 10−116.78 × 10−111.23 × 10−1315685
MOGWO-20243.45 × 10−138.91 × 10−134.56 × 10−1611292
QPSO-20235.67 × 10−121.45 × 10−118.91 × 10−1514588
F2ILGWO0.01560.02340.001216793
IGWO-20230.09230.14560.014523485
APSO-20240.4560.7890.089229872
HGA-20230.2340.5670.023426778
MOGWO-20240.1890.3450.018924582
QPSO-20230.3450.6780.045628975
F3ILGWO0.00240.0156015695
IGWO-20230.01890.02340.001218987
APSO-20240.2340.4560.023423478
HGA-20230.1560.2890.017827873
MOGWO-20240.1230.2340.004519882
QPSO-20230.1890.3670.028925680
F4ILGWO0.00120.0089013497
IGWO-20230.02340.04560.002316790
APSO-20240.1450.2890.015624575
HGA-20230.0890.1670.008922382
MOGWO-20240.0560.1230.003418985
QPSO-20230.0780.1450.006719883
F5ILGWO4.12 × 10−92.34 × 10−88.88 × 10−1489100
IGWO-20232.34 × 10−68.91 × 10−61.23 × 10−814588
APSO-20240.00450.01232.34 × 10−626770
HGA-20230.00230.00674.56 × 10−723475
MOGWO-20241.23 × 10−54.56 × 10−53.45 × 10−817885
QPSO-20231.23 × 10−74.56 × 10−72.34 × 10−912393
F6ILGWO1.2343.4560.023414592
IGWO-202312.3423.452.34518983
APSO-202445.6778.918.91229868
HGA-202334.5656.785.67826772
MOGWO-202423.4534.563.45623478
QPSO-202328.9145.674.56725675
Table 9. System parameter configuration.
Table 9. System parameter configuration.
Parameter CategoryParameter NameNumerical Value
UAVCPU frequency1.5–2.0 GHz
Battery capacity500–800 Wh
maximum power20–30 W
Flight speed15–25 m/s
Edge nodeCPU frequency20–40 GHz
Memory capacity32–64 GB
Coverage radius20 km
Network5G bandwidth100 MHz
Noise−174 dBm/Hz
Path-Loss Exponent2.0–4.0
Table 10. System performance under different task scales.
Table 10. System performance under different task scales.
Number of TasksParameter NameILGWOGWOPSOGA
100Average latency (ms)145.2168.7198.3234.5
Energy consumption ratio0.7320.8210.8890.923
Success rate (%)98.595.291.787.3
300Average latency (ms)287.4345.6412.8489.2
Energy consumption ratio0.7560.8470.9120.945
Success rate (%)96.892.186.479.8
500Average latency (ms)398.7487.3589.6698.4
Energy consumption ratio0.7790.8630.9340.967
Success rate (%)94.288.781.272.6
Table 11. Decision quality evaluation indicators.
Table 11. Decision quality evaluation indicators.
Indicator NameDefinitionCalculation FormulaEvaluation CriteriaTest Scenario
Accuracy of task priority adjustmentIn a dynamic environment, the system adjusts the proportion of task priorities correctlyAccuracy = number of correct adjustments/total adjustments × 100%A gold-standard dataset based on expert annotation, containing optimal decisions from 500 typical scenariosPriority ranking when multiple distress signals are received simultaneously in emergency rescue
Accuracy of dynamic resource allocationThe accuracy of the system in making optimal allocation decisions under resource constraintsAccuracy = number of optimal allocation schemes/total allocation schemes × 100%Compared with the theoretical optimal solution obtained by the linear programming solver, a deviation of less than 5% is considered correctDynamic allocation of computing resources among 5 edge nodes, considering load balancing and minimizing latency
Accuracy of handling abnormal situationsThe ability of the system to correctly identify and handle abnormal situationsAccuracy = number of correctly handled exceptions/total number of exceptions × 100%Based on the historical fault case library, including 300 abnormal scenarios such as communication interruption, equipment failure, and environmental mutationAutonomous decision-making and backup plan activation in case of communication interruption of unmanned aerial vehicles
Accuracy of collaborative path planningOptimal path planning for multi-UAV collaborative operationsAccuracy = number of quasi-optimal paths/total number of paths × 100%Compared with the optimal path length solved by the A* algorithm, a difference of less than 10% is considered accurateCollaborative search-and-rescue path planning of 20 UAVs in urban environments
Table 12. Comparison of decision quality before and after integration of LLM.
Table 12. Comparison of decision quality before and after integration of LLM.
Decision ScenarioTraditional RulesThis Article’s ProposalIncrease MarginNumber of Test SamplesStandard Deviation95% Confidence Interval
Task priority adjustment76.3% ± 3.2%94.7% ± 1.8%24.1%500 scenesσ1 = 3.2%, σ2 = 1.8%[70.0%, 82.6%] vs. [91.2%, 98.2%]
Dynamic allocation of resources68.9% ± 4.1%91.2% ± 2.3%32.4%200 allocation casesσ1 = 4.1%, σ2 = 2.3%[60.8%, 77.0%] vs. [86.7%, 95.7%]
Anomaly handling71.5% ± 3.7%93.8% ± 2.1%31.2%300 abnormal casesσ1 = 3.7%, σ2 = 2.1%[64.2%, 78.8%] vs. [89.7%, 97.9%]
Collaborative path planning74.2% ± 2.9%89.6% ± 2.4%20.8%100 path planning tasksσ1 = 2.9%, σ2 = 2.4%[68.5%, 79.9%] vs. [84.9%, 94.3%]
Average decision time (ms)234 ± 28156 ± 1533.3%1000 decision testsσ1 = 28 ms, σ2 = 15 ms[179, 289] vs. [127, 185]
Table 13. Comparison of experimental results in four scenarios.
Table 13. Comparison of experimental results in four scenarios.
MethodScenario 1: Target SearchScenario 2: Collaborative ReconnaissanceScenario 3: Obstacle Avoidance NavigationScenario 4: Resource Scheduling
Key MetricsSuccess Rate/Average time (s)Coverage Rate/Repetition RateSuccess Rate/Collision FrequencyOptimization Degree/Latency (ms)
DQN83.2%/24591.3%/18.5%79.8%/2.30.76/320
PPO85.6%/23293.2%/15.2%82.4%/1.80.79/298
SAC86.3%/22894.1%/14.3%84.1%/1.60.81/285
QMIX84.5%/23895.8%/11.2%81.2%/2.00.78/305
LLM-7B91.2%/19896.7%/8.5%89.3%/1.20.87/216
Table 14. Convergence comparison experimental results.
Table 14. Convergence comparison experimental results.
MethodConvergence RoundsFinal Success RateSample ComplexityTraining Time (h)
DQN35,00086.3% ± 2.1%3.5 × 107168
PPO28,00088.7% ± 1.8%2.8 × 107142
SAC25,00089.5% ± 1.6%2.5 × 107135
QMIX30,00087.2% ± 2.0%3.0 × 107185
LLM-7B (Fine-tuning)200091.2% ± 1.5%2.0 × 10524
LLM-7B (zero-shot)082.4% ± 2.3%00
Table 15. Comparison of generalization experimental results.
Table 15. Comparison of generalization experimental results.
MethodNew TerrainNew ClimateNew TaskScale AdjustmentAverage Retention Rate
DQN62.3%58.7%45.2%51.3%54.4%
PPO65.8%61.2%48.6%54.7%57.6%
SAC67.2%63.5%51.3%56.2%59.6%
QMIX64.5%60.8%47.9%53.4%56.7%
LLM-7B (fine-tuning)85.6%83.2%79.8%81.5%82.5%
LLM-7B (zero-shot)62.3%58.7%45.2%51.3%54.4%
Table 16. Real-time comparison experiment results.
Table 16. Real-time comparison experiment results.
MethodAverage LatencyP95 LatencyP99 LatencyStandard Deviation
DQN120 ± 1514516815.2
PPO135 ± 1816218518.3
SAC142 ± 2017519820.1
QMIX158 ± 2219021522.4
LLM-7B156 ± 1217218612.3
LLM-3B89 ± 8981058.1
Table 17. Experimental results of algorithm component ablation.
Table 17. Experimental results of algorithm component ablation.
ConfigurationAverage FitnessConvergent AlgebraSuccess RateDegradation
Complete ILGWO0.12478996.8%-
Remove Lévy flight0.143213489.3%14.8%
Remove adaptive weights0.138911891.7%11.4%
Remove elite learning0.135610793.2%8.7%
Remove chaos initialization0.13219894.5%5.9%
Table 18. Performance comparison of different LLM configurations.
Table 18. Performance comparison of different LLM configurations.
Model ConfigurationInference Delay (ms)Decision AccuracyMemory Usage (GB)Power Consumption (W)
GPT-4 Full Version125096.8%350450
7B compressed version15691.2%2785
3B Ultra Lightweight Edition8987.4%1235
Rule engine2368.9%28
Table 23. Analysis of cascade fault propagation.
Table 23. Analysis of cascade fault propagation.
Initial Number of Faulty NodesFinal Number of Faulty NodesLevel of ContactRecovery Time (s)Performance Retention Rate (%)
11.20.2018.494.7
22.70.3534.686.3
34.80.6052.175.8
47.30.8378.962.4
Table 24. Communication model validation results.
Table 24. Communication model validation results.
Environmental TypeAltitude (m)Distance (km)Measured Path Loss (dB)Model Prediction (dB)Relative Error (%)
City100292.390.81.6
City2005108.7107.21.4
Suburb150388.987.51.6
Countryside100485.284.11.3
Table 25. Calculation delay model validation.
Table 25. Calculation delay model validation.
Task TypeCalculation Complexity (G Cycles)Actual Delay (ms)Model Prediction (ms)Relative Error (%)
City1.5145.6142.32.3
City5.2487.9475.22.6
Suburb1.8167.4163.82.2
Countryside0.889.787.12.9
Table 26. Flight energy consumption model validation.
Table 26. Flight energy consumption model validation.
Flight Speed (m/s)Flight Time (min)Actual Energy Consumption (Wh)Model Prediction (Wh)Relative Error (%)
1530234.6228.92.4
2045421.8408.73.1
1530234.6228.92.4
2045421.8408.73.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, S.; Wan, P.; Lian, Z.; Wang, M.; Li, D.; Fan, C. Research on the Construction and Resource Optimization of a UAV Command Information System Based on Large Language Models. Drones 2025, 9, 639. https://doi.org/10.3390/drones9090639

AMA Style

Han S, Wan P, Lian Z, Wang M, Li D, Fan C. Research on the Construction and Resource Optimization of a UAV Command Information System Based on Large Language Models. Drones. 2025; 9(9):639. https://doi.org/10.3390/drones9090639

Chicago/Turabian Style

Han, Songyue, Pengfei Wan, Zhixuan Lian, Mingyu Wang, Dongdong Li, and Chengli Fan. 2025. "Research on the Construction and Resource Optimization of a UAV Command Information System Based on Large Language Models" Drones 9, no. 9: 639. https://doi.org/10.3390/drones9090639

APA Style

Han, S., Wan, P., Lian, Z., Wang, M., Li, D., & Fan, C. (2025). Research on the Construction and Resource Optimization of a UAV Command Information System Based on Large Language Models. Drones, 9(9), 639. https://doi.org/10.3390/drones9090639

Article Metrics

Back to TopTop