Next Article in Journal
Moisture Determination for Fine-Sized Copper Ore by Computer Vision and Thermovision Methods
Previous Article in Journal
CMOS Detector Staggered Array Module for Sub-Terahertz Imaging on Conveyor Belt System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ALIEN: Assisted Learning Invasive Encroachment Neutralization for Secured Drone Transportation System †

by
Simeon Okechukwu Ajakwe
,
Vivian Ukamaka Ihekoronye
,
Dong-Seong Kim
and
Jae-Min Lee
*
Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gyeongbuk 39253, Republic of Korea
*
Author to whom correspondence should be addressed.
This paper is an extended version of our conference papers: Adaptive Drone Identification and Neutralization Scheme for Real-Time Military Tactical Operations. In Proceedings of the 2022 International Conference on Information Networking (ICOIN), Jeju-si, Republic of Korea, 12–15 January 2022 and Tractable Minacious Drones Aerial Recognition and Safe-Channel Neutralization Scheme for Mission Critical Operations. In Proceedings of the IEEE 27th International Conference on Emerging Technologies and Factory Automation (ETFA), Stuttgart, Germany, 6–9 September 2022.
Sensors 2023, 23(3), 1233; https://doi.org/10.3390/s23031233
Submission received: 29 December 2022 / Revised: 16 January 2023 / Accepted: 19 January 2023 / Published: 20 January 2023

Abstract

:
Priority-based logistics and the polarization of drones in civil aviation will cause an extraordinary disturbance in the ecosystem of future airborne intelligent transportation networks. A dynamic invention needs dynamic sophistication for sustainability and security to prevent abusive use. Trustworthy and dependable designs can provide accurate risk assessment of autonomous aerial vehicles. Using deep neural networks and related technologies, this study proposes an artificial intelligence (AI) collaborative surveillance strategy for identifying, verifying, validating, and responding to malicious use of drones in a drone transportation network. The dataset for simulation consists of 3600 samples of 9 distinct conveyed objects and 7200 samples of the visioDECT dataset obtained from 6 different drone types flown under 3 different climatic circumstances (evening, cloudy, and sunny) at different locations, altitudes, and distance. The ALIEN model clearly demonstrates high rationality across all metrics, with an F1-score of 99.8 % , efficiency with the lowest noise/error value of 0.037 , throughput of 16.4 Gbps, latency of 0.021 , and reliability of 99.9 % better than other SOTA models, making it a suitable, proactive, and real-time avionic vehicular technology enabler for sustainable and secured DTS.

1. Introduction

Regardless of the plethora of potential advantages that prompted the creation of unmanned aerial vehicles (UAV), the proliferation of drones is likely to cause unprecedented disruption of future air-based transport ecosystems due to their susceptibility to societal and cybersecurity threats from non-state actors [1,2]. The author in [3] argues that security and privacy issues are frequently raised in relation to novel concepts propelled by innovation. Ironically, the more sophisticated an innovation becomes, the greater the need for introspective transparency in the enabling technology’s decision-making process [4] to ensure security. The viability and acceptability of innovation depend on the deployment of reliable methodologies and designs to address security concerns. A drone transportation system (DTS) is an emerging mobile cyber–physical system (CPS) that comprises the convergence of real-time control systems, distributed systems, embedded systems, and edge networks (wireless sensor networks) for the smart mobility of goods and services.
In recent years, unmanned aerial vehicles (UAV) and other intelligent autonomous systems (IAS) have become increasingly common, especially for “priority-based logistics,” which has raised concerns about the reliability of the technologies underpinning these systems and, consequently, the value placed on human lives [5,6]. As a result, if creative, adaptive strategies are not in place to prevent the misuse of UAV technology, an expansion in the intelligence and autonomy of UAVs will jeopardize their future use [7,8]. According to the authors in [9], a drone invasion is a calculated and planned attempt to sabotage operations, confuse workers, destroy installations, and perhaps leak important information. Such invasions are carried out through unauthorized access to restricted areas. According to the United States Federal Aviation Authority (FAA), there were 8344 drone-related violations between 2016 and 2019 [10] despite available counter-invasive technologies. Furthermore, there are already instances of invasions employing quad-copter UAVs (such as the DJI Phantom) for recreational, logistics, and consumer use to destroy industrial sites, transport dangerous goods over borders, and assault important installations [11,12]. Therefore, if sufficient proactive surveillance systems are not in place to confirm a drone’s legitimacy and appropriateness, the sight of a UAV deployed for logistics in a smart city over a drone transportation network (DTN) can overwhelm a civilian area, inflict distress, and cause severe emotional torture. Consequently, this chaotic scenario might endanger the viability of DTS as an intelligent vehicle transportation enabler.
The neutralization of an invasive drone encroachment in a DTS is a cooperative, multivariate, and cognitive decision-making process that depends on the quality of the data. Neutralization activities include deciding whether a drone is legitimate, determining its jurisdiction, and determining its integrity within the airspace transportation network. Sadly, there are not many trustworthy datasets available in this field. Furthermore, prior research focused on single-modal, license/authorization, flight path, or operating boundary drone detection, paying little to no attention to the attached objects, rendering its decision-making procedures ineffective [13,14,15]. Didactically, according to the authors in [5,6], a drone’s perceived risk in a DTN depends on how well one understands the source, the attached objects, and the drone’s network behavior. Additionally, due to its inherent shortcomings, using a single detection method to accomplish this purpose is ineffective. Moreover, the majority of neutralization techniques are either manual, militaristic, or reactive [10,16,17]. Undoubtedly, improper neutralization of invasive drones in a DTN, incorrect visual recognition of conveyed objects, and late or inaccurate detection can harm the future of DTS. Hence, a trustworthy tracking method should not only notify users that something is present in the network (detection) but also give a detailed description of the predetermined characteristics it is using (identification).
Therefore, to address various drone-related threat dynamics in a timely, accurate, efficient, and situation-aware manner, sustainable DTS requires a synergistic, scalable, and multifaceted networked-integration surveillance approach that makes use of 5G innovations and artificial intelligence (AI) capabilities [18]. This study is a novel attempt to address these problems by proposing a collaborative approach for determining a flown drone’s legitimacy using a fusion of a vision-based deep learning (DL) model and lidar technology. Assisted learning (AL) is an emerging machine learning framework that aims at autonomy, model privacy, data privacy, and unlimited access to local resources [19]. Unlike federated learning, the goal of AL is to provide protocols that significantly expand the learning capabilities of decentralized agents by assisting each other with their private modeling processes without sharing confidential information. Using neural networks and related technologies, assisted collaborative cognition in this context entails comprehending the activity and behavior of the drone, identifying its origin, and remembering its established relationships within a transportation network or route before choosing the best course of action among various dynamic scenarios from different network-based detection sources.
The specific contributions of this paper are:
  • To design a multimodal invasive drone detection network that can detect and classify various drones in a DTN operating in ambient environments, estimate their range, and identify the conveyed objects’ characterization in real-time.
  • To present a reliable and robust dataset for trustworthy and real-time malicious drones with attached object detection and elicitation that encompasses most real-life scenarios in a DTN.
  • To develop a mutivariate perceived danger analysis in tandem with the scenario’s current characterization required for cognitive data-centric decision making.
  • To formulate a multivariate situation-aware encroachment neutralization strategy for ascertaining the appropriate decision for any given dynamic state.
  • To evaluate the model’s performance with other state-of-the-art (SOTA) models and approaches.
This paper is organized into the following sections: Section 2 provides a brief overview of existing DTS surveillance methods. Section 3 presents the proposed system design, model architecture, and methodology. Section 4 highlights the experimental results, discussion, and performance evaluation, and Section 5 concludes with a promising research direction and open issues.

2. Overview of Invasive Drone Encroachment Techniques

This section provides an accessible overview of several research initiatives to stop recurrent intrusions and disturbances in a DTN using convergence approaches, as well as investigations into drone surveillance tactics in DTS.

2.1. Invasive Drone Encroachment Neutralization Techniques in DTS

The maintenance and security of drone transportation present a variety of surveillance issues due to the continued advancement of the drones’ underlying technology [1,20]. To conduct drone surveillance, it is necessary to communicate with, identify, authenticate, elicit, and disarm drones considered dangerous from a drone pool in a DTN. Effective drone surveillance entails the convergence or fusion of many technologies that work together for a specific purpose. Hence, the breakdown of the entire system implies a failure of the underlying technology in any of its components. As shown in Figure 1, a variety of methods (radio frequency, radar, thermal, acoustic, vision, and sniffing) are used to determine the location of a drone, the timing of its entry into a spatial area, and the appropriate divergent action (disable, disarm, or destroy) to take to keep the UAV within an authorized jurisdiction or destroy it [5,21,22]. Only the vision-based approach of these can provide an accurate visual description of the drone and the conveyed object with attendant weaknesses, which is essential for selecting the appropriate neutralizing response [23].

2.2. Collaborative Counter-Invasive Encroachment Approaches in DTS

Modern hybrid counter-invasive encroachment techniques combine sensor technology with hardware control systems to overcome the drawbacks of each detection method, expand the scope of detection, and enhance decision-making ability.
The use of multiple radio frequency (RF) scanners in detecting drones is prevalent these days due to its lower cost compared with radar detection technologies [3]. With multiple RF fusion detection technology, the control commands and other crucial data about multiple tracked drones and their operators in the network can be retrieved. However, newer drone designs typically outperform this sensor fusion surveillance strategy. A vision-acoustic sensor fusion strategy tends to improve detection precision by leveraging exact representation and long-range detection [24]. The effectiveness of the method depends on the auditory signals’ resistance to weather and other external factors as well as the precise visual description of the identified objects generated by electro-optical cameras. The inherent signal interference in a noisy environment is its drawback.
While many hybrid invasive drone encroachment and neutralization technologies integrate vision and radar, little attention is paid to integrating vision and light detection and ranging (LIDAR). Both RADAR and LIDAR detection aim at wide-area aerial object detection and identification. However, while RADAR uses radio waves to detect objects, LIDAR makes use of light waves. RADAR can detect objects at a distance of up to 30 km, but its capacity to do so is constrained by the size of its antenna. On the other hand, LIDAR offers 3D mapping with a high detection resolution accuracy of airborne objects and is the ideal alternative for a portable solution, which informs the deployment of LIDAR for 3D point cloud image generation. According to the authors in [25], LIDAR can now function at wavelengths above 1400 nm with a 500 m object detection range, 10% object reflectivity, and a 99% recognition confidence level. Consequently, recognition is a more difficult challenge than detection because it depends on the fine resolution and precision of real-time image processing, which is made possible by deep learning. Therefore, the drawbacks of the single application of each of these detection and recognition technologies can be addressed by integrating vision with the LIDAR approach, assisted by an AI learning model.
In reality, most drone surveillance architectures and configurations use network-based symbiotic multi-dimensional surveillance system implementations that rely on trustworthy AI models and data as well as operate on low-latency networks to make up for shortcomings in the current sensor fusion drone detection technology, as shown in Figure 2. Obtaining trustworthy data is currently difficult in this area. When collaborative learning is involved, the quality of decisions, the privacy of data, and the timeliness of responses become critical success factors in this mobile cyber–physical system. As a result, this paper’s goal is to enhance the existing vision-LIDAR fusion architecture to increase its detection speed, recognition accuracy, and range estimate. To do this, we developed a reliable drone-based dataset, proposed an effective underlying image-lidar-based detector, and formulated an assisted-learning strategy to determine the degree of invasion, evaluate the drone’s proximity to the targeted area, ascertain the impact of its threat(s), and choose the response strategy to use. This method of drone invasion defense will help the drone transportation system (DTS) become more acceptable and sustainable as a potential future “just-in-case” airborne vehicle type in smart cities.

3. Materials and Methods

3.1. ALIEN Design for Secured DTS

Managing complex and dynamic activities in real-time system designs demands both swiftness and accuracy. AI and system autonomy are inextricably linked. A scenario-based, adaptive-conscious, cognition-friendly AI model is at the heart of a data-centric, hard-based, real-time cyber–physical system. The ALIEN scheme therefore predicates and perpetuates situation-aware safe-channel neutralization, effective detection of variant UAVs, accurate visual identification of conveyed objects, and a timely interception in clusters, as conceptualized in the block diagram in Figure 1 and the system flowchart in Figure 2 respectively.
The objective function of the ALIEN scheme (as seen in Equation (1)) is an optimal solution to maximize DTS counter-invasive deployment through efficient resource utilization that is subject to time.
Mathematically, this is expressed as;
I m a x = t i D i + t i D r d λ i + t i D N i ;
where I m a x = objective function for the optimal solution; D i , D r d λ , and D N = decision variables representing drone detection (see Section 3.3), attached object recognition (see Section 3.5), and neutralization (see Section 3.7); subject to t i , the time taken to perform each of these tasks for each ith drone in the DTS network. Performing these tasks with a centralized procedure and feedback will imply poor timing or a delayed, counter-productive response.
There are parallel activities for detection and identification before neutralization (the ultimate decision). Identification involves performing a threat analysis to determine the position, legality, and danger of the drone and the object it is carrying. The model and associated item of the drone are determined by the detection task using morphological features. The system takes in drone images and coordinates, performs pattern discovery using the underlying detector, produces output, feeds the output for further analysis, and chooses the best response strategy to use using heterogeneous sensors (electro-optical camera and topographic lidars), databases, networks, etc. The robustness of the system is measured by how effectively it can distinguish between different drone types and sizes with discrete attached objects when operating in sunny, cloudy, and evening climatic conditions.

3.2. Camera-Lidar Sensor Calibration

For airborne object detection and ranging operations, an Ouster OSI-128 lidar and a digital camera (Logitech C922) are used. As shown in Figure 1, the lidar uses an infrared source to sense an object’s movement, track its direction and speed, and determine its elevation from the ground network. To calibrate the camera and lidar, it is necessary to determine how the sensors are oriented and placed relative to one another. These calibration parameters are transformed into sensor measurement data, which is essential when updating the coordinate system. Geometrical and optical characteristics of the camera, such as its distortion coefficient, primary point, and focal length, are required to estimate the camera’s translation and rotation with respect to the lidar. Using the camera calibration approach, these features can be approximated. To estimate the collection of feature points, this work uses a checkerboard target [26]. In calculating the camera’s intrinsic matrix, each feature point is related to many angles. The perspective-n-point (PnP) technique then obtains the extrinsic camera and lidar parameters.
The PnP algorithm minimizes the reprojection error in pose estimation between the corresponding 2D points in the camera image and the corresponding 3D points in the lidar. The appropriate 3D–2D points are carefully chosen by using the reflective map of the lidar measurements from several checkerboard targets. As a result, more calibration accuracy than when employing a single plane is recorded.

3.3. ALIEN Detector and Ranging Estimation ( D )

In all operational environments, an effective drone and range detector must distinguish drones from similar flying objects and offer a reliable range estimate of their distance from the detecting device. To do this, Figure 3 presents the proposed heterogeneous drone detection, range logic, and fusion procedure.
Object detection is carried out on the aerial images acquired by the electro-optical camera using the efficient drone detector network (ALIEN) model. To combine detection performance speed and accuracy in real-time, the ALIEN is an improved version of the DRONET [5] detector model that incorporates strip networks (SPP), focus, path aggregation networks (PANET), and other technologies, as illustrated in Figure 4.
The improvement in the detection model is the addition of extra bottleneck CSP layers at the backbone and neck layers, as captured in Table 1. Using a secure networked electro-optical camera, the “Backbone” receives the collected drone image (input), where feature extraction is carried out using a cross-stage partial network (CSP). After entering the “Neck” with PANET and the feature pyramid network (FPN) for feature fusion, the “Head” produces the real detection results, which include the position, score, size, and class. The bottleneck CSP at the backbone layer, combined with SPP and focus, reduces the complexity of large gradient information, truncates the gradient flow of the optimized network, and preserves feature extraction accuracy, as shown in Figure 5.
To do this, CSP separates the feature map of the base layers into two. While the first enters a dense block, the other gets integrated with the feature map and transferred to the next stage. The feed-forward propagation and weight update for this process are captured in Equations (2) and (3).
Y Y k = W k * [ Y 0 , Y 1 , . . . , Y k 1 ] Y t = W t * [ Y 0 , Y 1 , . . . , Y k ] Y u = W u * [ Y 0 , Y t , ]
W W k = f ( W k , g i 0 , g i 2 , g i 3 , . . . , g i k 1 ) W k = f ( W k , g i 0 , g i 2 , g i 3 , . . . , g i k ) W u = f ( W u , g i 0 , g i t )
with Y k being the input of the ( k + 1 ) th at the dense layer, g i being the network gradient information, and W being the weight. The backbone produces output with fewer channels, layers, and larger images.
Then, to further reduce the information path, enhance feature pyramid operations, and boost image localization accuracy, image instance segmentation is applied at the neck layer using PANET, which comprises four cardinal procedures. Table 1 summarizes the ALIEN convolutional structure, listing the specific convolutional parameters and values of its components.
First, a bottom-up path aggregation operation is performed to improve the localization capabilities of the feature extraction hierarchy by dispersing the low-level patterns. Then, feature levels with the same spatial size that corresponds to P 2 to P 5 are produced using bottom-up augmentation. Second, an new feature map with scales ranging from N 2 to N 5 is created. Each N i enters the convolution blocks in phases to further shrink the spatial size of the input. A fused feature map is then produced by fusing the reduced map and each P i . This enters one more convolution block and generates a new N i + 1 for the subsequent sub-network, which leads to the third step, adaptive feature pooling. To achieve adaptive pooling, each proposal is assigned to a different feature level. Then, the fourth operation is fully-connected fusion. Fully connected fusion is carried out by utilizing region of interest (ROI) alignment on the feature grids. ROI alignment distinguishes the properties of the foreground and background masks of the input by training more sample aerial images using the parameters of the fully linked layers, which makes actual prediction possible at the head component. Particularly for small objects, PANET dramatically improves the image feature process.
At the head layer, actual prediction outcomes are performed based on the neck layer’s characterization (drone, attached object, etc.). The YOLO head of the detector model applies the CNN concept to detect objects by using a single network to divide the image into regions of interest ( N * N grid), as seen in Figure 6.
Then, using the formula in Equation (4), the “head” predicts each bounding box region and probability before calculating the overlap known as the intersection of union (IOU).
I o u = A * 1 A ;
where A = area of intersection, and A = area of union. The I o u value ranges between 0 and 1 with a threshold > 0.5 . To enhance the performance of the traditional YOLOv5 model’s aerial object recognition and prediction, the highlighted network configuration settings in Table 1 are the key modifications made.
Thereafter, the PnP technique [27] is employed to extract global and local characteristics from the 3D points to acquire the visual aerial object detection from the ALIEN detector model with its associated lidar measurements. It is not essential to perform the vertical binning required for other representations because PnP converts 3D points into pillar representation. The 3D point target from the PnP algorithm and the 2D drone target from the ALIEN model is then transformed into a single coordinate system using the sensor calibration parameters, and the point clouds are then projected onto the detected drone images to represent the range estimate, object detection, and identification ( D r ) as shown in Figure 3.

3.4. Assisted Learning for Invasive Drone Interception ( D d )

To intercept invasive drone encroachment in an aerial transportation network, a cognitive networked-based real-time decision is taken by carrying out perceived danger or threat analysis to identify and intercept a perceived malicious drone or drones from among other drones and aerial objects in an aerial transportation network. Precision and promptness are essential elements.
Theorem 1.
Given that there are n d numbers of drones in a DTN, the estimated danger ( D r d ) to be carried out by a particular presumed malicious drone ( D r ) in the network is a function of the union of the defined metrics and elements in the universal set of its operation as expressed in Equation (5):
D r d = f { D o D f D l } ,
where D r d = overall drone’s perceived danger,
Dl = legality determinant based on proximity and range measurement,
Df = quantifier for physical feature of the drone, and
Do = conveyed object characteristics.
All transmitted messages in the DTN must be operated within a declared maximum load ( L m a x ), have a defined priority ( k p ), and be delivered within a set time ( t p ) since drone interception needs real-time communication among the participating nodes. Thus, minimizing detection error ( | ( t p + 1 ) ( t p 1 ) | ), protocol latency ( t p + 1 ), and transmission delay ( t p 1 ) is crucial in the network for effective interception. Based on the output of the model and related network media outputs, each of these parameters is evaluated. The procedures for intercepting a drone that is believed to be harmful in a DTN are outlined in Algorithm 1 with the d r o n e - r e p o r t as the expected output.
Algorithm 1: Steps for Invasive Drone Interception D r d .
Sensors 23 01233 i001

3.5. Harmful Object Recognition ( D o ) and Feature Quantification ( D f )

An inappropriate sense of panic can be generated by the sight of a UAV with a strange object attached. As shown in the network diagram in Figure 7, simultaneous visual identification of a drone and the object it is conveying in a DTN is a challenging object detection task (an N P hard issue).
The detection devices (A, B, and C) in the DTN distinguish between different detected aerial objects and choose the best state of each object based on dynamic characteristics. In AL, each cluster for invasive drone encroachment (say, device A) analyzes the drones within its sensing range (a,1; a,1; a,0; a,0; a,0) based on its unique characteristics. Then, with the help of its underlying detection model, it carries out feature extraction and learns from the detected object pattern. Then, the hub shares and communicates the learned or acquired knowledge with other clusters (devices B and C) in the DTN. With subsequent detection and learning, the acquired knowledge from each node is constantly updated in the DTN by the participating nodes/clusters thereby fostering connected intelligence for timely decision making. Therefore, this precise collaborative identification ensures proper and faster eliciting of hostile drones attempting to encroach into the network and differentiates them from hobbyist or logistics drones, thereby gauging their perceived threat in any given setting. To analyze this multivariate attached object recognition and elicitation scenario problem, we use Theorem 2.
Theorem 2.
Given that a detected drone is elicited to be malicious ( D r d ), the dynamic estimate of its threat ( D r d λ ) in a given environment is a measure of its attached object characterization ( D r d o ), the technique for object detection ( D r d f ), path planning/routing ( D r d p a t h ), and variability of its time ( D r d t i m e ). This is expressed as Equation (6):
D r d λ = f [ ( ( D r d o D r d f ) D r d p a t h ) D r d t i m e ] ,
where D r d o = O t = conveyed object threat induced by a drone represented as (a,1), (b,1), and (c,1);
D r d f = other physical feature quantifier = q;
D r d p a t h = the drone’s flight path = p; and
D r d t i m e = response time = t, represented in Figure 7 by (a, b), (a, c), and (b, c). The maximum value of D r d t i m e is set at α, β, and 1. In this study, less attention is given to D r d p a t h and D r d t i m e .
The other physical feature quantifier’s (q) estimation is dependent on the underlying detection technology deployed. Therefore, a malicious drone’s D r d f value can be a combination of weight/kinetic energy D r d ( o , k ) = k, noise level D r d ( o , n ) = n, loaded object D r d ( o , o ) = o, and scanability D r d ( o , s ) = s values as expressed in Equation (7).
D r d f = ( k , n , o , s ) ,
Generally, the mathematical formulae for deriving the loaded object (o), i.e., D o , o from an electro-optical sensor reading is given by Equation (8):
D r d ( o , o ) = o = ( l e v e l 4 ) * W l ,
where l e v e l is the drone’s proximity to the field of view as measured by the lidar point cloud, and W l is the perceived weight of the object being transported. For instance, prompt action and caution are triggered to reroute the drone’s movement if the recognized object ( D r d ( o , o ) = o) is found to be potentially dangerous and inside the line of sight. In general, the dynamic threat estimate of a malicious invasive drone in a DTN is determined by substituting Equation (6) for the enlarged expression of D f from Equation (7).
D r d λ = ( ( O t ( k , n , o , s ) ) D r d p a t h ) D r d t i m e * N d ;
where N d = number of swarming drones with conveyed objects within line of sight. However, keep in mind that the loaded object, D r d ( o , o ) , is a subset of D r d o , i.e., o O t . As a result of the law of association, Equation (9) is transformed into Equation (10):
D r d λ = ( ( k , n , o , s ) p ) t * N d ;
Each of these factors in a hybrid anti-drone model is obtained from several sensor metrics of the underlying detecting technology (video, acoustic, thermal, thermal, radio frequency, etc.). Because this strategy is vision-based, only the values used in deriving D ( o , o ; ) ( O t ) can be determined from the electro-optical and lidar range sensor readings. Equation (10) is so changed to:
D r d λ = f [ ( ( O t + Θ ) p ) t ] * N d r o n e s ;
where Θ represents other physical feature quantifiers associated with other detection technologies. In addition, p and t are not covered as they are outside the scope of this study. Thus, the total drone threat value is:
Σ i N d D T = m a x [ α ( D r d λ ) ] ;
where m a x and α values represent the maximum allowable drone’s conveyed object threat value; n d is the number of detected drones with conveyed objects. Algorithm 2 highlights the steps for attached object identification and elicitation in a DTN.
Algorithm 2: Harmful Object Recognition and Elicitation ( D r d λ ).
Sensors 23 01233 i002

3.6. Assessing Drone Encroachment Legality ( D l )

Aside from determining the harmful status of DTS insight based on object characteristics (see Equation (6)), we verify the drone’s encroachment status. The assumption that several UAVs operating in a specific mapped area are given true flight paths is kept. Adherence to defined pathways, aside from authentication, helps establish whether an incoming drone’s flight is legal and permitted to enter a restricted area.
Mathematically, Equation (13) defines whether or not a drone is legal.
D l = f A p A k ,
where A p = area of interest/the mapped area priority level, and A f = authorization permit/authentication key to operate within such a classified zone. Authenticating a drone’s authorization status in a DTN is an emerging research issue [28,29,30].
Mathematically, A p estimates the distance between the target drone and the predetermined restricted zone as expressed in Equation (14):
A p = L b = 1 1 + e x p ( d a D m a x 2 ) ,
where L b = the detection range/legality boundary, d a is the distance between the area of interest and the detected drone, and D m a x = the system’s allowable detection range based on the sensor (it is usually a constant value). An environment that is categorized or a significant spatial domain is indicated by a high-priority area value. Active and suitable neutralizing decisions are necessary as the drone moves closer (based on measurements from the point cloud’s lidar range). The step-by-step procedure for determining a drone’s closeness, legality, and authorization is summarized in Algorithm 3.
Algorithm 3: Proximity and Legality Assessment ( D l k ).
Sensors 23 01233 i003

3.7. Adaptive Neutralization ( D N )

In a real-time control system routine, an adaptive and cognitive response feature is essential to avoid the deployment of flawed regimens or schedulers that use preset routines and set off false response alerts. Before instantiating and responding to a given scenario, an adaptive neutralization strategy should:
  • Detect the peculiarity of the aerial object and its ranging measurement: drone ( D r ) or not drone, ( D r ) (see Section 3.3);
  • Identify the conveyed object: attached objects ( D o ) or no-attached-object, ( D o ) (see Section 3.5);
  • Access the harmfulness of conveyed objects: harmful ( D T ) or not harmful, ( D T ), through threat analysis (see Section 3.5);
  • Determine the legality of the drone’s route: legal ( D l ) or not legal, ( D l ) (see Section 3.6); and
  • Ascertain the authorization to operate: authorize ( A f ), or not authorize, ( A f ) (see Section 3.6),
Before taking the appropriate neutralization approach: destroy ( D e f e n s e M o d e ), disarm ( S a f e M o d e ), or direct ( R e r o u t e M o d e ). To analyze the neutralization response scenario, we use Theorem 3.
Theorem 3.
Assume that a given drone model D , conveying an identified object D ( o , o ) , with a classified status D ( T , T ) , flies into an environment based on legality route D ( l , l ) , and has authorization key to operate A ( f , f ) , the objective function of the neutralization response provided that flight path D p a t h is known and response time of the system, D t i m e is swift and is given by Equation (11):
D N m a x = ( D * D ( o , o ) * D ( T , T ) * D ( l , l ) * A ( f , f ) ) * N d r o n e s ; s . t . D p a t h , D t i m e
where D N m a x represents the optimal solution for maximum neutralization action, representing the three (3) possible responses (destroy, disarm, or direct/re-route) depending on the dynamic value or state of D N , and * is the weight multiplier effect of each counter-invasive activity subject to the time required to perform each activity and the path flight of the drone.
The ALIEN technique ensures proactive and automatic drone involvement in the network at changing dynamic intervals based on extracted and acquired features, behavioral traits, and other networked system information instead of rule-of-thumb heuristics. Algorithm 4 summarizes this proactive and situation-aware neutralization reaction.

3.8. Dataset Collection, Characterization, and Preprocessing

To assess UAV encroachment, two (2) datasets are used for simulation purposes; one for invasive UAV detection and the other for attached object identification. We created the visioDECT dataset for UAV detection, available on the IEEE Dataport [31]. VisioDECT comprises 20,924 drone samples taken from six (6) UAV models. Each of these UAV types represents a superclass. Each of these three superclasses contains three (3) scenarios (cloudy, evening, and sunny) of flown drones that represent the subclass. Then, each subclass represents the individual drone samples of different sizes flown at various locations, with different distances and altitudes and at different times of the day. The attached object recognition dataset comprises 3600 samples from nine (9) attached objects mounted to drones to represent the classes. Table 2 summarizes the dataset description and distribution.
Algorithm 4: Pseudocode for Situation-Aware Neutralization ( N ( d , s , r ) ).
Sensors 23 01233 i004
These drone types were chosen after considering research by [32] on their popularity in pertinent industries. Additionally, the need to consider both customized drones used in industry and by enthusiasts justifies the inclusion of these drones in the sample space. A detailed explanation of the visioDECT dataset generation and collected data can be accessed via [31]. The datasets were manually created by flying each UAV model under different weather conditions, at different times, and in different locations. As demonstrated in Figure 8, videos of the flown drones with attached objects (at heights ranging from 30 to 100 m) were captured using digital cameras and lidars.
Then, using the “Free Video to JPG” software, the captured video frames were converted into a series of sample frames. To ensure the accuracy of the data, the data frames were sorted to eliminate samples of frames with no drones in the background. Then, each sample frame was labeled by creating bounding boxes around the target objects. To generate ground truth values, bounding boxes were drawn around the target objects, as shown in Figure 6. This arduous AI labeling process was accomplished with the help of the “Make Sense” application. Only 7200 samples (34.4%) of the visioDECT were used for model simulation, as shown in Table 2 describing the sample size distribution of the dataset for the experimental setup.

3.9. Experimental Setup

For simulation, the datasets were split into three portions for model training, testing, and validation: 70%, 20%, and 10% to prevent model overfitting. All models were created using the same network depth-multiple and width-multiple values of 0.33 and 0.50, together with extra hyperparameters as given in Table 3.
Model ablation studies were carried out by varying the hyperparameters to validate the detection performance of the proposed model. To evaluate the capability of the proposed model for knowledge discovery and learning, transfer learning was implemented. The pre-trained weights were used to determine the point cloud for range measurement using PointPillars. Before model training, different image augmentations were applied to the images to lessen object misrepresentation. After that, the best and last weights were used to evaluate the inference of the model. The experimental simulation platform was carried out in a Python environment with PyTorch 1.10 framework on a computer running Windows 10 with the following specifications: Intel(R) Core(TM) i5-8500 CPU @ 3.00GHz, 6Core(s), NVIDIA GeForce GT 1030, GPU CUDA:0 (Tesla K80, 11441.1875MB), and 36GB RAM.

4. Experimental Results and Discussion

The experimental findings are presented in this section to assess how well the suggested ALIEN works to identify and elicit harmful drones and attached objects in a DTN. Additionally, a thorough performance evaluation comparison of the ALIEN model and SOTA approaches were performed using performance metrics, such as mean average precision (mAP), specificity, sensitivity, F1-score, G-mean, throughput (number of floating point operations per second (FLOPS)), and latency/response timeliness (frames per second). A further evaluation of the reliability and efficiency of the proposed model was performed to demonstrate that the scheme meets the criteria for a timely and efficient counter-invasive response.

4.1. Invasive Drone Detection and Elicitation by the ALIEN Model

Based on the visioDECT dataset [31], the findings in Table 4 illustrate the effectiveness of the ALIEN model for efficient drone detection and classification across scenarios.
Mean average precision (mAP) as a metric assesses the capacity of a model for accurate positive prediction and elicitation, expressed in Equation (16) as;
m A P = i = i k A v e . P r ( i ) K ,
with K = number of samples in the dataset, and A v e . P r ( i ) = average precision of each i sample. From Table 4, ALIEN achieved on average a 99.5% mAP value in detecting all UAV models for the cloudy and sunny scenarios, implying a high-level positive prediction capability. In addition, a closer look indicates that ALIEN performed relatively poorly in the evening due to obscured vision, with the lowest mAP of 12% for “Mavic-Air“ drones. Notwithstanding, a 99.5% mAP achieved in detecting miniature “Anafi-Ext” drones in the evening gives credence to the detection precision of ALIEN.
Sensitivity ( R c + ) measures the ability of a model to make accurate positive predictions. That is, the likelihood of a positive test reflects how well the model picks up true positives, written as seen in Equation (17):
S e n s i t i v i t y ( R c + ) = t p × 1 t p + f n ,
where t p = true positive predictions, and f n = false negative predictions by the model, respectively. A high sensitivity value denotes a good model performance. According to Table 4, the sensitivity value of the proposed model ranges from 75% to 100%, implying that the ALIEN model accurately detects various drones in different climatic conditions and at varying altitudes.
From Figure 9, a 100% sensitivity value achieved by the ALIEN model in detecting a distant and miniature drone, A n a f i - e x t e n d e d in the evening scenario is remarkable.
Furthermore, to assess the true negative prediction capacity of the ALIEN model for accurate invasive drone elicitation from other aerial objects, specificity ( R c ) is expressed as by Equation (18);
S p e c i f i c i t y ( R c ) = t n × 1 t n + f p ,
with t n being true negative errors in prediction (i.e., returning a right signal that the detected aerial object is not a drone), and f p representing false positive errors in prediction (i.e., returning a wrong signal that an aerial object is a drone when it is not) by ALIEN model. A low R c value denotes good model performance. The low average R c value of 42.5% achieved by the ALIEN model (see Table 4) indicates the capacity of the model to handle the difficult task of simultaneous detection and elicitation of multiple aerial objects in a DTN.
The moving average results for the evening and cloudy climatic conditions in Figure 10 reveal that the ALIEN model can sufficiently decide that the detected aerial object is not a drone ( D ) even in an obscure scene, which is necessary to prevent unjustified interruption of the airspace.
Finally, to further examine the prediction characteristics of the ALIEN model, particularly when there is an uneven distribution in the variability of the sample size of the dataset, we evaluate a trade-off between sensitivity ( R c + ) and specificity ( R c ), otherwise called the geometric mean ( G - m e a n ). The mathematical definition of G - m e a n is:
G - m e a n = ( R c + × R c ) 1 2 ,
The results in Table 4 affirm that the low G-mean value (between 26.01–70.67%) recorded by the ALIEN model indicates good prediction performance when there is an uneven sample distribution. At a glance, the result from Figure 11 affirms the prediction ability of the proposed model with an uneven sample distribution.

4.2. ALIEN Learning Capacity through Ablation Study

Finding a balance between the reliance of the model on batch gradient descent and stochastic gradient descent can determine the objectivity of the model in terms of error committal and accuracy for each full pass of the training algorithm (epoch) throughout the entire training set. To carry out the ablation study on the proposed model, the training set was divided into various mini-batch sizes and used to calculate the error and update its coefficients, as shown in Table 5. A mini-batch size ( s n ) refers to the portion of the training dataset ( S N ) that the CPU processes simultaneously. The training time increases with s n size.
The result of the ablation study shows a gradual decline in the error minimization by the ALIEN model, an indication of a good learning ability across all batch sizes (batch 8 = 0.037, batch 16 = 0.039, etc.), with little prediction error as the batch size grew at a specified epoch of 100 and learning rate of 0.005 as captured in the line graph of Figure 12.
The results from the performance evaluation analysis (sensitivity, specificity, irrational behavior (F1), etc.) in response to detecting drones and attached objects under dynamic environmental scenarios confirm the suitability of the ALIEN model as an efficient underlying model for multi-scale invasive drone and aerial object detection and status elicitation necessary for intercepting illegal drone operations in a DTN at different altitudes and distances, as shown in the detected samples in Figure 13 with their range estimates.

4.3. Attached Object Recognition by the ALIEN Model

In a DTN that is used for threat analysis and situation-aware neutralizing decisions, accurate and exact transmitted object identification and elicitation are essential to distinguish a hobby drone or logistics drone from a presumed malicious drone (see Section 3.5).
The values from Table 6 show that the ALIEN model can, to some extent, detect and differentiate various conveyed objects by each UAV type (see samples of recognized attached objects Figure 14) with mAP values ranging from 99.7% to 31.7%, maximum sensitivity value of 89.5%, and specificity value of 51.1%.
In addition, with a 99.7% accurate visual recognition of conveyed objects, D ( o , o ) (as shown in Figure 15), the harmful status or otherwise, D ( T , T ) , of the targeted drone can easily be ascertained through proper threat analysis as detailed in Section 3.5. This result validates the capacity of the ALIEN model for simultaneous drone detection and attached object recognition.

4.4. Performance Evaluation

Rationality and timeliness in decision making are crucial elements in any time-sensitive and precision-driven real-time system (such as an anti-drone system). This section compares and evaluates the performance of the ALIEN model with different YOLO models and other SOTA DL models.

4.4.1. ALIEN and YOLO variants

The results in Table 7 summarize the performance of the ALIEN model and variants of YOLO, highlighting the computational complexity analysis of each model.
The F1-score ( F 1 ) defined by Equation (20) measures the change in precision and sensitivity values of a model and quantifies how rationally the model behaves while performing a task.
F 1 = 2 × P r * R c + * 1 P r + R c + ,
where P r represents the precision of the model represented as mAP; and R c + is the recall/sensitivity of the model. In ML, F 1 s c o r e is preferred as a better performance evaluation metric than accuracy because an ML model needs to be rational while executing a task to minimize the false prediction or detection rate. The ALIEN model outperformed other YOLO models with F1-scores ( F 1 ) of 99.8% for drone detection and 80.1% for conveyed object recognition. This validates that the ALIEN model is more effective than other models for effective counter-invasive encroachment neutralization in a DTN.
The results in the line graph in Figure 16 show the mAP values of each model across the epochs, confirming the effectiveness of the ALIEN model for counter-invasive drone encroachment detection.
Furthermore, the latency of a model checks how quickly it responds to events (i.e., prediction time). By using the asynchronous execution method to measure delay, the results confirm that the ALIEN model outperformed other YOLO models with a latency value of 0.021s for both drone detection and conveyed object recognition.

4.4.2. ALIEN and SOTA Model Performance

For exhaustive performance evaluation, the results in Table 8 highlight the computational complexity analysis of the ALIEN model and other SOTA models.
Firstly, throughput defines how frequently each model receives service requests or the maximum number of input instances a neural network can handle in a specific time. This is expressed mathematically in Equation (21);
T h r o u g h p u t ( N t ) = N b * b n * ( 1 T t ) ,
with N b being the number of batches, b n is the batch size, and T t represents total time in seconds. According to Table 8, the ALIEN model achieved the best throughput of 16.1 GFLOPS which is closely followed by f p n + 5 v 5 with 16.20 GFLOPS, YOLOv5 with 16.50 GFLOPS, and then MobileNet with 18.40 GFLOPS. However, overall, the ALIEN model has a better prediction performance than other SOTA models using other evaluation metrics, thereby making the proposed model a preferable choice model for counter-invasive encroachment.
Secondly, the dependability or reliability of a model is the measure of its error minimization, otherwise called loss. When compared to other SOTA models, the ALIEN model achieved the lowest loss (0.0370) in Table 8. Though the other YOLO variants exhibited similar low error minimization values (such as 0.0426 for YOLOv3, 0.0421 for fpn+v5, etc.), the ALIEN model still had the least loss value, signifying its reliability in prediction performance.
Thirdly, the efficiency of a real-time system is measured as the ratio or point at which its precision ( m A P ) coincides with its sensitivity ( R c + ) in carrying out a particular task.
Thirdly, the efficiency of a real-time system is measured as the ratio or point at which its precision ( m A P ) coincides with its sensitivity R c + ) in carrying out a particular task. Equation (22) defines efficiency as;
E f f i c i e n c y ( ξ ) = P r R c + ,
From the result in Table 8, v3 has 97.20 89.50 ξ value, v5s has 96.50 100.0 ξ value, p2 + v5 has 98.30 100.0 ξ value, fpn+v5 has 100.0 70.80 ξ value, etc. However, the ALIEN model achieved a 99.50 100.0 ξ value validating the model with the highest efficiency in handling multi-scale counter-invasive drone with aerial objects encroachment as well as conveyed object recognition prediction under dynamic scenarios. This prediction efficiency is further verified by the confusion matrix in Figure 17 and the detected samples in Figure 13 and Figure 14, with a minimal degree of misclassification.

4.5. Adaptive Encroachment Neutralization Analysis

According to the analysis of the experimental data presented above, the proposed model is capable of achieving high drone detection (D) values of 99.8% and effective conveyed object recognition ( D r λ d ) value of 80.1%, which are required for perceived threat analysis, at a shorter response time ( D t i m e ) of 0.021s, which is required for prompt adaptive neutralization ( D N ( m a x ) ), and less memory consumption of 16.1%. The deployment of the ALIEN model as an underlying DL counter-invasion detection model in a mobile cyber–physical system operating with the hard-real-time control system principle can effectively carry out proper perceived threat analysis and drone authentication and trigger the appropriate neutralization strategy to stop or nullify a suspected malicious drone encroachment in a DTN before it disrupts the entire airspace transportation system, creates social apprehension, and creates public disapproval in an attempt to malign legal drone usage for critical emergency purposes and priority-based logistics.
The condition of a drone in the DTN can be determined at any time by combining these learnable DL model-based parameters (D, D r λ d , D t i m e , and D N ( m a x ) ) with other sensor-based and systems/network-derived parameters, such as legality authentication ( D l k ), flight path ( D p a t h ), map area priority ( g i ), and the appropriate authentication security infrastructure. This guarantees that the evolving anti-drone control system can adeptly instantiate, identify, and elicit perceived malicious invasive drone encroachment from drifting hobby and logistics drones before launching an automated counter-response that could endanger the development of drone technology as a viable priority-based freight carrier. The sustainability of DTS is therefore guaranteed by this approach to invasive drone encroachment detection, identification, and neutralization because it permits only authorized drone-based logistics, airborne transports, and hobby drones to operate along the approved routes in the DTN while circumspectly observing premeditated malicious drone activities aimed at disrupting the network.

5. Conclusions

This work proposes a robust approach to ascertaining safety and security in a drone transportation network by circumspectly determining the malicious status of a drone in the network using a multi-modal deep-learning approach. The approach collaboratively detects and elicits the harmful status or otherwise of a drone in flight in a drone transportation network by detecting the drone in the network, identifying the conveyed objects, assessing the legality boundary of the drone, and determining its authenticity and authorization to operate before deciding the appropriate counter-invasive encroachment response to initiate based on given dynamic metrics and feedback parameters. Performance evaluation and comparison with nine other SOTA models were performed. The experimental results validate the adequacy and inventiveness of the proposed approach in ensuring security and sanity in a drone-based intelligent transport system for the viability and sustainability of DTS through objectivity and circumspection in decision making before interfacing with perceived targets.
Due to their miniature size and concealment of conveyed objects, low detection precision was observed in the prediction results. Future work will tackle this and other emerging problems by developing a drone-based swarm optimization algorithm to enhance the learning of the detection network model and improve performance. In addition, consideration will be given to drone authentication and authorization in a DTN using semi-blockchain technology and a functional non-fungible token to interface with the evolving drone technology landscape and its diverse usage as a viable intelligent vehicle for a just-in-case supply chain.

Author Contributions

Conceptualization, S.O.A.; Data curation, S.O.A. and V.U.I.; Formal analysis, S.O.A.; Funding acquisition, D.-S.K. and J.-M.L.; Investigation, S.O.A.; Methodology, S.O.A. and J.-M.L.; Project administration, D.-S.K. and J.-M.L.; Resources, D.-S.K.; Supervision, D.-S.K. and J.-M.L.; Validation, S.O.A.; Visualization, S.O.A.; Writing—original draft, S.O.A.; Writing—review and editing, S.O.A. and V.U.I. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Priority Research Centers Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science, and Technology (2018R1A6A1A03024003) and by the National Research Foundation of Korea (NRF) grant funded by the Ministry of Science and ICT (MSIT), Korea (1711175292/2022-IT-RD-0084-01).and Grand Information Technology Research Center support program (IITP-2023-2020-0-01612) supervised by the Institute for Information communications Technology Planning Evaluation (IITP).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset used in this study is the VisioDect Dataset which can be accessed from the IEEE Dataport using the https://ieee-dataport.org/documents/visiodect-dataset-aerial-dataset-scenario-based-multi-drone-detection-and-identification (accessed on 20 December 2022).

Acknowledgments

This work was supported by the Priority Research Centers Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science, and Technology (2018R1A6A1A03024003) and by the National Research Foundation of Korea (NRF) grant funded by the Ministry of Science and ICT (MSIT), Korea (1711175292/2022-IT-RD-0084-01).and Grand Information Technology Research Center support program (IITP-2023-2020-0-01612) supervised by the Institute for Information communications Technology Planning Evaluation (IITP).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AcronymMeaning
AIArtificial Intelligence
ALAssisted Learning
ALIENAssisted Learning Invasive Encroachment Neutralization
CNNConvolution Neural Network
CPSCyber Physical System
CSPCross Stage Partial Network
DLDeep Learning
DTNDrone Transportation Network
DTSDrone Transportation System
EDDNEfficient Drone Detector Network
FLOPSFloating point Operations Per Seconds
FPNFeature Pyramid Network
IASIntelligent Autonomous System
IOUIntersection Of Union
LIDARLight Detection and Ranging
mAPMean Average Precision
MLMachine Learning
PANETPath Aggregation Network
RADARRadio Detection and Ranging
SOTAState-of-the-Art
SSDSingle Shot Detector
TEASTimely, Efficient, Accurate, Situation-aware
UAVUnmanned Aerial Vehicles
VGGVisual Geometry Group
YOLOYou Only Look Once

References

  1. Khan, M.A.; Menouar, H.; Eldeeb, A.; Abu-Dayya, A.; Salim, F.D. On the Detection of Unauthorized Drones—Techniques and Future Perspectives: A Review. IEEE Sens. J. 2022, 22, 11439–11455. [Google Scholar] [CrossRef]
  2. Alwateer, M.; Loke, S.W. Emerging Drone Services: Challenges and Societal Issues. IEEE Technol. Soc. Mag. 2020, 39, 47–51. [Google Scholar] [CrossRef]
  3. Shi, X.; Yang, C.; Xie, W.; Liang, C.; Shi, Z.; Chen, J. Anti-Drone System with Multiple Surveillance Technologies: Architecture, Implementation, and Challenges. IEEE Commun. Mag. 2018, 56, 68–74. [Google Scholar] [CrossRef]
  4. David, B.; Gavin, M. Resilient Hermeneutics: Using Simulations in Decision-Centric and Information Rich Environments. In Proceedings of the 2022 Future of War, Amsterdam, The Netherlands, 5–7 October 2022; pp. 1–20. [Google Scholar]
  5. Ajakwe, S.O.; Ihekoronye, V.U.; Kim, D.S.; Lee, J.M. DRONET: Multi-Tasking Framework for Real-Time Industrial Facility Aerial Surveillance and Safety. Drones 2022, 6, 46. [Google Scholar] [CrossRef]
  6. Haviv, H.; Elbit, E. Drone Threat And CUAS Technology: White Pape. Elbit Syst. 2019, 1–19. Available online: https://www.tweedekamer.nl/downloads/document?id=c6b69754-8a63-4772-a90c-9377aff2e248 (accessed on 17 October 2022).
  7. Wild, G.; Murray, J.; Baxter, G. Exploring Civil Drone Accidents and Incidents to Help Prevent Potential Air Disasters. Aerospace 2016, 3, 22. [Google Scholar] [CrossRef] [Green Version]
  8. Swinney, C.J.; Woods, J.C. A Review of Security Incidents and Defence Techniques Relating to the Malicious Use of Small Unmanned Aerial Systems. IEEE Aerosp. Electron. Syst. Mag. 2022, 37, 14–28. [Google Scholar] [CrossRef]
  9. Ajakwe, S.O.; Ihekoronye, V.U.; Akter, R.; Kim, D.S.; Lee, J.M. Adaptive Drone Identification and Neutralization Scheme for Real-Time Military Tactical Operations. In Proceedings of the 2022 International Conference on Information Networking (ICOIN), Jeju-si, Republic of Korea, 12–15 January 2022; pp. 380–384. [Google Scholar] [CrossRef]
  10. Castrillo, V.U.; Manco, A.; Pascarella, D.; Gigante, G. A Review of Counter-UAS Technologies for Cooperative Defensive Teams of Drones. Drones 2022, 6, 65. [Google Scholar] [CrossRef]
  11. Park, S.; Kim, H.T.; Lee, S.; Joo, H.; Kim, H. Survey on Anti-Drone Systems: Components, Designs, and Challenges. IEEE Access 2021, 9, 42635–42659. [Google Scholar] [CrossRef]
  12. Phillips, C.; Gaffey, C. Most french Nuclear Plants ‘should be shutdown’ over Drone Threat. Elbit Syst. 2015, 1–3. Available online: https://www.newsweek.com/2015/03/06/most-frenchnuclear-plants-should-be-shut-down-over-drone-threat-309019.html (accessed on 18 September 2021).
  13. Yong, S.P.; Chung, A.L.W.; Yeap, W.K.; Sallis, P. Motion Detection Using Drone’s Vision. In Proceedings of the 2017 Asia Modelling Symposium (AMS), Kota Kinabalu, Malaysia, 4–6 December 2017; pp. 108–112. [Google Scholar] [CrossRef]
  14. Nalamati, M.; Kapoor, A.; Saqib, M.; Sharma, N.; Blumenstein, M. Drone Detection in Long-Range Surveillance Videos. In Proceedings of the 2019 16th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Taipei, Taiwan, 18–21 September 2019; pp. 1–6. [Google Scholar] [CrossRef]
  15. Meng, W.; Tia, M. Unmanned Aerial Vehicle Classification and Detection Based on Deep Transfer Learning. In Proceedings of the 2020 International Conference on Intelligent Computing and Human-Computer Interaction (ICHCI), Sanya, China, 4–6 December 2020; pp. 280–285. [Google Scholar] [CrossRef]
  16. Chamola, V.; Kotesh, P.; Agarwal, A.; Naren; Gupta, N.; Guizani, M. A Comprehensive Review of Unmanned Aerial Vehicle Attacks and Neutralization Techniques. Ad Hoc Netw. 2021, 111, 102324. [Google Scholar] [CrossRef]
  17. Lykou, G.; Moustakas, D.; Gritzalis, D. Defending Airports from UAS: A Survey on Cyber-Attacks and Counter-Drone Sensing Technologies. Sensors 2020, 20, 3537. [Google Scholar] [CrossRef] [PubMed]
  18. Ajakwe, S.O.; Nwakanma, C.I.; Kim, D.S.; Lee, J.M. Key Wearable Device Technologies Parameters for Innovative Healthcare Delivery in B5G Network: A Review. IEEE Access 2022, 10, 49956–49974. [Google Scholar] [CrossRef]
  19. Xian, X.; Wang, X.; Ding, J.; Ghanadan, R. Assisted Learning: A Framework for Multi-Organization Learning. arXiv 2020, arXiv:2004.00566. [Google Scholar]
  20. Abro, G.E.M.; Zulkifli, S.A.B.M.; Masood, R.J.; Asirvadam, V.S.; Laouti, A. Comprehensive Review of UAV Detection, Security, and Communication Advancements to Prevent Threats. Drones 2022, 6, 284. [Google Scholar] [CrossRef]
  21. Fraga-Lamas, P.; Ramos, L.; Mondéjar-Guerra, V.; Fernández-Caramés, T.M. A Review on IoT Deep Learning UAV Systems for Autonomous Obstacle Detection and Collision Avoidance. Remote Sens. 2019, 11, 2144. [Google Scholar] [CrossRef] [Green Version]
  22. Gopal, V. Developing an Effective Anti-Drone System for India’s Armed Forces. pp. 1–20. Available online: https://www.orfonline.org/research/developing-an-effective-anti-drone-system-for-indias-armed-forces-68001/ (accessed on 20 December 2022).
  23. Ajakwe, S.O.; Ihekoronye, V.U.; Kim, D.S.; Lee, J.M. Tractable Minacious Drones Aerial Recognition and Safe-Channel Neutralization Scheme for Mission Critical Operations. In Proceedings of the 2022 IEEE 27th International Conference on Emerging Technologies and Factory Automation (ETFA), Stuttgart, Germany, 6–9 September 2022; pp. 1–8. [Google Scholar] [CrossRef]
  24. Chang, X.; Yang, C.; Wu, J.; Shi, X.; Shi, Z. A Surveillance System for Drone Localization and Tracking Using Acoustic Arrays. In Proceedings of the 2018 IEEE 10th Sensor Array and Multichannel Signal Processing Workshop (SAM), Sheffield, UK, 8–11 July 2018; pp. 573–577. [Google Scholar] [CrossRef]
  25. Rangwala, S. The LiDAR Range Wars-Mine Is Longer Than Yours. Forbes Magazine, 2022 October 22; 1–12. [Google Scholar]
  26. Zhang, Z. A Flexible New Technique for Camera Calibration. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 1330–1334. [Google Scholar] [CrossRef]
  27. Tu, J.; Wang, P.; Liu, F. PP-RCNN: Point-Pillars Feature Set Abstraction for 3D Real-time Object Detection. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–8. [Google Scholar] [CrossRef]
  28. Ajakwe, S.O.; Ihekoronye, V.U.; Kim, D.; Lee, J.M. Pervasive Intrusion Detection Scheme to Mitigate Sensor Attacks on UAV Networks. In Proceedings of the 2022 Korean Institute of Communication and Sciences Summer Conference, Gumi, Republic of Korea, 22–24 June 2022; pp. 1267–1268. [Google Scholar]
  29. Samanth, S.; V, P.K.; Balachandra, M. Security in Internet of Drones: A Comprehensive Review. Cogent Eng. 2022, 9, 2029080. [Google Scholar] [CrossRef]
  30. Seo, S.; Han, S.; Kim, D. D-CEWS: DEVS-Based Cyber-Electronic Warfare M&S Framework for Enhanced Communication Effectiveness Analysis in Battlefield. Sensors 2022, 22, 3147. [Google Scholar] [CrossRef] [PubMed]
  31. Ajakwe, S.O.; Ihekoronye, V.U.; Mohtasin, G.; Akter, R.; Aouto, A.; Kim, D.S.; Lee, J.M. VisioDECT Dataset: An Aerial Dataset for Scenario-Based Multi-Drone Detection and Identification. IEEE Dataport 2022. [Google Scholar] [CrossRef]
  32. Grossman, N. Drones and Terrorism: Assymetric Warfare and the Threat to Global Security; I.B. Tauris: New York, NY, USA, 2018. [Google Scholar]
Figure 1. Architecture of collaborative cognitive invasive encroachment scheme highlighting the detection and classification, identification and unique recognition, threat analysis, and situation-aware neutralization interactions.
Figure 1. Architecture of collaborative cognitive invasive encroachment scheme highlighting the detection and classification, identification and unique recognition, threat analysis, and situation-aware neutralization interactions.
Sensors 23 01233 g001
Figure 2. System flow of ALIEN procedure to cooperative counter-invasion in DTN.
Figure 2. System flow of ALIEN procedure to cooperative counter-invasion in DTN.
Sensors 23 01233 g002
Figure 3. Logic flow of the efficient drone and ranging detector (EDRD) indicating how the drone images from the electro-optical camera and the ranging estimates from the lidar converge for efficient detection estimation and prediction.
Figure 3. Logic flow of the efficient drone and ranging detector (EDRD) indicating how the drone images from the electro-optical camera and the ranging estimates from the lidar converge for efficient detection estimation and prediction.
Sensors 23 01233 g003
Figure 4. ALIEN model with its the underlying structures. The backbone layer comprises cross partial network (CSP) and focus for feature extraction. The neck layer consist of path aggregation network for feature aggregation, while the head layer has the YOLOv5 for actual prediction.
Figure 4. ALIEN model with its the underlying structures. The backbone layer comprises cross partial network (CSP) and focus for feature extraction. The neck layer consist of path aggregation network for feature aggregation, while the head layer has the YOLOv5 for actual prediction.
Sensors 23 01233 g004
Figure 5. BottleNeck CSP block showing its constituent convolution blocks.
Figure 5. BottleNeck CSP block showing its constituent convolution blocks.
Sensors 23 01233 g005
Figure 6. Drone feature extraction process highlighting the principles for determining outcome.
Figure 6. Drone feature extraction process highlighting the principles for determining outcome.
Sensors 23 01233 g006
Figure 7. Assisted learning scheme for invasive encroachment neutralization in a DTN highlighting the various hubs of surveillance devices.
Figure 7. Assisted learning scheme for invasive encroachment neutralization in a DTN highlighting the various hubs of surveillance devices.
Sensors 23 01233 g007
Figure 8. (a) Images of dataset capturing and flown drones at different altitudes and climate; (b) images of drones with attached objects.
Figure 8. (a) Images of dataset capturing and flown drones at different altitudes and climate; (b) images of drones with attached objects.
Sensors 23 01233 g008
Figure 9. Sensitivity graph of ALIEN performance for detecting UAV models across all scenarios.
Figure 9. Sensitivity graph of ALIEN performance for detecting UAV models across all scenarios.
Sensors 23 01233 g009
Figure 10. Specificity graph of ALIEN performance for detecting UAV models across all scenarios.
Figure 10. Specificity graph of ALIEN performance for detecting UAV models across all scenarios.
Sensors 23 01233 g010
Figure 11. G-mean graph of ALIEN performance showing the trade-off between R c + and R c in detecting different UAVs across all scenarios.
Figure 11. G-mean graph of ALIEN performance showing the trade-off between R c + and R c in detecting different UAVs across all scenarios.
Sensors 23 01233 g011
Figure 12. Box loss graph across different batch sizes indicating ALIEN’s learning capability in predicting results on different operation quanta.
Figure 12. Box loss graph across different batch sizes indicating ALIEN’s learning capability in predicting results on different operation quanta.
Sensors 23 01233 g012
Figure 13. Drone detection samples by the ALIEN model across drone models at different heights and scenarios.
Figure 13. Drone detection samples by the ALIEN model across drone models at different heights and scenarios.
Sensors 23 01233 g013
Figure 14. Samples of recognition of conveyed objects by ALIEN.
Figure 14. Samples of recognition of conveyed objects by ALIEN.
Sensors 23 01233 g014
Figure 15. Graph showing the mAP, sensitivity, specificity, and G-mean values of ALIEN in identifying different conveyed objects by the drones.
Figure 15. Graph showing the mAP, sensitivity, specificity, and G-mean values of ALIEN in identifying different conveyed objects by the drones.
Sensors 23 01233 g015
Figure 16. Hyperparameter graph showing the average precision of different models in identifying conveyed objects by UAVs.
Figure 16. Hyperparameter graph showing the average precision of different models in identifying conveyed objects by UAVs.
Sensors 23 01233 g016
Figure 17. ALIEN confusion matrix.
Figure 17. ALIEN confusion matrix.
Sensors 23 01233 g017
Table 1. ALIEN convolutional structure.
Table 1. ALIEN convolutional structure.
ALIEN Model Configuration Specification
LayerOutput ShapeDescriptions
depth_multiple0.33model depth
multiple
width_multiple0.50layer channel
multiple
[−1, 1, Focus, [64, 3]]
[−1, 1, Conv, [128, 3, 2]]
[−1, 3, BottleneckCSP, [128]]
[−1, 1, Conv, [256, 3, 2]]
[−1, 9, BottleneckCSP, [256]]
[−1, 1, Conv, [512, 3, 2]]
[−1, 9, BottleneckCSP, [512]]
[−1, 1, Conv, [1024, 3, 2]]
[−1, 1, SPP, [1024, [5, 9, 13]]]
Backbone[−1, 3, BottleneckCSP, [1024, False]]CSP performs
the feature
extraction
on the
acquired
aerial
images
[−1, 1, Conv, [512, 1, 1]]
[−1, 1, nn.Upsample, [None, 2, `nearest’]]
[[−1, 6], 1, Concat, [1]]
[−1, 3, BottleneckCSP, [512, False]]
[−1, 1, Conv, [256, 1, 1]]
[−1, 1, nn.Upsample, [None, 2, `nearest’]]
[[−1, 4], 1, Concat, [1]]
[−1, 3, BottleneckCSP, [256, False]]
[−1, 1, Conv, [256, 3, 2]]
[[−1, 14], 1, Concat, [1]]
[−1, 3, BottleneckCSP, [512, False]]
[−1, 1, Conv, [512, 3, 2]]
[[−1, 10], 1, Concat, [1]]
[−1, 3, BottleneckCSP, [1024, False]]
Neck and
Head
[[17, 20, 23], 1, Detect, [nc, anchors]]Image
segmentation,
Feature
Aggregation,
and eventual
prediction are
carried out
Table 2. Dataset Characterization.
Table 2. Dataset Characterization.
Dataset Description
UAV
Model
ScenarioSample
Size
Conveyed
Objects
Sample
Size
Anafi-ExtSunny200Gun400
Evening200Medical
Supplies
500
Cloudy200Spy camera400
DJI-PhantomSunny200Sealed
Package
300
Evening200Containers350
Cloudy200Food Items500
DJI-FPVSunny200Explosives240
Evening200Missile400
Cloudy200Total Sample ( S N ) 3600
EFTE410SSunny200
Evening200
Cloudy200
Mavic-EntSunny200
Evening200
Cloudy200
Mavic-AirSunny200
Evening200
Cloudy200
Total Sample ( S N ) 7200
Table 3. Model hyperparameters.
Table 3. Model hyperparameters.
Model Parameters
No.ParametersValues
1Batch size8, 16, 24, 32, 48, 64, 128, 160
2Box loss0.05
3Epoch100
4Input size416 × 416 × 3
5Learning rate0.01 (0.005)
6Weight-decay0.0005
7Warmup-epochs3.0
8Warm-momentum0.8
Table 4. UAV model detection and classification performance.
Table 4. UAV model detection and classification performance.
Proposed Model Detection and Classification Performance
UAV
Models
ScenariomAP
%
Sensitivity
( R c + ) %
Specificity
( R c ) %
G-Mean
%
Anafi-Extsunny99.5075.0049.8661.15
evening99.50100.0046.1067.89
cloudy99.5085.0045.2061.98
DJI-Phantomsunny99.50100.0047.2168.70
evening94.5095.0049.7168.72
cloudy99.50100.0046.1367.91
DJI-FPVsunny99.50100.0044.8566.97
evening24.90100.0047.5068.92
cloudy99.50100.0044.1566.45
EFTE410Ssunny99.50100.0046.1267.91
evening44.6075.0052.2562.59
cloudy90.80100.0049.1370.09
Mavic-Entsunny80.0082.3048.5063.17
evening66.7061.6049.6855.48
cloudy81.0082.5043.7560.08
Mavic-Airsunny99.50100.0044.1266.42
evening12.00100.0049.9570.67
cloudy99.5015.0045.1226.01
Table 5. Model ablation based on hyperparameter tuning.
Table 5. Model ablation based on hyperparameter tuning.
ALIEN Model Ablation Performance.
Batch SizeEpochmAP
%
Sensitivity
( R c + ) %
Box_Loss
810099.51000.037
1610099.51000.039
2410099.51000.044
3210099.61000.046
4810099.61000.050
6410099.81000.052
9610099.61000.053
12810099.51000.059
14410099.51000.062
16010099.51000.063
Table 6. Conveyed objects status recognition.
Table 6. Conveyed objects status recognition.
Recognition of Conveyed Objects by the ALIEN Model
Conveyed
Objects
mAP
%
Sensitivity
( R c + ) %
Specificity
( R c ) %
G-Mean
%
Medical Supplies48.565.543.353.3
Spy Camera33.959.140.849.1
Sealed Packages30.236.830.133.3
Containers95.289.551.167.6
Guns31.741.211.221.5
Food Items84.584.349.664.6
Missile99.768.544.555.2
Explosives80.045.234.539.5
Table 7. Models performance evaluation I.
Table 7. Models performance evaluation I.
UAV Detection PerformanceConveyed Object Recognition Performance
ModelsBackF1TimeTrain.SpaceF1TimeTrain.Space
Bone(%)(fps)TimeGFLOPs(%)(fps)TimeGFLOPs
v3Darknet95.50.016 s10 m 57 s23.468.90.015 s1 m 8 s23.3
v5sTiny-s98.10.021 s5 m 27 s16.561.50.022 s2 m 54 s16.4
p2 + v5sPPF99.10.023 s16 m 52 s19.274.50.024 s3 m 45 s19.1
fpn+v5FPN82.90.022 s13 m 23 s16.277.70.023 s2 m 55 s16.3
ALIENHybrid99.80.021 s03 m 0 s16.180.10.021 s1 m 8 s16.1
Table 8. Models performance evaluation II.
Table 8. Models performance evaluation II.
Overall Model Performance (Detection and Identification)
ModelsmAP R c + F1TimeRel.Space Used
(%)(%)(%)(fps)LossGFLOPS
v397.2089.5095.500.016s0.042623.30
v5s96.50100.098.100.022s0.044016.50
p2 + v598.30100.099.100.024s0.058719.10
fpn+v5100.070.8082.900.023s0.042116.20
SqueezeNet83.4185.1088.500.030s0.092122.50
GoogleNet89.5588.7589.100.035s0.068524.10
VGG-1689.2583.2086.100.039s0.078126.50
MobileNet V274.5072.0574.010.476s0.095118.40
ResNet89.5390.1089.810.040s0.098321.70
ALIEN99.50100.099.800.021s0.037016.10
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ajakwe, S.O.; Ihekoronye, V.U.; Kim, D.-S.; Lee, J.-M. ALIEN: Assisted Learning Invasive Encroachment Neutralization for Secured Drone Transportation System. Sensors 2023, 23, 1233. https://doi.org/10.3390/s23031233

AMA Style

Ajakwe SO, Ihekoronye VU, Kim D-S, Lee J-M. ALIEN: Assisted Learning Invasive Encroachment Neutralization for Secured Drone Transportation System. Sensors. 2023; 23(3):1233. https://doi.org/10.3390/s23031233

Chicago/Turabian Style

Ajakwe, Simeon Okechukwu, Vivian Ukamaka Ihekoronye, Dong-Seong Kim, and Jae-Min Lee. 2023. "ALIEN: Assisted Learning Invasive Encroachment Neutralization for Secured Drone Transportation System" Sensors 23, no. 3: 1233. https://doi.org/10.3390/s23031233

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop