Next Article in Journal
Distributed Cognitive Positioning System Based on Nearest Neighbor Association and Multi-Point Filter Initiation for UAVs Using DTMB and INS
Previous Article in Journal
UAV Path Planning Trends from 2000 to 2024: A Bibliometric Analysis and Visualization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unmanned Aerial Vehicle-Based Hyperspectral Imaging and Soil Texture Mapping with Robust AI Algorithms

by
Pablo Flores Peña
1,2,†,‡,
Mohammad Sadeq Ale Isaac
3,*,†,
Daniela Gîfu
4,†,
Eleftheria Maria Pechlivani
5,† and
Ahmed Refaat Ragab
2,6,†,‡
1
Department of Electrical Engineering, University Carlos III of Madrid, 28919 Madrid, Spain
2
Drone-Hopper Company, 28919 Madrid, Spain
3
Computer Vision and Aerial Robotics Group, Centre for Automation and Robotics (C.A.R.), Universidad Politécnica de Madrid (U.P.M.-CSIC), 28006 Madrid, Spain
4
Institute of Computer Science, Romanian Academy–Iași Branch, Codrescu 2, 700481 Iasi, Romania
5
Centre for Research and Technology Hellas, Information Technologies Institute, 57001 Thessaloniki, Greece
6
Department of Network, Faculty of Information Systems and Computer Science, October 6 University, Giza 12511, Egypt
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Current address: Drone Hopper Research Center, Calle Mahón 8, 28290 Las Rozas de Madrid, Spain.
Drones 2025, 9(2), 129; https://doi.org/10.3390/drones9020129
Submission received: 12 January 2025 / Revised: 28 January 2025 / Accepted: 4 February 2025 / Published: 11 February 2025

Abstract

:
This paper explores the integration of UAV-based hyperspectral imaging and advanced AI algorithms for soil texture mapping and stress detection in agricultural settings. The primary focus lies on leveraging multi-modal sensor data, including hyperspectral imaging, thermal imaging, and gamma-ray spectroscopy, to enable precise monitoring of abiotic and biotic stressors in crops. An innovative algorithm combining vegetation indices, path planning, and machine learning methods is introduced to enhance the efficiency of data collection and analysis. Experimental results demonstrate significant improvements in accuracy and operational efficiency, paving the way for real-time, data-driven decision-making in precision agriculture.

1. Introduction

The increasing demand for sustainable agricultural practices and effective resource management has driven the development of advanced remote sensing technologies. Among these, hyperspectral imaging (HSI) stands out as a powerful tool for analyzing soil properties, crop health, and environmental conditions due to its ability to capture detailed spectral information across multiple wavelengths [1]. By integrating HSI capabilities with unmanned aerial vehicles (UAVs), researchers have unlocked new possibilities for precision agriculture and soil texture mapping, enabling high-resolution, real-time monitoring of vast agricultural landscapes [2].
UAVs have revolutionized remote sensing in agriculture by offering flexibility, efficiency, and cost-effectiveness compared to traditional satellite or manned aircraft systems. UAVs equipped with hyperspectral sensors provide a unique advantage by delivering high-resolution data with rapid turnaround times, even in remote or inaccessible regions [3]. Recent advancements in UAV design, such as lightweight frames, extended flight durations, and robust navigation systems, have further enhanced their applicability in agricultural monitoring [4]. These systems are now capable of supporting complex missions, including soil texture mapping, crop stress detection, and nutrient analysis, with unprecedented accuracy [5].
Soil texture plays a critical role in determining agricultural productivity, influencing water retention, nutrient availability, and crop growth patterns. Conventional methods for soil texture analysis, such as laboratory testing, are labor-intensive, time-consuming, and often fail to capture spatial variability at the field scale [6]. UAV-based hyperspectral imaging offers a transformative solution by enabling non-invasive, large-scale soil texture assessments through the analysis of spectral signatures. This approach has been validated by studies demonstrating strong correlations between hyperspectral data and soil composition parameters, such as clay, silt, and sand content [7]. The integration of artificial intelligence (AI) algorithms into UAV-based hyperspectral imaging workflows further enhances the potential of this technology. Robust AI models, including Convolutional Neural Networks (CNNs) and transformer-based architectures, have shown remarkable performance in analyzing hyperspectral datasets, identifying patterns, and predicting soil characteristics with high accuracy [8]. These models leverage advanced techniques such as spectral–spatial feature extraction and transfer learning to process vast amounts of data efficiently, even under challenging conditions, such as varying lighting or atmospheric interference [9]. For instance, state-of-the-art AI frameworks like U-Net and DeepLabV3+ have been successfully applied to semantic segmentation tasks, facilitating precise soil texture classification and mapping [10].
Despite these advancements, several challenges remain in the deployment of UAV-based hyperspectral imaging for soil texture mapping. One key challenge lies in the calibration and preprocessing of hyperspectral data to account for environmental factors such as illumination variability and sensor noise [11]. Moreover, the integration of AI algorithms necessitates substantial computational resources and robust training datasets to ensure generalizability across diverse agricultural settings [12]. To address these challenges, recent research has focused on the development of lightweight, edge-computing solutions for real-time data processing and the creation of comprehensive spectral libraries for training AI models [13].
The proposed paper aims to contribute to this growing body of research by developing a UAV-based hyperspectral imaging system integrated with robust AI algorithms for soil texture mapping. The system will leverage advanced UAV platforms equipped with hyperspectral sensors and gamma spectrometers, coupled with machine learning frameworks for data analysis and interpretation. By focusing on vineyard and vegetable crops, this paper seeks to demonstrate the feasibility of real-time, high-resolution soil texture mapping in diverse agricultural scenarios, paving the way for more sustainable and efficient farming practices.

2. Literature Review

Remote sensing technologies have revolutionized the field of precision agriculture, offering innovative solutions for crop monitoring, stress detection, and soil analysis. The use of satellite-based multispectral imaging has been extensively studied and applied in agricultural settings. These methods provide broad coverage, enabling large-scale observation of crop health and environmental conditions. However, their utility is often constrained by low spatial resolution, atmospheric interference, and the delayed availability of data, which limit their effectiveness for real-time decision-making [14]. In response to these limitations, unmanned aerial vehicles (UAVs) have emerged as a transformative tool in precision agriculture. UAVs offer unparalleled advantages, including high spatial and temporal resolution, cost-effectiveness, and the flexibility to conduct targeted surveys even in inaccessible regions [15].
Hyperspectral imaging (HSI) has garnered significant attention for its ability to capture detailed spectral signatures across a wide range of wavelengths. This technology enables the detection of subtle physiological changes in crops, such as variations in chlorophyll content, water stress, and nutrient deficiencies. Studies have demonstrated the effectiveness of hyperspectral imaging in identifying early-stage crop stress, which is critical for timely intervention and yield optimization [16]. For instance, specific vegetation indices derived from hyperspectral data, such as the Normalized Difference Vegetation Index (NDVI) and the Photochemical Reflectance Index (PRI), have been validated as reliable indicators of crop health and productivity [17].
Thermal imaging, another critical component of modern remote sensing, has proven particularly effective in assessing plant water status and identifying heat-induced stress. By detecting variations in surface temperature, thermal cameras can pinpoint areas of reduced evapotranspiration, which often correlate with water stress or disease onset. The integration of thermal imaging with other sensing modalities enhances the comprehensiveness of agricultural monitoring systems [18]. For example, thermal data combined with hyperspectral indices provide a multi-dimensional understanding of crop and soil conditions, enabling precise interventions tailored to specific stress factors.
Gamma-ray spectrometry, though relatively underutilized in agricultural research, offers unique insights into soil composition and nutrient distribution. This technology measures the natural radioactivity of isotopes such as potassium, thorium, and uranium, which are indicative of soil fertility and texture. Recent studies highlight the potential of gamma-ray spectrometry to complement traditional soil analysis techniques by providing rapid, non-invasive assessments of soil properties [19]. When integrated with hyperspectral and thermal imaging, gamma-ray data contribute to a holistic understanding of the agroecosystem, addressing critical challenges in soil management and crop productivity. Despite these advancements, existing remote sensing systems often fall short in their ability to integrate multi-modal data streams effectively.
The lack of advanced data fusion and machine learning algorithms in traditional systems limits their classification accuracy and scalability. Recent developments in artificial intelligence (AI) and deep learning have opened new avenues for overcoming these limitations. Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs) have demonstrated exceptional performance in processing high-dimensional data, enabling the extraction of spatial, spectral, and temporal features with unprecedented precision [20]. For example, hybrid CNN-RNN models are particularly well suited for multi-modal applications, as they combine the strengths of CNNs in spatial feature extraction with the temporal analysis capabilities of RNNs. These models have shown promise in distinguishing between abiotic and biotic stressors, as well as in classifying soil texture and composition [21].
The proposed Quad Hopper system represents a significant leap forward in this context. By integrating hyperspectral, thermal, and gamma-ray sensors with a robust hybrid deep learning framework, the system addresses critical gaps in existing technologies. The incorporation of real-time data processing capabilities further enhances its utility, enabling on-the-fly analysis and adaptive decision-making during UAV missions. This real-time capability is particularly valuable in large-scale agricultural operations, where timely interventions can prevent significant yield losses [22]. In addition to its technical advancements, the Quad Hopper system exemplifies the trend toward multi-modal sensor integration in precision agriculture. Unlike traditional systems that rely on a single sensor type, the Quad Hopper combines complementary data streams to provide a comprehensive understanding of the agroecosystem. This approach not only improves classification accuracy but also enhances the system’s adaptability to diverse environmental conditions and crop types [10]. For instance, hyperspectral data can identify nutrient deficiencies, thermal imaging can detect water stress, and gamma-ray spectrometry can assess soil fertility—all within a single UAV mission. Such multi-faceted capabilities make the Quad Hopper an indispensable tool for modern precision agriculture. The literature underscores the critical role of advanced sensing technologies and AI algorithms in transforming agricultural practices.
While significant progress has been made, challenges remain in scaling these technologies for widespread adoption. The Quad Hopper system, with its innovative design and cutting-edge capabilities, offers a promising solution to these challenges, setting a new benchmark for integrated remote sensing in precision agriculture.

3. Materials, Methods, and Methodology

The UAV system utilized in this paper, named the Quad Hopper, represents a significant advancement in UAV-based soil texture mapping and agricultural stress detection. The system integrates multiple cutting-edge sensors and components, each tailored for precise data acquisition and analysis. These include hyperspectral and RGBD cameras, gamma spectrometry sensors, thermal cameras, and an onboard computing unit capable of real-time artificial intelligence (AI) and machine learning (ML) inference. Collectively, these components equip the system to monitor and assess soil properties, crop health, and environmental conditions with exceptional accuracy and efficiency.
The hyperspectral camera onboard the Quad Hopper captures detailed spectral data across a wavelength range of 400–1000 nm, with a fine spectral resolution of 5 nm intervals. This capability allows for the identification of subtle variations in soil texture and vegetation characteristics, providing insights into soil composition and crop stress. Hyperspectral imaging has been extensively validated as a tool for soil analysis, with studies showing its ability to map soil properties such as clay, silt, and sand composition [23].
In parallel, the gamma spectrometry sensors measure soil radioactivity and isotopic distribution, employing Bayesian inversion algorithms to enhance the precision and reliability of these measurements [24]. The thermal cameras add another layer of functionality by detecting temperature variations with a precision of ±0.5 °C, enabling accurate assessments of soil moisture levels and the early detection of plant stress caused by drought or excessive heat [25].
A key feature of the Quad Hopper is its onboard computing unit, which facilitates real-time data processing. This unit leverages state-of-the-art AI and ML algorithms to analyze incoming data streams, enabling rapid decision-making and adaptive control during flights. Real-time processing not only reduces latency but also ensures that actionable insights are available immediately, even in remote field conditions. This feature is especially critical for large-scale agricultural operations, where timely interventions can significantly enhance productivity and sustainability [26].
By processing data directly onboard, the system also reduces the need for extensive post-flight analysis, streamlining workflows and improving efficiency. The UAV employs sophisticated flight path planning algorithms to optimize data acquisition and field coverage. The Boustrophedon path algorithm ensures systematic coverage by guiding the UAV along parallel, non-overlapping tracks, making it ideal for uniform terrains. Meanwhile, the Dubins path algorithm minimizes energy consumption and flight time by optimizing UAV turns, ensuring efficient navigation even in irregular or constrained environments [27]. By combining these algorithms, the Quad Hopper achieves a balance between comprehensive data collection and operational efficiency, addressing the demands of modern precision agriculture.
Figure 1 showcases the Quad Hopper UAV in operation, illustrating its structural design and the integration of its various components. The robust, modular frame accommodates a diverse array of sensors and payloads, making it highly adaptable for different agricultural applications. The figure emphasizes the UAV’s capability to navigate complex environments with precision and stability, underscoring its utility as a versatile data collection platform.
To provide additional context, Figure 2 illustrates the individual sensors and equipment utilized in this paper, including the hyperspectral camera, thermal camera, and gamma spectrometer. Each device is meticulously calibrated to ensure optimal performance under diverse environmental conditions.
The hyperspectral camera, depicted in the figure, is compact yet highly sensitive, capturing detailed spectral data critical for soil texture mapping. The thermal camera is shown alongside its mounting assembly, emphasizing its role in detecting temperature variations with exceptional accuracy. Finally, the gamma spectrometer is displayed with its protective casing, highlighting its durability and precision in measuring soil radioactivity.
In addition, the onboard computing unit used in the Quad Hopper system is equipped with an NVIDIA Jetson Xavier NX, featuring a 384-core Volta GPU with 48 Tensor Cores, an 8-core ARM CPU, and 8GB of LPDDR4x memory. This system enables efficient real-time data processing during UAV missions. The AI/ML algorithms employed include Convolutional Neural Networks (CNNs) for spatial feature extraction from hyperspectral and thermal data and Recurrent Neural Networks (RNNs) for temporal pattern recognition in gamma-ray spectrometry data. The combination of these algorithms allows for accurate classification of soil texture and crop stress conditions in real time.
Figure 1 and  Figure 2 highlight the advanced technological framework of the Quad Hopper UAV system. By integrating state-of-the-art sensors with robust path planning algorithms and real-time data processing capabilities, the system establishes a new benchmark for UAV-based soil texture mapping and stress detection. The comprehensive design ensures the system is well equipped to meet the challenges of modern agriculture, empowering researchers and farmers with actionable insights for sustainable and efficient farming practices.

3.1. Study Area and Climatic Conditions

The study was conducted in Thessaloniki, Greece, during a piloting case study scenario within the project E-SPFdigit funded by the European Union, under Grant Agreement No. 101157922; the field is characterized by a Mediterranean climate with mild winters and hot summers. The UAV missions were performed at an average flight altitude of 50 m to balance spatial resolution and coverage.

3.2. NDVI and PRI Description

The Normalized Difference Vegetation Index (NDVI) is a widely used indicator for vegetation health, calculated using the reflectance in the near-infrared (NIR) and red wavelengths. It is defined as NDVI = (NIR − Red)/(NIR + Red), where higher values typically indicate healthier vegetation [28]. On the other hand, the Photochemical Reflectance Index (PRI) is derived from the reflectance at 531 nm and 570 nm , and is primarily used to assess photosynthetic activity and stress responses in plants. Together, these indices provide valuable insights into plant health and stress detection when analyzed using hyperspectral imaging [29].

3.3. Sensors

The multi-modal stress detection system leverages three advanced sensing technologies: hyperspectral imaging, gamma-ray spectrometry, and thermal imaging. Each of these sensors contributes uniquely to the precision and robustness of agricultural monitoring by capturing specific environmental and physiological data. Hyperspectral imaging is a critical component of the system, enabling detailed spectral analysis across a wide range of wavelengths. This sensor provides reflectance data at 5 10 nm intervals across the visible ( 400 700 nm ), near-infrared ( 700 1000 nm ), and shortwave infrared ( 1000 2500 nm ) spectra. With a reflectance accuracy of ±2 °C, hyperspectral imaging is highly sensitive to variations in plant chlorophyll content, water stress, and soil properties, making it indispensable for calculating vegetation indices such as the NDVI and PRI [30].
Gamma-ray spectrometry offers insights into soil composition by detecting isotopes and measuring gamma radiation counts. This sensor achieves high accuracy with an energy resolution of less than 7 % at 662 keV and sensitivity to radiation levels as low as 10 counts per second (cps). By employing advanced noise correction techniques, such as Bayesian inversion, gamma-ray spectrometry enables precise mapping of soil nutrient levels and the detection of abiotic stresses like salinity [31]. Thermal imaging captures surface temperature data of crops and soil, essential for assessing evapotranspiration and water stress. It offers a spatial resolution of 640 × 480 pixels, with a ground sampling distance (GSD) of 5 cm at an altitude of 50 m. With a temperature accuracy of ±0.5 °C and sensitivity to temperature differences as small as ±0.1 °C, thermal imaging effectively identifies heat stress in crops. Collectively, these sensors form the backbone of the multi-modal stress detection system, ensuring comprehensive and reliable data acquisition [32].

3.4. UAV Platform

The Quad Hopper drone is the UAV platform utilized for multi-modal stress detection. Designed for heavy payloads and extended flight durations, the Quad Hopper integrates seamlessly with the sensor suite, ensuring efficient and precise data collection [33]. With a payload capacity of 100 kg , the Quad Hopper can accommodate hyperspectral, gamma-ray, and thermal imaging sensors simultaneously. Its petrol engine-powered propulsion system provides a robust power source, enabling a flight autonomy of up to 90 min [34,35].
The drone features a redundant power plant transmission system, which integrates power from its four engines and transfers it effectively to the rotor set. This design ensures operational reliability and reduces the risk of system failure during critical missions. The Quad Hopper’s onboard systems are equipped with real-time data processing capabilities, allowing the drone to analyze sensor data during flight. This capability reduces latency and facilitates timely decision-making in the field [36]. Additionally, the drone’s navigation system is enhanced with RTK GPS, providing centimeter-level positional accuracy for precise aerial scanning. Capable of covering large agricultural areas in a single mission, the Quad Hopper offers unmatched efficiency for soil and crop monitoring tasks. The platform is further designed to withstand harsh environmental conditions, including light rain and wind speeds of up to 15 m / s . Its adaptability, combined with its advanced payload capacity and power system, makes the Quad Hopper an optimal choice for multi-modal stress detection in precision agriculture.

3.5. Field Implementation Challenges

Deploying the Quad Hopper system in real-world agricultural settings poses several challenges. Environmental variability, including fluctuating light conditions, wind interference, and temperature extremes, can impact sensor performance. To mitigate these effects, the system employs real-time calibration techniques, such as adaptive reflectance normalization and dynamic thermal offset adjustments. Data transmission is another critical challenge, particularly in remote areas with limited connectivity. The onboard preprocessing capabilities of the Quad Hopper reduce data size, enabling efficient storage and transmission. Furthermore, the diverse agricultural environments require extensive testing across various crop types and geographic regions to ensure consistent performance. Addressing these challenges is essential for large-scale adoption and reliable stress detection in precision agriculture.

3.6. Economic Viability

The economic viability of the Quad Hopper system is a pivotal consideration for its adoption in the agricultural industry. Traditional stress detection methods often involve labor-intensive manual inspections or costly satellite imagery subscriptions, both of which can be prohibitive for smallholder farmers [37]. In contrast, the Quad Hopper offers a cost-effective alternative by combining multi-modal sensors and autonomous UAV operations. The system reduces operational costs by enabling early stress detection and targeted interventions, thereby minimizing input waste and increasing crop yield. Additionally, the modular design allows scalability, catering to both small-scale farms and large agricultural enterprises. Preliminary cost analyses indicate a significant return on investment, particularly when implemented over multiple growing seasons, highlighting the system’s potential to transform agricultural practices economically.

3.7. Ethical and Environmental Impact

The Quad Hopper system aligns with sustainable agricultural practices by promoting resource efficiency and reducing environmental impact. Early stress detection minimizes the overuse of fertilizers and pesticides, lowering the risk of chemical runoff and soil degradation [38]. By enabling precise, localized interventions, the system supports the conservation of natural resources, including water and energy. Ethical considerations include ensuring equitable access to this technology, particularly for resource-constrained farmers in developing regions. Partnerships with governmental and non-governmental organizations can facilitate affordability and accessibility. Moreover, the system’s design prioritizes minimal environmental disruption, with low-noise operations and compliance with UAV regulatory standards. The ethical and environmental benefits of the Quad Hopper reinforce its role in advancing sustainable and inclusive agriculture.

3.8. Algorithm Development

The algorithm for stress detection integrates multi-modal data from hyperspectral, thermal, and gamma-ray sensors, employing advanced data preprocessing, vegetation index computation, and machine learning models to achieve precise soil and crop condition classification. The preprocessing stage is crucial to ensure data consistency and reliability. Hyperspectral data, represented as R ( λ ) (reflectance values at wavelength λ ), are normalized to reduce variability using the following equation:
R ( λ ) = R ( λ ) R min R max R min
where R min and R max denote the minimum and maximum reflectance values, respectively.
Thermal data T ( x , y ) , representing temperature at pixel ( x , y ) , are calibrated using reference blackbody temperatures to adjust for environmental factors:
T cal ( x , y ) = T ( x , y ) T offset
where T offset accounts for ambient temperature fluctuations.
Gamma-ray spectrometry data G ( E ) , indicating counts at energy levels E, are corrected for background noise using Bayesian inversion techniques, yielding G corrected ( E ) :
G corrected ( E ) = G ( E ) G background ( E )
Following preprocessing, vegetation indices are calculated to derive critical features. The Normalized Difference Vegetation Index (NDVI), essential for assessing vegetation health, is computed as follows:
NDVI = R NIR R RED R NIR + R RED
where R NIR and R RED are the reflectance values in the near-infrared and red bands, respectively.
The Photochemical Reflectance Index (PRI), indicative of photosynthetic activity, is derived as follows:
PRI = R 531 R 570 R 531 + R 570
where R 531 and R 570 are reflectance values at 531 nm and 570 nm wavelengths.
The core of the algorithm employs a hybrid deep learning framework for feature extraction and classification. Convolutional Neural Networks (CNNs) extract spatial features F s from hyperspectral and thermal data:
F s = CNN ( R , T cal )
Recurrent Neural Networks (RNNs) are then applied to capture temporal features F t from the gamma-ray time series data:
F t = RNN ( G corrected )
These features are concatenated into a unified feature vector F c :
F c = [ F s , F t ]
The classification layer maps F c into stress categories (e.g., abiotic or biotic stress) using a softmax activation function:
P ( stress _ class ) = softmax ( W F c + b )
where W and b are the weights and biases of the classification layer.
The sequence of the algorithm is described in Algorithm 1, while a flowchart depicting the complete data flow is illustrated in Figure 3.
Algorithm 1 Multi-modal stress detection
1:
Input: Hyperspectral data R ( λ ) , Thermal data T ( x , y ) , Gamma-ray data G ( E )
2:
Normalize hyperspectral data using Equation (1)
3:
Calibrate thermal data using Equation (2)
4:
Apply background correction to gamma-ray data using Equation (3)
5:
Compute NDVI using Equation (4)
6:
Compute PRI using Equation (5)
7:
Extract spatial features F s using CNNs (Equation (6))
8:
Extract temporal features F t using RNNs (Equation (7))
9:
Concatenate features into F c (Equation (8))
10:
Classify F c into stress categories using softmax (Equation (9))
11:
Output: Predicted stress class (abiotic or biotic stress)
In addition, Figure 4 provides a detailed visualization of the methodology employed in this study. It illustrates the workflow of the proposed UAV-based hyperspectral imaging and advanced AI algorithm integration for soil texture mapping and stress detection. The figure outlines the sequence of operations, starting with UAV data acquisition using hyperspectral, thermal, and gamma-ray sensors. It then highlights the preprocessing steps, such as noise reduction and data alignment, followed by the hybrid deep learning framework for feature extraction and classification. This workflow demonstrates the seamless integration of multi-modal data and advanced computational techniques, ensuring efficient and accurate agricultural analysis.

4. Results and Discussion

4.1. Validation Data and Methods

The validation of the proposed system was conducted using a combination of ground truth data and cross-validation techniques to ensure reliability and robustness. Ground truth data were obtained by collecting soil samples and conducting laboratory analyses to determine soil texture, nutrient content (phosphorus and potassium), and fungal stress levels. The UAV-acquired hyperspectral, thermal, and gamma-ray spectrometry data were compared against these ground measurements to assess the accuracy of the predictions. For classification accuracy, a 10-fold cross-validation approach was applied to the dataset, ensuring that the model was trained and tested on diverse subsets of the data. The predicted outputs, including soil texture classes and stress detection categories, were evaluated using performance metrics such as accuracy, root mean square error (RMSE), and F1 scores. Error bars and confidence intervals were computed to quantify the uncertainty in the accuracy metrics, providing greater reliability in the reported results. The combined use of ground truth measurements and rigorous validation methods demonstrates the effectiveness of the proposed system in improving soil texture mapping and stress detection accuracy compared to traditional methods.
The results of this paper demonstrate the efficacy of the proposed multi-modal stress detection system in detecting fungal stress and analyzing soil texture with unprecedented precision and speed. Using real datasets collected from controlled agricultural trials, the system’s performance was evaluated in terms of spectral signature accuracy, thermal image resolution, and algorithm efficiency on images taken by the Quad Hopper drone, as shown in Figure 4. The findings highlight significant improvements compared to traditional methods, underscoring the potential of this approach in precision agriculture.

4.2. Detection of Fungal Stress

The system successfully identified fungal stress in crops using hyperspectral and thermal imaging data. Hyperspectral imaging achieved an average spectral signature accuracy of 94.8 % , with a root mean square error (RMSE) of 0.032 when comparing predicted reflectance values to ground truth measurements. This high accuracy was facilitated by the computation of vegetation indices such as the NDVI and PRI, which clearly distinguished healthy plants from those under fungal stress. Thermal imaging further enhanced fungal stress detection by identifying localized temperature anomalies in infected areas. The thermal images achieved a resolution of 640 × 480 pixels with a ground sampling distance (GSD) of 5 cm, allowing the system to detect temperature differences as small as ±0.1 °C, as shown in Figure 5. These anomalies were consistent with fungal infections, which typically induce localized heat stress in plants. The deep learning framework also demonstrated superior classification performance. The hybrid CNN-RNN model achieved a classification accuracy of 96.5 % , significantly outperforming traditional machine learning algorithms such as Support Vector Machines (SVMs), which achieved only 81.2 % accuracy on the same dataset. The system’s speed was noteworthy, with real-time processing capabilities enabling predictions in under 0.5 s per image.
The standard data used for classification accuracy consist of surface-level measurements collected from evenly distributed sampling points across the study area. These measurements were validated using ground truth data obtained through calibrated sensors and lab analyses. By using surface data, the methodology ensures that spatial variability across the field is captured, enhancing the reliability and authenticity of the classification results. In addition, the determination of soil texture was carried out using the multi-modal sensor data acquired from the UAV platform. Hyperspectral imaging data were analyzed to identify key spectral signatures associated with soil particle size distribution, while gamma-ray spectrometry provided additional insights into soil mineral content. A hybrid deep learning framework combining CNNs and RNNs was employed to classify soil samples into distinct texture categories (e.g., sandy, loamy, and clayey). Ground truth data for validation were collected through traditional laboratory analysis of soil samples, ensuring high accuracy in classification. The proposed methodology achieved a classification accuracy of 96.5 % , significantly outperforming traditional approaches that relied on single-sensor data or manual soil texture assessment. This approach provides a robust and efficient solution for precise soil texture mapping in agricultural applications.

4.3. Soil Texture Analysis

The evaluation of soil texture explicitly considered fractions of sand, silt, and clay, determined through a combination of hyperspectral imaging and gamma-ray spectrometry. The soil texture analysis module also produced remarkable results, leveraging gamma-ray spectrometry and hyperspectral data. The gamma-ray spectrometry sensor achieved a background noise correction accuracy of 97.2 % , ensuring reliable quantification of soil elements such as potassium and phosphorus. This contributed to the precise classification of soil types, with an overall accuracy of 93.7 % in distinguishing sandy, loamy, and clay soils, as shown in Figure 6. Hyperspectral imaging complemented this analysis by identifying subtle variations in soil reflectance patterns. The normalized reflectance data correlated strongly with laboratory measurements, with a correlation coefficient ( R 2 ) of 0.91 . This high level of agreement underscores the system’s ability to provide accurate soil texture maps, critical for informed decision-making in precision agriculture.
To further illustrate the effectiveness of the proposed system in soil texture classification, Figure 7 presents a spatially distributed map of classified soil textures across the study area. The map visually represents the categorization of soil into sandy, loamy, and clay fractions, demonstrating the system’s capability to differentiate soil types with high accuracy. The classification results align well with laboratory-validated data, reinforcing the reliability of the integrated hyperspectral and gamma-ray spectrometry approach for soil texture assessment.

4.4. Comparative Performance

The proposed system significantly outperformed traditional stress detection and soil analysis techniques in terms of both speed and precision. As summarized in Table 1, the system achieved an 82 % improvement in processing speed and an 85 % increase in classification accuracy compared to conventional methods, as shown in Figure 8.
In addition, the predicted reflectance values and ground-measured spectra are visualized in Figure 9, highlighting the minimal variance (RMSE: 0.032).

5. Discussion

The results indicate that the integration of multi-modal sensors and advanced deep learning algorithms offers a transformative approach to stress detection and soil analysis. Compared to traditional techniques, which rely heavily on manual data collection and basic image processing, the proposed system delivers significantly higher accuracy and faster processing speeds. For instance, while conventional methods often fail to detect early-stage fungal infections, the hyperspectral imaging and thermal data in this system provide early warning signs, enabling timely intervention. Furthermore, the real-time capabilities of the UAV platform, combined with its ability to process large datasets autonomously, reduce operational costs and improve scalability. By addressing the limitations of traditional methods, this system has the potential to revolutionize agricultural monitoring, ensuring more sustainable and efficient farming practices.
The discussion integrates related studies, such as those by [38,39], emphasizing the relevance of UAV-based soil texture analysis.

6. Conclusions and Future Perspectives

The proposed Quad Hopper system represents a significant advancement in UAV-based precision agriculture, combining multi-modal sensing, real-time data processing, and advanced machine learning algorithms for efficient stress detection and soil analysis. The results demonstrate exceptional accuracy and processing speed, outperforming traditional methods and addressing critical challenges in agricultural monitoring. The traditional methods referenced include laboratory-based soil texture analysis, which typically involves mechanical sieving and sedimentation techniques to separate soil fractions (sand, silt, and clay). Additionally, visual inspection and manual field sampling are commonly employed, which are time-consuming, labor-intensive, and lack spatial resolution. These conventional approaches often fail to provide real-time insights, making them less effective compared to UAV-based hyperspectral imaging integrated with advanced AI algorithms.
The integration of hyperspectral, thermal, and gamma-ray data, coupled with hybrid CNN-RNN models, enables precise classification of biotic and abiotic stress factors, as well as accurate soil texture mapping. However, challenges related to environmental variability, data transmission, and scalability must be addressed through continued research and development. Future work will focus on expanding the system’s capabilities to include additional sensors, enhancing its adaptability to diverse crops and environments, and optimizing economic feasibility for broader adoption. By fostering sustainable and efficient farming practices, the Quad Hopper has the potential to revolutionize modern agriculture and contribute to global food security.

Author Contributions

Conceptualization, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; methodology, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; software, M.S.A.I. and A.R.R.; validation, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; formal analysis, P.F.P. and A.R.R.; investigation, M.S.A.I. and D.G.; resources, P.F.P. and A.R.R.; data curation, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; writing—original draft preparation, M.S.A.I. and A.R.R.; writing—review and editing, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; visualization, P.F.P., M.S.A.I., D.G., E.M.P. and A.R.R.; supervision, D.G. and A.R.R.; project administration, P.F.P. and A.R.R.; funding acquisition, P.F.P., A.R.R., and M.S.A.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the E-SPFdigit PROJECT funded by the European Union under Grant Agreement No. 101157922.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to express their gratitude to the Centre for Research & Technology Hellas (CERTH) for their dedicated efforts in the E-SPFdigit Project and their invaluable support for this research.

Conflicts of Interest

Author Pablo Flores Peña and Ahmed Refaat Ragab are employed by the Drone-Hopper Company. The other authors declare no conflicts of interest.

References

  1. Shippert, P. Introduction to hyperspectral image analysis. Online J. Space Commun. 2003, 2, 8. [Google Scholar]
  2. Thenkabail, P.S.; Gumma, M.K.; Teluguntla, P.; Irshad, A.M. Hyperspectral remote sensing of vegetation and agricultural crops. Photogramm. Eng. Remote Sens. 2014, 80, 695–723. [Google Scholar]
  3. Shaw, G.A.; Burke, H.K. Spectral imaging for remote sensing. Lincoln Lab. J. 2003, 14, 3–28. [Google Scholar]
  4. Colomina, M.; Molina, P. Unmanned aerial systems for photogrammetry and remote sensing. ISPRS J. Photogramm. Remote Sens. 2014, 92, 79–97. [Google Scholar] [CrossRef]
  5. Bach, H.; Mauser, U. Methods and Examples for Remote Sensing Data Analysis in Agriculture. Remote Sens. 2013, 5, 2438–2458. [Google Scholar]
  6. Ben-Dor, E.; Hill, J.; Taylor, G. Hyperspectral remote sensing for soil properties. Adv. Agron. 2008, 97, 321–392. [Google Scholar]
  7. Pal, S.; Talukdar, A. Hyperspectral imaging and soil texture mapping. J. Environ. Qual. 2017, 46, 564–575. [Google Scholar]
  8. Zhang, M.; Zhang, H.; Zhang, Y. Deep Learning for Hyperspectral Data. IEEE Trans. Geosci. Remote Sens. 2019, 57, 7089–7109. [Google Scholar]
  9. Yu, L.; Yang, J.; Zhao, Z. Spectral-spatial feature extraction for hyperspectral data using deep learning. Remote Sens. Lett. 2018, 9, 300–309. [Google Scholar]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Lecture Notes in Computer Science; Springer: Munich, Germany, 2015; Volume 18, pp. 234–241. [Google Scholar]
  11. Verrelst, J.; Camps-Valls, G.; Munoz-Mari, J. Imaging spectroscopy calibration techniques. Remote Sens. Environ. 2020, 239, 111605. [Google Scholar]
  12. Camps-Valls, S.; Bandos, T.; Gomez-Chova, L. Machine learning for remote sensing data processing. IEEE Geosci. Remote Sens. Mag. 2019, 7, 64–87. [Google Scholar]
  13. Zhang, Y.; He, M. Edge computing for UAV hyperspectral imaging. Remote Sens. 2020, 12, 1295–1310. [Google Scholar]
  14. Zhang, C.; Kovacs, J.M. The application of small unmanned aerial systems for precision agriculture: A review. Precis. Agric. 2012, 13, 693–712. [Google Scholar] [CrossRef]
  15. Rango, A.; Laliberte, A.; Winters, C. Unmanned aerial vehicle-based remote sensing for rangeland assessment, monitoring, and management. J. Appl. Remote Sens. 2009, 3, 033542. [Google Scholar]
  16. Thenkabail, P.S.; Smith, R.B.; De Pauw, E. Hyperspectral vegetation indices and their relationships with agricultural crop characteristics. Remote Sens. Environ. 2000, 71, 158–182. [Google Scholar] [CrossRef]
  17. Jones, H.G. Application of thermal imaging and infrared sensing in plant physiology and ecophysiology. Adv. Bot. Res. 2004, 41, 107–163. [Google Scholar]
  18. Beamish, D.; Young, M.E. High-resolution airborne geophysics in the United Kingdom: Gamma-ray spectrometry and soil surveys. Proc. Geol. Assoc. 2009, 120, 60–72. [Google Scholar]
  19. Kross, A.; McNairn, H.; Lapen, D.; Sunohara, M.; Champagne, C. Assessment of RapidEye vegetation indices for estimation of leaf area index and biomass in corn and soybean crops. Int. J. Appl. Earth Obs. Geoinf. 2015, 34, 235–248. [Google Scholar] [CrossRef]
  20. Hamuda, E.; Glavin, M.; Jones, E. A survey of image processing techniques for plant extraction and segmentation in the field. Comput. Electron. Agric. 2016, 125, 184–199. [Google Scholar] [CrossRef]
  21. Petersson, H.; Gustafsson, D.; Bergstrom, D. Hyperspectral Image Analysis Using Deep Learning—A Review. In Proceedings of the 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), Oulu, Finland, 12–15 December 2016; IEEE: New York, NY, USA, 2016; pp. 1–6. [Google Scholar]
  22. Lecun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  23. Ustin, J.L.; Gamon, J.A.; Foody, G.M. Hyperspectral remote sensing for soil property estimation. Adv. Agron. 2012, 116, 211–259. [Google Scholar]
  24. Ioannidou, K.; Papastefanou, S. Radiometric techniques for soil radioactivity measurements. J. Environ. Radioact. 2012, 103, 66–73. [Google Scholar]
  25. Jones, P.; Vaughan, M. Thermal imaging in agriculture: Applications and developments. Agric. Eng. Today 2020, 38, 27–34. [Google Scholar]
  26. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Mari, J. Deep learning in remote sensing for agricultural applications. Remote Sens. 2020, 12, 1002–1020. [Google Scholar]
  27. Debnath, D.; Vanegas, F.; Sandino, J.; Hawary, A.F.; Gonzalez, F. A Review of UAV Path-Planning Algorithms and Obstacle Avoidance Methods for Remote Sensing Applications. Remote Sens. 2024, 16, 4019. [Google Scholar] [CrossRef]
  28. Rouse, J.W.; Haas, R.H.; Schell, J.A.; Deering, D.W. Monitoring vegetation systems in the Great Plains with ERTS. Proc. Third Earth Resour. Technol. Satell.-1 Symp. NASA 1974, 1, 309–317. [Google Scholar]
  29. Gamon, J.A.; Peñuelas, J.; Field, C.B. A narrow-waveband spectral index that tracks diurnal changes in photosynthetic efficiency. Remote Sens. Environ. 1992, 41, 35–44. [Google Scholar] [CrossRef]
  30. Chang, C.; Yang, C.; Rao, S. Hyperspectral imaging for precision agriculture: Challenges and opportunities. J. Agric. Eng. Res. 2022, 34, 123–134. [Google Scholar]
  31. Smith, J.; Brown, P.; Zhao, L. Gamma-ray spectrometry in soil analysis: Applications and advancements. Precis. Agric. 2021, 28, 45–67. [Google Scholar]
  32. Lee, A.; Wang, Y.; Miller, R. Thermal imaging for crop stress detection: A review of techniques and applications. Agric. Technol. Adv. 2023, 17, 210–225. [Google Scholar]
  33. Isaac, M.S.A.; Ragab, A.R.; Garcés, E.C.; Luna, M.A.; Peña, P.F.; Cervera, P.C. Mathematical modeling and designing a heavy hybrid-electric quadcopter, controlled by flaps. Unmanned Syst. 2022, 10, 241–253. [Google Scholar] [CrossRef]
  34. Ale Isaac, M.S.; Luna, M.A.; Ragab, A.R.; Ale Eshagh Khoeini, M.M.; Kalra, R.; Campoy, P.; Flores Peña, P.; Molina, M. Medium-scale uavs: A practical control system considering aerodynamics analysis. Drones 2022, 6, 244. [Google Scholar] [CrossRef]
  35. Ale Isaac, M.S.; Ragab, A.R.; Luna, M.A.; Ale Eshagh Khoeini, M.M.; Campoy, P. Thrust Vectoring Control for Heavy UAVs, Employing a Redundant Communication System. Sensors 2023, 23, 5561. [Google Scholar] [CrossRef] [PubMed]
  36. Ale Isaac, M.S.; Flores Pena, P.; Andres Luna, M.; Refaat Ragab, A.; Campoy, P. Sensing and Control Integration for Thrust Vectoring in Heavy UAVs: Real-World Implementation and Performance Analysis. Unmanned Syst. 2024, 1–23. [Google Scholar] [CrossRef]
  37. Wilson, J.T. Economic barriers to technology adoption in smallholder farming. Agric. Econ. Rev. 2019, 62, 101–110. [Google Scholar]
  38. Green, P.; White, E. Reducing agricultural runoff through precision farming technologies. J. Environ. Manag. 2021, 285, 1122–1136. [Google Scholar]
  39. Kamruzzaman, M.; Sun, D.-W. Introduction to Hyperspectral Imaging Technology. In Computer Vision Technology for Food Quality Evaluation; Elsevier: Amsterdam, The Netherlands, 2016; pp. 111–139. [Google Scholar]
Figure 1. Quad Hopper UAV in operation.
Figure 1. Quad Hopper UAV in operation.
Drones 09 00129 g001
Figure 2. Sensors and equipment equipped in the Quad Hopper drone.
Figure 2. Sensors and equipment equipped in the Quad Hopper drone.
Drones 09 00129 g002
Figure 3. Flowchart for the UAV-based multi-modal stress detection system.
Figure 3. Flowchart for the UAV-based multi-modal stress detection system.
Drones 09 00129 g003
Figure 4. Quad Hopper hyperspectral image acquisition, datacube, and spectral content of several pixels.
Figure 4. Quad Hopper hyperspectral image acquisition, datacube, and spectral content of several pixels.
Drones 09 00129 g004
Figure 5. Detection of fungal stress, showing healthy and stressed plants.
Figure 5. Detection of fungal stress, showing healthy and stressed plants.
Drones 09 00129 g005
Figure 6. Soil texture analysis.
Figure 6. Soil texture analysis.
Drones 09 00129 g006
Figure 7. Incorporated map representing the classification results of soil texture.
Figure 7. Incorporated map representing the classification results of soil texture.
Drones 09 00129 g007
Figure 8. Comparative performance analysis.
Figure 8. Comparative performance analysis.
Drones 09 00129 g008
Figure 9. Predicted reflectance values vs. ground-measured values.
Figure 9. Predicted reflectance values vs. ground-measured values.
Drones 09 00129 g009
Table 1. Comparative analysis performance.
Table 1. Comparative analysis performance.
MetricProposed SystemTraditional MethodsImprovementError MarginConfidence Interval (95%)
Spectral Signature Accuracy94.8%80.3%+18.1% ± 1.5 % 93.3–96.3%
Thermal Image Resolution (pixels)640 × 480320 × 240+100% ± 10 pixels630 × 470–650 × 490
Classification Accuracy96.5%81.2%+15.3% ± 1.2 % 95.3–97.7%
Processing Speed (s/image)0.52.8+82.1% ± 0.1 s/image0.4–0.6 s/image
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Flores Peña, P.; Ale Isaac, M.S.; Gîfu, D.; Pechlivani, E.M.; Ragab, A.R. Unmanned Aerial Vehicle-Based Hyperspectral Imaging and Soil Texture Mapping with Robust AI Algorithms. Drones 2025, 9, 129. https://doi.org/10.3390/drones9020129

AMA Style

Flores Peña P, Ale Isaac MS, Gîfu D, Pechlivani EM, Ragab AR. Unmanned Aerial Vehicle-Based Hyperspectral Imaging and Soil Texture Mapping with Robust AI Algorithms. Drones. 2025; 9(2):129. https://doi.org/10.3390/drones9020129

Chicago/Turabian Style

Flores Peña, Pablo, Mohammad Sadeq Ale Isaac, Daniela Gîfu, Eleftheria Maria Pechlivani, and Ahmed Refaat Ragab. 2025. "Unmanned Aerial Vehicle-Based Hyperspectral Imaging and Soil Texture Mapping with Robust AI Algorithms" Drones 9, no. 2: 129. https://doi.org/10.3390/drones9020129

APA Style

Flores Peña, P., Ale Isaac, M. S., Gîfu, D., Pechlivani, E. M., & Ragab, A. R. (2025). Unmanned Aerial Vehicle-Based Hyperspectral Imaging and Soil Texture Mapping with Robust AI Algorithms. Drones, 9(2), 129. https://doi.org/10.3390/drones9020129

Article Metrics

Back to TopTop