3.1. UAV System
For the AgriAdapt experiment, we developed a custom UAV system featuring a hexa-rotor design with a maximum take-off weight of up to 6 kg (as shown in
Figure 1). This UAV is equipped with a Plug and Play System that allows for quick switching between different payloads. It supports real-time mission management and Autonomous Waypoint navigation, facilitated by the Pixhawk PX4 flight controller, manufactured by Dronecode Foundation, San Francisco, CA, USA. The PX4 integrates an Inertial Measurement Unit (IMU), which includes accelerometers and gyroscopes, along with a Global Positioning System (GPS) receiver.
Table 2 summarizes the key hardware components of the UAV platform and their indicative single-unit prices, highlighting its affordability and suitability for low-cost precision agriculture applications.
Our UAV configuration includes an Nvidia Jetson Nano, powered by an independent battery supply, and a radio transceiver module for remote communication with the ground control station. A schematic overview of the entire configuration is depicted in
Figure 2. The Jetson Nano connects to the PX4 flight controller via a serial UART protocol, allowing access to telemetry data. The payload features an Arducam autofocus camera equipped with a Sony IMX519 sensor (Sony Corporation, Tokyo, Japan), offering 16 MP spatial resolution. To ensure high-quality image acquisition, the camera is mounted on a 2-axis gimbal with brushless motors for stabilization.
This system is capable of a maximum flight time of up to 20 min. The Italian Civil Aviation Authority (ENAC) has certified the design, ensuring compliance with both Italian and European (EASA) regulations.
The payload system is designed to integrate high-performance components that facilitate real-time weed detection and geotagging in agricultural fields. It includes an RGB camera, a GPS module, a processing unit, and a communication transceiver, all working together to ensure precise data collection, processing, and storage.
The Arducam 16 MP Autofocus Camera (Arducam Technology Co., Ltd., Hong Kong, China) plays a critical role in capturing high-resolution images of the agricultural field. With its 16 Megapixel resolution (4608 × 3456), the camera provides the detailed imagery necessary for precise analysis. The autofocus feature was enabled to maintain sharp image quality even when the UAV was in motion, ensuring that vegetation contours remained well-defined. The camera operated with fixed exposure and color parameters, as radiometric calibration and automatic white balance were not applied. This configuration ensured consistent image characteristics across flights, minimizing illumination-induced variations while preserving radiometric comparability. The camera is mounted on a 2-axis gimbal, which stabilizes the images during flight by compensating for UAV movements.
The NEO-M9N GPS Module (u-blox AG, Thalwil, Switzerland) provides accurate geolocation data. Supporting multiple GNSSs (Global Navigation Satellite Systems) including GPS, GLONASS, Galileo, and BeiDou, it enhances positioning accuracy. The GPS module offers horizontal accuracy within 2.0 m and vertical accuracy within 4.0 m, essential for precise mapping. While this level of accuracy is not sufficient for individual plant localization, it is adequate for field-level referencing and flight-path logging, as image overlap and relative positioning ensure spatial consistency within each field. Its high update rate, up to 10 Hz, ensures timely and continuous location updates during UAV operations, providing reliable data for geotagging.
Central to the payload system is the NVIDIA Jetson Nano, which performs real-time image analysis and processing using onboard AI algorithms. The Jetson Nano features an NVIDIA Maxwell architecture GPU with 128 CUDA cores, delivering the computational power required for deep learning tasks. Its quad-core ARM Cortex-A57 CPU supports efficient multitasking and data processing, while the 4 GB 64-bit LPDDR4 memory ensures smooth operation of multiple neural networks in parallel. Connectivity options, including Gigabit Ethernet, USB 3.0 ports, and MIPI CSI-2 for the camera interface, facilitate the integration of various sensors and peripherals, making the Jetson Nano a powerful processing unit for the payload system.
The Jetson Nano was selected as the onboard computing platform because it provides a favorable balance between processing capability and energy efficiency, which is critical for real-time UAV operation. The board is capable of running modern deep learning inference tasks while consuming as little as 5–10 W of power, making it suitable for battery-powered aerial platforms. Its small form factor and low thermal profile further support seamless integration into lightweight UAV payloads.
The LoRa HM-TRLR-D TTL-868 Transceiver (Hope Microelectronics Co., Ltd., Shenzhen, China) ensures robust long-range communication between the UAV and the ground station. Operating at a frequency of 868 MHz, the transceiver supports data rates of up to several hundred kbps, suitable for transmitting essential status updates and operational data. The LoRa transceiver maintains stable communication over several kilometers in line-of-sight conditions, ensuring reliable connectivity during UAV missions.
The Ground Station Control Software (custom-developed) is crucial for managing real-time monitoring and control of the UAV and its payload. Developed using Python and PyQt, the software provides a user-friendly interface for operators. It allows for starting, pausing, and stopping data acquisition, as well as monitoring the system’s status, including battery levels and live data feeds. This software ensures that UAV and payload operations are managed efficiently, facilitating seamless data collection and analysis.
3.2. Dataset
We evaluated our approach on two UAV-based weed detection datasets: an initial small-scale dataset for preliminary validation and a new, dedicated dataset that constitutes one of the main contributions of this work. The preliminary dataset consisted of several hundred raw UAV images collected with consumer-grade drones over mixed test fields containing different crop species. This dataset was used exclusively for early validation of the processing pipeline.
Building upon this, we acquired a new salad-weed dataset using the UAV system described in
Section 3.1. The dataset was collected over salad fields at different growth stages (approximately 25–65 days old). Flights were conducted at an average altitude of 5 m, resulting in a ground sampling distance of 0.4 cm/pixel. All images were captured at a resolution of
pixels using the UAV’s stabilized RGB camera.
Figure 3 shows representative UAV images of salad rows interspersed with weeds, while
Figure 4 illustrates the manually annotated masks, where weeds are highlighted in red overlay and all other areas—including crop (salad) rows—are considered as background.
To facilitate systematic use, the dataset is organized into two main partitions: Field_ID_1 and Field_ID_2.
Field_ID_1: Acquired during three consecutive short flights over adjacent salad fields. The original collection comprised 397 images. After manual inspection, images with no useful crop content (e.g., field edges, empty soil regions) were discarded, resulting in 322 annotated images.
Field_ID_2: Acquired more than one month later over a larger, adjacent field. The original set included 408 images, of which 321 were retained after quality control.
Table 3 summarizes the dataset partitions, including acquisition dates, locations, and environmental conditions. In total, the dataset contains 643 high-quality images across the two partitions. Each image was manually labeled in YOLOv7 PyTorch format, distinguishing between salad crops and weeds. Pixel data were standardized by removing EXIF-based orientation metadata and resizing all images to
pixels, using bilinear interpolation. Because the original UAV images were captured at 1280 × 1280 pixels with a square aspect ratio, resizing to 640 × 640 did not require stretching, padding, or other aspect-ratio adjustments. No additional pre-processing was applied, preserving the raw visual information.
The annotation process was performed manually by two trained annotators using the Roboflow online platform [
32]. Each image was independently segmented at the pixel level to distinguish between weed and crop regions. For morphologically similar cases (e.g., salad seedlings vs. low-growing broadleaf weeds), annotators relied on visible leaf-edge serration and venation patterns to ensure consistent differentiation. To ensure consistency, the two annotators cross-checked and refined each other’s labels under supervision of the dataset curator. Final masks were exported in
YOLOv7-compatible PyTorch format, ensuring standardized input for all network training procedures. This manual double-annotation process ensured high-quality ground truth data and minimized labeling bias.
This dataset is made publicly available (Dataset available at:
https://app.roboflow.com/agriadaptweeddetection/agriadapt-uex2n/overview (accessed on 25 September 2025)) and serves as a benchmark for evaluating real-time weed detection in salad crops. Also, because it was acquired using the same hardware pipeline deployed for inference experiments, it provides a realistic testbed for developing and validating lightweight models suitable for embedded UAV platforms.