1. Introduction
The integration of unmanned aerial vehicles (UAVs) with artificial intelligence (AI) is rapidly transforming agricultural systems through enhanced data acquisition, improved monitoring efficiency, and support for informed, data-driven decision-making [
1]. These technologies form the foundation of modern agriculture, enabling non-destructive approaches, scalable solutions, and progress toward real-time analytics [
2,
3].
In modern agricultural research, automated tools such as UAVs and high-resolution cameras are widely used to rapidly and non-destructively capture large-scale trait data [
4]. These image-based technologies enable researchers to simultaneously assess structural, physiological, and biochemical traits across time and space, thereby providing a comprehensive view of plant performance and variability. Moreover, this approach reduces subjectivity by minimizing human bias and supports real-time or near-real-time monitoring of crop development throughout the growing season [
5,
6]. At the same time, advances in web-based platforms powered by AI offer promising alternatives for crop monitoring by improving data analysis efficiency and equipping end users with accessible, user-friendly tools. For example, these platforms facilitate automated trait extraction from UAV-acquired images through intuitive interfaces, eliminating the need for advanced image-processing expertise and enabling broader implementation in both research and commercial contexts [
2,
7,
8]. By combining AI algorithms, pre-trained computer vision models, and cloud-based processing, such platforms can reliably generate object counts, estimates of canopy size, plant height, object dimensions, and other phenotypic parameters from standard RGB or multispectral imagery [
1,
8]. These innovations have been widely applied in large-scale cropping systems (e.g., corn, soybeans, and cotton). However, their application in specialty crops remains limited [
9]. This makes such technologies particularly valuable in specialty crop systems, where high phenotypic diversity and labor-intensive measurements constrain the feasibility of traditional monitoring methods.
As digital tools become more integrated into agricultural management, ensuring their accessibility and usability, particularly for smallholder farmers and non-expert users, has become essential [
10,
11]. One example is Solvi [
12], a drone-based analytics platform designed for field trial assessments. Founded in 2015 with the goal of making advanced drone technology accessible and easy to use, Solvi originated as a pilot project with the Swedish University of Agricultural Sciences. Today, it is widely used by farmers, agronomists, and researchers to support more efficient and sustainable crop management. Due to its capabilities, Solvi has been applied in a few studies for image-based analysis, including the determination of in-season nitrogen rates [
13] and prediction of grain protein concentration in winter wheat [
14], as well as the estimation of leaf nitrogen content, leaf area, berry yield [
15], and monitoring and prediction of wild blueberry [
16]. However, despite these advantages in data analysis and time-saving potential, its application remains limited in the literature, particularly with respect to biometric and phenotypic data in specialty crops. This gap is especially evident in perennial and vegetable crops such as pecan and onion, which, despite their high economic value and spatially stable planting structures, have received considerably less attention than annual field crops [
17].
Biometric and phenotypic data play a central role in modern crop management. In perennial crops such as pecan, canopy attributes such as area and tree height serve as key indicators of plant vigor, productivity potential, and overall orchard health. These measurements also enable producers to manage inputs (e.g., water, chemicals, fertilizers) more efficiently [
18,
19]. In vegetable crops such as onions, bulb size is directly linked to market classification standards and yield estimation, and strongly influences consumer preferences [
20]. For both pecan and onion, conventional trait analysis still relies heavily on manual measurements. The integration of AI-based tools that require minimal user input can therefore broaden accessibility for growers, researchers, and Extension agents, while offering real-time decision support across diverse production environments [
7,
11].
Therefore, based on the premise that UAV- and AI-driven digital agriculture platforms can provide reliable phenotypic measurements across diverse crop morphologies while offering a user-friendly and scalable alternative to traditional field methods, we aimed to evaluate the performance of a commercial AI-powered digital agriculture web platform (Solvi) for automated biometric analysis in specialty crops (pecan and onion) using UAV-acquired imagery.
2. Materials and Methods
2.1. Study Environment
The study was conducted in two experimental fields at the University of Georgia, Georgia, USA (
Figure 1A). The fields were designated as Field 1 (Pecan; 31°30′31.0″ N 83°39′06.2″ W) and Field 2 (Onions; 32°01′08.6″ N 82°13′19.6″ W), selected to represent a range of crop types, growth architectures, and phenotyping challenges. Field 1 was located at Ponder Farms in Tifton, GA, USA (
Figure 1B), and the data collection was performed on 28 October 2024. It consisted of a pecan orchard with trees at three distinct developmental stages: young (1–5 years after transplanting), mid-age (5–10 years), and mature (over 10 years). Trees were transplanted in a 14 × 14 m grid. Field 2 was located at the Vidalia Onion and Vegetable Research Center in Reidsville, GA, USA (
Figure 1C), and the data collection was performed on 15 May 2025. The onions were cultivated on beds spaced 1.8 m center-to-center. Seedlings were transplanted at 0.3 m between rows and 0.1 m between plants, with four rows of onions per bed, totaling 222,000 plants per hectare.
The climate of the region is classified as humid subtropical, with an average annual precipitation of 1400 mm and an average temperature of 20 °C. Moreover, cultivation practices were conducted in accordance with regional environmental conditions and the specific agronomic requirements of the crops.
2.2. UAV Image Data Collection
For both fields, RGB images were acquired using a multirotor UAV (DJI Mavic 3 Multispectral, Shenzhen, China) equipped with a dual-camera system. The primary RGB camera features a 4/3-inch CMOS sensor with an effective resolution of 20 megapixels (5280 × 3956 pixels), a field of view (FOV) of 84°, and a 24 mm equivalent focal length (approximately 35 mm in full-frame format). The UAV also includes a 1/2.8-inch multispectral camera with five bands (Green, Red, RedEdge, and NIR), each with a resolution of 5 megapixels, a 25 mm equivalent focal length, and a 73.91° FOV. To ensure image quality and radiometric consistency, the UAV is equipped with a 3-axis mechanical gimbal (tilt, roll, and pan) and a sunlight sensor that records real-time irradiance for automatic radiometric correction during post-processing. In this study, only the RGB images were used. The UAV also integrates an RTK-GNSS module capable of receiving signals from GPS, GLONASS, BeiDou, and Galileo constellations. During the flights, positioning accuracy was maintained within 5 cm horizontally and vertically.
The flights were planned and executed using a mobile app (DJI Pilot 2, Shenzhen, China). All flights were conducted around solar noon (±1 h) under clear sky conditions to minimize shadow effects. Flight parameters were adjusted for each field to capture high-resolution imagery of the respective targets. For Field 1, the UAV was operated at a speed of 3 m/s, an altitude of 74 m above ground level (AGL), and with 80% frontal and side image overlap, resulting in a ground sample distance (GSD) of approximately 2.5 cm/pixel. For Field 2, the UAV was also operated at 3 m/s, but at an altitude of 12 m AGL, and with 85% frontal and side image overlap, yielding a GSD of 0.33 cm/pixel.
2.3. Ground-Truth Data Collection
Field 1: For tree counting, we considered all the trees in the field, totaling 242 trees. For tree height and canopy area measurements, we selected 18 pecan trees (6 for each development stage;
Section 2.1) (
Figure 2). Tree height was measured manually using a measuring tape, considering the vertical distance from ground level to the highest vegetative point. Canopy area was determined using a photogrammetry-driven approach with structure-from-motion software (Agisoft Metashape Professional 2.1.1, St. Petersburg, Russia). High-resolution UAV imagery was processed to generate orthomosaics. The canopy area was then calculated by manually delineating the horizontal projection of each tree canopy using the software’s built-in tools. This combination of manual height measurements and computer-aided canopy analysis was treated as ground-truth data. The pecan ground-truth measurements were later compared to the outputs produced by the AI-driven web platform (Solvi), which automatically counted and estimated tree height and canopy area from UAV imagery.
Field 2: For bulb counting, we delineated five plots of 10 m
2 and counted all onions within them, totaling 697 onions. For diameter measurements, 20 georeferenced sample points were established across the field. At each point, ten onions were measured and labeled (
Figure 3). Bulb diameter was measured using digital calipers at the equatorial point of each analyzed bulb. No soil removal was performed; onions remained under standard field conditions throughout data collection. Bulb diameter was also used for market classification: medium (<76.2 mm), jumbo (76.2–95.3 mm), and colossal (>95.3 mm) [
21]. According to GDA [
21] standards, there is a 5% tolerance for onions smaller than the minimum specified diameter and a 10% tolerance for larger onions. The diameter measurements and market classifications were calculated and later compared to the AI-driven web platform (Solvi).
2.4. AI-Driven Web Platform (Solvi)
UAV images collected in this study were processed using an AI-driven web platform (Solvi [
12], Gothenburg, Sweden). The software supports imagery from a wide range of UAV-mounted sensors, including both RGB and multispectral cameras. After uploading, individual images were automatically stitched into georeferenced orthomosaics, with digital surface models (DSMs) generated prior to orthorectification; alternatively, users can upload their own DSMs. For tree height estimation, each detected tree is buffered by 1 m, and the lowest point within this buffer is defined as the ground level. Tree height is then calculated as the difference between the highest and lowest elevation points within the buffered area.
For automatic boundary creation, the user needs to provide the platform with a set of hand-drawn (annotation) examples of plant boundaries. The platform then fine-tunes its pre-trained model using these examples to improve measurement accuracy. The model architecture is based on Mask R-CNN with a ResNet101 backbone and Feature Pyramid Network (FPN), trained with large-scale jittering. Hand-drawn examples were created by manually outlining the visible canopy contours, with the operator zooming in and out to ensure precise edge delineation and complete boundary closure. In general, a greater number of examples results in better detection. For the pecan field, we created 12 examples, totaling 50 identified trees. For the onion field, we created 6 examples, totaling 298 identified onions. The number of examples was adjusted iteratively based on performance: initial inputs did not yield satisfactory results; therefore, additional examples were added until the outputs stabilized. Results were visualized as interactive maps, allowing for comparisons among plots and classification based on user-defined variables. All processed data can be exported in standard formats (e.g., CSV, SHP, TIFF) for downstream statistical analysis or accessed programmatically via an application programming interface (API). In this study, we used these segmentation and measurement tools to derive plant-level metrics from onion and pecan datasets, enabling direct comparisons with ground-truth observations (
Figure 4).
2.5. Data Analysis
For counting measurement, we used the precision, recall, and F1 Score metrics to assess the platform’s efficiency. Additionally, for tree height, canopy area, and bulb diameter, linear regression models were performed to model estimated and observed values. The performance of the models was assessed using the coefficient of determination (R2), mean absolute error (MAE), root mean squared error (RMSE), and mean absolute percentage error (MAPE). For better validation, pecan tree analysis was stratified by developmental stage (young, mid-age, and mature). Conversely, onion analysis was grouped by market class (medium, jumbo, and colossal). Moreover, supplementary statistical analyses (interquartile range (IQR) and one-way ANOVA, p < 0.05) were performed to identify potential outliers and to evaluate differences between observed and estimated measurements, respectively. All statistical computations and data visualizations were conducted using Microsoft Excel.
4. Discussion
Several studies have highlighted the advantages and limitations of UAV-based analysis and explored recent advances in AI [
22,
23,
24]. However, most research has focused on field crops, leaving a gap in applications tailored to specialty crops. In particular, the use of commercially available AI-driven web platforms for UAV-based data remains largely underexplored in these systems. Our findings demonstrated that such platforms can accurately count the number of trees and onions, estimate tree height, canopy area, and bulb diameter under real-world field conditions, underscoring their potential for practical implementation in precision horticulture.
In pecan, the overall regression metrics were strong (R
2 > 0.98), and errors were <23.16%. However, the model efficiency varied by developmental stage. For instance, young trees exhibited higher errors for height (19.43%) and canopy area (46.70%) measurement. Although MAE and RMSE were numerically smaller for young trees due to their limited size, relative errors (MAPE) were higher, indicating proportionally lower model efficiency rather than absolute error magnitude. Likely, their smaller stature and reduced pixel representation made the tree detection and further measurement more challenging. These findings are consistent with previous studies, which have shown that limited structural definition in early growth stages impairs UAV-based feature extraction [
18,
25]. Moreover, tree height estimates may have been affected by GNSS positional accuracy, particularly in altitude, which is generally associated with higher error. Similar findings have been reported in previous studies [
26,
27]. Conversely, canopy area estimation followed a similar pattern, with model performance improving as trees matured. MAPE decreased from 46.7% in young trees to just 5.5% in mature ones, indicating enhanced precision when applied to well-developed canopies with clearly defined geometry and greater visual separability from the background. To verify that the higher MAPE value was not driven by abnormal values, an outlier analysis using the interquartile range (IQR) method was performed, revealing no outliers in the dataset (
Figures S1 and S2). Therefore, no data removal or cleaning was applied, and all analyses were conducted using the original dataset.
Canopy-related traits such as height and canopy area are agronomically critical for managing tree vigor, input allocation, and productivity in perennial nut crops. In Pecan orchards, these attributes guide key decisions including irrigation scheduling, disease control, and yield potential modeling [
18,
19,
25]. These attributes also contribute to the decision-making for chemical applications. The ability to extract these traits rapidly and non-destructively using UAV-based tools provides substantial value for both research and commercial orchard management. Perennial crops such as pecan present both opportunities and specific challenges for the application of precision agriculture technologies [
28]. Their fixed spatial layout is well-suited to the use of UAVs and stationary sensors, but complicates tasks such as yield monitoring and fruit detection due to the three-dimensional variability of the canopy [
17]. Within the precision agriculture framework proposed by Bhakta et al. [
29], comprising data collection, data analysis, decision-making, and variable rate application, the platform evaluated in this study directly addresses the first two layers. It enables high-throughput trait acquisition and real-time analysis via cloud-based processing, offering an accessible and scalable alternative to more technically demanding phenotyping systems.
Onion is a high-value horticultural crop whose bulb size is a key determinant of both yield and market classification. Market classes such as medium, jumbo, and colossal are directly tied to consumer preferences and pricing strategies, making size prediction a valuable indicator of economic potential. Estimating market class distribution enables stakeholders to make informed decisions regarding logistics and field-level variability [
20]. In this study, size estimation errors varied across market classes. For instance, jumbo bulbs had the lowest relative error (4.9%), while medium-sized bulbs showed more variability (9.9%). This trend matches previous findings suggesting larger, more regularly shaped objects are easier to segment and measure using RGB imagery. In contrast, smaller or less defined targets may suffer from lower resolution and reduced object detection accuracy [
30]. However, in this study, larger bulbs tended to be slightly underestimated (
Figure S3). This effect is likely associated with partial soil coverage at the bulb perimeter and the irregular curvature of the exposed bulb surface, which can cause the visible portion of the bulb in RGB images to appear smaller than its true equatorial diameter. Additionally, mild shadowing and saturation effects in bright field conditions may have contributed to reduced pixel contrast at the bulb edges, further influencing underestimation.
Our results expand this body of work by demonstrating that UAV-derived imagery, processed through an AI-driven platform, can directly estimate bulb diameter and classify onions into the official market categories. Such in-field estimations of market class distribution provide actionable information for growers, supporting early decisions on harvest scheduling, storage allocation, and marketing strategies. Anticipating the proportion of medium, jumbo, or colossal onions before harvest enables producers to align with contractual demands and optimize postharvest handling. Furthermore, mapping within-field variability in bulb size creates opportunities for targeted management interventions, such as site-specific fertilization or irrigation, to enhance uniformity and reduce losses. Collectively, these applications highlight the added value of UAV-based monitoring in onion production, where profitability is closely linked to size grading and market class compliance.
To the best of our knowledge, this is the first study to quantitatively evaluate the performance of the AI-driven web platform Solvi for tree counting and biometric estimations. Previous research has primarily relied on alternative image-processing pipelines, such as image stitching [
13,
14,
15,
16]. Consequently, direct comparisons with Solvi-based studies are not yet possible, which limits cross-study benchmarking but underscores the novelty and relevance of the present work.
Finally, our results underscore the potential of the AI-driven platforms to democratize digital monitoring of vegetable crops. Their intuitive interfaces, cloud-based automation, and reliable trait extraction make them particularly useful for growers, extension professionals, and researchers with limited access to technical infrastructure. However, caution is warranted when applying such tools to early growth stages or small structures.