Next Article in Journal
Rubber Plantation Expansion Leads to Increase in Soil Erosion in the Middle Lancang-Mekong River Basin During the Period 2003–2022
Previous Article in Journal
A Category–Pose Jointly Guided ISAR Image Key Part Recognition Network for Space Targets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Succulent-YOLO: Smart UAV-Assisted Succulent Farmland Monitoring with CLIP-Based YOLOv10 and Mamba Computer Vision

1
Graduate School of Frontier Sciences, The University of Tokyo, Kashiwa 277-8561, Japan
2
Department of Math and Applied Mathematics, China University of Petroleum-Beijing, Beijing 102249, China
3
Department of Environmental Health Sciences, University of California, Los Angeles, CA 90095, USA
4
Graduate School of Global Environmental Studies, Sophia University, Tokyo 102-8554, Japan
5
Data Science and AI Innovation Research Promotion Center, Shiga University, Hikone 522-8522, Japan
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(13), 2219; https://doi.org/10.3390/rs17132219 (registering DOI)
Submission received: 23 April 2025 / Revised: 18 June 2025 / Accepted: 20 June 2025 / Published: 28 June 2025

Abstract

Recent advances in unmanned aerial vehicle (UAV) technology combined with deep learning techniques have greatly improved agricultural monitoring. However, accurately processing images at low resolutions remains challenging for precision cultivation of succulents. To address this issue, this study proposes a novel method that combines cutting-edge super-resolution reconstruction (SRR) techniques with object detection and then applies the above model in a unified drone framework to achieve large-scale, reliable monitoring of succulent plants. Specifically, we introduce MambaIR, an innovative SRR method leveraging selective state-space models, significantly improving the quality of UAV-captured low-resolution imagery (achieving a PSNR of 23.83 dB and an SSIM of 79.60%) and surpassing current state-of-the-art approaches. Additionally, we develop Succulent-YOLO, a customized target detection model optimized for succulent image classification, achieving a mean average precision (mAP@50) of 87.8% on high-resolution images. The integrated use of MambaIR and Succulent-YOLO achieves an mAP@50 of 85.1% when tested on enhanced super-resolution images, closely approaching the performance on original high-resolution images. Through extensive experimentation supported by Grad-CAM visualization, our method effectively captures critical features of succulents, identifying the best trade-off between resolution enhancement and computational demands. By overcoming the limitations associated with low-resolution UAV imagery in agricultural monitoring, this solution provides an effective, scalable approach for evaluating succulent plant growth. Addressing image-quality issues further facilitates informed decision-making, reducing technical challenges. Ultimately, this study provides a robust foundation for expanding the practical use of UAVs and artificial intelligence in precision agriculture, promoting sustainable farming practices through advanced remote sensing technologies.

1. Introduction

Succulent plants are a unique type of drought-resistant vegetation that can store large amounts of water in their leaves, stems, or roots, making them highly adaptable to various environments [1]. Globally, succulents have diverse purposes—from ornamental uses to food sources, fibers, beverages, and natural products [2]. Studies have shown that they also have potential in the field of bioenergy [3]. Due to their resilience in the face of global climate change, succulents are highly valued for their role in habitat protection, urban environments, and ecosystem services [2]. Particularly in ecological agriculture, succulents play a crucial role in arid and semi-arid regions, which contributes to water conservation, soil erosion control, and ecosystem stability [4].
The research on succulent plant classification holds significant ecological and economic value. Accurate classification benefits biodiversity conservation and plant resource management and deepens our understanding of different species [2]. Moreover, it supports the horticultural industry, enabling hobbyists and businesses to identify and select varieties effectively, thereby optimizing cultivation and sales [5]. Given the diversity of succulent species and the challenges in their identification, particularly for non-experts, large-scale classification studies are essential [6]. Such studies facilitate precision agriculture and sustainable ecological practices, which can boost farmer incomes and promote agricultural technology.
In recent years, computer vision (CV) technology has been widely applied for agriculture, which uses advanced image acquisition tools and intelligent algorithms to obtain accurate monitoring of crop growth conditions [7,8]. For example, CV has been used to detect grape clusters [9], assist in palm oil plantation management [10], and determine wheat growth stages [11]. CV also plays a key role in succulent plant detection. By analyzing succulent images, CV can identify plant varieties, offering valuable management insights to optimize growing conditions and enhance both the quality and ornamental value of succulents [12].
The combination of CV and unmanned aerial vehicles (UAVs) has become a leading method in precision agriculture since it has high detection accuracy, cost-efficiency, full-field coverage, ease of use, and the ability to support decision-making [13,14,15,16]. CV and UAV-based techniques have been used to study farmland–soil temperature relationships [17], evaluate agronomic parameters in winter wheat [18], and estimate cabbage biomass [19]. However, traditional digital image processing (DIP) methods such as color, texture, and shape feature extraction can be influenced by observer subjectivity, reducing classification accuracy [20]. Furthermore, succulents have very little morphological variation, making them difficult to classify by traditional methods [21].
Deep learning-based image processing, especially convolutional neural networks (CNNs), has gained attention for its ability to extract high-level features from images automatically, improving object detection and classification accuracy [22]. CNN-based models have been used in various crop detection tasks, including disease detection [23] and crop classification optimization [24]. CNNs have evolved from simple plant recognition [25] to more complex tasks such as fruit detection using Faster Region-based Convolutional Neural Network (Faster R-CNN) [26]. However, the fine-grained classification of succulent plants remains challenging due to minimal differences between classes and significant variability within the same class, which are further compounded by limited datasets and suboptimal image quality [21].
Data augmentation (DA) can help address the problem of limited datasets by generating more training samples, although it may sometimes impact model accuracy if the augmented data fails to preserve the original characteristics [27]. At the same time, UAVs as a data collection tool can provide real and diverse data sources to further supplement the actual scene information that cannot be generated by DA. UAV-based remote sensing, combined with deep learning, has facilitated precise crop classification, yield estimation, and water stress monitoring [28,29,30]. However, due to vibrations and altitude constraints, UAVs face limitations in image resolution and quality, which compromises detection efficiency [31,32].
Super-resolution reconstruction (SRR) techniques, especially deep learning-based models, offer potential solutions to the UAV low-resolution problem [33]. SRR can enhance image resolution and restore or enhance detailed information in an image by transforming low-resolution (LR) images to high-resolution (HR) images. SRR algorithms can also improve model performance [34]. Traditional approaches contain interpolation, reconstruction-based methods, and more recently, learning-based techniques [35]. Among them, deep learning-based SRR algorithms have demonstrated superior performance by learning complex image patterns through deep learning models [36,37].
The SRR model based on MambaIR integrates state-space modeling, balancing global receptive fields and efficient computation while restoring image details [38]. These models have been applied in fields like medical image reconstruction, but their potential for enhancing succulent plant classification remains largely untapped.
This study introduces an innovative approach that combines SRR with an object detection model to improve succulent plant detection and classification. By integrating MambaIR-based SRR with YOLOv10, this approach enhances detection accuracy in low-resolution images. Additionally, we utilize the contrastive language-image pre-training (CLIP) image-text encoder [39] to enhance feature extraction and classification capabilities. We further analyzed the effects of scaling factors, Gaussian blur, and noise on image degradation and reconstruction and evaluated the robustness of the proposed detection model using these newly generated test sets.
The structure of this paper is outlined below: Section 2 outlines the research area, details regarding data acquisition, preprocessing procedures, and the algorithms utilized for both SRR and object detection. Section 3 presents experimental outcomes and compares multiple SRR methods. Section 4 explores the effects of magnification scale, various detection architectures, Gaussian blur, and noise on detection performance. Finally, Section 5 concludes the paper.
The outline of the current study is organized as follows: Section 2 introduces the survey site, image acquisition and preprocessing procedures, as well as the reconstruction and identification algorithms employed. Section 3 presents the experimental results, including a comparative analysis of different SRR algorithms. Section 4 discusses the degradation of the test datasets and evaluates the robustness of the proposed detection model. Section 5 concludes the study.

2. Materials and Methods

The workflow of the current research is shown in Figure 1:
Step 1: High-definition images are captured using UAVs for data collection. As a data preprocessing step, the Scale-Invariant Feature Transform (SIFT) [40] algorithm is applied to process these images and generate panoramic images for mapping the distribution of succulent plants.
Step 2: For image SRR, the panoramic images of succulent plants are cropped into 600 × 600 pixel blocks to construct the HR dataset. The HR dataset is then downsampled to generate an LR dataset of 150 × 150 pixels. Several deep learning-based SRR methods, including MambaIR, are utilized on the LR dataset to enhance the resolution to 600 × 600 pixels. Additionally, various augmentation strategies are applied to enrich the dataset and enhance its variability.
Step 3: In the object detection stage, the HR dataset is divided into training and validation subsets. Various detectors, including Faster R-CNN, DEtection Transformer (DETR), Ghost Network (GhostNet), the You Only Look Once (YOLO) series, and the Succulent-YOLO model, are applied. These models are then used on the SRR dataset to produce prediction results.
Step 4: The visualization stage primarily involves result visualization and further analysis. First, Gradient-weighted Class Activation Mapping (Grad-CAM) technology is employed to generate heatmaps that highlight the key parts within the images that the detector focuses on. The subsequent analysis centers on several points: the impact of the SRR scaling ratio, the performance of different detectors on degraded test sets, and the robustness of the proposed model. The workflow of this study is shown in Figure 1. The entire process is divided into four main stages: data collection, super-resolution reconstruction, object detection, and quantification.

2.1. Survey Site

A field survey was conducted at a succulent plant cultivation base in Yuxi City, China, as shown in Figure 2. The city’s unique geographic and climatic conditions make it a key production area for succulent plants. As an important flower planting base in Yunnan Province, Tonghai County has shown a good development trend in succulent planting driven by policy support and market demand. In 2024, the planting area of succulent plants in Tonghai County was 9600 acres, and the output was 81.58 million pots. The floral industry makes a significant contribution to agricultural output, with both planting area and yield showing an upward trend.

2.2. Data Collection and Preprocessing

In recent years, UAVs have become essential tools for agricultural surveys since they offer efficiency, precision, and flexibility in crop image collection [41,42,43,44]. This study employed the DJI Mini 3 UAV for data collection. The device features a compact structure, high-resolution imaging capabilities, and longer working hours, which contribute to the accurate capture of succulent plant images. The compact structure and intelligent navigation further enhance its operability in confined environments, optimizing work efficiency. To ensure comprehensive coverage of the collected data, the Mini 3 follows a curved flight path, allowing it to capture multi-angle images without the need to adjust the camera angle. This minimized redundant data, stabilized the camera to prevent blurring, and ultimately improved image quality. The full data collection procedure is completed within approximately 30 min.
The UAV acquired high-resolution images at 600 × 600 pixels. The captured images were subsequently used to generate a panoramic visualization covering the full succulent planting region. To achieve this, the SIFT algorithm was adopted for stitching multiple image frames, as depicted in Figure 2d,e. Figure 2d is intended to show the orthophoto image of the target area obtained by the drone. Geometric distortion was corrected by using tie points generated through SIFT-like feature matching, followed by iterative bundle block adjustment (BBA), which refined internal and external camera parameters and distortion coefficients to minimize reprojection error. Orthorectification was performed using DEM or point cloud-based terrain elevation references to project corrected images onto a planar map, applying a parallel (orthogonal) projection into the WGS 84/UTM Zone 47N coordinate system. The integration process included identifying shared regions across successive frames, aligning consistent visual features, and combining them to form a seamless panoramic image of the succulent area. As shown in Figure 2f and Figure 3, the Labeling 1.8.1 tool was employed to manually annotate each plant type, with the annotated instances categorized into six categories. The annotated sections were subsequently cropped into high-resolution patches with dimensions of 600 × 600 pixels.

2.3. Super-Resolution Reconstruction Using Deep Learning Techniques

Driven by advancements in computational power, SRR methods based on deep learning have witnessed substantial progress in recent years [45,46]. These approaches utilize neural networks, typically taking LR images—obtained via downsampling or synthetic degradation—as input and reconstructing HR outputs [36]. The LR images are processed through convolutional layers for feature extraction, followed by upsampling stages to produce HR results. The selection of loss functions plays a pivotal role in determining SRR performance, as lower loss values generally reflect better model generalization. Frequently employed loss functions include pixel-level losses (L1, L2, Charbonnier), mean squared error (MSE), perceptual loss, content loss, and adversarial loss. A well-chosen loss function can not only enhance reconstruction quality but also facilitate faster convergence during training [36].

2.3.1. Architectures of SRR Models

Deep learning-based SRR utilizes various network architectures, including CNN-based models (e.g., Super-Resolution Convolutional Neural Network, Very Deep Super-Resolution, Enhanced Super-Resolution Generative Adversarial Network), Transformer-based models, and state-space models like MambaIR, each offering distinct advantages in their respective domains. These models have accelerated the rapid development of deep learning, driving its application across various fields.
Early SRR models, such as Super-resolution convolutional neural network (SRCNN) and Deeply-recursive convolutional network (DRCN), used CNNs to learn mappings from LR to HR images. While these models are simple and easy to implement, SRCNN struggles with complex image content and fine details. With the increasing demand for higher-quality image reconstruction, CNN-based models face challenges in preserving image details and quality. To address these challenges, new architectures like GANs, Transformers, and MambaIR introduce innovative solutions, significantly improving image quality. Enhanced super-resolution generative adversarial network (ESRGAN) and Super-Resolution Generative Adversarial Network (SRGAN) employ adversarial training to enhance image quality, improve textures and details, thereby generating more realistic HR images. Transformers, with their self-attention mechanisms, excel at capturing global information and handling long-range dependencies, which makes them ideal for large-scale image reconstruction. MambaIR employs lightweight attention modules and hierarchical design to improve both efficiency and performance.
As shown in Figure 4, this paper evaluates traditional bicubic interpolation and SRR models based on deep learning, emphasizing their improvements in detail preservation, computational efficiency, and structural complexity. SRCNN, the first to use a shallow network for LR-to-HR mapping, lays the foundation for future work. Enhanced deep residual network (EDSR) improves deep feature extraction by removing batch normalization in ResBlocks. Very deep residual channel attention network (RCAN) introduces channel attention to better restore fine details. Real-world blind super-resolution with pure synthetic data (Real-ESRGAN) refines the degradation model and produces higher-quality images. Swin Transformer for Image Restoration (SwinIR) and Transformer-based models combine convolution and self-attention for high-quality restoration with fewer parameters, reducing noise and artifacts. MambaIR integrates state-space modeling, further enhancing global perception and significantly improving restoration performance.

2.3.2. Network Optimization and Training Process

Training SRR networks involves inputting paired LR and HR images into the model. Next, the model evaluates the difference between the predicted and true HR images through a loss function. By iteratively adjusting the model parameters, the model gradually learns how to transform LR images into SR images that closely match the real HR images. Over time, this optimization process enhances the model’s performance, ultimately generating clearer and more detailed HR images. For CNN-based models (e.g., SRCNN, Very Deep Super Resolution), L1 loss is commonly used. For GAN-based models (e.g., Super-Resolution Generative Adversarial Network, ESRGAN), optimization includes perceptual and adversarial losses. For MambaIR-based models, L1 and Charbonnier losses are employed.

2.3.3. SRR Evaluation Metrics

With the growing use of super-resolution (SR) techniques, assessing the fidelity of reconstructed images has become increasingly significant. As noted by Sara et al. [47], the evaluation of image quality often involves measuring how closely the generated HR image resembles the original HR version. To evaluate the impact of SRR methods, this paper adopts two widely recognized quantitative evaluation indicators: Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index (SSIM).
P S N R is a commonly used objective metric in super-resolution tasks, which reflects the difference between the reconstructed image and the original high-resolution image. A greater P S N R value indicates superior image quality. The formula for calculation is as follows:
P S N R = 10 × log 10 M A X I 2 M S E
where M A X I indicates the highest possible pixel intensity in the image, and M S E denotes the mean of the squared differences between corresponding pixel values.
S S I M is another widely used image quality metric that considers not only pixel differences but also brightness, contrast, and structural information. The formula for calculation is as follows:
S S I M x , y = 2 μ x μ y + C 1 2 σ x y + C 2 μ x 2 + μ y 2 + C 1 σ x 2 + σ y 2 + C 2
where μ x and μ y are the mean values of the images, σ x 2 and σ y 2 are the variances, σ x y is the covariance, and C 1 and C 2 are constants.

2.4. Object Detection

2.4.1. Network Architectures for Object Recognition

Object detection methods based on deep learning are classified into two types: two-stage and one-stage. Two-stage methods start by generating candidate regions, then classify and regress them to determine the object’s category and location. Algorithms like the RCNN series (e.g., RCNN, Fast-RCNN) fall under this category. While they offer high accuracy, they are slower. In contrast, one-stage methods directly detect objects across the entire image without generating candidate regions, making them faster. Notable examples include the YOLO series and Single Shot Multibox Detector (SSD). The YOLO series is favored for its efficiency, accuracy, and lightweight design. By adopting a one-stage approach, YOLO achieves a simpler architecture and easier training and deployment, making it ideal for devices with limited computational resources. YOLO was originally proposed by Joseph Redmon as a single-stage regression problem [48]. In 2024, Wang et al. proposed YOLOv10, which incorporates a unified dual-assignment training strategy, removing the requirement for non-maximum suppression (NMS), reducing latency, and enhancing deployment efficiency. It also optimizes component design to maintain accuracy while minimizing computational overhead [49]. Saving computational resources is crucial for succulent plant detection, which can improve the real-time system, reduce hardware requirements, and significantly reduce the consumption of computational resources while guaranteeing detection accuracy and promote the practical application and promotion of succulent plant detection technology.
However, succulent plants exhibit significant diversity in shape, color, and texture, which makes accurate recognition challenging. The varying leaf structures—some tightly clustered, others loosely spread—further complicate image detection. To overcome these difficulties, this study presents the Succulent-YOLO model, which is based on YOLOv10 and incorporates three improvements: (1) a text-image encoder: contrastive language-image pre-training (CLIP); (2) a dynamic group convolution structure (DGCST); and (3) a multi-scale fusion neck network: semantics and detail infusion (SDI). These improvements enable Succulent-YOLO to achieve high-precision, fast classification in complex environments, making it ideal for succulent plant detection tasks. The YOLOv10 improved model is shown in Figure 5.
This study uses manually annotated label data and image features as input, employing YOLOv10 to extract features and obtain target region images. These images and labels are input into an optimisation module, where NMS operations are performed to select key target feature regions using CLIP image and text matching non-maximum suppression. The selected target features are then combined to form the final training data. The enhanced dataset using Mosaic is used to train the optimised model. During the training phase, as shown in Figure 6, images and data labels are input into the YOLOv10 backbone and CLIP text encoder, respectively. The backbone is then used as the main embedding, constrained by CLIP-based semantic information, along with the embedded category label text. CLIP text matching is used to achieve dual embedding, and the CLIP-based classification branch aligns with the zero-shot detection (ZSD) of the CLIP model for classification transfer learning. The YOLOv10 backbone network is enhanced with the proposed DGCST structure to improve its ability to detect unknown object features. The enhanced features are then input into the neck network for feature fusion. During the inference phase, images of size 600 × 600 are processed through the backbone and neck feature pyramid networks for feature extraction. The generated features are input into the detection head, where they undergo regression and classification branches to obtain the bounding box positions of detected objects and the semantic information of embedded image regions, respectively. The object categories are obtained through similarity matching with the text generated from the data labels.

2.4.2. Model Training and Refinement

At the initial stage of training, the model uses its backbone network to capture critical features from the input images. Subsequently, to further enhance detection accuracy, Succulent-YOLO employs a dual-part loss function consisting of classification and regression losses. Specifically, the classification loss is determined using the Binary Cross-Entropy (BCE) method, effectively measuring discrepancies between the predicted outcomes and corresponding ground truth annotations, thereby guiding the model toward improved classification performance. The formula for calculation is as follows:
L B C E = 1 N i = 1 N [ y i log p i + 1 y i log 1 p i ]
Here, N refers to the total count of predictions made, y i indicates the true label for the i t h instance, and p i corresponds to the probability estimated by the model. The output of this loss function reflects the difference between the predicted probability distribution and the ground truth labels, acting as an indicator of the model’s prediction accuracy. A smaller BCE loss value implies that the predicted probabilities are closer to the actual labels, indicating better model performance.
The regression loss combines Distribution Focal Loss ( D F L ) and Complete IoU ( C I o U ) loss. D F L addresses the issue of sample imbalance by accounting for the distribution of the object, directing more attention to hard-to-classify samples and thereby improving detection performance. C I o U considers the overlapping area between bounding boxes while incorporating additional factors to further enhance detection accuracy. Their calculation formulas are as follows:
L D F L = α 1 y i γ log p i
Here, y i represents the ground truth label, p i is the predicted probability by the model, α is a coefficient for balancing positive and non-positive samples, and γ is a hyperparameter that adjusts the weight of easy and hard samples.
L C I o U = 1 C I o U
C I o U = I o U ρ 2 b , b g t c 2 + α v
α = v 1 I o U + v
v = 4 π 2 arctan w g t h g t arctan w h 2
Here, I o U represents the intersection-over-union between the predicted bounding box and the ground truth bounding box. b denotes the center coordinates of the predicted bounding box, while b g t represents the center coordinates of the ground truth bounding box. ρ 2 indicates the Euclidean distance between the center points of the predicted and ground truth bounding boxes. C represents the diagonal length of the smallest enclosing box covering both the predicted bounding box and the corresponding ground truth. Additionally, v denotes the loss associated with aspect ratio differences, quantifying how closely the predicted bounding box matches the ground truth in terms of width-to-height proportion.
By combining BCE, D F L , and C I o U loss functions, the Succulent-YOLO model can be trained in a balanced manner. The total loss is calculated as follows:
L T o t a l = λ 1 L B C E + λ 2 L D F L + λ 3 L C I o U
Here, λ 1 , λ 2 , and λ 3 are the weights used to balance the BCE, D F L , and C I o U loss functions. The Adam optimizer is used to optimize this combined loss function, significantly boosting the training effectiveness and final performance in succulent plant detection, which is characterized by small inter-class differences and large intra-class variations.

2.4.3. Metrics for Evaluating Object Detection

The following evaluation metrics are used: Precision, Recall, Average Precision ( A P ), Mean Average Precision ( m A P ), and Intersection over Union ( I o U ). Precision measures the proportion of correctly identified positive instances relative to the total number of predicted positives ( T P + F P ). Recall evaluates the proportion of correctly detected true positives ( T P ) relative to all actual objects ( T P + F N ). AP evaluates performance by averaging precision across various recall levels. mAP represents the mean AP across all categories in multi-class detection. I o U measures the extent of intersection between predicted bounding boxes and their corresponding ground truth annotations.
The corresponding formulas are as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
A P y , y * = 1 N C a r e a ( P r )
m A P = 1 n i = 1 n A P i
I o U = A r e a   o f   O v e r l a p A r e a   o f   U n i o n
Here, T P , F N , and F P refer to True Positives, False Negatives, and False Positives, respectively. T P refers to the number of succulent plant samples correctly identified as belonging to a specific category. F N refers to the count of succulent plant samples that should be in a certain category but are misclassified as different ones. F P denotes the count of instances that are incorrectly assigned to a specific class. A model with high precision indicates strong capability in identifying true positive cases while minimizing incorrect positive predictions. A high recall indicates that the model can detect most actual positive samples, with fewer missed detections. A higher A P value signifies strong precision performance across different recall levels, demonstrating overall robustness. The higher the m A P , the better the model performs across all categories. I o U evaluates the accuracy of predicted bounding boxes—by setting a threshold (e.g., 0.5), a prediction is considered a true positive if the I o U exceeds the threshold. A higher I o U value reflects increased overlap between the predicted and ground truth boxes, resulting in improved localization accuracy.

3. Results

3.1. Dataset Details

This study utilized a dataset collected by a drone to train each succulent plant detection model. The dataset consists of 597 succulent plant images with a resolution of 600 × 600 pixels. To enhance diversity and improve the model’s generalization ability and robustness, geometric transformations such as flipping, rotation, and cropping were applied [27]. By applying image augmentation techniques, the training set contains 30 times more images than before, resulting in a training dataset of 18,000 images, ultimately producing HR images.
To train the super-resolution reconstruction model, a composite degradation model was used to generate LR images from HR images, enabling the model to learn image degradation patterns that closely resemble real-world scenarios [36]. The downsampling model is formulated as:
I L R = I H R h s b i c u b i c + η
Let I L R and I H R denote the low-resolution and high-resolution images, respectively. The variable h indicates the point spread function (PSF) corresponding to uniform linear motion, while the symbol denotes the convolution operation. The operator s b i c u b i c signifies bicubic downsampling with a scaling factor of s . η is additive Gaussian white noise, employed to replicate the randomness found in real-world imaging conditions.
LR sub-images of size l s u b × l s u b pixels are generated from the LR images by cropping method, so that it can increase the diversity of the dataset and improve the training efficiency. Then, to ensure accurate alignment, HR sub-images with dimensions of s l s u b × s l s u b pixels are extracted from identical coordinates within the HR images. These sub-images are then paired with their corresponding LR counterparts to construct the training dataset. After preprocessing, 18,000 high-resolution images are separated into training, validation, and testing subsets according to a 7:1:2 distribution.

3.2. Experiment Design

In this work, seven SRR models were investigated for their effectiveness in enhancing images of succulent plants. Initially, various SRR architectures were trained and subsequently applied to restore high-resolution (SR) images from their LR counterparts. The reconstructed image quality was quantitatively evaluated using PSNR and SSIM. Following this, the enhanced Succulent-YOLO model was developed using both the training and test partitions of the original succulent plant image dataset. The detection performance of the trained model was then measured using standard evaluation metrics, including precision, recall, and mAP, across datasets such as HR, Bicubic, and SR. To further analyze the impact of SRR techniques on plant detection performance, Grad-CAM was employed to produce visual explanations via heatmaps, thereby enhancing the interpretability of the detection outcomes.
The experiment environment parameters are shown in Table 1. Input images for this experiment are of size 600 × 600, with an initial learning rate of 0.01. The learning rate was optimized through the stochastic gradient descent (SGD) algorithm, set with a momentum factor of 0.937 and a weight decay of 0.0005. Additionally, Mosaic data augmentation was implemented during the training phase, randomly selecting four images per batch and applying transformations such as flipping, scaling, and rotation before merging them, thus significantly diversifying the training set and enhancing the robustness of detection. To prevent overfitting and boost generalization, label smoothing is applied with a value of 0.01. Model training was conducted over 500 epochs using a batch size of 32, with 32 worker threads allocated for data loading.

3.3. Assessment of Super-Resolution Techniques

To analyze the capability of various SRR approaches in improving the resolution of succulent plant imagery, eight models were implemented and their outputs were quantitatively evaluated using PSNR and SSIM metrics, as presented in Table 2. Table 3 shows comparison of SRR architectures by parameters, training time, and inference time. The experimental results indicate that SRR models based on deep learning architectures consistently outperform the traditional bicubic interpolation approach in terms of both PSNR and SSIM, thereby demonstrating superior performance in high-fidelity image restoration. Nevertheless, the SRR images generated by the Real-ESRGAN method show lower PSNR and SSIM values, likely due to its use of perceptual loss, which focuses on high-level feature similarity between generated and real images, enhancing details and textures [50]. Among all methods, MambaIR achieved the highest performance, with PSNR and SSIM values of 23.83 dB and 79.6%, respectively. MambaIR has a higher SSIM but lower PSNR for two main reasons. On the one hand, the higher SSIM of MambaIR in terms of the structural features of the model indicates that it can identify fine features such as leaf edges, vein patterns, and surface irregularities. In addition, since the features of MambaIR and SwinIR are different, it is more appropriate to use MambaIR for fine textures, thus distinguishing similar succulent species based on surface structure. For smoother images, it is better to use SwinIR. On the other hand, there are slight differences in the curvature and edges of the leaves of different species as seen in the features of the images, and the surface pattern and vein distribution may not be captured in the low-resolution images, which may also lead to such differences in MambaIR.
From a technical perspective, unlike SwinIR, which relies on hierarchical shifted window attention, MambaIR employs the state-space model, enabling it to effectively model global context with a linear computational complexity. This architectural advantage translates to reduced computational cost and enhanced performance on our specific task, especially when handling larger image patches or scenarios requiring extensive spatial correlation modeling. While SwinIR performs well through localized attention, its reliance on window-based operations can limit its effectiveness for tasks demanding global coherence across extensive image regions. In contrast, MambaIR’s global-context approach aligns closely with our task objectives—precisely reconstructing high-resolution features from images with subtle yet significant details. Therefore, considering computational efficiency, accuracy, and suitability for our super-resolution reconstruction task, MambaIR was deemed the more appropriate choice.
Figure 7 illustrates the super-resolution processing applied to the input images, which are already pre-processed through object detection. The Bicubic method generates a blurry image, failing to capture the complex textures of succulent plants. Conversely, deep learning-based methods produce sharper images with more precise textures and edges, closely matching the original HR image. However, CNN-based methods, due to the locality of convolution operations, tend to generate blurry edges and struggle with high-frequency details [51]. While generative adversarial network (GAN)-based methods provide sharper details, they may overemphasize certain features or create unnatural textures, particularly in succulent plants, which have soft, natural textures. Transformer-based methods improve clarity and texture but still fall short of MambaIR. MambaIR excels in preserving high-frequency details, maintaining the complex textures and edges of succulent images. Its unique reconstruction mechanism avoids the blurriness common in CNNs and the unrealistic detail generation seen in GANs.
Despite significant improvements in image clarity and detail restoration, deep learning methods still face challenges, such as excessive edge smoothing that can result in unnatural textures. This is particularly important in succulent plant texture detection, where leaf structures and surface details are crucial for accurate identification. Nevertheless, SRR approaches powered by deep learning have demonstrated notable advancements compared to conventional techniques, underscoring their effectiveness in achieving high-fidelity image reconstruction.

3.4. Grad-CAM Visualization of Object Detection Results

A performance assessment was carried out on the Bicubic approach alongside six deep learning-driven SRR models by employing the Succulent-YOLO network. Table 4 lists the mAP values for the HR test set and other methods (Bicubic and SRR), reflecting the detection effectiveness and overall performance of the different methods. Significantly, the mAP value for the HR test set stands at the highest, 87.8, indicating that Succulent-YOLO achieves very high accuracy in succulent plant detection. Among the deep learning-based SRR methods, MambaIR achieved an mAP of 85.1, the highest among the methods, which is 24.4% higher than the Bicubic method and only 3.1% lower than the HR test set. This shows that deep learning-based SRR methods lead to a substantial improvement in target detection performance, demonstrating their effectiveness and feasibility in both image reconstruction and target detection. Additionally, it suggests the potential for precise target detection in scenarios with lower computational costs, limited data collection, and more flexible applications, thus reducing the requirements for data acquisition.
However, the detection performance of the Real-ESRGAN method is the lowest, with an mAP value of 72.4%. This may result from the generation of unnatural textures during the reconstruction, which compromised the image’s authenticity and detection accuracy. The mAP values of the other SRR methods are fairly similar and greatly exceed the Bicubic method, showing an average improvement of about 13.5%. As shown in Figure 8a,b, the improvements in SRR methods are accompanied by higher confidence scores, indicating a correlation between the enhancements in SRR methods and the improvement in image quality.
To assess the performance of different SRR algorithms in succulent plant image detection and their impact on the Succulent-YOLO model, we applied the Grad-CAM technique [52]. Grad-CAM visualizes the decision-making process of CNNs by generating class activation maps (heatmaps), which highlight areas of the image the model focuses on. These maps improve model interpretability and provide insights for further optimization.
Figure 8c,d present the heatmaps superimposed on the object detection results. The Grad-CAM results reveal that the model’s focus extends beyond the bounding boxes, covering the entire image. The Succulent-YOLO model obtained the highest score of 0.84 for the HR test set, with heatmaps highlighting the intricate leaf textures of succulent plants. In contrast, the Bicubic method, with a confidence score of 0.46, displayed highlighted areas deviating from the actual location of the plants, indicating significant differences between reconstruction algorithms.
SRR models based on deep learning exhibited superior performance compared to the Bicubic method, with confidence scores ranging from 0.46 to 0.84. However, their heatmaps varied, reflecting differences in detail restoration and overall structure. For instance, the heatmap for Real-ESRGAN displayed multiple red zones, indicating the presence of unrealistic textures and false detections, caused by the model’s attention being drawn to background elements such as soil and pots.
CNN-based algorithms also exhibited similar issues. Although their heatmaps emphasized succulent plant leaves, the convolution operation’s local receptive field caused the model to focus on irrelevant background areas, reducing detection accuracy. Interestingly, SRCNN, as an early SRR model, displayed a more stable performance. Its heatmap showed fewer red highlights, suggesting the model focused more on relevant regions rather than background noise. While SRCNN’s simpler structure had limitations in resolution improvement, it laid the foundation for subsequent SRR advancements.
The Transformer-based SwinIR algorithm excelled in the Succulent-YOLO model, surpassing CNNs in terms of image reconstruction. However, akin to SRCNN, SwinIR’s attention was partially diverted to the background, indicating its challenge in capturing global information.
Finally, MambaIR demonstrated significant advantages in succulent plant image reconstruction, achieving a confidence score equivalent to the HR test set, reaching 0.84. MambaIR effectively utilized a wider receptive field to preserve rich textures, optimizing image quality through lighting and color adjustments. Its heatmap showed fewer red highlights, indicating the model successfully removed irrelevant details and focused on key succulent features. This balance between global and local features provides valuable insights for future SRR algorithm improvements in succulent plant detection.

3.5. Ablation Experiment for Succulent-YOLO

Detection of succulent plants using drone-captured images is challenging due to environmental interference, the diversity of plant morphologies, individual variations, and image quality issues. To address these challenges, we enhanced the original YOLOv10 [49] model to improve succulent plant detection. Table 5 summarizes the incremental improvements made to YOLOv10. By cumulatively adding the CLIP, DGCST, and SDI modules, we calculated the corresponding metrics.
In the updated model, the text-image encoder CLIP was incorporated, resulting in the YOLOv10-S1 model. This enhancement boosts feature extraction and classification by leveraging CLIP’s image-text matching [39]. NMS is used to identify key target areas, ensuring precise feature extraction. The dual embedding of image and text further strengthens the model’s semantic alignment, and integrating a CLIP-based classification branch through transfer learning improves detection performance, particularly for unknown objects and complex features. These modifications enhance YOLOv10’s ability to generate richer and more accurate feature representations, resulting in improved detection accuracy. The model achieved precision, recall, and mAP values of 79.0%, 76.9%, and 83.0%, respectively.
Additionally, a new prediction layer was added for small objects (150 × 150 pixels) to improve detection accuracy. This new layer incorporates the SDI module, which utilizes layered feature maps and attention mechanisms to preserve positional information and details of small objects. The module enables more accurate detection, especially in occlusion and multi-object scenarios. The neck network enhances small object feature expression by upsampling low-resolution maps to align with high-resolution maps, further improving detection accuracy. These improvements led to the creation of the YOLOv10-S1-2-3 model, also known as the Succulent-YOLO model, which achieved a mAP improvement of 87.8% compared to the original YOLOv10.
By progressively optimizing the YOLOv10 model, its performance has been significantly enhanced, overcoming the challenges in drone-based succulent plant detection, and laying the foundation for precise and efficient agricultural detection applications.

3.6. Effects of the Super-Resolution Scaling Parameter

Many studies have highlighted the importance of the magnification factor in determining the final reconstruction quality of SR images [53,54,55]. This parameter dictates how much the low-resolution input image needs to be expanded and how many new pixels must be generated to restore image details. A larger magnification factor increases the reconstruction difficulty, often leading to blurring or distortion, as the model must generate more details from limited information. In contrast, a smaller magnification factor better preserves the original texture and details. To determine the ideal magnification factor for succulent plant detection, we adopted the following experimental procedure: first, the HR images were preprocessed and subsequently split into training and validation subsets in a 4:1 ratio. Then, we applied four-fold downsampling to the HR images to create the LR dataset. Finally, we applied deep learning-based SRR methods to the LR dataset. Based on previous findings, we selected the top-performing MambaIR model and evaluated the effect of different magnification factors (2–5) on SRR performance.
Initially, the × 2_SR model is trained using the LR training data alongside its corresponding HR counterparts. Subsequently, this trained model is applied to the × 1_LR test samples (150 × 150 pixels) to produce the × 2_SR test images with a resolution of 300 × 300 pixels. Similarly, the same process is used to create the × 3_SR (450 × 450), × 4_SR (600 × 600), and × 5_SR (750 × 750) test sets. Next, the Succulent-YOLO model is trained using the training and validation sets of the succulent plant HR dataset and applied to the SR images reconstructed at different magnification factors. Finally, mAP and confidence scores are used to evaluate the detection performance of these SRR-reconstructed images. Table 6 shows the recognition accuracy of the Succulent-YOLO model for succulent plant images at different magnification factors.
Figure 9 shows the detection results of succulent plant images at different magnification factors. As the magnification factor grows, the detection outcomes more accurately reflect the ground truth labels, leading to higher recognition accuracy. However, higher magnification factors also increase computational complexity and processing time, impacting the model’s efficiency. For instance, the computation times for magnification factors of × 4 are 24% longer than for × 1. The results indicate that the recommended magnification of SRR for succulent target detection is 4. Furthermore, this finding generally aligns with the results of other target detection studies using SRR, supporting its validity.

4. Discussion

4.1. Influence of Different Object Detection Architectures on SRR

To investigate how various object detection approaches affect succulent plant identification under SRR settings, we trained a total of 12 detection models on a high-resolution succulent plant dataset captured via UAVs. The evaluated models include Faster R-CNN and multiple YOLO-based variants. Table 7 provides these models’ detection performance for three test sets: HR, Bicubic, and MambaIR.
Our results show that Succulent-YOLO performs the best across all test sets. Succulent-YOLO achieves an mAP 5.7% higher than the second-best model on the HR test set, showcasing its superior performance with high-resolution images. In the Bicubic test set, although all models experienced a decline in performance due to reduced image resolution, Succulent-YOLO still attained the top mAP score, highlighting its strong capability for handling low-resolution images. Equally, in the MambaIR test set, Succulent-YOLO’s mAP is 6.5% higher than YOLOv7, showcasing its superiority in super-resolution reconstruction.
The results indicate that for images reconstructed by MambaIR, two-stage detectors like Faster R-CNN cannot surpass certain single-stage models such as the YOLO series. Despite its larger model size, Faster R-CNN is less efficient than Succulent-YOLO in all metrics, primarily due to architectural differences. Being a two-stage detection framework, Faster R-CNN first generates region proposals, which are subsequently classified and refined, resulting in a relatively complex and computationally intensive pipeline [56]. This structure limits its applicability in real-time scenarios. Conversely, single-stage detectors eliminate the proposal generation phase by directly performing dense predictions across the entire image, thereby significantly enhancing inference speed and computational efficiency [48].
Within the Bicubic test set, Faster R-CNN exhibited the lowest mAP score, highlighting its limitations when processing images with reduced resolution. Several YOLO-based models, such as YOLOX-s and YOLOv7, similarly demonstrated decreased detection performance. YOLOX improved upon the YOLOv3 and YOLOv4 backbone but replaced the anchor-based design with an anchor-free detection paradigm, which can blur targets and cause localization errors [57]. Therefore, it can lead to the emergence of problems such as missed detection, false detection and degradation of localization accuracy in UAV image detection of succulents. Similarly, YOLOv7’s complex deep feature aggregation strategy (E-ELAN structure) enhances local information extraction but performs poorly on Bicubic images, as it only interpolates neighborhood pixels without restoring high-frequency details [58]. In contrast, Succulent-YOLO combines high accuracy and robustness with a compact model size and minimal parameters, making it ideal for high-performance detection tasks, particularly in resource-constrained environments. Its lightweight design enables efficient operation, expanding its potential for real-time crop monitoring when deployed on platforms like UAVs or smart agricultural machinery. Currently, applications of real-time monitoring of succulent plants include intelligent succulent plant management systems based on wireless networks [59], mobile monitoring systems with cameras [60], and so on. Real-time monitoring technology provides intelligent and refined management tools for succulent plant cultivation, promoting agricultural modernization and sustainable development.
Figure 10 illustrates the interpretive analysis of the detection results of the various networks on the high-resolution images using Grad-CAM. Succulent-YOLO shows a clear advantage, obtaining the highest confidence score of 0.84 on the test images and showing the widest range of highlighted areas in the heatmap.

4.2. Robustness Against Degraded Input Data

Gaussian blur is a smoothing technique that applies a Gaussian kernel to image pixels, where the kernel size is crucial for the super-resolution effect. This method is typically used to reduce image details and create a softer image. Gaussian noise refers to the introduction of random noise that follows a Gaussian distribution, used to test the model’s stability in noisy environments. Gaussian blur is mainly used to generate realistic low-resolution data, while Gaussian noise enhances the model’s noise resistance and robustness. The integration of both factors strengthens the model’s ability to process low-quality data, leading to better performance in various situations.
To evaluate the robustness of the proposed model under realistic imaging conditions, we introduced varying levels of Gaussian blur and Gaussian noise to the input data and analyzed the impact on model performance. These perturbations simulate common challenges in UAV-based image acquisition, such as vibration, motion blur, sensor noise, and atmospheric interference.
As shown in Figure 11a, we applied Gaussian blur with increasing standard deviation (σ = 0 to 2) across 12 different detection models. CNN-based architectures exhibited significant performance degradation under blurring, with their mAP scores declining sharply as blur intensity increased. In contrast, models employing the GhostNet-P6 backbone demonstrated greater robustness, maintaining relatively stable detection accuracy. Notably, our proposed Succulent-YOLO model achieved a minimal performance drop of only 4.32%, indicating its strong tolerance to visual degradation caused by blur.
Figure 11b illustrates the impact of additive Gaussian noise (standard deviation ranging from 0 to 20) on model performance. The results show that Succulent-YOLO maintains high resilience, with only a 7.73% reduction in mAP under the most severe noise level. This highlights the model’s ability to generalize well in noisy environments commonly encountered in precision agriculture UAV deployments.
These findings confirm that the architectural choices—such as the integration of CLIP embeddings and super-resolved inputs via MambaIR—contribute not only to improved baseline accuracy but also to enhanced robustness under adverse visual conditions. The model’s relatively compact parameter count further ensures that this robustness is achieved without incurring excessive computational overhead, making it suitable for real-world UAV platforms.

4.3. Proposals for Future Work

While the Succulent-YOLO approach demonstrates significant advantages in succulent farmland monitoring through the integration of CLIP-based YOLOv10 and Mamba computer vision methods, several limitations must be acknowledged.
First, our dataset, although sufficiently diverse for this initial validation, remains limited in size and variability. The images were collected under relatively controlled environmental conditions, which may limit the model’s performance when generalized to broader field scenarios characterized by variable illumination, seasonal changes, shadows, and weather conditions. Future studies could enrich the dataset with more extensive sampling across diverse environments and conditions to enhance robustness and generalizability. Second, the computational complexity of combining super-resolution reconstruction (SRR) and deep learning-based detection models may pose challenges for real-time processing on typical UAV onboard hardware or mobile devices. Future work should explore model optimization strategies, such as quantization, pruning, lightweight architectures, and edge-computing deployment, enabling practical, real-time implementation on resource-constrained platforms common in precision agriculture. Third, although the model shows strong performance specifically for succulent species, its adaptability and scalability to other plant species or agricultural crops were not explicitly evaluated. Further research is necessary to assess the transferability of the Succulent-YOLO pipeline to broader agricultural contexts, ensuring applicability and wider adoption across diverse agricultural scenarios. Finally, integrating the Succulent-YOLO system with real-time decision-support tools or agricultural management platforms represents a promising avenue for future development. Such integration could significantly enhance farmer decision-making capabilities, facilitate proactive crop monitoring, and streamline precision agriculture workflows.
Addressing these limitations and exploring these potential directions in future research will further advance the practicality, efficiency, and robustness of Succulent-YOLO, contributing to sustainable agricultural practices and technological advancements in UAV-assisted farmland monitoring.

5. Conclusions

This study presents a practical and integrated strategy for large-scale classification of succulent plant imagery by combining UAV-based data acquisition with advanced deep learning techniques. We introduced a super-resolution reconstruction (SRR) module, MambaIR, which enhances the clarity of low-resolution UAV images. Compared to several baseline SRR models, MambaIR achieved competitive performance, with a PSNR of 23.83 dB and SSIM of 79.60%, enabling better downstream detection results.
Additionally, we developed a customized detection architecture, Succulent-YOLO, based on YOLOv10 and tailored for the fine-grained recognition of succulent species. The model achieved an mAP of 87.8% on high-resolution inputs and maintained strong performance (85.1%) even when applied to images reconstructed via MambaIR, indicating the robustness and synergy of the integrated pipeline.
Grad-CAM visualizations further demonstrated the model’s ability to focus on key morphological features of succulents, supporting its interpretability and potential for precision agriculture applications. Moreover, an appropriate SR scaling factor was determined to balance image quality with computational efficiency, which is essential for real-world UAV deployment scenarios.
Despite these promising results, several limitations remain. The dataset was limited in species diversity and environmental variation, which may affect generalization to other crop types or conditions. Additionlly, the current model’s computational demands may restrict deployment on lightweight UAV platforms without further optimization. Future work will focus on extending the dataset across more plant species and phenological stages, evaluating cross-domain transferability, and optimizing the model for real-time, on-board inference. These efforts aim to enhance the system’s scalability and practical value in supporting data-driven precision agriculture.

Author Contributions

Conceptualization, F.Z.; methodology, H.L., F.Z. and F.X.; software, H.L., F.Z., J.W., G.Z. and D.X.; validation, H.L., F.Z., F.X., J.W., Y.L., Y.C., J.T. and D.X.; formal analysis, H.L.; investigation, H.L., F.Z., F.X., Y.L., Q.W., J.T. and G.Z.; resources, F.Z., Y.C., J.T., J.C. and H.H.K.; data curation, F.Z., J.W. and Y.C.; writing—original draft preparation, H.L., F.Z. and F.X.; writing—review and editing, F.Z., J.W., Y.L., Y.C., Q.W., J.T., G.Z., J.C. and H.H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by JST FOREST Program, Grant Number: JPMJFR211B.

Data Availability Statement

The data utilized for this research are available at https://pan.baidu.com/s/1Uh9Pk9uffa0mSibrlb_IEg with the password: 0nep.(accessed on 27 June 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CVComputer vision
UAVsUnmanned Aerial Vehicles
DIPDigital image processing
CNNsConvolutional Neural Networks
Faster R-CNNFaster Region-based Convolutional Neural Network
DAData augmentation
SRRSuper-resolution reconstruction
LRLow-resolution
HRHigh-resolution
SIFTScale-Invariant Feature Transform
DETRDEtection Transformer
GhostNetGhost Network
YOLOYou Only Look Once
Grad-CAMGradient-weighted Class Activation Mapping
BBABundle block adjustment
MSEMean squared error
SRCNNSuper-Resolution Convolutional Neural Network
DRCNDeeply-Recursive Convolutional Network
ESRGANEnhanced Super-Resolution Generative Adversarial Network
SRGANSuper-Resolution Generative Adversarial Network
EDSREnhanced Deep Residual Network
RCANVery deep residual channel attention networks
Real-ESRGANReal-world blind super-resolution with pure synthetic data
SwinIRSwin Transformer for Image Restoration
VDSRVery Deep Super Resolution
SRSuper-resolution
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index
SSDSingle Shot Multibox Detector
NMSNon-maximum suppression
CLIPContrastive language-image pre-training
DGCSTDynamic group convolution structure
SDISemantics and detail infusion
ZSDZero-shot detection
BCEBinary Cross-Entropy
DFLDistribution Focal Loss
CIoUComplete IoU
APAverage Precision
mAPMean Average Precision
IoUIntersection over Union
TPTrue positives
FNFalse Negatives
FPFalse Positives
PSFPoint spread function
SGDstochastic gradient descent
GANgenerative adversarial network

References

  1. Von Willert, D.J. Life Strategies of Succulents in Deserts: With Special Reference to the Namib Desert; CUP Archive: Cambridge, UK, 1992. [Google Scholar]
  2. Grace, O.M. Succulent plant diversity as natural capital. Plants People Planet 2019, 1, 336–345. [Google Scholar] [CrossRef]
  3. Borland, A.M.; Griffiths, H.; Hartwell, J.; Smith, J.A.C. Exploiting the potential of plants with crassulacean acid metabolism for bioenergy production on marginal lands. J. Exp. Bot. 2009, 60, 2879–2896. [Google Scholar] [CrossRef]
  4. Burgess, T.L.; Shmida, A.; Whitehead, E.E.; Hutchinson, C.F.; Timmermann, B.N.; Varady, R.G. Succulent growth-forms in arid environments. In Arid Lands Today and Tomorrow; Westview Press: Boulder, Colorado, USA, 1988. [Google Scholar]
  5. Ong, A.K.S.; Prasetyo, Y.T.; De Leon, L.A.S.; Ayuwati, I.D.; Nadlifatin, R.; Persada, S.F. Plantitas/Plantitos preference analysis on succulents attributes and its market segmentation: Integrating conjoint analysis and K-means clustering for gardening marketing strategy. Sustainability 2022, 14, 16718. [Google Scholar] [CrossRef]
  6. Mata-Montero, E.; Carranza-Rojas, J. Automated plant species identification: Challenges and opportunities. In ICT for Promoting Human Development and Protecting the Environment, Proceedings of the 6th IFIP World Information Technology Forum, WITFOR 2016, San José, Costa Rica, 12–14 September 2016; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; Volume 6, pp. 26–36. [Google Scholar]
  7. Mana, A.A.; Allouhi, A.; Hamrani, A.; Rehman, S.; El Jamaoui, I.; Jayachandran, K. Sustainable AI-based production agriculture: Exploring AI applications and implications in agricultural practices. Smart Agric. Technol. 2024, 7, 100416. [Google Scholar] [CrossRef]
  8. Kumar, V.; Sharma, K.V.; Kedam, N.; Patel, A.; Kate, T.R.; Rathnayake, U. A comprehensive review on smart and sustainable agriculture using IoT technologies. Smart Agric. Technol. 2024, 8, 100487. [Google Scholar] [CrossRef]
  9. Pérez-Zavala, R.; Torres-Torriti, M.; Cheein, F.A.; Troni, G. A pattern recognition strategy for visual grape bunch detection in vineyards. Comput. Electron. Agric. 2018, 151, 136–149. [Google Scholar] [CrossRef]
  10. Fahmi, F.; Andayani, U.; Siregar, B. SENSING OF VEGETATION STATUS USING ORTOPHOTOS IMAGE GENERATED WITH UAV. Abdimas Talent. J. Pengabdi. Kpd. Masy. 2018, 3, 54–58. [Google Scholar] [CrossRef]
  11. Sadeghi-Tehran, P.; Sabermanesh, K.; Virlet, N.; Hawkesford, M.J. Automated method to determine two critical growth stages of wheat: Heading and flowering. Front. Plant Sci. 2017, 8, 252. [Google Scholar] [CrossRef]
  12. López-Jiménez, E.; Vasquez-Gomez, J.I.; Sanchez-Acevedo, M.A.; Herrera-Lozada, J.C.; Uriarte-Arcia, A.V. Columnar cactus recognition in aerial images using a deep learning approach. Ecol. Inform. 2019, 52, 131–138. [Google Scholar] [CrossRef]
  13. Ghazal, S.; Munir, A.; Qureshi, W.S. Computer vision in smart agriculture and precision farming: Techniques and applications. Artif. Intell. Agric. 2024, 13, 64–83. [Google Scholar] [CrossRef]
  14. Rahman, M.F.F.; Fan, S.; Zhang, Y.; Chen, L. A comparative study on application of unmanned aerial vehicle systems in agriculture. Agriculture 2021, 11, 22. [Google Scholar] [CrossRef]
  15. Zhao, F.; Liu, Y.; Wang, J.; Chen, Y.; Xi, D.; Shao, X.; Tabeta, S.; Mizuno, K. Riverbed litter monitoring using consumer-grade aerial-aquatic speedy scanner (AASS) and deep learning based super-resolutio n reconstruction and detection network. Mar. Pollut. Bull. 2024, 209, 117030. [Google Scholar] [CrossRef] [PubMed]
  16. Yang, L.; Lu, B.; Schmidt, M.; Natesan, S.; McCaffrey, D. Applications of Remote Sensing for Crop Residue Cover Mapping. Smart Agric. Technol. 2025, 11, 100880. [Google Scholar] [CrossRef]
  17. Raeva, P.L.; Šedina, J.; Dlesk, A. Monitoring of crop fields using multispectral and thermal imagery from UAV. Eur. J. Remote Sens. 2019, 52 (Suppl. S1), 192–201. [Google Scholar] [CrossRef]
  18. Schirrmann, M.; Giebel, A.; Gleiniger, F.; Pflanz, M.; Lentschke, J.; Dammer, K.H. Monitoring agronomic parameters of winter wheat crops with low-cost UAV imagery. Remote Sens. 2016, 8, 706. [Google Scholar] [CrossRef]
  19. Moeckel, T.; Dayananda, S.; Nidamanuri, R.R.; Nautiyal, S.; Hanumaiah, N.; Buerkert, A.; Wachendorf, M. Estimation of vegetable crop parameter by multi-temporal UAV-borne images. Remote Sens. 2018, 10, 805. [Google Scholar] [CrossRef]
  20. Aguero, H.P. Review of the Current Technologies and Applications of Digital Image Processing. J. Biomed. Sustain. Healthc. Appl. 2022, 2, 148–158. [Google Scholar] [CrossRef]
  21. Feng, Y.; Liang, S.; Tong, H. Research on succulent plant classification and recognition method based on contrastive learning. J. Henan Agric. Sci. 2023, 52, 154. [Google Scholar]
  22. Traore, B.B.; Kamsu-Foguem, B.; Tangara, F. Deep convolution neural network for image recognition. Ecol. Inform. 2018, 48, 257–268. [Google Scholar] [CrossRef]
  23. Ahmad, A.; Saraswat, D.; El Gamal, A. A survey on using deep learning techniques for plant disease diagnosis and recommendations for development of appropriate tools. Smart Agric. Technol. 2023, 3, 100083. [Google Scholar] [CrossRef]
  24. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. Deep learning techniques to classify agricultural crops through UAV imagery: A review. Neural Comput. Appl. 2022, 34, 9511–9536. [Google Scholar] [CrossRef] [PubMed]
  25. Lee, S.H.; Chan, C.S.; Wilkin, P.; Remagnino, P. Deep-plant: Plant identification with convolutional neural networks. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 452–456. [Google Scholar]
  26. Bargoti, S.; Underwood, J. Deep fruit detection in orchards. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 3626–3633. [Google Scholar]
  27. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  28. Divyanth, L.G.; Guru, D.S.; Soni, P.; Machavaram, R.; Nadimi, M.; Paliwal, J. Image-to-image translation-based data augmentation for improving crop/weed classification models for precision agriculture applications. Algorithms 2022, 15, 401. [Google Scholar] [CrossRef]
  29. Yang, Q.; Shi, L.; Han, J.; Zha, Y.; Zhu, P. Deep convolutional neural networks for rice grain yield estimation at the ripening stage using UAV-based remotely sensed images. Field Crops Res. 2019, 235, 142–153. [Google Scholar] [CrossRef]
  30. Su, J.; Coombes, M.; Liu, C.; Zhu, Y.; Song, X.; Fang, S.; Guo, L.; Chen, W.H. Machine learning-based crop drought mapping system by UAV remote sensing RGB imagery. Unmanned Syst. 2020, 8, 71–83. [Google Scholar] [CrossRef]
  31. Dahlin Rodin, C.; de Alcantara Andrade, F.A.; Hovenburg, A.R.; Johansen, T.A. A survey of practical design considerations of optical imaging stabilization systems for small unmanned aerial systems. Sensors 2019, 19, 4800. [Google Scholar] [CrossRef]
  32. Singh, K.K.; Frazier, A.E. A meta-analysis and review of unmanned aircraft system (UAS) imagery for terrestrial applications. Int. J. Remote Sens. 2018, 39, 5078–5098. [Google Scholar] [CrossRef]
  33. Zhao, F.; Ren, Z.; Wang, J.; Wu, Q.; Xi, D.; Shao, X.; Liu, Y.; Chen, Y.; Mizuno, K. Smart UAV-assisted rose growth monitoring with improved YOLOv10 and Mamba restoration techniques. Smart Agric. Technol. 2025, 10, 100730. [Google Scholar] [CrossRef]
  34. Park, S.C.; Park, M.K.; Kang, M.G. Super-resolution image reconstruction: A technical overview. IEEE Signal Process. Mag. 2003, 20, 21–36. [Google Scholar] [CrossRef]
  35. Siu, W.C.; Hung, K.W. Review of image interpolation and super-resolution. In Proceedings of the 2012 Asia Pacific Signal and Information Processing Association Annual Summit and Conference, Hollywood, CA, USA, 3–6 December 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 1–10. [Google Scholar]
  36. Wang, P.; Bayram, B.; Sertel, E. A comprehensive review on deep learning based remote sensing image super-resolution methods. Earth-Sci. Rev. 2022, 232, 104110. [Google Scholar] [CrossRef]
  37. Zhao, F.; Huang, J.; Liu, Y.; Wang, J.; Chen, Y.; Shao, X.; Ma, B.; Xi, D.; Zhang, M.; Tu, Z.; et al. A Deep Learning Approach Combining Super-Resolution and Segmentation to Identify Weed and Tobacco in UAV Imagery. In Proceedings of the 2024 9th International Conference on Electronic Technology and Information Science (ICETIS), Hangzhou, China, 17–19 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 594–597. [Google Scholar]
  38. Guo, H.; Li, J.; Dai, T.; Ouyang, Z.; Ren, X.; Xia, S.T. Mambair: A simple baseline for image restoration with state-space model. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 222–241. [Google Scholar]
  39. Radford, A.; Kim, J.W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. Learning transferable visual models from natural language supervision. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; PmLR. pp. 8748–8763. [Google Scholar]
  40. Lindeberg, T. Scale invariant feature transform. Scholarpedia 2012, 7, 10491. [Google Scholar] [CrossRef]
  41. Huang, Y.; Thomson, S.J.; Hoffmann, W.C.; Lan, Y.; Fritz, B.K. Development and prospect of unmanned aerial vehicle technologies for agricultural production management. Int. J. Agric. Biol. Eng. 2013, 6, 1–10. [Google Scholar]
  42. Tsouros, D.C.; Bibi, S.; Sarigiannidis, P.G. A review on UAV-based applications for precision agriculture. Information 2019, 10, 349. [Google Scholar] [CrossRef]
  43. Olson, D.; Anderson, J. Review on unmanned aerial vehicles, remote sensors, imagery processing, and their applications in agriculture. Agron. J. 2021, 113, 971–992. [Google Scholar] [CrossRef]
  44. Velusamy, P.; Rajendran, S.; Mahendran, R.K.; Naseer, S.; Shafiq, M.; Choi, J.G. Unmanned Aerial Vehicles (UAV) in precision agriculture: Applications and challenges. Energies 2021, 15, 217. [Google Scholar] [CrossRef]
  45. Chen, M.; Ke, Y.; Bai, J.; Li, P.; Lyu, M.; Gong, Z.; Zhou, D. Monitoring early stage invasion of exotic Spartina alterniflora using deep-learning super-resolution techniques based on multisource high-resolution satellite imagery: A case study in the Yellow River Delta, China. Int. J. Appl. Earth Obs. Geoinf. 2020, 92, 102180. [Google Scholar] [CrossRef]
  46. Sathya, K.; Rajalakshmi, M. RDA-CNN: Enhanced Super Resolution Method for Rice Plant Disease Classification. Comput. Syst. Sci. Eng. 2022, 42, 33–47. [Google Scholar] [CrossRef]
  47. Sara, U.; Akter, M.; Uddin, M.S. Image quality assessment through FSIM, SSIM, MSE and PSNR—A comparative study. J. Comput. Commun. 2019, 7, 8–18. [Google Scholar] [CrossRef]
  48. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  49. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  50. Wang, X.; Xie, L.; Dong, C.; Shan, Y. Real-esrgan: Training real-world blind super-resolution with pure synthetic data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 1905–1914. [Google Scholar]
  51. Li, Z.; Liu, F.; Yang, W.; Peng, S.; Zhou, J. A survey of convolutional neural networks: Analysis, applications, and prospects. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 6999–7019. [Google Scholar] [CrossRef]
  52. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. Int. J. Comput. Vis. 2020, 128, 336–359. [Google Scholar] [CrossRef]
  53. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef] [PubMed]
  54. Gao, X.; Zhang, K.; Tao, D.; Li, X. Image super-resolution with sparse neighbor embedding. IEEE Trans. Image Process. 2012, 21, 3194–3205. [Google Scholar]
  55. Yue, L.; Shen, H.; Li, J.; Yuan, Q.; Zhang, H.; Zhang, L. Image super-resolution: The techniques, applications, and future. Signal Process. 2016, 128, 389–408. [Google Scholar] [CrossRef]
  56. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  57. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  58. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  59. Yue, X.; Wang, W.; Yang, C.; Kang, H.; Wang, J.; Ma, S. Intelligent succulent plant management system based on wireless network. In Proceedings of the 2019 IEEE Symposium Series on Computational Intelligence (SSCI), Xiamen, China, 6–9 December 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 2863–2868. [Google Scholar]
  60. Yanlin, W.; Yu, C.; Qingquan, C.; Xin, G. Research and design of mobile monitoring system for muilt-section Succulent Plants cultivation. In Proceedings of the 2022 China Automation Congress (CAC), Xiamen, China, 25–27 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 4590–4593. [Google Scholar]
Figure 1. Procedure of the proposed approach for succulent detection.
Figure 1. Procedure of the proposed approach for succulent detection.
Remotesensing 17 02219 g001
Figure 2. (a) Map showing Yunnan Province’s geographic position; (b) General view of Yuxi City; (c) Map highlighting Tonghai County’s position within Yuxi; (d) Schematic diagram of UAV-based imaging approaches (orthophoto image of the target area); (e) Sample panoramic image of succulent plants; (f) Diagram showing succulent classification. Six categories, including a miscellaneous group. The six categories were obtained through surveys of local farmers to determine the general classification range, and then classified manually based on the shape and length of the leaves.
Figure 2. (a) Map showing Yunnan Province’s geographic position; (b) General view of Yuxi City; (c) Map highlighting Tonghai County’s position within Yuxi; (d) Schematic diagram of UAV-based imaging approaches (orthophoto image of the target area); (e) Sample panoramic image of succulent plants; (f) Diagram showing succulent classification. Six categories, including a miscellaneous group. The six categories were obtained through surveys of local farmers to determine the general classification range, and then classified manually based on the shape and length of the leaves.
Remotesensing 17 02219 g002
Figure 3. Labelimg 1.8.1 label examples. The green dots indicate the labeling of succulent plants in the figure, while “B, D, F” are the labeled categories.
Figure 3. Labelimg 1.8.1 label examples. The green dots indicate the labeling of succulent plants in the figure, while “B, D, F” are the labeled categories.
Remotesensing 17 02219 g003
Figure 4. Various SRR model architectures employing different design approaches. The input image resolution is 150 × 150, and the output image resolution is 600 × 600.
Figure 4. Various SRR model architectures employing different design approaches. The input image resolution is 150 × 150, and the output image resolution is 600 × 600.
Remotesensing 17 02219 g004
Figure 5. Diagram of the improved overall model of YOLOv10.
Figure 5. Diagram of the improved overall model of YOLOv10.
Remotesensing 17 02219 g005
Figure 6. Flow of CLIP methodology training.
Figure 6. Flow of CLIP methodology training.
Remotesensing 17 02219 g006
Figure 7. Comparison of visualization of detection images based on 8 reconstruction methods. “D” refers to the category of succulent plants.
Figure 7. Comparison of visualization of detection images based on 8 reconstruction methods. “D” refers to the category of succulent plants.
Remotesensing 17 02219 g007
Figure 8. (a) Results of succulent plant detection using different test sets; (b) Results of succulent plant detection using different test sets; (c) Heat map of detection results from different test sets; (d) Heat map of detection results from different test sets.
Figure 8. (a) Results of succulent plant detection using different test sets; (b) Results of succulent plant detection using different test sets; (c) Heat map of detection results from different test sets; (d) Heat map of detection results from different test sets.
Remotesensing 17 02219 g008
Figure 9. Detection results of succulent images from SRR test sets obtained with different magnification factors.
Figure 9. Detection results of succulent images from SRR test sets obtained with different magnification factors.
Remotesensing 17 02219 g009
Figure 10. Interpretable detection results of different detection networks for HR images using Grad-CAM.
Figure 10. Interpretable detection results of different detection networks for HR images using Grad-CAM.
Remotesensing 17 02219 g010
Figure 11. (a) Comparison of Gaussian blur for different models; (b) comparison of Gaussian noise for different models.
Figure 11. (a) Comparison of Gaussian blur for different models; (b) comparison of Gaussian noise for different models.
Remotesensing 17 02219 g011
Table 1. Experimental environment parameters.
Table 1. Experimental environment parameters.
Experiment ConfigurationType or Version
Operating systemUbuntu 18.04
Bit architecture32 bit
Random Access Memory64GB
GPUGeForce RTX 4090
Memory24GB
CPUIntel(R) Xeon(R) Platinum i9-13900k
Pytorch1.10
CUDA11.1
Cudnn8.0.4
Table 2. Evaluation metrics of various SRR methods on the LR test set with the magnification factor 4.
Table 2. Evaluation metrics of various SRR methods on the LR test set with the magnification factor 4.
MetricsBicubicReal-ESRGANSRCNNRCANEDSRSwinIRMambaIR
PSNR (dB)21.2019.7722.6323.8424.1224.2823.83
SSIM (%)62.4562.7071.9776.7577.9178.3379.60
Table 3. Comparison of SRR Architectures by Parameters, Training Time, and Inference Time.
Table 3. Comparison of SRR Architectures by Parameters, Training Time, and Inference Time.
MetricsBicubicReal-ESRGANSRCNNRCANEDSRSwinIRMambaIR
Parameter Counts-16.7 M0.01 M16 M43 M11.8 M21 M
Training Time (h)-31.251.25152522.525
Inference Time
(ms per image)
35130254030
Table 4. Evaluation metrics of detection results of the test set reconstructed by different SRR algorithms.
Table 4. Evaluation metrics of detection results of the test set reconstructed by different SRR algorithms.
MetricsHRBicubicReal-ESRGANSRCNNRCANEDSRSwinIRMambaIR
mAP@5087.868.472.476.780.682.884.485.1
Table 5. Ablation experiments of different detection networks.
Table 5. Ablation experiments of different detection networks.
ModelCLIPDGCSTSDIP (%) ± SDR (%) ± SDmAP50 (%) ± SDSize (MB)
YOLOv10s×××78.9 ± 0.373.8 ± 0.680.6 ± 0.416.6
YOLOv10s-1××79.0 ± 0.276.9 ± 0.583.0 ± 0.317.9
YOLOv10s-1-2×81.8 ± 0.478.2 ± 0.385.9 ± 0.417.7
YOLOv10s-1-2-385.0 ± 0.378.1 ± 0.687.8 ± 0.527.4
Hyperparameter settings: batch size = 32, epoch = 500 (√ indicates the corresponding improvement strategy module is introduced, × indicates it is not used). To ensure the robustness of the proposed models, each variant was trained and evaluated over five independent runs. This table presents the average performance along with the standard deviation.
Table 6. Evaluation metrics of detection results of the test set reconstructed by different magnification factors.
Table 6. Evaluation metrics of detection results of the test set reconstructed by different magnification factors.
Metrics (%)x1_LRx2_SRx3_SRx4_SRx5_SR
Precision45.677.279.480.379.6
Recall46.167.372.075.075.1
mAP@5042.777.882.285.183.9
mAP@0.5:0.9529.453.857.159.158.0
Table 7. The detection results of competitive networks on various test sets.
Table 7. The detection results of competitive networks on various test sets.
DifferentSizeHR (%)Bicubic (%)MambaIR (%)
Models(MB)PRmAP@50mAP@
0.5:0.95
PRmAP@50mAP@
0.5:0.95
PRmAP@50mAP@
0.5:0.95
Faster RCNN113.682.975.483.155.973.720.533.620.584.464.375.448.9
RT-DETR18.272.077.377.946.953.353.451.433.672.069.774.747.1
MobileNetv2-SSD17.775.577.878.247.264.251.052.334.971.971.871.143.6
GhostNet-P619.478.179.982.655.258.751.856.238.372.771.475.148.1
YOLOX-s36.081.977.579.349.389.428.241.625.879.656.365.339.5
YOLO-World27.177.376.379.849.757.046.851.834.274.771.776.049.9
YOLOv6s32.977.476.780.150.359.547.253.636.277.969.178.453.2
YOLOv774.979.477.882.454.749.648.547.429.977.975.279.954.8
YOLOv8s22.577.975.078.747.962.954.557.139.475.376.073.245.5
YOLOv9t6.178.873.281.653.143.255.648.831.177.269.077.652.3
YOLOv10s16.679.073.680.551.752.350.049.732.177.670.777.351.5
Succulent-YOLO27.485.078.087.860.564.468.168.447.880.375.085.159.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Zhao, F.; Xue, F.; Wang, J.; Liu, Y.; Chen, Y.; Wu, Q.; Tao, J.; Zhang, G.; Xi, D.; et al. Succulent-YOLO: Smart UAV-Assisted Succulent Farmland Monitoring with CLIP-Based YOLOv10 and Mamba Computer Vision. Remote Sens. 2025, 17, 2219. https://doi.org/10.3390/rs17132219

AMA Style

Li H, Zhao F, Xue F, Wang J, Liu Y, Chen Y, Wu Q, Tao J, Zhang G, Xi D, et al. Succulent-YOLO: Smart UAV-Assisted Succulent Farmland Monitoring with CLIP-Based YOLOv10 and Mamba Computer Vision. Remote Sensing. 2025; 17(13):2219. https://doi.org/10.3390/rs17132219

Chicago/Turabian Style

Li, Hui, Fan Zhao, Feng Xue, Jiaqi Wang, Yongying Liu, Yijia Chen, Qingyang Wu, Jianghan Tao, Guocheng Zhang, Dianhan Xi, and et al. 2025. "Succulent-YOLO: Smart UAV-Assisted Succulent Farmland Monitoring with CLIP-Based YOLOv10 and Mamba Computer Vision" Remote Sensing 17, no. 13: 2219. https://doi.org/10.3390/rs17132219

APA Style

Li, H., Zhao, F., Xue, F., Wang, J., Liu, Y., Chen, Y., Wu, Q., Tao, J., Zhang, G., Xi, D., Chen, J., & Kobayashi, H. H. (2025). Succulent-YOLO: Smart UAV-Assisted Succulent Farmland Monitoring with CLIP-Based YOLOv10 and Mamba Computer Vision. Remote Sensing, 17(13), 2219. https://doi.org/10.3390/rs17132219

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop