Next Article in Journal
All-Ceramic Fiber Fabry–Perot Cavity High-Temperature Pulsating Pressure Sensor Based on HTCC
Previous Article in Journal
Moon Sensor Station to Improve the Performance of Lunar Satellite Navigation Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Review of Research on Fruit and Vegetable Picking Robots Based on Deep Learning

Institute of Data Science and Agricultural Economics, Beijing Academy of Agriculture and Forestry Sciences, Beijing 100097, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(12), 3677; https://doi.org/10.3390/s25123677
Submission received: 26 March 2025 / Revised: 28 May 2025 / Accepted: 4 June 2025 / Published: 12 June 2025
(This article belongs to the Section Sensors and Robotics)

Abstract

Fruit and vegetable picking robots are considered an important way to promote agricultural modernization due to their high efficiency, precision, and intelligence. However, most of the existing research has sporadically involved single application areas, such as object detection, classification, and path planning, and has not yet comprehensively sorted out the core applications of deep learning technology in fruit and vegetable picking robots, the current technological bottlenecks faced, and future development directions. This review summarizes the key technologies and applications of deep learning in the visual perception and target recognition, path planning and motion control, and intelligent control of end effectors of fruit and vegetable picking robots. It focuses on the optimization strategies and common problems related to deep learning and explores the challenges and development trends of deep learning in improving the perception accuracy, multi-sensor collaboration, multimodal data fusion, adaptive control, and human–computer interaction of fruit and vegetable picking robots in the future. The aim is to provide theoretical support and practical guidance for the practical application of deep learning technology in fruit and vegetable picking robots.

1. Introduction

With the continuous advancement of global agricultural modernization, fruits and vegetables are essential agricultural products whose planting area and production are in-creasing yearly. According to statistics, the global annual output of fruits and vegetables has reached billions of tons, becoming an important part of the agricultural economy [1,2,3]. However, in stark contrast to the increase in production, the harvesting of fruits and vegetables has long relied on manual operation, labor intensity, low efficiency, and rising costs, especially in countries and regions with a shortage of labor resources; the labor cost occupies a significant part of the total cost of fruit and vegetable harvesting. This status quo not only puts pressure on farmers’ economic benefits but also challenges the stability of the supply chain [4,5,6]. In this context, the automation of fruit and vegetable picking has become an inevitable trend in the development of modern agriculture, and picking robots, because of their high efficiency, precision, and intelligence, are considered to be an important way to solve the above problems, which is of great significance in promoting agricultural modernization.
In recent years, with the continuous development of artificial intelligence technology, fruit and vegetable picking robots have ushered in important technological breakthroughs, especially the rapid development of deep learning technology, which has injected new vitality into fruit and vegetable picking robots. In terms of fruit detection, deep neural network structures such as convolutional neural networks (CNNs) have been widely used in fruit and vegetable target recognition and localization tasks, such as the application of the R-CNN family of methods in target detection proposed by Girshick et al. [7], as well as the advantages of YOLO and SSD algorithms in terms of real-time and accuracy [8,9], both of which have been introduced to fruit and vegetable picking scenarios by several research teams. For example, Sa et al. [10] applied Faster R-CNNs to detect apples in real time with a recognition accuracy of over 90%. In addition, deep reinforcement learning (DRL) gradually shows greater potential in path planning and control. By jointly optimizing the sensing results with the robotic arm motion decision, researchers can achieve adaptive adjustment of picking strategies in dynamic environments. For example, Yuqi Liu et al. proposed a deep reinforcement learning strategy combined with expert experience guidance for the path planning of a fruit and vegetable picking robotic arm. The method improves the learning efficiency and performance of the model by introducing expert experience at the early stage of training [11]. Regarding data enhancement and environment simulation, generative adversarial networks (GANs) provide a new means of synthesizing fruit and vegetable image data, thus alleviating the problem of difficult labeling and insufficient samples of agricultural visual data [12].
At present, although the fruit and vegetable picking robot research has realized the initial transformation from “can pick” to “will pick” and “pick well”, it still faces many challenges and technical bottlenecks. First, many kinds of fruits and vegetables have different forms. The picking environment is complex and variable (such as changes in light, shading, overlap, and differences in the maturity of fruits and vegetables), which leads to the recognition algorithms based on deep learning in the generalization ability and the robustness of the recognition algorithms still having large room for improvement. Secondly, most of the existing picking robot systems are laboratory versions or have small-scale validation under specific conditions, and it is not easy to realize stable and efficient applications in large-scale agricultural scenarios. Some of the research focuses too much on algorithmic index (e.g., mAP, IoU) enhancement, while neglecting the evaluation of the overall performance of the whole picking system in real agricultural scenarios in terms of synergistic efficiency, energy consumption, cost, and other comprehensive performances. In addition, there is still a lack of systematic deep learning solutions for multi-sensor fusion, 3D vision reconstruction, and precise control of fruit grasping and picking movements. For example, data synchronization and feature fusion methods between multimodal data (RGB-D, LIDAR, IR, etc.) are not yet mature; robotic arm trajectory planning often relies on rules or predefined models, failing to combine visual information to achieve dynamic feedback control effectively, and most models have high requirements on computational resources and data volume during training, which is not conducive to deployment under edge computing conditions. Therefore, this paper provides a systematic review of the research on fruit and vegetable picking robots based on deep learning, comprehensively summarizes its latest progress in fruit detection, environmental adaptability, multi-sensor fusion, etc., analyzes the advantages and shortcomings of deep learning technology, and explores possible future research directions. This not only helps to provide an important reference for the research in related fields but also provides theoretical support and practical guidance for promoting the practical application of deep learning technology in fruit and vegetable picking robots.
In order to reflect the latest research results, this paper focuses on research results after 2020. More than 325 relevant scientific papers were retrieved from the Web of Science database using keywords such as deep learning, agriculture, and picking robots. Through further research and exclusion, 122 papers related to this topic were selected for in-depth study. Section 2 of the article provides an overview of deep learning technology; Section 3 analyzes the key technologies and applications of deep learning in picking robots in the literature; Section 4 focuses on optimization strategies and common problems related to deep learning; and Section 5 concludes the whole article with a summary and discussion of the challenges and future trends of deep learning technology.

2. Deep Learning Overview

Deep learning was proposed and developed during the research of artificial neural networks. Deep learning is one of the core technologies in the field of artificial intelligence, and its essence is an algorithmic framework based on multilayer neural networks, which can simulate the hierarchical abstraction and processing of data by the human brain. By constructing multilayer nonlinear computational units, deep learning can automatically extract features from large-scale data and perform pattern recognition, which is suitable for solving complex perception, classification, and prediction tasks [13,14]. In 2006, Hinton et al. [15] proposed an unsupervised greedy layer-by-layer training algorithm based on deep belief nets (DBNs). This proved that the deep neural network could be trained layer-by-layer after pre-training. After pre-training, it can achieve far better results than other methods in handwritten digit recognition tasks, which lays the foundation for the development of deep learning technology. Lecun et al. [16] proposed that convolutional neural networks, as an innovative multilayer learning architecture, are unique in effectively utilizing spatial positional relationship information. They thereby significantly reduce the number of network parameters and achieve a breakthrough improvement in backpropagation training efficiency. Currently, deep learning techniques mainly include convolutional neural networks (CNNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), and deep reinforcement learning (DRL). Each of these algorithms has its own characteristics, with CNNs excelling in image processing and RNNs having an advantage in time-series data modeling; GANs are used for data generation and augmentation, and DRL excels in solving complex decision-making problems (Table 1). Listed below are several deep learning methods commonly used in picking robots.

2.1. Convolutional Neural Networks

A convolutional neural network (CNN) is a type of neural network that specializes in processing grid topology data. The network has representation learning ability, which can realize hierarchical feature extraction and representation learning of images through operations such as local connectivity, weight sharing, and spatial dimensionality reduction. The convolutional neural network can learn and extract effective features automatically. It can fuse image information of different scales and extract multigranularity features through the multi-scale convolutional kernel and pooling window, which enhances its ability to understand complex scenes [17,18]. At the same time, the convolutional neural network has good generalization ability, using local connections and weight-sharing mechanisms, which greatly reduce the number of network parameters and model complexity. The detailed workflow of the convolutional neural network (CNN) is illustrated in Figure 1. However, the convolutional neural network relies on many labeled data. The computation required in training and inference is huge, resulting in high computational resources and storage space requirements. It is easily affected by noise, occlusion, and adversarial attacks in feature learning [19]. Despite some limitations, its powerful feature learning capability and excellent performance make it the model of choice in the field of image recognition. Future research directions include reducing the dependence on labeled data, improving the model’s interpretability and robustness, and reducing the computation and storage overhead to expand the application scope of CNNs further. In picking robots, CNNs are mainly used for visual perception, and most researchers have used CNNs to improve picking accuracy and efficiency around target detection and localization, fruit ripeness detection, picking path planning, and fruit quality grading [20,21,22,23]. Commonly used CNN architectures include AlexNet [24], VGGNet [25], GoogLeNet [26], and ResNet [27].

2.2. Object Detection

Target detection algorithms belong to an important branch in computer vision; by learning a large number of standard samples, they can accurately recognize and detect the category of the target object and where it is located and provide its bounding box edges [28]. The detailed workflow of the object detection is illustrated in Figure 2. Deep learning has spawned two mainstream algorithms in the field of object detection, including two-stage detection algorithms and one-stage detection algorithms [29]. The two-stage algorithm belongs to segment-to-segment target detection; firstly, it uses various convolutional neural networks as the backbone network and performs the feature extraction to generate the candidate region; then, secondly, it classifies and regresses the candidate region; the algorithm has high accuracy and is more robust to occlusion, deformation, etc., but the computational speed is slow. Mainstream two-stage algorithms mainly include RCNN [30], Fast R-CNN [31], Faster R-CNN [32], FPN [33], etc. The one-stage algorithm belongs to the end-to-end target detection. It uses CNN convolutional features to directly regress the object’s category probability and location coordinate values, which is less accurate than the two-stage algorithm, but it is simplified. The algorithm is less accurate than the two-stage algorithm, but it simplifies the process and is faster. Mainstream one-stage algorithms mainly include the YOLO series [34], SSD [35], etc. Target detection algorithms can quickly and accurately identify and localize mature targets in complex agricultural production environments for picking robots, providing support for picking decisions and path planning. It also plays an active role in recognizing obstacles and guiding the robot end effector to avoid obstacles. The commonly used target detection frameworks for general picking robots are SSD, YOLO, Faster R-CNN, etc. [36,37,38,39]. The generalization and robustness of the current target detection algorithms in terms of scene lighting problems, shooting angle, fruit shape, etc., need to be further optimized.

2.3. Deep Reinforcement Learning

Deep reinforcement learning is a neural network-based machine learning method that combines deep learning and reinforcement learning to define the problem and optimization objective with reinforcement learning; it uses deep learning to solve the problems of state representation, policy representation, value function modeling, etc., and deep neural networks to approximate reinforcement learning value functions or policy functions, thus realizing end-to-end autonomous learning and decision making [40,41]. Deep reinforcement learning can adaptively extract helpful features in the face of complex, high-dimensional environments. With strong generalization ability, it can deal with the problem of continuous action space and be applied to intelligent robotics, natural language processing, image recognition, and other fields. The detailed workflow of the deep reinforcement learning is illustrated in Figure 3. The application in fruit and vegetable picking robots mainly plays an important research role in vision servo control, path planning strategy optimization, flexible end effector control, etc. Researchers have utilized deep reinforcement learning algorithms, such as DQN [42], DDPG [43], SAC [44], etc., combined with computer vision and multi-sensor fusion technology, to significantly improve the perception, decision making, and execution of picking robots’ ability to pick [45,46,47,48,49]. However, deep reinforcement learning algorithms require a large amount of environmental interaction data in practical applications; the training and learning process is slow, easily affected by hyperparameters, initialization, and other factors, and the model interpretability is poor, which limits its application effect. Future research needs to improve the efficiency and accuracy of the technique in methods such as data enhancement, multi-intelligence body collaboration, model interpretability, and combination with migration learning to better adapt to different tasks and environments.

2.4. Semantic Segmentation

Semantic segmentation is an important application of deep learning in image processing, aiming to classify each pixel in an image, thereby determining the region or class it belongs to [50]. The detailed workflow of the semantic segmentation is illustrated in Figure 4. Commonly used semantic segmentation networks include FCN [51], SegNet [52], U-Net [53], DeepLab [54], etc., which usually adopt the network structure of encoder–decoder. The encoder is responsible for extracting the high-level semantic features of an image, and the decoder categorizes each pixel according to the feature mapping. Picking robots incorporating semantic segmentation can accurately identify and localize the features and attributes of target picks, which can be applied to target detection and localization, ripeness recognition, fruit and vegetable counting, obstacle recognition, and other aspects to improve the robot’s perceptual ability and picking accuracy [55,56,57,58]. However, semantic segmentation techniques have problems such as inaccurate segmentation accuracy and poor handling of fuzzy boundaries for targets with small areas or low frequency of occurrence, and enhancement strategies such as data augmentation and cost-sensitive learning are needed to improve the accuracy. In addition, since the end-to-end segmentation model relies on a large number of pixel-level sample data, the computational volume is large, and the real-time performance is poor, which makes it less efficient in practical applications. In the future, the robustness and efficiency of the technique can be improved by investigating lightweight models, small-sample learning training, and multimodal fusion.

3. Key Technology and Application of Deep Learning in Picking Robots

This section focuses on the key technologies and applications around deep learning in visual perception and target recognition, path planning and motion control, and intelligent control of end effectors for picking robots.

3.1. Visual Perception and Target Recognition

Visual perception and target recognition technology are the key to realizing intelligent picking robots. An efficient visual perception and target recognition system can help fruit and vegetable picking robots to quickly locate and recognize ripe fruits in complex and changing agricultural environments and provide reliable sensing information for subsequent path planning and end effector control, thus improving picking efficiency and success rate. This section focuses on the research progress and key technologies of deep learning techniques in fruit detection and classification and stem and leaf segmentation. Table 2 summarizes the latest research progress of deep learning in visual perception and object recognition.

3.1.1. Fruit Detection and Classification

Fruit detection and classification aims to locate the spatial position of target fruits from images or video streams and further determine their attributes, such as variety category and maturity. Currently, mainstream deep learning target detection algorithms, such as Faster R-CNN, YOLO, SSD, etc., have been widely used in fruit detection and have been improved and optimized for agricultural scenarios, such as the introduction of multi-scale feature fusion, attention mechanism, etc., to improve the detection accuracy and robustness of the algorithm. In addition, some lightweight target detection models, such as YOLO-Tiny and MobileNet, are also used to build an efficient embedded visual perception system to meet the real-time demand of picking robots.
Researchers have improved and optimized the classical deep convolutional neural network target detection algorithms to meet the special characteristics of agricultural scenarios. Zhang et al. [59,60] proposed the HGCA-YOLO and YOLO5-Spear models to achieve high-precision real-time localization of white asparagus tips by designing modules to enhance the sensitivity of the target in complex backgrounds and a lightweight network structure. Xiao et al. [61] improved the YOLOv4-Tiny network by adopting multi-scale feature fusion, cavity convolution, and an anchor-free strategy to improve the detection capability of citrus small targets while taking into account the computational efficiency. Liu et al. [39] introduced the coordinate attention mechanism and BiFPN structure based on the YOLOv5s and optimized the detection head design to form the YOLOv5s-BC model, which significantly improved the apple detection accuracy. Chai et al. [62] were the first to integrate YOLOv7 with augmented reality technology for the purpose of strawberry ripeness detection in agricultural applications, achieving real-time detection at a rate of 18 ms per frame while maintaining high accuracy. Zeeshan et al. [23] used a general real-time dataset and convolutional neural network technology to accurately recognize orange targets. The model uses the RMSprop optimizer and ReLU activation function and adjusts the hyperparameters (e.g., batch size, number of layers, neuron number) to optimize the model performance, which achieves efficient detection and localization of orange fruits in dynamic environments.
The research has also explored multimodal perception fusion technologies to improve fruit detection and attribute analysis, enhancing the ability to perceive the internal attributes and spatial positioning of fruits. Ge et al. [21] combined the internal sensing system of a multi-view gripper with a lightweight convolutional neural network to realize a comprehensive assessment of strawberry ripening by fusing the front and back images of fruits to overcome the limitations of single-view perception. Liu et al. [38] proposed a multi-sensor fusion method based on ORB-SLAM3, YOLO-V5, and Lidar to achieve high-precision 3D localization of apple targets in dynamic environments through machine vision localization, image detection, and point cloud processing. These studies utilize the complementary characteristics of different sensors to obtain more comprehensive fruit information, expand the dimension of traditional visual perception, and effectively improve the precision control level of fruit picking robots.
To address the challenges of sample scarcity and scene variability in agricultural object detection, researchers have explored the application of small-sample learning and data augmentation techniques in this field. Lee et al. [63] evaluated the effect of RandAugment (RA) data augmentation in optimizing the crop detection model, and experiments have shown that applying an RA strategy to small fruit datasets, such as tomato and apple datasets, can effectively improve the model’s performance. Škrabánek et al. [64] combined data enhancement with the DenseNet network structure in a grape variety identification task to alleviate the small-sample problem while ensuring the generalization of the model. These works show that data augmentation and domain-adaptive small-sample learning strategies are effective ways to alleviate the problem of data scarcity in agriculture and can provide algorithmic support for the practical deployment of fruit picking robots.
The real-time performance of the visual perception module plays an important role in fruit detection. In order to meet the resource constraints of embedded devices, researchers have explored lightweight design methods for deep learning models. Ge et al. [21] proposed a MiniNet convolutional neural network, which reduces the inference time to 6.5 ms by optimizing the depth of the network and the size of the convolution kernel while guaranteeing the accuracy of maturity regression. Zhang et al. [60] designed the YOLO5-Spear model with a highly converged lightweight backbone network and optimized detection head structure to achieve 97.8% accuracy and real-time performance on the white asparagus tip detection task. Qi et al. [65] proposed a highly efficient detection architecture based on CSPDenseNet and a recursive feature pyramid for chamomile detection, achieving 47.23 FPS with 92.49% average accuracy to achieve a 47.23 FPS inference speed. The improved YOLO model of Brasilia et al. [66] can detect tree fruits in real time at a rate of more than 20 FPS. Kang et al. [67] achieved high-precision recognition and segmentation of apple fruits by introducing a lightweight Mobile-DasNet backbone in combination with the PointNet network, significantly reducing computational complexity while ensuring recognition accuracy. These works show that the lightweight network structure design and computational optimization techniques can effectively balance the accuracy and speed of the fruit detection task and provide an algorithmic foundation for the real-time sensing of fruit picking robots.

3.1.2. Stem and Leaf Segmentation

In addition to the fruits themselves, the picking robot also needs to accurately recognize and localize the stems and leaves of the fruits and vegetables in order to achieve obstacle avoidance and accurate picking. Semantic segmentation plays a key role here, where each pixel in the image is labeled with a specific category. CNN-based models like FCN, U-Net, and DeepLab are widely adopted, and their performance has been enhanced using context modules and attention mechanisms. Research has also explored instance segmentation to handle overlapping leaves and branches, allowing more precise structural understanding.
Fujinaga et al. [68] realized multitasking for strawberry picking and stem pruning by combining DeepLabV3 with semantic segmentation models; the F1 scores for ripe and unripe fruit detection were 0.96 and 0.88 for ripe and unripe fruit detection, and the F1 scores were 0.93 and 0.86 for picking and pruning cutting point detection, respectively. Lemsalu et al. [69] developed a real-time, lightweight visual perception system based on YOLOv5 and realized strawberry and fruit stalk detection at 30 fps on a Jetson AGX Xavier edge device, with average accuracies of 91.5%, and Coll Ribes et al. [22] proposed an innovative method that combines monocular depth estimation with convolutional neural networks (CNNs) to achieve precise detection and localization of grape clusters and fruit stems in the complex environments of vineyards. This method was tested on two different vineyard scene datasets, and the results showed that its performance was superior to the current state-of-the-art methods. Li et al. [70] proposed a strawberry recognition method based on improved YOLOv7 and RGB-D sensing and introduced the integrated GhostConv module and CBAM attention mechanism module on the basis of YOLOv7, which optimized the detection performance of small targets, and the recognition accuracy was improved by 11–14.8%. Wang et al. [71] proposed a parallel network structure, the DualSeg model, which fused the advantages of the CNN and Transformer through the parallel double-branch structure and the features of the Transformer. Through the parallel dual-branch structure and feature fusion, it can segment grape bunches and fruit stalks more accurately in complex vineyard environments, and the IoU value of the grape stalk segmentation of this model is 72.1%, which is more than 3.9% higher than that of other models.
Researchers applied instance segmentation to the separation of overlapping leaves and branches to obtain finer structural information. Li et al. [72] proposed a multitask perceptual network, MTA-YOLACT, which added detection and instance segmentation branches to the pre-trained YOLACT model and used the shared ResNet-101 backbone network and loss function based on the uncertainty weights to achieve fruit bunch detection, fruit stalk main stem segmentation, and its subordination decision. Peng et al. [57] proposed a depth information-based region growing method, DRG, for segmenting overlapping grape bunches, which can effectively deal with the complex situation of multiple overlapping bunches and obtain a finer contour, with an average recall rate of 89.2% and an average precision rate of 87.5%.
Multitask learning approaches, which combine target detection, semantic segmentation, and pose estimation, have been explored to offer more comprehensive spatial and structural perception of fruits and vegetables. Jing et al. [73] proposed a three-dimensional localization method for kiwifruit with an end-to-end stereo matching network LaC-Gwc Net combined with a target detection model, YOLOv8m, and a better result was obtained for kiwifruit and its calyx. Good results were obtained with an average accuracy of 93.1% and a detection speed of 7.0 ms, which is especially important for sensing obstacle avoidance of small obstacles such as branches and wires. Kim et al. [74] introduced the PAF method in the field of human pose estimation into the pose estimation of a tomato fruit stalk system for the first time; the structural information of multiple tomatoes could be acquired from a single image simultaneously, and the average detection rates of the key points and segments were 92.9% and 92.9%, respectively. Miao et al. [75] proposed a visual perception method integrating YOLOv5, ROI extraction, and an adaptive weighting strategy for tomato picking, which combines the advantages of image processing and deep learning and realizes the estimation of tomato ripeness and accurate stalk localization under different light conditions, and the developed picking robot achieves the picking efficiency of one cluster in 9 s and the success rate of 95%. The developed picking robot achieves a picking efficiency of 9 s per cluster and a success rate of 95%.
Table 2. Comparison of different applications of deep learning in visual perception and object recognition.
Table 2. Comparison of different applications of deep learning in visual perception and object recognition.
TypeRef.CropYearTaskNetwork Framework and AlgorithmsF1 ScorePrecisionmAPOther Metric
Fruit detection and classification[59]White asparagus2024Object detectionHGCA-YOLO92.4%-95.2%
[60]White asparagus2022Object detectionYOLO5-Spear96.8%-1
[61]Citrus2024Object detectionYOLOv4-Tiny-93.5%93.25%Recall: 93.73%
[62]Strawberry2023Object detectionYOLOv7-Multi-Scale92%-89%
[39]Apple2024Object detectionYOLOv5s-BC84.32%99.8%88.7%
[23]Orange2023Object detectionCNN96.5%98%-Recall: 94.8%
[21]Strawberry2023Object detectionMiniNet CNN---MAE: 4.8%
[38]Apple2023target locationYOLO-V5, ORB-SLAM3---RMSE: 26 mm
[63]Tomato2023Object detectionYOLOv3, YOLOv8, EfficientDet-76.42%93.73%
[64]Grapevine2023Image segmentationDenseNet-98.00 ± 0.13%-
[65]Tea Chrysanthemum2021Object detectionTC-YOLO--92.49%
[66]Apple, pear2019Object detectionCNN81%95%-
[67]Apple2020Object detectionMobile-DasNet85.1%-86.3%
Stem and leaf segmentation[68]Strawberries2024Cutting point detectionDeepLabV3+, ResNet-5091%--MBF: 74.2%
[69]Strawberries2022Object detectionYOLOv5m89.4%89.0%91.5%Recall: 89.8%
[22]Table Grapes2023Object detectionMask R-CNN94.7%95.6%-Recall: 93.8%
[70]Strawberries2024Object detectionYOLOv797.9%99.0%99.8%Recall: 96.9%
[72]Cherry tomato 2023Image segmentationMTA-YOLACT, ResNet-10195.4%98.9%45.3%Recall: 92.1%
[57]Grape2021Image segmentationDeepLabV3+98.63%87.5%-Recall: 89.2%
[73]Kiwifruit2024Object detectionLaC-Gwc Net, YOLOv8m--93.1%
[74]Tomato2023Object detectionYOLOv5--91.7%

3.2. Path Planning and Motion Control

The autonomous navigation of harvesting robots in orchards requires overcoming the challenges of environmental complexity and dynamic changes. Optimized path planning algorithms and smooth motion control are key to achieving this goal. By utilizing deep learning algorithms for optimization, harvesting robots can adjust their travel trajectory in real time in complex environments, reducing the risk of collisions. In addition, the optimized path can cover the entire work area, minimizing the possibility of omissions and repeated picking. Table 3 summarizes the latest research progress of deep learning in path planning and motion control.

3.2.1. Deep Reinforcement Learning-Based Path Planning

In the field of picking robots, the core task of path planning is to comprehensively consider environmental information and task requirements and find a collision-free, optimal or suboptimal path for the robot from the starting point to the target point. The limitations of traditional path planning algorithms are becoming increasingly prominent in dynamic and complex environments. Algorithms such as Dijkstra’s often struggle to meet performance requirements in terms of real-time performance and robustness, resulting in severe constraints on robots’ adaptability and fault tolerance. In recent years, deep reinforcement learning (DRL) has shown great potential in the field of robot path planning by virtue of its powerful environment awareness and policy optimization capabilities. Through end-to-end learning, deep reinforcement learning can realize direct mapping from raw observation data to action output without designing complex rules manually and shows excellent adaptability and robustness in dynamic, uncertain environments. More and more scholars have started to apply deep reinforcement learning to the path planning task of fruit and vegetable picking robots.
Lin et al. [47] used a recurrent deep reinforcement learning (recurrent DDPG) algorithm, combined with a long short-term memory network (LSTM), to realize accurate trajectory prediction in complex scenarios by utilizing historical state information. The success rate of the collision-free path planning of recurrent DDPG was 90.90%, which was better than that of the traditional DDPG (19.32%) and the bi-directional RRT algorithm and two-way RRT algorithm (88.18%). Li et al. [48] proposed an intermittent stop point planning method based on the deep deterministic policy gradient (DDPG) to improve the picking efficiency and optimize the motion path of a dual-arm tomato picking robot in greenhouse environments. By generating a sequence of stop points, this method significantly reduces redundant movements of the robotic arms between multiple target points and increases the number of effective picks per unit time. Wang et al. [46] designed a path planning strategy that combines deep reinforcement learning. The research team first maps the position coordinates of kiwifruit to a two-dimensional plane, thereby obtaining visual information about the distribution of fruits in the orchard. Next, they transformed the path coverage problem in this scenario into a mathematical model of the traveling salesman problem (TSP). In order to optimize the traversal order of regions, they further proposed an improved DQN algorithm. This approach effectively shortens the coverage path and improves the efficiency of target point coverage. Overall, deep reinforcement learning empowers the picking robot with human-like autonomous decision making and adaptive ability. It also provides new ideas and methods for flexible path planning in an unstructured farmland environment. In the future, further improvement of the generalization ability and balancing the exploration and utilization of the strategy while ensuring high harvesting efficiency will be an urgent challenge to be tackled in this direction.

3.2.2. Multimodal Environment Perception and Semantic Understanding

As the orchard scene is complex and dynamic, it is often difficult for a single sensor to perceive the surrounding environment comprehensively and accurately. Therefore, integrating the advantages of multiple heterogeneous sensors and extracting semantically rich multimodal feature representations is the key to the accurate perception and control of fruit and vegetable picking robots. The RGB camera directly simulates the human eye’s vision with high image clarity. Chen et al. [76] proposed a set of dynamic visual servo control frameworks and localized fruits in real time through the YOLOv4 target detection network, combined with the self-built map, to achieve stable and efficient autonomous positioning. Combined with the self-built map and autonomous localization, it achieves stable and efficient fruit approximation and continuous operation. At the same time, the depth camera can obtain the distance information of the target to solve the scale uncertainty problem. Wang et al. [77] designed a multitask convolutional network A3N, which skillfully integrates the pixel-level features of RGB images and depth maps, realizes the segmentation of fruit instances and the regression of the grasping position, and integrates them into the robotic perceptual decision-making system, which greatly improves the performance of the robot in complex scenarios. The success rate of grasping complex scenes is greatly improved. Lidar has a larger detection range and accuracy. Xiong et al. [78] constructed a point cloud processing pipeline on this basis through the combination of threshold segmentation, depth target detection, accurate analysis of the distribution of obstacles around the fruits, and the generation of the optimal fruit separation path for robotic arm motion planning. It can be seen that multimodal environment perception requires deep learning technology and traditional image vision to cooperate and synergize their advantages. However, it still faces many challenges, such as sparse samples, variable scenes, and limited computation. A lightweight and efficient multimodal feature learning framework and autonomous evolution of the online learning paradigm will be an important research direction in the future.

3.2.3. Deep Learning-Based Machine Vision Localization

Accurate target localization is a prerequisite for picking robots to realize automated operation. However, the orchard environment has a complex background and large changes in light, and the fruits are often obscured and clustered, which makes it difficult to adapt to traditional methods such as color thresholding and contour detection. Deep learning, with its powerful feature extraction and classification capabilities, provides a new solution to the fruit localization problem. Kok et al. [79] proposed a new method for apple pose estimation. They first used the Mask R-CNN to realize the fine segmentation of fruits, then combined this with trunk detection to eliminate the background interference, and finally cleverly utilized the circular a priori of apples to construct a 2D to 3D pose mapping model; finally, they realized the spatial localization of apples under the monocular camera. For the occlusion problem in dense fruit clusters, Xiong et al. [78] combined deep target detection and point cloud postprocessing organically, returned to the 3D bounding box directly through end-to-end learning, and designed a point cloud clustering optimization strategy, which effectively improved the localization accuracy in dense scenes. Currently, deep learning is reshaping the traditional machine vision paradigm to explore end-to-end object localization and pose estimation methods. Although promising progress has been made, the non-rigid and fragile nature of the orchard environment puts forward higher requirements for localization accuracy and speed. Lightweight and efficient detection and tracking algorithms, multi-view joint optimization strategies, and position estimation models with good interpretability are the main directions for the future.

3.2.4. Multitask-Oriented Robot Motion Planning

After grasping the global information of the orchard environment, it is necessary to further formulate the optimal picking motion strategy according to the constraints of fruit distribution and robotic arm structure. This is a complex combinatorial optimization problem involving multi-objective constraints in kinematics and dynamics. In recent years, deep learning has injected new vitality into traditional motion planning methods. Li et al. [80] proposed an integrated model (DRL-MPC-GNNs) based on deep reinforcement learning (DRL), model predictive control (MPC), and graph neural networks (GNNs), which solves the multi-robot system optimization problem by combining the powerful feature extraction capability of deep learning and the decision-making mechanism to solve the multitask motion planning problem of multi-robot systems in complex dynamic environments. Martini et al. [81] proposed a PIC 4 rl-gym modular framework combined with DRL implementation for training robots to perform multitask autonomous navigation planning in complex environments. Li et al. [82] proposed a new framework for collaborative multi-arm picking scheduling optimization, which first applies the improved Faster R-CNN detection network to realize multi-angle fruit recognition and localization and construct a global perceptual map accordingly. On this basis, a multi-arm task allocation algorithm based on deep reinforcement learning is designed. Through the end-to-end learning collaboration strategy, multiple robotic arms are guided to autonomously divide the work, minimize the overall operation time while avoiding collision and interference, reduce the repetitive movement and idling waiting time in the dense fruit area and narrow operation space, and improve the overall picking efficiency. In order to further improve the coverage picking efficiency, Wang et al. [46] divided the orchard into several regions and optimized the traversal order of each sub-region using the re-DQN network. By modeling it as a traveler’s problem that solves and leverages the high-dimensional approximation capability of deep reinforcement learning, the method shortens the global path length while also dramatically reducing the motion search complexity and improves the amount of picking per unit time by reducing unnecessary path backtracking and conflicts. With the improvement of the intelligence level of orchard operation, deep learning is reshaping the traditional robot motion control paradigm, injecting autonomous learning and optimization capabilities. However, ensuring the real-time aspects and robustness of planning while meeting various constraints, balancing local path optimization and global long-term gain, and organically integrating operations research optimization and deep learning are still challenges that need to be overcome in this field.
Table 3. Comparison of different applications of deep learning in path planning and motion control.
Table 3. Comparison of different applications of deep learning in path planning and motion control.
Ref.CropYearTaskNetwork Framework and AlgorithmsF1 ScoremAPAccOther Metric
[47]Guava2021Path PlanningRNN, DDPG---Success Rate: 90.90%
[48]Cherry Tomato2024Path PlanningYOLOv5, DDPG95.7%--Detection speed: 16.5 FPS
[76]Cherry Tomato2024Object detectionYOLOv4, ORB-SLAM3---Recall:
87.5%
[77]Apple2022Object detectionYOLACT, Apple 3D Network89%--IoU: 87.3%
[78]Strawberry2020Object detectionMask R-CNN---Success Rate: 65.1%
[79]Apple2024Object detectionDaSNet-v2, HigherHRNet-89.11%-
[80]-2024Path PlanningDRL-MPC-GNNs92.91%-96.82%Recall:
91.47%
[82]Apple2023Path PlanningMulti-task DCNN71.0%67.3%92.0%Recall:
74.4%

3.3. Intelligent Control of End Effector

End effector intelligent control technology is the key for fruit and vegetable picking robots to realize accurate, efficient, and lossless picking operations. The current research on intelligent end effector control mainly focuses on perception and understanding, motion planning, force control, learning and adaptation, and task planning. Combining traditional control theory with artificial intelligence technology is the development direction of this field to continuously improve the autonomy, flexibility, and robustness of the robot end effector. This section focuses on grasping state classification and anomaly detection, optimal grasping attitude detection, obstacle avoidance path planning, and collision-free control, as well as target attribute sensing and force prediction. Table 4 summarizes the latest research progress of deep learning in intelligent control of the end effector.

3.3.1. Target Attribute Extraction and Grip Force Prediction for Haptic Perception

Haptic perception plays an indispensable role in the process of fruit and vegetable picking. In recent years, researchers have begun to introduce deep learning into haptic perception to adaptively extract target attribute features from massive haptic data. Huang et al. [83] constructed a large-scale haptic dataset for agricultural applications. They proposed a deep neural network based on the Transformer, which extracts the key features of the haptic sequence in both spatial and temporal dimensions to realize target attribute perception and strength prediction in the process of grasping. Ma et al. [84] proposed a deep neural network based on the Transformer, which extracts the key features of haptic sequences from both spatial and temporal dimensions and realizes the target attributes and strength prediction in the grasping process. Ma et al. [84], Lin et al. [85], and Li et al. [86] investigated fruit hardness and grasping force from the perspectives of robotic opto-tactile sensors, active deformation of flexible fingers, the self-attentive capsule network, etc. Han et al. [87] explored the visual–tactile multimodal information fusion, further improving the grasping success rate and prediction accuracy. Zhang et al. [88] proposed a method targeting crops with inflorescence peduncles, wherein the Mask R-CNN model was integrated with a low-cost all-in-one gripper to enable simultaneous gripping and cutting of crop stems without contacting the fruit flesh. These studies utilize deep learning models to extract sparse, multi-scale, and cross-domain feature representations from tactile data, which, to a certain extent, overcome the difficulties of complex surface textures and large material differences between fruits and vegetables. However, the current research mainly focuses on the laboratory environment and how to expand it to the real orchard scene, which still faces many challenges; the future also needs to focus on overcoming the key challenges of active tactile strategy learning, tactile–motor control integration, fine classification of the grasping state, and more.

3.3.2. Optimal Gripping Posture Detection by Fusing Visual Information

Determining the appropriate grasping angle is crucial for fruit and vegetable harvesting. Deep learning technology has recently been widely applied in grasping posture detection. By calculating the optimal grasping posture to guide the robotic arm’s picking action, the end effector’s positioning accuracy has been effectively improved, thereby increasing the success rate of harvesting. Ma et al. [89] improved the GG-CNN2 network; they proposed a method to predict the optimal grasping angle of kiwifruit, guided the manipulator to safely approach, grasp, and detach the kiwifruit, and combined it with the YOLO v4 target detection and achieved an 88.7% picking success rate. Sun et al. [90] proposed the FPVP, which is the most effective way to integrate visual information into grasping gesture detection. The FPENet multitask learning network proposed by Sun et al. [90] can simultaneously localize the position of the navel point and predict the 3D rotation vector of citrus. The CM-YOLOv5s-CSPVoVnet improved network designed by Wang et al. [91] can accurately localize the yellow peach gripping point in complex environments. Hewen Zhang et al. [92] used the YOLOv5s target detection algorithm combined with a depth camera to realize the transformation from a pixel coordinate system to a robotic arm-based coordinate system. Ni et al. [93] proposed a multimodal representation combining two visual modalities, the RGB image and depth map, and applied a two-way CNN to classify the grasping poses. These works mainly start from RGB images or RGB-D data and realize the accurate prediction of grasping parameters through end-to-end deep convolutional neural networks. Despite the good experimental results, there are more specialized detection models designed for different fruit and vegetable varieties and scenes, and there is an urgent need to develop unified frameworks and general algorithms and make breakthroughs in multimodal sensing information fusion and grasping gesture dynamic optimization.
Table 4. Comparison of different applications of deep learning in intelligent control of end effector.
Table 4. Comparison of different applications of deep learning in intelligent control of end effector.
Ref.CropYearTaskNetwork Framework and AlgorithmsPrecisionAccRecallOther Metric
[83]15 types of fruits2023Force PredictionTransformer
DNN
97.33%97.33%-MAE: 0.216
[84]Peach2025Force PredictionCNN, CNN-LSTM---R2: 94.2%
[85]Tomato, Nectarine2023Force PredictionCNN-LSTM, ResNet95.7%97.8%98.1%IoU: 90.2%
[87]Tomato2024Force PredictionCNN, LSTM95.7%97.8%98.1%-
[89]Kiwifruit2022Grab detectionGG-CNN2, YOLO v476.0%--Success rate: 88.7%
[91]Yellow Peaches2023Object detectionCM-YOLOv5s-CSPVoVnet87.8%-94.2%-
[93]-2018Grab detectionTwo-stream CNNs93.4%-77.2%-

3.3.3. Collision-Free Control of Robotic Arm Attitude

Realizing obstacle avoidance grasping in complex environments is another challenge for fruit and vegetable picking. In recent years, reinforcement learning has provided new ideas for solving this challenge. Li et al. [45] proposed a deep reinforcement learning-based collision-free control method of robotic arm attitude for tomato picking in complex environments with sequential obstacle avoidance grasping problems. The method combines the optimal operating attitude plane and the HER-SAC strategy. It dynamically generates collision-free robotic arm motion sequences through the improvement of attitude constraints, heuristics, and dynamic gain modules and achieves a picking success rate of 85.5% in greenhouse experiments, and analyzed from the perspective of the picking experiments and efficiency, the method significantly reduces the time delay due to the collision detection and replanning. Tabakis et al. [94], Lin et al. [95], and Choi et al. [96] proposed a deep reinforcement learning-based path planning method for a robotic arm by training in a simulated environment; the robotic arm was able to find the optimal path in a complex environment while avoiding collisions. This work shows that deep reinforcement learning can adaptively generate obstacle avoidance motion sequences through end-to-end policy learning, providing a new solution for fruit and vegetable picking in complex environments. In the future, it is necessary to improve the algorithm’s sampling efficiency and generalization ability further and make breakthroughs in multi-intelligent body cooperative obstacle avoidance, dynamic environment perception, and prediction.

3.3.4. Deep Learning-Based Collaborative Energy Efficiency Optimization for Multiple Robots

From the perspective of energy efficiency optimization, deep learning techniques show great potential in the cooperative control of the end effector of multi-picking robots. Aiming at the problem of collaborative energy efficiency optimization of the end effector in a multi-robot system, Zhao et al. [97] proposed a deep reinforcement learning framework for multi-robot collaboration based on proximal policy optimization (PPO); they systematically analyzed the effects of different types, magnitudes, and distributions of perturbations on the learning efficiency and laid a foundation for constructing a highly energy-efficient end effector collaborative system adapted to the uncertainties of real orchard environments. Ao et al. [98] proposed a dual-stream attention actor-critic (DSAAC) algorithm based on multi-intelligence deep reinforcement learning, which enables the system to automatically learn the optimal strategies in complex environments, strike a balance between dynamic allocation of communication loads and trajectory planning, and effectively solve the energy-constrained problem in emergency communication scenarios. Wu and Wu et al. [99] proposed a communication embedded deep Q-network (CE-DQN) for distributed multi-robot control, which enables robots to make autonomous decisions on when to share local information and effectively solves the energy- and time-consumption problems in multi-robot systems. Xiao et al. [100] proposed a collaborative inference framework for mobile edge computing based on multi-intelligent deep reinforcement learning (DMECI), which enables mobile devices to self-adaptively select partition points of deep learning models and collaborative edge servers, combined with a learning experience exchange mechanism to achieve knowledge sharing among neighboring devices, enabling intelligent collaborative computing resource allocation and energy efficiency optimization without relying on inference performance models. These deep learning-based approaches improve the energy efficiency of multi-end effector systems and enhance the adaptability of the end effector to complex fruit and vegetable picking environments. Future research can explore more advanced deep learning architectures, such as deep spatio-temporal graph networks and multimodal Transformers, to enhance the end effector’s intelligence and energy efficiency in fruit and vegetable picking systems.

3.4. Summary

In summary, deep learning technology shows great potential and practical value in the three modules of visual recognition, path planning, and end effector control of fruit and vegetable picking robots. In visual recognition, it is recommended to adopt advanced convolutional neural network models such as YOLO and Faster R-CNN combined with attention mechanism and image enhancement technology to achieve high-precision recognition and ripeness determination of different fruit and vegetable categories in complex environments and, at the same time, introduce migration learning and multimodal fusion strategies (e.g., RGB-D images and LiDAR point clouds) to enhance the model’s generalization ability and 3D spatial localization accuracy. We also adopt lightweight models such as MobileNet or EfficientNet to balance model performance and deployment efficiency. Regarding path planning and motion control, traditional planning methods are difficult to adapt to the dynamics of unstructured orchard environments. It is recommended to introduce deep reinforcement learning (e.g., PPO, DDPG, etc.) to achieve self-learning and self-adaptive path optimization and to cooperate with simulation platforms for model training and imitation learning to enhance the generalization ability of the strategy and to incorporate SLAM technology and graph neural networks for environment mapping and localization to further improve the navigation accuracy and dynamic obstacle avoidance ability. In terms of end effector control, due to the wide variety of fruits and vegetables and large structural differences, it is necessary to have flexible and fine control capabilities. It is recommended to train a multimodal perception network by integrating visual and tactile sensors and introduce deep reinforcement learning to realize a dynamic force–control strategy based on feedback, so that the robot can adaptively adjust the grasping posture and strength according to the shape and growth state of the fruits and thus effectively reduce the damage rate of the fruits.
In the future, with the development of algorithms and hardware technology, the introduction of deep learning will significantly promote the fruit and vegetable picking robot from “usable” to “efficient and reliable”, providing strong technical support for modern intelligent agriculture. Table 5 compares and analyzes the key technologies of deep learning-based harvesting robots.

4. Training and Optimization of Deep Learning Models in Picking Robots

4.1. Dataset Construction and Labeling Methods

A high-quality dataset is a prerequisite for training high-performance deep learning models. Constructing high-quality datasets faces many challenges when studying fruit and vegetable picking robots. First of all, there are many kinds of fruits and vegetables, and different kinds and varieties of fruits and vegetables have significant differences in shape, size, color, etc.; so, it is necessary to collect data samples that cover a variety of variations. Second, the growing environment of fruits and vegetables is complex and variable; light, shade, background, and other factors will affect the data quality. In addition, the picking robot needs to judge the ripeness and integrity of fruits and vegetables, which requires the dataset to contain samples of fruits and vegetables at different growth stages and with different degrees of integrity.
To address the above challenges, researchers have proposed various dataset construction methods. Zhang et al. [60] proposed a multi-scale combined slicing and resampling-based image transformation enhancement method for selective harvesting of white asparagus and constructed datasets in rainy and non-rainy day scenarios to expand the training test samples. Wang et al. [75] utilized a depth camera and an RGB-D camera for the training tests on a rainy day and a non-rainy day scenario. RGB-D cameras, 768 RGB-D images and 1132 color images were collected at different times and distances in an apple orchard for training the model; Hwang et al. [101] investigated the creation of the HWANGMOD dataset and generated a large number of 3D citrus images in virtual and real environments through Blender and BlenderProc model data to provide support for model training. Zhang et al. [102] collected tomato image data at five ripening stages to create a dataset through image annotation and grading. Geometric transformations and added noise were used for data enhancement, and the t-SNE method was used to verify the data distribution and clean up the unqualified samples; Qi et al. [103] proposed the Tea Chrysanthemum generative adversarial network (TC-GAN) for generating high-resolution images of Tea Chrysanthemums in a complex and unstructured environment (512 × 512); Zeeshan et al. [23] created a diverse dataset containing actual conditions such as different lighting, shadows, and blade overlaps, enhancing the algorithm’s adaptability and robustness in real-world scenarios. Sa et al. [104] integrated a public dataset with a self-constructed sweet pepper dataset and employed a generative adversarial network (GAN) to synthesize near-infrared (NIR) images from RGB inputs. By incorporating these synthesized images into a four-channel fruit detection framework based on YOLOv5, they provided highly adaptable training data for multimodal detection tasks. Table 6 lists the datasets and their processing methods used in some articles.
Regarding data annotation, the mainstream methods include bounding box annotation and semantic segmentation annotation. Bounding box annotation is to frame the location of target fruits and vegetables in the image and label their categories, such as apples, oranges, etc. The semantic segmentation annotation method achieves object recognition by labeling the category of each image point. Specifically, researchers annotate the collected raw data pixel by pixel when processing crop images to distinguish target objects such as fruits and vegetables from the background. Compared with bounding box annotation, semantic segmentation annotation can provide finer target contour information, but the annotation cost is also higher. Dai et al. [105] captured 960 images of tobacco leaves, increased the image data by rotating, scaling, and luminance change hoods, and labeled four different maturity levels of tobacco leaves. Kim et al. [106] proposed an automated data generation method to generate tomato datasets containing different maturity levels and bit positions through a 3D simulator, which avoids the cost of manual labeling and significantly improves the efficiency of training data generation. Wang et al. [107] captured 1260 images of Shine Muscat and Megapolis grapes. They used the professional annotation software Labelme 5.0.1 to label both the grapes and their stems, aiming to extract more comprehensive grape features.

4.2. Model Training Strategy

In terms of training strategy, fine-tuning is an effective means of improving the performance of fruit and vegetable detection. Selecting an ImageNet pre-trained backbone network (e.g., ResNet) and fine-tuning the self-constructed fruit and vegetable dataset can be significantly better than training from scratch. Secondly, some training techniques, such as data-balanced sampling, multi-scale training, learning rate warmup, etc., also help to further improve the performance and accelerate the convergence. In the training process, the selection of hyperparameters, such as learning rate, batch size, regularization parameter, and so on, is crucial. In order to optimize the hyperparameters, some researchers adjust the batch size, the number of layers, the number of neurons, etc., to optimize the model performance and search for the optimal combination of hyperparameters [23]. Sa et al. [10] adopted a two-phase training strategy: first, pre-training the backbone network on ImageNet to learn the general-purpose features, then fine-tuning it on a specific fruit dataset to extract the fine-grained features. In addition, hard example mining and online bootstrapping are also introduced for hard sample mining and training sample equalization to improve the detection accuracy. Suchet Bargoti and James Underwood [108] designed an iterative training strategy to augment the training data with the a priori knowledge of fruits and to iteratively update the detector.
When the self-constructed dataset is limited, a migration learning strategy can be considered to migrate knowledge from related domains and reduce labeling dependency. With models pre-trained on large-scale datasets (e.g., ImageNet), generic visual feature representations can be learned. In target tasks (e.g., fruit and vegetable detection), the feature extraction capability of the pre-trained models can be utilized to fine-tune them on smaller datasets for better performance. For example, Károly et al. [109] introduced a novel “Filling the Reality Gap” (FTRG) approach using automatically generated synthetic data to improve the fine-tuning phase of transfer learning in a target segmentation task, and Nguyen et al. [110] proposed a forward progressive learning (FPL) approach to achieve reliable transfer learning in robotics applications.

4.3. Model Optimization and Deployment

Although deep learning models have performed well in fruit and vegetable picking tasks, they still face challenges such as high model complexity, slow inference speed, and high resource consumption in practical deployment. To solve these problems, researchers have proposed several methods for optimizing models, including model lightweight, edge computing, and testing and evaluation using multiple models.
Model lightweight is a commonly used optimization method to reduce the storage and computation overhead of the model by reducing the number of parameters and the computation of the model. In the fruit and vegetable detection task, MobileNet [36,111] and other lightweight network structures are widely used; through the depth separable convolution and other techniques, the number of parameters of the model and the computational effort is greatly reduced while maintaining high precision. By improving the main network of CSPDenseNet, embedding the feature fusion module, and optimizing the multi-scale detection head, the number of parameters and the computational cost of the model are significantly reduced while guaranteeing the detection accuracy, and a lightweight and highly efficient Camellia sinensis detection network is finally realized.
In terms of model deployment, edge computing is a promising deployment method. Compared with cloud deployment, edge computing can deploy the model on edge devices close to the user, such as mobile devices, embedded devices, etc., which reduces the delay and cost of data transmission and improves the real-time aspects and privacy of the model [112]. Researchers have explored various edge computing schemes in fruit and vegetable picking robots. For example, Xiao et al. [100] tested real-time inference by putting the deep learning model on an NVIDIA Jetson TX2 edge device(made in NVIDIA Corporation, Santa Clara, CA, USA), with an inference time of about 0.1 s per image of 512 × 512 resolution, laying the foundation for realizing rapid deployment and updating of the model on robots in the future. Zhang et al. [113] proposed using a deep learning model using fine-grained computational partitioning to distribute the computational tasks of deep neural networks to edge devices and the cloud to optimize the end-to-end latency.
In addition to model optimization and deployment, model testing and evaluation are crucial to ensure model performance and reliability. In the fruit and vegetable picking task, researchers have used a variety of evaluation metrics and testing methods. For example, Wang et al. [114] proposed an evaluation framework based on a confusion matrix, which comprehensively evaluates the model’s performance by analyzing its performance in different categories and scenarios. In contrast, Tituaña et al. [115] proposed a testing method based on adversarial attacks, which evaluates the robustness and security of the model by generating adversarial samples.

4.4. Analysis of Model Adaptability and Scalability

Although deep learning technology has significantly promoted the intelligent development of fruit and vegetable picking robots, their adaptability and scalability in complex and changing orchard environments still face challenges. The spatial and temporal dynamics and heterogeneity of the orchard (e.g., light changes, crop growth differences, shading, and interference) significantly affect the robot’s recognition and operation performance, and it has become a research priority to improve the system’s robustness and generalization ability under multi-environmental conditions. First, the spatial distribution of crops significantly affects the visibility and occlusion of fruits, and deep learning can realize accurate detection of target fruits in complex backgrounds by convolutional neural networks (CNNs). For example, RGB images combined with Mask R-CNN networks were used for apples hanging high in the sky to achieve stable recognition under different lighting conditions [116]. For strawberries growing close to the ground and often occluded by foliage, it is necessary to introduce an attention mechanism and an instance segmentation network to improve the robustness and accuracy of occlusion. The SGSNet model developed by Wang et al. combines a lightweight network structure and an attention mechanism to effectively improve the visibility and accuracy of strawberries growing in the ground. The SGSNet model developed by Wang et al. combines a lightweight network structure and an attention mechanism, which effectively improves the detection accuracy of strawberry growth stages [117]. Second, the ripening process of different crops exhibits varied visual feature changes, such as color, texture, and fruit surface state. Deep learning effectively supports fine-grained discrimination of strawberry and grape ripening stages by modeling fruit ripening dynamics through time-series modeling (e.g., LSTM or Transformer) or combining hyperspectral images for non-invasive analysis. For example, Su et al. realized real-time estimation of strawberry ripeness using hyperspectral imaging combined with a deep learning model with a classification accuracy of more than 84% [118]. At the execution level, deep neural networks can be used for robotic gripping point prediction and path planning, combined with visual feedback to achieve adaptive clamping control for different fruit shapes (e.g., spherical for apples, bunches of grapes, and conical for strawberries) to minimize mechanical damage. In order to improve the generalization ability of the robot under different geographical and seasonal conditions, migration learning and incremental learning mechanisms are also widely used, so that its perceptual model can be quickly adapted to new crop varieties or new environmental conditions.
For cost analysis, although picking robots show great potential in enhancing agricultural automation, their commercialization faces significant cost challenges, especially for small and medium-sized farms, where higher initial investment and ongoing operation and maintenance costs are the main barriers to their popularity. Therefore, optimizing the overall cost structure through deep learning technology is particularly important. At the hardware level, the initial acquisition cost of the system can be reduced by choosing cost-effective sensors (e.g., consumer-grade high-resolution cameras and low-cost LIDAR) and adopting multiple alternative technologies. At the software level, optimizing deep learning models is one of the core strategies to reduce the computational resource requirements and operating costs. Current mainstream lightweight neural network architectures, such as MobileNet and EfficientNet, have been widely used in embedded platforms and resource-constrained smart devices. MobileNet significantly reduces the model parameters and computation volume by introducing deep separable convolution to efficiently support mobile vision tasks [119]. EfficientNet further proposes a compound scaling strategy to achieve a balanced scaling between model depth, width, and input resolution, which achieves a better accuracy–efficiency balance across multiple vision tasks [120]. This type of network architecture is suitable for deployment in agricultural picking robots, helping to reduce their dependence on high-power GPUs or the cloud, thus saving operational energy and costs. In addition, the maintenance cost of picking robots can be controlled by introducing AI-based fault prediction and diagnosis mechanisms to monitor the device operational status and predict potential failures in real time, thus extending the device lifecycle and reducing unplanned downtime.

5. Challenges and Future Trends

The rapid development of deep learning technology has injected new vitality into the research of fruit and vegetable picking robots; however, its application to real-world scenarios still faces many technical challenges. The complex and variable agricultural environment puts forward strict requirements on the robustness and generalization ability of sensing algorithms; the differences in heterogeneous sensor data from multiple sources exacerbate the difficulty of multimodal information fusion; the scarcity of agricultural scene samples contradicts the demand for deep learning data; and the key technologies such as semantic comprehension of the agricultural context and human–robot collaboration are not yet mature. For example, in the multi-vehicle cooperative fruit and vegetable picking task, some studies try to introduce a comprehensive multi-population genetic algorithm for task allocation, but in the real drift scenario, the lack of environment modeling and multi-objective trade-off imbalance results in the algorithm frequently falling into the local optimum in the high occlusion and fruit and vegetable dense area, which makes it difficult to achieve efficient operation [121]. In the face of the above challenges, the future development of fruit and vegetable picking robots needs to seek transformative breakthroughs in perception, integration, learning, planning, and interaction. By exploring the self-supervised learning paradigm to enhance the perception robustness, optimizing the multi-sensor cooperative perception framework to improve the quality of information fusion, exploring the new frontier of small-sample learning to ease the data demand, developing data-driven planning methods for dynamic environments to enhance the decision-making intelligence, and constructing an intelligent interaction mode in the context of agriculture to promote human–machine collaboration, we can encourage the fruit and vegetable picking robots to make leapfrog progress for the transformation of intelligent agriculture and empower the future development of fruit and vegetable picking robots.

5.1. Robust Perception Algorithms for Complex Environments

The research focus of future fruit and vegetable picking robot perception algorithms will be to improve the robustness and adaptability in complex and dynamic farmland environments. On the one hand, self-supervised learning, continuous learning, and other paradigms should be explored in depth, and unstructured environmental data should be used to learn more generalized visual feature expressions to reduce dependence on large-scale labeled data. On the other hand, combining the knowledge of the agricultural field to build a finer a priori model of the scene guides the deep learning algorithm to adapt to different farmland environments quickly. At the same time, the active learning strategy is used to optimize the sampling process dynamically, focus on complex cases for key learning, and continuously improve the environmental adaptation ability of the perception algorithm. For example, in the case of perceptual failure under hazy weather in an orchard, it was found that introducing an active sampling method based on a confidence regression mechanism effectively reduced perceptual errors [49].

5.2. Deep Multimodal Information Fusion Framework

Efficient fusion of visual, depth, near-infrared, and other multi-source sensory information is the key to improving the ability to understand fruit and vegetable picking robots. Future research should focus on constructing a unified deep multimodal fusion framework, realizing the entire interaction of multimodal data in the feature learning stage and capturing the intrinsic semantic association between different modalities. An attention mechanism should be introduced to dynamically adjust the importance of various modalities and suppress redundant information. At the decision-making level, it is appropriate to use probabilistic graphical modeling and other methods for global inference of multimodal information to obtain semantically consistent and robust fusion representations. Global reasoning, such as Bayesian graphical models, can be used at the decision-making level. For example, in an existing citrus picking experiment, a multimodal fusion approach based on graph neural networks reduced the false recognition rate by 18% in the occluded region compared to the traditional weighted average strategy, demonstrating its advantages in coping with uncertainty in agricultural environments [122]. Considering the specificity of agricultural scenarios, the framework should be fault-tolerant to incomplete and noisy inputs to ensure the continuity and stability of the sensing process.

5.3. New Paradigm for Small-Sample Learning in Agriculture

The development of a new paradigm of small-sample learning for agricultural applications is a meaningful way to break through the bottleneck of rapid deployment of fruit and vegetable picking robots. Meta-learning is a promising idea that acquires task-independent prior knowledge through a two-layer optimization process and then quickly adapts to new tasks according to the support set. Combining meta-learning with agricultural scenarios, constructing a priori libraries of farmland environments, and using a small number of target fruit and vegetable samples to fine-tune and form an exclusive model of picking are expected to improve the generalization of perception algorithms greatly. Knowledge distillation is another direction worth exploring. It can condense the feature representations learned from different fruit and vegetable picking models into a shared teacher network and then distil them into a small student network to improve reasoning efficiency while alleviating the data demand.

5.4. Data-Driven Autonomous Operation Planning

New data-driven job planning methods are urgently needed for the farmland’s complex and changing environment. Deep reinforcement learning (DRL) shows great potential for learning optimal harvesting strategies. By designing a reasonable reward function and embedding safety constraints, the intelligent body can autonomously learn obstacle avoidance path planning and picking action sequence optimization skills in the simulation environment and then transfer them to the actual robot. At the same time, the modeling capability should be further strengthened, and the simulation environment should be continuously optimized by using accurate data to narrow the gap between reality and simulation. At present, we still face the problem of the “simulation-reality migration gap”. Taking the multi-vehicle task assignment as an example, in the drifting orchard environment, there is “right-of-way conflict” and “path duplication” when using the reinforcement learning strategy to cooperate with multi-vehicles, resulting in local congestion; so, we should further strengthen the modeling ability, use real data to continuously optimize the simulation environment, and reduce the gap between reality and simulation. Therefore, the modeling capability should be further strengthened, and the simulation environment should be continuously optimized with the measured data to narrow the gap between reality and simulation. At the execution level, the planning system should be able to process dynamic feedback in real time, dynamically adjust the planning path according to the growth state of fruits and vegetables, the location of obstacles, etc., and ensure the robustness of autonomous operation with intelligent adaptability.

5.5. New Mode of Human–Computer Interaction in Intelligent Agriculture

Accompanied by the rapid development of natural language processing, knowledge mapping and other artificial intelligence technologies, fruit and vegetable picking robot human–machine interaction is expected to move towards a more intelligent and natural new stage. In the future, we should focus on the specific semantic understanding of agriculture, integrate the knowledge base of the agricultural field with the robot execution model, and realize the accurate recognition and mapping of agricultural terms. In terms of visualization, we should explore introducing knowledge mapping, virtual reality, and other technologies into the design of human–machine interfaces to provide users with a more intuitive image of multi-dimensional information presentation. At the level of human–robot collaboration and research on the active interaction paradigm, the robot actively seeks human help and dynamically adjusts the original strategy according to the feedback.

Author Contributions

Y.T. proposed research topics, designed article frameworks, and drafted articles; X.L. and J.Z. revised the article and ultimately completed it; Y.W. and Y.H. collected and organized data and research, organized literature, and prepared tables. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program Project (CN)2023YFD2201805, the Beijing Smart Agriculture Innovation Consortium Project (CN)BAIC10-2024, and the Beijing Science and Technology Plan (CN)Z231100003923005.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to acknowledge the anonymous reviewers for their valuable comments and members of the editorial team for their careful proofing.

Conflicts of Interest

The authors declared that there are no conflicts of interest.

References

  1. Xiao, F.; Wang, H.; Li, Y.; Cao, Y.; Lv, X.; Xu, G. Object Detection and Recognition Techniques Based on Digital Image Processing and Traditional Machine Learning for Fruit and Vegetable Harvesting Robots: An Overview and Review. Agronomy 2023, 13, 639. [Google Scholar] [CrossRef]
  2. Xiong, Y.; Ge, Y.; Grimstad, L.; From, P.J. An autonomous strawberry-harvesting robot: Design, development, integration, and field evaluation. J. Field Robot. 2019, 37, 202–224. [Google Scholar] [CrossRef]
  3. Khosro Anjom, F.; Vougioukas, S.G.; Slaughter, D.C. Development of a linear mixed model to predict the picking time in strawberry harvesting processes. Biosyst. Eng. 2018, 166, 76–89. [Google Scholar] [CrossRef]
  4. Zhang, K.; Lammers, K.; Chu, P.; Li, Z.; Lu, R. System design and control of an apple harvesting robot. Mechatronics 2021, 79, 102644. [Google Scholar] [CrossRef]
  5. Xie, H.; Zhang, D.; Yang, L.; Cui, T.; He, X.; Zhang, K.; Zhang, Z. Development, Integration, and Field Evaluation of a Dual-Arm Ridge Cultivation Strawberry Autonomous Harvesting Robot. J. Field Robot. 2024; early view. [Google Scholar] [CrossRef]
  6. Bac, C.W.; van Henten, E.J.; Hemming, J.; Edan, Y. Harvesting Robots for High-value Crops: State-of-the-art Review and Challenges Ahead. J. Field Robot. 2014, 31, 888–911. [Google Scholar] [CrossRef]
  7. Girshick, R.B.; Donahue, J.; Darrell, T.; Malik, J.; Berkeley, U. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014. [Google Scholar]
  8. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A.; Recognition, P. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar] [CrossRef]
  9. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  10. Sa, I.; Ge, Z.; Dayoub, F.; Upcroft, B.; Perez, T.; McCool, C. DeepFruits: A Fruit Detection System Using Deep Neural Networks. Sensors 2016, 16, 1222. [Google Scholar] [CrossRef] [PubMed]
  11. Liu, Y.; Gao, P.; Zheng, C.; Tian, L.; Tian, Y. A Deep Reinforcement Learning Strategy Combining Expert Experience Guidance for a Fruit-Picking Manipulator. Electronics 2022, 11, 311. [Google Scholar] [CrossRef]
  12. Frid-Adar, M.; Klang, E.; Amitai, M.M.; Goldberger, J.; Greenspan, H. Synthetic data augmentation using GAN for improved liver lesion classification. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; Volume 83, pp. 289–293. [Google Scholar] [CrossRef]
  13. Talaei Khoei, T.; Ould Slimane, H.; Kaabouch, N. Deep learning: Systematic review, models, challenges, and research directions. Neural Comput. Appl. 2023, 35, 23103–23124. [Google Scholar] [CrossRef]
  14. Espinoza, S.; Aguilera, C.; Rojas, L.; Campos, P.G. Analysis of Fruit Images With Deep Learning: A Systematic Literature Review and Future Directions. IEEE Access 2024, 12, 3837–3859. [Google Scholar] [CrossRef]
  15. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  16. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  17. Vargas-Hakim, G.-A.; Mezura-Montes, E.; Acosta-Mesa, H.-G. A Review on Convolutional Neural Network Encodings for Neuroevolution. IEEE Trans. Evol. Comput. 2022, 26, 12–27. [Google Scholar] [CrossRef]
  18. Yao, G.; Lei, T.; Zhong, J. A review of Convolutional-Neural-Network-based action recognition. Pattern Recognit. Lett. 2019, 118, 14–22. [Google Scholar] [CrossRef]
  19. Chen, F.; Li, S.; Han, J.; Ren, F.; Yang, Z. Review of Lightweight Deep Convolutional Neural Networks. Arch. Comput. Methods Eng. 2023, 31, 1915–1937. [Google Scholar] [CrossRef]
  20. Yu, Y.; Zhang, K.; Yang, L.; Zhang, D. Fruit detection for strawberry harvesting robot in non-structural environment based on Mask-RCNN. Comput. Electron. Agric. 2019, 163, 104846. [Google Scholar] [CrossRef]
  21. Ge, Y.; From, P.J.; Xiong, Y. Multi-view gripper internal sensing for the regression of strawberry ripeness using a mini-convolutional neural network for robotic harvesting. Comput. Electron. Agric. 2024, 216, 108474. [Google Scholar] [CrossRef]
  22. Coll-Ribes, G.; Torres-Rodríguez, I.J.; Grau, A.; Guerra, E.; Sanfeliu, A. Accurate detection and depth estimation of table grapes and peduncles for robot harvesting, combining monocular depth estimation and CNN methods. Comput. Electron. Agric. 2023, 215, 108362. [Google Scholar] [CrossRef]
  23. Zeeshan, S.; Aized, T.; Riaz, F. The Design and Evaluation of an Orange-Fruit Detection Model in a Dynamic Environment Using a Convolutional Neural Network. Sustainability 2023, 15, 4329. [Google Scholar] [CrossRef]
  24. Attri, I.; Awasthi, L.K.; Sharma, T.P.; Rathee, P. A review of deep learning techniques used in agriculture. Ecol. Inform. 2023, 77, 102217. [Google Scholar] [CrossRef]
  25. Simonyan, A.Z.K. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations, Ababa, Ethiopia, 24–28 April 2025; p. 1556. [Google Scholar]
  26. Yang, L.; Yu, X.; Zhang, S.; Long, H.; Zhang, H.; Xu, S.; Liao, Y. GoogLeNet based on residual network and attention mechanism identification of rice leaf diseases. Comput. Electron. Agric. 2023, 204, 107543. [Google Scholar] [CrossRef]
  27. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5987–5995. [Google Scholar]
  28. Tong, K.; Wu, Y.; Zhou, F. Recent advances in small object detection based on deep learning: A review. Image Vis. Comput. 2020, 97, 103910. [Google Scholar] [CrossRef]
  29. Xiuling, Z.; Huijuan, W.; Yu, S.; Gang, C.; Suhua, Z.; Quanbo, Y. Starting from the structure: A review of small object detection based on deep learning. Image Vis. Comput. 2024, 146, 105054. [Google Scholar] [CrossRef]
  30. Sumit; Shrishti, B.; Sunita, J.; Urvi, R. Comprehensive Review of R-CNN and its Variant Architectures. Int. Res. J. Adv. Eng. Hub 2024, 2, 959–966. [Google Scholar] [CrossRef]
  31. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  32. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Washington, DC, USA, 6 June 2016; pp. 1137–1149. [Google Scholar] [CrossRef]
  33. Lin, T.-Y.; Dollar, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  34. Sirisha, U.; Praveen, S.P.; Srinivasu, P.N.; Barsocchi, P.; Bhoi, A.K. Statistical Analysis of Design Aspects of Various YOLO-Based Deep Learning Models for Object Detection. Int. J. Comput. Intell. Syst. 2023, 16, 126. [Google Scholar] [CrossRef]
  35. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, Y. SSD: Single Shot MultiBox Detector. In Proceedings of the Computer Vision—ECCV, Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar] [CrossRef]
  36. Rong, J.; Wang, P.; Yang, Q.; Huang, F. A Field-Tested Harvesting Robot for Oyster Mushroom in Greenhouse. Agronomy 2021, 11, 1210. [Google Scholar] [CrossRef]
  37. Zhong, M.; Han, R.; Liu, Y.; Huang, B.; Chai, X.; Liu, Y. Development, integration, and field evaluation of an autonomous Agaricus bisporus picking robot. Comput. Electron. Agric. 2024, 220, 108871. [Google Scholar] [CrossRef]
  38. Liu, T.; Kang, H.; Chen, C. ORB-Livox: A real-time dynamic system for fruit detection and localization. Comput. Electron. Agric. 2023, 209, 107834. [Google Scholar] [CrossRef]
  39. Liu, J.; Liu, Z. YOLOv5s-BC: An improved YOLOv5s-based method for real-time apple detection. J. Real Time Image Process. 2024, 21, 88. [Google Scholar] [CrossRef]
  40. Le, N.; Rathour, V.S.; Yamazaki, K.; Luu, K.; Savvides, M. Deep reinforcement learning in computer vision: A comprehensive survey. Artif. Intell. Rev. 2021, 55, 2733–2819. [Google Scholar] [CrossRef]
  41. Arulkumaran, K.; Deisenroth, M.P.; Brundage, M.; Bharath, A.A. Deep Reinforcement Learning: A Brief Survey. IEEE Signal Process. Mag. 2017, 34, 26–38. [Google Scholar] [CrossRef]
  42. Mnih, V.; Kavukcuoglu, K.; Silver, D.; Rusu, A.A.; Veness, J.; Bellemare, M.G.; Graves, A.; Riedmiller, M.; Fidjeland, A.K.; Ostrovski, G.; et al. Human-level control through deep reinforcement learning. Nature 2015, 518, 529–533. [Google Scholar] [CrossRef] [PubMed]
  43. Lillicrap, T.P.; Hunt, J.J.; Pritzel, A.; Heess, N.M.O.; Erez, T.; Tassa, Y.; Silver, D.; Wierstra, D.J.C. Continuous control with deep reinforcement learning. arXiv 2015, arXiv:1509.02971. [Google Scholar]
  44. Haarnoja, T.; Zhou, A.; Hartikainen, K.; Tucker, G.; Ha, S.; Tan, J.; Kumar, V.; Zhu, H.; Gupta, A.; Abbeel, P.; et al. Soft Actor-Critic Algorithms and Applications. arXiv 2018, arXiv:1812.05905. [Google Scholar]
  45. Li, Y.; Feng, Q.; Zhang, Y.; Peng, C.; Ma, Y.; Liu, C.; Ru, M.; Sun, J.; Zhao, C. Peduncle collision-free grasping based on deep reinforcement learning for tomato harvesting robot. Comput. Electron. Agric. 2024, 216, 108488. [Google Scholar] [CrossRef]
  46. Wang, Y.; He, Z.; Cao, D.; Ma, L.; Li, K.; Jia, L.; Cui, Y. Coverage path planning for kiwifruit picking robots based on deep reinforcement learning. Comput. Electron. Agric. 2023, 205, 107593. [Google Scholar] [CrossRef]
  47. Lin, G.; Zhu, L.; Li, J.; Zou, X.; Tang, Y. Collision-free path planning for a guava-harvesting robot based on recurrent deep reinforcement learning. Comput. Electron. Agric. 2021, 188, 106350. [Google Scholar] [CrossRef]
  48. Li, Y.; Feng, Q.; Zhang, Y.; Peng, C.; Zhao, C. Intermittent Stop-Move Motion Planning for Dual-Arm Tomato Harvesting Robot in Greenhouse Based on Deep Reinforcement Learning. Biomimetics 2024, 9, 105. [Google Scholar] [CrossRef]
  49. Yang, J.; Ni, J.; Li, Y.; Wen, J.; Chen, D. The Intelligent Path Planning System of Agricultural Robot via Reinforcement Learning. Sensors 2022, 22, 4316. [Google Scholar] [CrossRef]
  50. Yi, Z.; Chang, T.; Li, S.; Liu, R.; Zhang, J.; Hao, A. Scene-Aware Deep Networks for Semantic Segmentation of Images. IEEE Access 2019, 7, 69184–69193. [Google Scholar] [CrossRef]
  51. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar] [CrossRef]
  52. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. In Proceedings of the IEEE Transactions on Pattern Analysis and Machine Intelligence, Seoul, Republic of Korea, 11–12 June 2017; pp. 2481–2495. [Google Scholar] [CrossRef]
  53. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Lecture Notes in Computer Science. pp. 234–241. [Google Scholar]
  54. Chen, L.-C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef] [PubMed]
  55. Li, Q.; Jia, W.; Sun, M.; Hou, S.; Zheng, Y. A novel green apple segmentation algorithm based on ensemble U-Net under complex orchard environment. Comput. Electron. Agric. 2021, 180, 105900. [Google Scholar] [CrossRef]
  56. Rasheed, M.; Jasim, W.M.; Farhan, R. Enhancing robotic grasping with attention mechanism and advanced UNet architectures in generative grasping convolutional neural networks. Alex. Eng. J. 2024, 102, 149–158. [Google Scholar] [CrossRef]
  57. Peng, Y.; Zhao, S.; Liu, J. Segmentation of Overlapping Grape Clusters Based on the Depth Region Growing Method. Electronics 2021, 10, 2813. [Google Scholar] [CrossRef]
  58. Kang, S.; Li, D.; Li, B.; Zhu, J.; Long, S.; Wang, J. Maturity identification and category determination method of broccoli based on semantic segmentation models. Comput. Electron. Agric. 2024, 217, 108633. [Google Scholar] [CrossRef]
  59. Zhang, P.; Dai, N.; Liu, X.; Yuan, J.; Xin, Z. A novel lightweight model HGCA-YOLO: Application to recognition of invisible spears for white asparagus robotic harvesting. Comput. Electron. Agric. 2024, 220, 108852. [Google Scholar] [CrossRef]
  60. Zhang, P.; Liu, X.; Yuan, J.; Liu, C. YOLO5-spear: A robust and real-time spear tips locator by improving image augmentation and lightweight network for selective harvesting robot of white asparagus. Biosyst. Eng. 2022, 218, 43–61. [Google Scholar] [CrossRef]
  61. Xiao, X.; Jiang, Y.; Wang, Y. A Method of Robot Picking Citrus Based on 3D Detection. IEEE Instrum. Meas. Mag. 2024, 27, 50–58. [Google Scholar] [CrossRef]
  62. Chai, J.J.K.; Xu, J.-L.; O’Sullivan, C. Real-Time Detection of Strawberry Ripeness Using Augmented Reality and Deep Learning. Sensors 2023, 23, 7639. [Google Scholar] [CrossRef]
  63. Lee, G.; Yonrith, P.; Yeo, D.; Hong, A. Enhancing detection performance for robotic harvesting systems through RandAugment. Eng. Appl. Artif. Intell. 2023, 123, 106445. [Google Scholar] [CrossRef]
  64. Škrabánek, P.; Doležel, P.; Matoušek, R. RGB images-driven recognition of grapevine varieties using a densely connected convolutional network. Log. J. IGPL 2023, 31, 618–633. [Google Scholar] [CrossRef]
  65. Qi, C.; Gao, J.; Pearson, S.; Harman, H.; Chen, K.; Shu, L. Tea chrysanthemum detection under unstructured environments using the TC-YOLO model. Expert Syst. Appl. 2022, 193, 116473. [Google Scholar] [CrossRef]
  66. Bresilla, K.; Perulli, G.D.; Boini, A.; Morandi, B.; Corelli Grappadelli, L.; Manfrini, L. Single-Shot Convolution Neural Networks for Real-Time Fruit Detection Within the Tree. Front. Plant Sci. 2019, 10, 611. [Google Scholar] [CrossRef]
  67. Kang, H.; Zhou, H.; Wang, X.; Chen, C. Real-Time Fruit Recognition and Grasping Estimation for Robotic Apple Harvesting. Sensors 2020, 20, 5670. [Google Scholar] [CrossRef]
  68. Fujinaga, T. Strawberries recognition and cutting point detection for fruit harvesting and truss pruning. Precis. Agric. 2024, 25, 1262–1283. [Google Scholar] [CrossRef]
  69. Lemsalu, M.; Bloch, V.; Backman, J.; Pastell, M. Real-Time CNN-based Computer Vision System for Open-Field Strawberry Harvesting Robot. IFAC Pap. 2022, 55, 24–29. [Google Scholar] [CrossRef]
  70. Li, Y.; Wang, W.; Guo, X.; Wang, X.; Liu, Y.; Wang, D. Recognition and Positioning of Strawberries Based on Improved YOLOv7 and RGB-D Sensing. Agriculture 2024, 14, 624. [Google Scholar] [CrossRef]
  71. Wang, J.; Zhang, Z.; Luo, L.; Wei, H.; Wang, W.; Chen, M.; Luo, S. DualSeg: Fusing transformer and CNN structure for image segmentation in complex vineyard environment. Comput. Electron. Agric. 2023, 206, 107682. [Google Scholar] [CrossRef]
  72. Li, Y.; Feng, Q.; Liu, C.; Xiong, Z.; Sun, Y.; Xie, F.; Li, T.; Zhao, C. MTA-YOLACT: Multitask-aware network on fruit bunch identification for cherry tomato robotic harvesting. Eur. J. Agron. 2023, 146, 126821. [Google Scholar] [CrossRef]
  73. Jing, X.; Jiang, H.; Niu, S.; Zhang, H.; Gilbert Murengami, B.; Wu, Z.; Li, R.; Zhou, C.; Ye, H.; Chen, J.; et al. End-to-end stereo matching network with two-stage partition filtering for full-resolution depth estimation and precise localization of kiwifruit for robotic harvesting. Comput. Electron. Agric. 2024, 225, 109333. [Google Scholar] [CrossRef]
  74. Kim, T.; Lee, D.-H.; Kim, K.-C.; Kim, Y.-J. 2D pose estimation of multiple tomato fruit-bearing systems for robotic harvesting. Comput. Electron. Agric. 2023, 211, 108004. [Google Scholar] [CrossRef]
  75. Miao, Z.; Yu, X.; Li, N.; Zhang, Z.; He, C.; Li, Z.; Deng, C.; Sun, T. Efficient tomato harvesting robot based on image processing and deep learning. Precis. Agric. 2022, 24, 254–287. [Google Scholar] [CrossRef]
  76. Chen, M.; Chen, Z.; Luo, L.; Tang, Y.; Cheng, J.; Wei, H.; Wang, J. Dynamic visual servo control methods for continuous operation of a fruit harvesting robot working throughout an orchard. Comput. Electron. Agric. 2024, 219, 108774. [Google Scholar] [CrossRef]
  77. Wang, X.; Kang, H.; Zhou, H.; Au, W.; Chen, C. Geometry-aware fruit grasping estimation for robotic harvesting in apple orchards. Comput. Electron. Agric. 2022, 193, 106716. [Google Scholar] [CrossRef]
  78. Xiong, Y.; Ge, Y.; From, P.J. An obstacle separation method for robotic picking of fruits in clusters. Comput. Electron. Agric. 2020, 175, 105397. [Google Scholar] [CrossRef]
  79. Kok, E.; Chen, C. Occluded apples orientation estimator based on deep learning model for robotic harvesting. Comput. Electron. Agric. 2024, 219, 108781. [Google Scholar] [CrossRef]
  80. Li, Z.; Shi, N.; Zhao, L.; Zhang, M. Deep reinforcement learning path planning and task allocation for multi-robot collaboration. Alex. Eng. J. 2024, 109, 408–423. [Google Scholar] [CrossRef]
  81. Martini, M.; Eirale, A.; Cerrato, S.; Chiaberge, M. PIC4rl-gym: A ROS2 Modular Framework for Robots Autonomous Navigation with Deep Reinforcement Learning. In Proceedings of the 2023 3rd International Conference on Computer, Control and Robotics, Zhangjiajie, China, 24–26 March 2023; pp. 198–202. [Google Scholar] [CrossRef]
  82. Li, T.; Xie, F.; Zhao, Z.; Zhao, H.; Guo, X.; Feng, Q. A multi-arm robot system for efficient apple harvesting: Perception, task plan and control. Comput. Electron. Agric. 2023, 211, 107979. [Google Scholar] [CrossRef]
  83. Huang, R.; Zheng, W.; Zhang, B.; Zhou, J.; Cui, Z.; Zhang, Z. Deep learning with tactile sequences enables fruit recognition and force prediction for damage-free grasping. Comput. Electron. Agric. 2023, 211, 107985. [Google Scholar] [CrossRef]
  84. Ma, C.; Ying, Y.; Xie, L. Development of a visuo-tactile sensor for non-destructive peach firmness and contact force measurement suitable for robotic arm applications. Food Chem. 2025, 467, 142282. [Google Scholar] [CrossRef]
  85. Lin, J.; Hu, Q.; Xia, J.; Zhao, L.; Du, X.; Li, S.; Chen, Y.; Wang, X. Non-destructive fruit firmness evaluation using a soft gripper and vision-based tactile sensing. Comput. Electron. Agric. 2023, 214, 108256. [Google Scholar] [CrossRef]
  86. Li, S.; Sun, W.; Liang, Q.; Liu, C.; Liu, J. Assessing fruit hardness in robot hands using electric gripper actuators with tactile sensors. Sens. Actuators A Phys. 2024, 365, 114843. [Google Scholar] [CrossRef]
  87. Han, Y.; Yu, K.; Batra, R.; Boyd, N.; Mehta, C.; Zhao, T.; She, Y.; Hutchinson, S.; Zhao, Y. Learning Generalizable Vision-Tactile Robotic Grasping Strategy for Deformable Objects via Transformer. IEEE/ASME Trans. Mechatron. 2024, 34, 554–566. [Google Scholar] [CrossRef]
  88. Zhang, T.; Huang, Z.; You, W.; Lin, J.; Tang, X.; Huang, H. An Autonomous Fruit and Vegetable Harvester with a Low-Cost Gripper Using a 3D Sensor. Sensors 2019, 20, 93. [Google Scholar] [CrossRef] [PubMed]
  89. Ma, L.; He, Z.; Zhu, Y.; Jia, L.; Wang, Y.; Ding, X.; Cui, Y. A Method of Grasping Detection for Kiwifruit Harvesting Robot Based on Deep Learning. Agronomy 2022, 12, 3096. [Google Scholar] [CrossRef]
  90. Sun, Q.; Zhong, M.; Chai, X.; Zeng, Z.; Yin, H.; Zhou, G.; Sun, T. Citrus pose estimation from an RGB image for automated harvesting. Comput. Electron. Agric. 2023, 211, 108022. [Google Scholar] [CrossRef]
  91. Wang, Y.; Wu, H.; Zhu, Z.; Ye, Y.; Qian, M. Continuous picking of yellow peaches with recognition and collision-free path. Comput. Electron. Agric. 2023, 214, 108273. [Google Scholar] [CrossRef]
  92. Zhang, H.; Li, X.; Wang, L.; Liu, D.; Wang, S. Construction and Optimization of a Collaborative Harvesting System for Multiple Robotic Arms and an End-Picker in a Trellised Pear Orchard Environment. Agronomy 2023, 14, 80. [Google Scholar] [CrossRef]
  93. Ni, P.; Zhang, W.; Bai, W.; Lin, M.; Cao, Q. A New Approach Based on Two-stream CNNs for Novel Objects Grasping in Clutter. J. Intell. Robot. Syst. 2018, 94, 161–177. [Google Scholar] [CrossRef]
  94. Tabakis, I.-M.; Dasygenis, M. Deep Reinforcement Learning-Based Path Planning for Dynamic and Heterogeneous Environments. In Proceedings of the 2024 Panhellenic Conference on Electronics & Telecommunications (PACET), Thessaloniki, Greece, 28–29 March 2024; pp. 1–4. [Google Scholar]
  95. Lin, J. Path planning based on reinforcement learning. Appl. Comput. Eng. 2023, 5, 853–858. [Google Scholar] [CrossRef]
  96. Choi, J.; Lee, G.; Lee, C. Reinforcement learning-based dynamic obstacle avoidance and integration of path planning. Intell. Serv. Robot. 2021, 14, 663–677. [Google Scholar] [CrossRef] [PubMed]
  97. Zhao, W.; Queralta, J.P.; Li, Q.; Westerlund, T.; Engineering, A. Towards Closing the Sim-to-Real Gap in Collaborative Multi-Robot Deep Reinforcement Learning. In Proceedings of the 2020 5th International Conference on Robotics and Automation Engineering (ICRAE), Singapore, 20–22 November 2020; pp. 7–12. [Google Scholar] [CrossRef]
  98. Ao, T.; Zhang, K.; Shi, H.; Jin, Z.; Zhou, Y.; Liu, F. Energy-Efficient Multi-UAVs Cooperative Trajectory Optimization for Communication Coverage: An MADRL Approach. Remote Sens. 2023, 15, 429. [Google Scholar] [CrossRef]
  99. Wu, B.; Suh, S.J.A. Deep Reinforcement Learning for Decentralized Multi-Robot Control: A DQN Approach to Robustness and Information Integration. arXiv 2024, arXiv:2408.11339. [Google Scholar] [CrossRef]
  100. Xiao, Y.; Xiao, L.; Wan, K.; Yang, H.; Zhang, Y.; Wu, Y.; Zhang, Y. Reinforcement Learning Based Energy-Efficient Collaborative Inference for Mobile Edge Computing. IEEE Trans. Commun. 2023, 71, 864–876. [Google Scholar] [CrossRef]
  101. Hwang, H.-J.; Cho, J.-H.; Kim, Y.-T. Deep Learning-Based Real-Time 6D Pose Estimation and Multi-Mode Tracking Algorithms for Citrus-Harvesting Robots. Machines 2024, 12, 642. [Google Scholar] [CrossRef]
  102. Zhang, L.; Jia, J.; Gui, G.; Hao, X.; Gao, W.; Wang, M. Deep Learning Based Improved Classification System for Designing Tomato Harvesting Robot. IEEE Access 2018, 6, 67940–67950. [Google Scholar] [CrossRef]
  103. Qi, C.; Gao, J.; Chen, K.; Shu, L.; Pearson, S. Tea Chrysanthemum Detection by Leveraging Generative Adversarial Networks and Edge Computing. Front. Plant Sci. 2022, 13, 850606. [Google Scholar] [CrossRef] [PubMed]
  104. Sa, I.; Lim, J.Y.; Ahn, H.S.; MacDonald, B. deepNIR: Datasets for Generating Synthetic NIR Images and Improved Fruit Detection System Using Deep Learning Techniques. Sensors 2022, 22, 4721. [Google Scholar] [CrossRef]
  105. Dai, Y.; Zhao, P.; Wang, Y. Maturity discrimination of tobacco leaves for tobacco harvesting robots based on a Multi-Scale branch attention neural network. Comput. Electron. Agric. 2024, 224, 109133. [Google Scholar] [CrossRef]
  106. Kim, J.; Pyo, H.; Jang, I.; Kang, J.; Ju, B.; Ko, K. Tomato harvesting robotic system based on Deep-ToMaToS: Deep learning network using transformation loss for 6D pose estimation of maturity classified tomatoes with side-stem. Comput. Electron. Agric. 2022, 201, 107300. [Google Scholar] [CrossRef]
  107. Wang, W.; Shi, Y.; Liu, W.; Che, Z. An Unstructured Orchard Grape Detection Method Utilizing YOLOv5s. Agriculture 2024, 14, 262. [Google Scholar] [CrossRef]
  108. Bargoti, S.; Underwood, J. Deep Fruit Detection in Orchards. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 3626–3633. [Google Scholar] [CrossRef]
  109. Károly, A.I.; Tirczka, S.; Gao, H.; Rudas, I.J.; Galambos, P. Increasing the Robustness of Deep Learning Models for Object Segmentation: A Framework for Blending Automatically Annotated Real and Synthetic Data. IEEE Trans. Cybern. 2024, 54, 25–38. [Google Scholar] [CrossRef]
  110. Nguyen, H.-T.; Cheah, C.C.; Toh, K.-A. An analytic layer-wise deep learning framework with applications to robotics. Automatica 2022, 135, 110007. [Google Scholar] [CrossRef]
  111. Song, C.; Wang, K.; Wang, C.; Tian, Y.; Wei, X.; Li, C.; An, Q.; Song, J. TDPPL-Net: A Lightweight Real-Time Tomato Detection and Picking Point Localization Model for Harvesting Robots. IEEE Access 2023, 11, 37650–37664. [Google Scholar] [CrossRef]
  112. Ayranci, A.A.; Erkmen, B. Edge Computing and Robotic Applications in Modern Agriculture. In Proceedings of the 2024 International Congress on Human-Computer Interaction, Optimization and Robotic Applications (HORA), Istanbul, Turkey, 23–25 May 2024; pp. 1–6. [Google Scholar]
  113. Zhang, X.; Cao, Z.; Dong, W. Overview of Edge Computing in the Agricultural Internet of Things: Key Technologies, Applications, Challenges. IEEE Access 2020, 8, 141748–141761. [Google Scholar] [CrossRef]
  114. Wang, C.; Zou, X.; Tang, Y.; Luo, L.; Feng, W. Localisation of litchi in an unstructured environment using binocular stereo vision. Biosyst. Eng. 2016, 145, 39–51. [Google Scholar] [CrossRef]
  115. Tituaña, L.; Gholami, A.; He, Z.; Xu, Y.; Karkee, M.; Ehsani, R. A small autonomous field robot for strawberry harvesting. Smart Agric. Technol. 2024, 8, 100454. [Google Scholar] [CrossRef]
  116. Chu, P.; Li, Z.; Lammers, K.; Lu, R.; Liu, X.J.A. DeepApple: Deep Learning-based Apple Detection using a Suppression Mask R-CNN. arXiv 2020, arXiv:2010.09870. [Google Scholar]
  117. Li, Z.; Wang, J.; Gao, G.; Lei, Y.; Zhao, C.; Wang, Y.; Bai, H.; Liu, Y.; Guo, X.; Li, Q. SGSNet: A lightweight deep learning model for strawberry growth stage detection. Front. Plant Sci. 2024, 15, 91706. [Google Scholar] [CrossRef]
  118. Su, Z.; Zhang, C.; Yan, T.; Zhu, J.; Zeng, Y.; Lu, X.; Gao, P.; Feng, L.; He, L.; Fan, L. Application of Hyperspectral Imaging for Maturity and Soluble Solids Content Determination of Strawberry with Deep Learning Approaches. Front. Plant Sci. 2021, 12, 736334. [Google Scholar] [CrossRef]
  119. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H.J.A. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. arXiv 2017, arXiv:1704.04861. [Google Scholar] [CrossRef]
  120. Tan, M.; Le, Q.V.J.A. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2019, arXiv:1905.11946. [Google Scholar] [CrossRef]
  121. Woehrle, T.; Sivakumar, A.N.V.; Uppalapati, N.K.; Chowdhary, G.J.A. MetaCropFollow: Few-Shot Adaptation with Meta-Learning for Under-Canopy Navigation. arXiv 2024, arXiv:2411.14092. [Google Scholar] [CrossRef]
  122. Ghadirzadeh, A.; Chen, X.; Poklukar, P.; Finn, C.; Björkman, M.; Kragic, D. Systems. Bayesian Meta-Learning for Few-Shot Policy Adaptation Across Robotic Platforms. In Proceedings of the IEEE/RJS International Conference on Intelligent Robots and Systems, Prague, Czech Republic, 27 September–1 October 2021; Volume 96, pp. 1274–1280. [Google Scholar] [CrossRef]
Figure 1. Convolutional neural network workflow diagram.
Figure 1. Convolutional neural network workflow diagram.
Sensors 25 03677 g001
Figure 2. Object detection workflow diagram.
Figure 2. Object detection workflow diagram.
Sensors 25 03677 g002
Figure 3. Deep reinforcement learning workflow diagram.
Figure 3. Deep reinforcement learning workflow diagram.
Sensors 25 03677 g003
Figure 4. Semantic segmentation workflow diagram.
Figure 4. Semantic segmentation workflow diagram.
Sensors 25 03677 g004
Table 1. Summary of key deep learning methods commonly used in picking robots.
Table 1. Summary of key deep learning methods commonly used in picking robots.
Method CategoryTypical NetworksMechanismAdvantageLimitation
Convolutional neural networkLeNet, AlexNet VGGNet, ResNetHierarchical feature extraction and representation learning of images through local connectivity, weight sharing, and spatial dimensionality reduction.Automatically learns rich features from a large amount of image data and performs well in tasks such as fruit detection and ripeness determination.Dependent on a large amount of labeled data; the amount of computation required in training and inference is huge, leading to high requirements for computational resources and storage space; it is susceptible to noise, occlusion, and adversarial attacks in feature learning.
Target detectionYOLO Series, SSD RCNN, FastR-CNNThrough end-to-end feature learning, candidate region generation, classification, and regression, the mapping from images to target bounding boxes and categories is realized.Quickly and accurately identifies and locates mature targets in complex agricultural production environments to support picking decisions and path planningGeneralization and robustness to be further optimized with respect to scene lighting issues, shooting angle, fruit shape, etc.
Deep reinforcement learningDQN, DDPG, SACApproximating the value function or policy function of reinforcement learning through deep neural networks, thus realizing end-to-end autonomous learning and decision making.Adaptive extraction of useful features in the face of complex, high-dimensional environments. Strong generalization ability, able to deal with problems such as continuous action space.Requires a large amount of environmental interaction data, slower training and learning process, easily affected by hyperparameters, initialization and other factors, poor model interpretability, limiting its application effect.
Semantic segmentationFCN, SegNet, U-Net, DeepLabEach pixel within the image is systematically classified to ascertain the specific region or category to which it belongs.It is capable of achieving extremely high target recognition and classification accuracy at the pixel level, accurately distinguishing between different target classes.Due to the reliance on extensive pixel-level sample data, the computational requirements are substantial, leading to suboptimal real-time performance and reduced efficiency in practical applications. For targets with small areas or low frequencies, issues such as inaccurate segmentation and inadequate handling of fuzzy boundaries arise.
Table 5. Summary of comparative analysis of key technologies based on deep learning in harvesting robots.
Table 5. Summary of comparative analysis of key technologies based on deep learning in harvesting robots.
Technical CategoryTechnology TypeMain AdvantagesMain DisadvantagesRepresentative Technologies (Frameworks/Methodologies)
Visual perception and target recognitionCNN-based Classification/DetectionStrong feature representation, adaptable to various fruitsLarge model size, slow inference, difficult to deployResNet, EfficientNet
YOLO Series Object DetectionFast detection, suitable for real-time systemsPoor accuracy for small/occluded targetsYOLOv5, YOLOv7, YOLO-NAS
Attention Mechanisms (Spatial/Channel)Focuses on key areas, improves recognition in cluttered scenesIncreased complexity, harder to trainCBAM, SE, Transformer Attention
Multimodal Fusion (RGB + D)Strong spatial understanding, better in occluded scenariosComplex calibration, high hardware costDenseFusion, PointNet, RGB-DNet
Self-Supervised/Few-Shot LearningLess dependence on annotations, fast adaptationWeak generalization, unstable trainingSimCLR, Meta-RCNN, FSL-YOLO
Path planning and motion controlDeep Reinforcement Learning (DRL)Learns adaptive paths and obstacle avoidanceSlow convergence, needs large-scale simulationPPO, DDPG, SAC, TD3
GNN + Reinforcement LearningLearns orchard topology, enables intelligent decision makingComplex architecture, hard to trainGNN-RL, GraphNav, GAT-RL
Imitation LearningEfficient training from expert demonstrationsLacks exploration, cannot self-correct errorsBehavior Cloning, DAgger
Deep Visual NavigationNo map needed, end-to-end instruction predictionLow fault tolerance, image quality dependentDeep Visual Planner, VNNet
Learning-Based SLAMSupports dynamic environment mapping and navigationComplex fusion, hard to deployDL-SLAM, Neural SLAM, DROID-SLAM
Intelligent control of end effectorDeep Learning-Based Force ControlAdjusts gripping force automatically, reduces damageRequires accurate sensors and dataDeepForce, GripNet, TactileCNN
Multimodal Fusion (Vision + Tactile + Force)High adaptability, robust graspingComplex system integration, real-time challengesVisuo-Tactile Fusion Net, GelSightNet
Reinforcement Learning for GraspingLearns grasping strategies for different shapesHigh training cost, low sample efficiencyQT-Opt, VPG, RL-GraspNet
Generative Models for Grasping (VAE/GAN)Predicts grasp strategies under incomplete observationsUnstable inference, hard to trainGraspGAN, VAE-Grasp
Graph-Based Grasping Strategy (GNN)Suitable for fruit clusters, captures structural featuresComplex modeling, limited real-world applicationGGCNN, GAT-Grasp
Table 6. The datasets and their processing methods used in some articles.
Table 6. The datasets and their processing methods used in some articles.
DatasetData VolumeData Acquisition MethodsData Processing Methods
White asparagus dataset6498 images (4248 non-rain enhanced + 2250 rain enhanced)Image acquisition was performed using an industrial camera (Basler acA2500-14gc, made in Basler AG, Ahrensburg, Germany) with an image resolution of 2540 × 1920 pixels.Multi-scale combined images, resampling-based image transformation
Apple orchards dataset768 RGB-D images, 1132 color imagesUsing the Intel RealSense D435 RGB-D camera (made in Intel Corporation, Santa Clara, CA, USA)Spatial conversion, color distortion, point cloud processing
Citrus Orchard Dataset10,000 imagesUsing the Intel RealSense D435 RGB-D Camera (made in Intel Corporation, Santa Clara, CA, USA)Using BlenderProc in conjunction with Blender to generate virtual datasets
Tomato Dataset200 imagesManual shootingGeometric transformations, random noise
Tea Chrysanthemum Dataset26,432 imagesUsing an Apple X phone (made in Apple Inc., Cupertino, CA, USA) with an image resolution of 1080 × 1920TC-GAN
Orange Dataset2000 imagesUsing the Azure Kinect RGB-D camera (made in Microsoft Corporation, Washington, DC, USA) with a resolution of 1920 × 1080Flip, rotate, and change lighting on images
Bell Pepper Dataset1615 pairs of RGB + NIR imagesUsing the multispectral camera acquisitionSynthesis of NIR image from RGB image using GAN
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tan, Y.; Liu, X.; Zhang, J.; Wang, Y.; Hu, Y. A Review of Research on Fruit and Vegetable Picking Robots Based on Deep Learning. Sensors 2025, 25, 3677. https://doi.org/10.3390/s25123677

AMA Style

Tan Y, Liu X, Zhang J, Wang Y, Hu Y. A Review of Research on Fruit and Vegetable Picking Robots Based on Deep Learning. Sensors. 2025; 25(12):3677. https://doi.org/10.3390/s25123677

Chicago/Turabian Style

Tan, Yarong, Xin Liu, Jinmeng Zhang, Yigang Wang, and Yanxiang Hu. 2025. "A Review of Research on Fruit and Vegetable Picking Robots Based on Deep Learning" Sensors 25, no. 12: 3677. https://doi.org/10.3390/s25123677

APA Style

Tan, Y., Liu, X., Zhang, J., Wang, Y., & Hu, Y. (2025). A Review of Research on Fruit and Vegetable Picking Robots Based on Deep Learning. Sensors, 25(12), 3677. https://doi.org/10.3390/s25123677

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop