Next Article in Journal
Grazing Reduces Field Bindweed Infestations in Perennial Warm-Season Grass Pastures
Previous Article in Journal
Structural Heterogeneity of Biochar Modulates’ Soil Hydraulic Properties and Nutrient Migration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Empowering Smart Soybean Farming with Deep Learning: Progress, Challenges, and Future Perspectives

1
School of Mechanical and Electrical Engineering, Huainan Normal University, Huainan 232038, China
2
School of Mechanical Engineering, Hefei University of Technology, Hefei 230009, China
3
Human-Computer Collaborative Robot Joint Laboratory of Anhui Province, Huainan 232038, China
4
College of Land Science and Technology, China Agricultural University, 17 Qinghua East Road, Haidian, Beijing 100083, China
5
International College Beijing, China Agricultural University, 17 Qinghua East Road, Haidian, Beijing 100083, China
6
Artie McFerrin Department of Chemical Engineering, Texas A&M University, College Station, TX 77843-3122, USA
7
Department of Crop and Soil Sciences, College of Agriculture and Environmental Sciences, University of Georgia, Tifton, GA 31793, USA
8
College of Engineering, China Agricultural University, Qinghua East Road No. 17, Haidian, Beijing 100083, China
*
Authors to whom correspondence should be addressed.
Agronomy 2025, 15(8), 1831; https://doi.org/10.3390/agronomy15081831
Submission received: 3 June 2025 / Revised: 23 July 2025 / Accepted: 26 July 2025 / Published: 28 July 2025
(This article belongs to the Section Precision and Digital Agriculture)

Abstract

This review comprehensively examines the application of deep learning technologies across the entire soybean production chain, encompassing areas such as disease and pest identification, weed detection, crop phenotype recognition, yield prediction, and intelligent operations. By systematically analyzing mainstream deep learning models, optimization strategies (e.g., model lightweighting, transfer learning), and sensor data fusion techniques, the review identifies their roles and performances in complex agricultural environments. It also highlights key challenges including data quality limitations, difficulties in real-world deployment, and the lack of standardized evaluation benchmarks. In response, promising directions such as reinforcement learning, self-supervised learning, interpretable AI, and multi-source data fusion are proposed. Specifically for soybean automation, future advancements are expected in areas such as high-precision disease and weed localization, real-time decision-making for variable-rate spraying and harvesting, and the integration of deep learning with robotics and edge computing to enable autonomous field operations. This review provides valuable insights and future prospects for promoting intelligent, efficient, and sustainable development in soybean production through deep learning.

1. Introduction

Soybean is one of the world’s primary sources of vegetable oil [1]. It possesses high nutritional value, being rich in protein and essential micronutrients beneficial to human health, such as potassium, magnesium, iron, and vitamins B and E. The unsaturated fatty acids present in soybeans also contribute positively to cardiovascular health [2], making soybean-derived products a common component of daily diets worldwide [3]. Beyond its nutritional merits, soybean serves diverse industrial purposes, including the production of hydraulic oil, plastics, and adhesives [4]. Moreover, it is regarded as a natural “gene reservoir,” playing a vital role in crop improvement [5]. Over the past few decades, global soybean production and demand have continued to rise steadily [6], and its applications have expanded across various fields, as illustrated in Figure 1.
As illustrated in Figure 1, global soybean consumption has surged from less than 50 million tonnes in 1961 to over 350 million tonnes in 2023—an increase of more than sevenfold. During this period, the composition of soybean usage has undergone a marked transformation. Processing applications (including animal feed, vegetable oil, and biofuel) have become the dominant use, accounting for approximately 75% of total consumption in 2023. Although the share of direct soybean consumption as food has declined, processing applications—including vegetable oil and animal feed production—ultimately support human and animal nutrition, thus remaining a key component of global soybean consumption. In contrast, the share of soybeans consumed directly as food for human consumption (e.g., tofu, soy milk) has declined to below 10%, and the proportion used directly as animal feed has also decreased. Since 2000, driven by biofuel policies and rapidly rising demand for meat in emerging markets, soybean processing has accelerated significantly.
With the continuous growth of the global population, agriculture faces mounting pressure to meet the rising demand for food production [7]. As a major cash crop, soybean is experiencing increasing market demand. Despite the rising dominance of industrial processing applications, soybeans continue to play a fundamental role in global food and feed systems. However, its yield potential is severely constrained by a range of biotic and abiotic factors. Among these, pests and diseases pose significant threats to soybean production in key growing regions [8,9]. To combat these challenges, significant efforts have been made in breeding disease-resistant and stress-tolerant soybean varieties, which have shown promising results in maintaining yield stability. In parallel, soybean has also become a key target crop in the advancement of precision agriculture technologies. Common diseases such as S c l e r o t i n i a stem rot, P h y t o p h t h o r a root rot, and soybean cyst nematode [10], along with newly emerging conditions like the stay-green syndrome [11], have been reported. In addition, insect vectors, such as soybean aphids and thrips, continue to spread these pathogens and disrupt production [12,13,14]. Abiotic stresses also play a critical role. Low temperatures can delay germination and reduce yields, while high temperatures may impair pollen viability, decrease pod set, and deteriorate grain quality, thus necessitating precise temperature regulation [15]. Drought stress reduces soybean antioxidant capacity, increases membrane lipid peroxidation, disrupts osmotic regulation, and ultimately hinders plant growth and yield formation [16]. Photoperiod is another influential factor, as soybean is a short-day plant and prolonged daylight delays flowering [17]. Nutrient management further complicates cultivation: excessive nitrogen application can inhibit nodule nitrogen fixation, phosphorus deficiency can directly reduce yields, and over-fertilization may lead to soil compaction [18]. Given these complex challenges, the adoption of advanced technologies is urgently needed to enhance soybean management efficiency and yield performance.
In recent years, the application of deep learning in agriculture has advanced rapidly, giving rise to a series of innovative breakthroughs. Compared with traditional machine learning methods, deep learning offers notable advantages [19,20,21]. Its ability to automatically extract features enables more accurate identification of crop images, while its end-to-end learning architecture facilitates hierarchical feature learning without relying heavily on manual feature engineering [22,23,24,25]. Moreover, techniques such as data augmentation help mitigate the impact of noise commonly present in agricultural datasets, thereby enhancing model robustness [26,27]. The adoption of transfer learning, particularly the use of pretrained models, has further enabled deep learning to maintain strong performance even with limited data, a scenario in which conventional approaches often struggle to generalize or adapt to the complex and dynamic nature of agricultural environments [28,29]. Importantly, deep learning has not only made significant progress in agriculture but has also achieved widespread success in remote sensing and other fields [30,31,32]. These cross-domain advancements provide a solid technological foundation for its continued development in agriculture. As a result, deep learning is expected to play a central role in driving the future of agriculture toward greater intelligence and precision.
In summary, soybean occupies a pivotal role in the global agricultural system. However, its production is frequently challenged by a range of biotic and abiotic stressors. In recent years, the widespread adoption of deep learning in agriculture has introduced novel solutions to address these complex issues. Leveraging deep learning to enable intelligent optimization across key stages of soybean production has emerged as a critical pathway toward agricultural modernization. To systematically review the research progress and technical landscape in this domain, the structure of this paper is arranged as follows: Section 2 illustrates the literature search and selection strategy based on PRISMA. Section 3 presents an overview of the soybean industry chain and analyzes the core demands driving its transition toward intelligent systems. Section 4 details the applications of deep learning in the soybean domain, focusing on three key areas: pest and disease diagnosis, growth status monitoring and phenotypic analysis, and weed detection and management. Section 5 examines the mainstream deep learning models used in soybean production, explores current model optimization strategies, and evaluates their adaptability across various application scenarios. Section 6 discusses the practical challenges facing current technologies and outlines potential future research directions and development trends. Section 7 concludes the paper with a summary and outlook, aiming to provide theoretical insights and technical guidance for subsequent studies in this field.

2. Literature Search and Selection Strategy Based on PRISMA

This review follows the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) methodology to ensure transparency and reproducibility in the literature selection process. A comprehensive search was conducted in four major scientific databases: Web of Science, Scopus, IEEE Xplore, and ScienceDirect. The search covered publications from 2010 to July 2025 and used combinations of keywords such as “soybean,” “deep learning,” “CNN,” “phenotyping,” “weed detection,” and “precision agriculture.”
Studies were included if they met the following criteria: (1) focused on the application of deep learning in any stage of soybean production; (2) peer-reviewed articles written in English; (3) provided technical details on models and datasets used. Exclusion criteria included the following: (1) non-English articles, (2) reviews unrelated to soybean or deep learning, and (3) conference abstracts without full texts.
A total of 1032 records were initially identified through four major databases. After removing 216 duplicates, 816 unique articles remained. Titles and abstracts of these 816 records were screened, resulting in the exclusion of 542 irrelevant or ineligible studies. The full texts of the remaining 274 articles were assessed for eligibility, and 76 were excluded due to reasons such as methodological insufficiency or lack of relevance to soybean deep learning applications. Ultimately, 156 studies were included in the final synthesis. The entire selection process is illustrated in Figure 2 following the PRISMA flow diagram.

3. Soybean Industry Chain and the Demand for Intelligent Development

The cultivation of soybean dates back to the earliest stages of Chinese agricultural civilization [33]. In response to growing global demands for increased agricultural productivity, all segments of the soybean industry chain are becoming increasingly reliant on intelligent technologies. As a complex system composed of interrelated stages—namely sowing, cultivation, field management, harvesting, and processing—each phase significantly influences the final yield and quality of soybeans.
The sowing stage marks the starting point of the soybean production cycle and serves as the foundation for subsequent growth and output. Optimizing sowing time is critical; early sowing may reduce plant density, whereas delayed sowing can increase protein content but potentially reduce yield. Regional variations in optimal sowing periods are largely governed by environmental conditions [34]. In seed selection, recent studies have explored the use of deep learning for more precise classification. For example, Huang et al. [35] proposed a lightweight deep learning method (SNet) that integrates Mask R-CNN with a mixed feature recalibration module to extract key features of healthy seeds effectively.
The cultivation stage is crucial for ensuring healthy crop growth. Deep learning models, such as CNNs and LSTMs, have been widely adopted to analyze multi-source data, including soil moisture and satellite meteorological images, to facilitate intelligent irrigation and variable-rate fertilization decisions [36]. Chandel et al. [37], for instance, found GoogLeNet to outperform other models in detecting water stress in soybean crops.
The field management stage is the longest and most technically complex part of the production cycle. The integration of deep learning with UAV technology has significantly improved the identification of pests and diseases in soybean fields, thereby reducing manual labor [38]. Lightweight CNN models have also been employed in weed detection systems to enable precision weeding [39]. Despite achieving high accuracy in disease recognition, weed detection, and yield prediction, current research still faces challenges such as limited data diversity and constrained model generalizability [40].
The harvesting stage is pivotal in determining the final output. Deep learning shows promising potential in maturity detection. For example, Zhang et al. [41] developed DS-SoybeanNet, which combines UAV remote sensing and deep learning, outperforming traditional machine learning in detection accuracy. Recent studies have also explored the integration of deep learning into robotic harvesting systems, using multiple sensors (such as monocular RGB cameras, LiDAR, and thermal imagers) to enhance precision and adaptability [42].
In the processing stage, deep learning surpasses traditional machine learning by automatically extracting complex features (e.g., texture, color, shape) from large-scale image or audio data without manual intervention [27,43]. Zhu et al. [44], for example, proposed a deep learning-based quantitative prediction model to assess soybean processing potential for soy milk, enabling intelligent evaluation and value enhancement of agricultural products.
At a broader level, smart agriculture aims to simulate human cognition and decision-making using artificial intelligence technologies such as deep learning to address complex agricultural challenges [45]. The core demands of intelligent agriculture can be summarized into three dimensions: high throughput, automation, and precision decision-making.
  • High-throughput agriculture relies on high-throughput phenotyping platforms that integrate multi-source sensors (e.g., UAVs, LiDAR, multispectral imaging) and AI algorithms to automate environmental data acquisition and analysis. Examples include LQ-FieldPheno [46], HT3P [47], and UIISScan 1.1 [48], which have significantly improved the efficiency and precision of crop monitoring.
  • Automation seeks to minimize human intervention in agricultural tasks through deep learning-based autonomous systems, thereby reducing labor costs and improving management efficiency and decision accuracy [49].
  • Precision decision-making is driven by high-resolution, real-time agricultural data analyzed through advanced deep learning models. Techniques such as convolutional neural networks, combined with IoT, remote sensing, and transfer learning, have been employed to optimize machinery operations and decision-making processes [50,51].
Historically, soybean has been at the forefront of integrating geospatial technologies into crop management. The NAPPFAST (North Carolina State University’s Automated Pest and Pathogen Forecasting System) platform was among the first geoweb tools designed for pest and disease monitoring, utilizing GIS-based modeling to provide early warnings and improve decision-making. Furthermore, HYSPLIT (Hybrid Single-Particle Lagrangian Integrated Trajectory) models were successfully employed to simulate the long-distance aerial dispersion of fungal spores attacking soybean fields, offering insights into epidemic spread patterns. These early implementations demonstrated how spatial modeling and environmental forecasting could directly contribute to yield protection and informed field interventions. The integration of such traditional models with modern deep learning frameworks—by fusing weather, spatial, and image data—holds promise for further enhancing predictive accuracy and operational efficiency in soybean production.
In summary, deep learning plays a pivotal role across all stages of soybean production and serves as a fundamental driver of agricultural intelligence. Its advantages lie not only in end-to-end learning capabilities, automatic feature extraction, strong transferability, and robustness, but also in its flexible architectures that can be tailored to a wide range of agricultural tasks. Representative models include stacked autoencoders (SAEs), deep belief networks (DBNs), deep Boltzmann machines (DBMs), convolutional neural networks (CNNs), graph neural networks (GNNs) [52], and generative adversarial networks (GANs) [53]. CNNs excel in pest and disease detection from crop images [54,55]; GNNs are effective in modeling spatial relationships within fields for yield prediction [56]; and GANs show considerable promise in image enhancement and remote sensing data generation [57,58].

4. The Applications of Deep Learning in Soybean Production

4.1. Diagnosis of Soybean Diseases and Pests

Pests and diseases have long been a major constraint on agricultural productivity and crop yield [59]. Traditional manual identification methods are inefficient and prone to errors, often resulting in substantial crop losses. With the rapid advancement of deep learning and its widespread application in image recognition, intelligent detection and monitoring of crop pests and diseases have emerged as a key area of modern agricultural research.
Deep learning offers powerful capabilities in processing image and sensor data, enabling effective integration with agricultural hardware systems. For instance, Park et al. [60] developed a soybean pest detection and forecasting platform based on unmanned ground vehicles (UGVs) and multiple deep learning algorithms (e.g., MRCNN, YOLOv3, Detectron2). Their system enabled early pest monitoring via image recognition and included a lightweight web-based application to support agricultural management. Similarly, Abed et al. [42] employed camera-equipped robots for in-field acquisition of soybean leaf images, which were processed by a U-Net and ResNet34-based segmentation model, followed by classification using DenseNet121 to distinguish healthy and diseased leaves. Nguyen et al. [61] further demonstrated the potential of hyperspectral imaging (400–1000 nm) combined with 2D/3D CNNs to detect grapevine virus infections at asymptomatic stages, offering a non-invasive solution for early disease warning. Notably, Sharma et al. [62] proposed SoyaTrans, a diagnostic model for soybean leaf diseases based on the Swin Transformer architecture integrated with CNN-derived features. The model achieved a 94% accuracy on a custom soybean leaf dataset. Figure 3a illustrates the data collection process, and Figure 3b shows the SoyaTrans architecture.
Fara et al. [63] applied transfer learning using a pretrained VGG19 model optimized with RMSprop to construct a classification network for three types of soybean leaf diseases. Omole et al. [64] utilized VGG16 and VGG19 models to detect soybean diseases and validated the efficacy of CNNs for early diagnosis through k-fold cross-validation and standard preprocessing. Goshika et al. [65] trained a YOLOv5-based model on a large-scale image dataset to classify five severity levels of soybean diseases, achieving a mAP of 92%. Yu et al. [66] introduced a soybean leaf disease identification method using an enhanced Residual Attention Network (RANet) that incorporated OTSU segmentation and data augmentation, reaching an accuracy of 98.49%. Building on this, Yu et al. [67] extended the task to four disease categories, employing stricter data splitting protocols, batch normalization, learning rate decay, and a transfer-learning-enhanced ResNet18 (TRNet18), which improved the accuracy by an additional 1.04%. The final test set accuracies for five models were as follows: AlexNet (76.54%), ResNet18 (98.42%), TRNet18 (99.53%), ResNet50 (95.89%), and TRNet50 (98.42%). Figure 4 illustrates some samples of soybean diseases.
These findings underscore the growing accuracy of deep learning models in pest and disease identification. However, the pursuit of higher accuracy often results in increasingly complex models with substantial computational demands, hindering practical deployment in agricultural environments. Lightweight models, which reduce parameter counts and computational overhead, offer a promising alternative by lowering storage, computation, and energy requirements—making them well-suited for deployment on resource-constrained edge devices while maintaining high performance [69]. Such solutions are particularly relevant for small-scale farms where cost and energy efficiency are critical, enhancing operational efficiency without compromising diagnostic accuracy.
It is worth noting that open-access datasets specific to soybean pest and disease detection remain scarce. To address this gap, our previous research introduced a general pest and disease classification model called CRE, based on masked image modeling and contrastive learning. We also constructed a large-scale open-source dataset containing 205,371 images, including 9613 images of soybean-related pests and diseases [9], providing essential data resources for future research in this domain.
According to the studies mentioned above, the primary challenges in current soybean diseases and pests diagnosis research can be summarized as follows:
  • Limited model generalizability, which hampers adaptability to variations in region, climate, and field conditions;
  • Underdeveloped multimodal data fusion frameworks, with an over-reliance on RGB imagery and insufficient integration of hyperspectral, thermal infrared, and environmental sensor data;
  • Lack of open-access, high-quality, and standardized image repositories for soybean pest and disease detection, restricting model reproducibility and comparative evaluation;
  • Unoptimized trade-offs between model deployment and energy consumption, resulting in limited practicality and scalability.
To address these challenges, future research should focus on the following directions:
  • Development of multimodal recognition frameworks that integrate RGB, hyperspectral, thermal imaging, and LiDAR data to enhance model adaptability across diverse agricultural scenarios;
  • Incorporation of multi-temporal remote sensing and UAV imagery to enable dynamic modeling and early warning of pest and disease progression over time;
  • Advancement of lightweight models for edge deployment, tailored to meet the agricultural sector’s requirements for low cost, low energy consumption, and minimal latency;
  • Construction of standardized and diversified soybean pest and disease image databases to improve model transferability and robustness across different environments;
  • Promotion of cross-disciplinary research, such as integrating GANs, digital twins, sim2real transfer, and domain adaptation techniques into agricultural vision tasks, thereby mitigating challenges related to high annotation costs and sample imbalance.
It is worth noting that soybean production occurs across diverse geographic regions with substantial variation in climate, temperature, and precipitation. These environmental differences can significantly affect model performance and generalizability. Some studies have demonstrated that environmental variability—rather than model complexity alone—plays a key role in recognition accuracy. For example, a CNN model trained in temperate zones may perform poorly in tropical climates without domain adaptation.
Moreover, the incorporation of vegetation indices (VIs), such as NDVI, SAVI, and EVI, into deep learning pipelines has shown the potential to enhance pest and disease detection, especially under conditions where RGB imagery lacks a sufficient spectral resolution. However, further research is needed to systematically compare the performance of different VIs in soybean-specific scenarios.
In parallel, recent work has explored the use of hyperspectral and multispectral data to capture crop spectral signatures across growth stages. These spectral features, spanning visible to near-infrared wavelengths, enable more accurate modeling of physiological traits and stress symptoms. Integrating such spectral data with deep learning models could significantly improve early diagnosis and crop health monitoring.

4.2. Soybean Growth Status and Phenotypic Trait Detection

Phenotypic trait detection and growth monitoring in plants heavily rely on image-based analysis to distinguish features and extract key characteristics at the individual plant level [70,71,72,73]. In the case of soybean, subtle yet significant variations in morphological structure, textural details, and spectral features arise across different growth stages, health conditions, and varietal attributes. These variations often pose challenges for traditional image processing methods, which struggle to deliver efficient and precise recognition [74,75]. In contrast, deep learning, with its powerful image analysis capabilities and multi-level feature extraction, has emerged as a crucial technology in soybean phenotyping, providing strong support for breeding, trait improvement, and physiological assessment [76,77,78].
To enhance physiological state perception, Yang et al. [79] developed a CNN model enhanced with an attention mechanism (AM), coupled with dimensionality reduction (DR) techniques and machine learning classifiers, to classify salt stress levels in soybeans using RGB images. For drought stress detection, Xu et al. [80] proposed MSAFNet, a ResNet50-based architecture integrating squeeze-and-excitation (SE) attention and multi-scale feature fusion modules. This model achieved an accuracy of 90.2% and an F1-score of 91.3% in simulated tests while maintaining a real-time inference speed of 13.6 fps. Figure 5a illustrates the image acquisition process under varying salt stress levels, and Figure 5b presents the MSAFNet model architecture.
In terms of trait identification, Grinblat et al. [81] used deep convolutional neural networks (DCNNs) to model leaf venation morphology, significantly improving classification accuracy across three legume species and visualizing critical venation features. Wang et al. [82] demonstrated that fusing deep features from upper, middle, and lower leaf images of soybean plants enhances variety classification performance. Lin et al. [83] developed an online seed classification system using multi-scale Retinex segmentation and a lightweight CNN model (SoyNet), effectively improving seed recognition efficiency.
Yield estimation, particularly via counting tasks, has become a focal point in recent research. Li et al. [84] proposed a two-column CNN (TCNN) for accurate soybean pod modeling and seed count estimation, significantly outperforming traditional counting approaches. Zhang et al. [74] built a high-precision phenotypic extraction model based on YOLOv8, achieving pod and seed recognition with an R 2 of 0.96 and RMSE values of 2.89 and 6.90, respectively, laying the groundwork for downstream yield prediction. Liu et al. [85] introduced SmartPod, a Transformer-based model employing semi-supervised learning for field pod counting, attaining an average accuracy of 94.1%. The data collection and annotation process is shown in Figure 6.
Additionally, Sun et al. [76] combined UAV-based hyperspectral imaging with a prototype contrastive learning network for soybean yield estimation and lodging classification, boosting accuracy by 48% and demonstrating the potential of remote sensing fused with deep learning in complex agricultural tasks. Feng et al. [86] used a UGV to collect field video data of soybean pods and developed a high-throughput yield prediction system, achieving 83% accuracy while reducing data collection costs by 32% compared to traditional methods. The collection and identification process is illustrated in Figure 7.
Despite these advancements, several challenges persist:
  • Limited model generalizability, particularly under complex and dynamic field conditions;
  • Insufficient multimodal fusion, as most current models rely on single-modality inputs (e.g., RGB or hyperspectral images), limiting their ability to leverage diverse sources of information;
  • Scarcity of high-quality, multi-annotated public datasets, which hampers training and benchmarking of models.
Future research should focus on the following key directions:
  • Multimodal data fusion: Integrate RGB, hyperspectral, LiDAR, and thermal images to improve model adaptability and robustness in complex field scenarios.
  • Multi-temporal remote sensing modeling: Use UAV imagery to dynamically monitor the full soybean growth cycle from sowing to harvest, enabling comprehensive modeling of growth stages, stress responses, and yield fluctuations.
  • Lightweight model design and deployment optimization: Develop models suitable for edge computing to ensure efficient deployment in resource-constrained farm environments.
  • Construction of standardized large-scale phenotyping databases: Promote the creation of open-access, high-quality soybean image datasets to enhance model comparability and reproducibility.
In summary, deep learning has shown strong potential in tasks such as soybean phenotyping, stress detection, and yield estimation. However, its widespread practical deployment requires further exploration in multimodal integration, deployment feasibility, and data resource development.

4.3. Weed Detection and Management

Effective weed management remains a critical factor in ensuring soybean yield and quality in modern agriculture [87]. However, traditional manual weeding methods are labor-intensive, costly, and inadequate for the demands of large-scale cultivation [88]. With the rapid advancement of computer vision and deep learning, numerous intelligent detection and automated weeding solutions have emerged, offering new strategies for soybean field management [23].
In complex and dynamic field environments, accurate detection of soybean plants and their center points is essential. Deep learning models, often combined with crop-specific features, enable precise crop-weed differentiation and provide vital data for path planning and control in automated weeding. For example, Jiang et al. [89] employed fluorescence imaging and multi-view localization to achieve 96.7% accuracy in automatic soybean seedling positioning. Su et al. [90] further integrated fluorescence imaging with 3D plant localization to distinguish between inter-row soybean plants and weeds, reaching 97% accuracy. These high-precision localization techniques provide a foundation for both mechanical and laser-based weeding and support the extraction of accurate navigation lines for UGVs.
Following plant localization, weed detection and classification can be achieved through semantic segmentation, object detection, or multi-task learning frameworks. These methods can distinguish between grass and broadleaf weeds and identify their growth stages [91,92], facilitating stage-specific weeding strategies and tailored herbicide application plans. For instance, early-stage weeds may be best managed through mechanical weeding, while late-stage or spatially dense weed populations can be targeted with precision spraying.
By integrating crop center point data with weed spatial distribution, inter- and intra-row weeding robots can autonomously plan paths and control various weeding actuators. Multiple execution modules—including mechanical, pneumatic, laser, and flame-based tools—can be integrated and deployed based on spatial crop-weed information to achieve high-precision, low-damage field operations. These robots can dynamically adjust the trajectory of weeding arms to avoid soybean plants while maximizing weeding efficiency. In our previous study on lettuce [88], we developed a pneumatic intra-row weeding robot (as shown in Figure 8), whose design and control principles could inform future development of soybean-specific weeding platforms. Additionally, Ye et al. [93] proposed a resilient comb-type intra-row soybean weeding robot based on PLC and laser ranging sensors (as shown in Figure 9), demonstrating the potential of hardware innovation in real-world applications. However, the system has not yet incorporated deep learning; future iterations could explore advanced models such as Mamba [94] to further enhance task accuracy and decision intelligence.
For weed populations requiring chemical treatment, object detection results can be combined with GPS/RTK data to enable geospatially aware path planning and differentiated precision spraying. Such intelligent spraying systems can reduce pesticide use and environmental impact. Meanwhile, visual navigation technologies are vital for maintaining operational continuity and stability. Recent research has integrated visual odometry, inertial navigation units (IMUs), and LiDAR into robotic platforms, alongside deep learning modules for object detection and semantic understanding, enabling autonomous navigation in environments with hills, lodged crops, or discontinuous ridges [95,96]. The application of Simultaneous Localization and Mapping (SLAM) further allows robots to perform mapping and localization simultaneously, significantly improving system flexibility and intelligence [97]. Looking ahead, incorporating Active Disturbance Rejection Control (ADRC) could enhance path planning and motion control of UGVs or UAVs, offering better stability and precision under complex environmental disturbances [98].
Despite significant progress, several challenges remain:
  • Insufficient real-time multimodal data fusion: Most current systems rely on single-sensor inputs and lack efficient mechanisms for integrating multimodal data, limiting their adaptability to environmental changes.
  • Limited model generalization and deployment adaptability: Deep learning models often struggle to generalize across diverse field conditions, regions, and growth stages due to limited training data and transfer difficulties.
  • Lack of dynamic decision-making in weeding strategies: Many weeding robots still follow preset paths or fixed parameters, lacking the flexibility to adjust strategies based on real-time conditions.
  • Balancing system complexity and economic feasibility: Achieving large-scale deployment requires managing hardware complexity and costs without compromising system performance.
Future research should focus on the following directions:
  • Develop lightweight, multi-task learning models optimized for real-time execution on edge devices;
  • Introduce self-supervised and reinforcement learning mechanisms to enhance model adaptability in complex environments;
  • Build open-access, multimodal field datasets for soybean to support high-quality model training and benchmarking;
  • Explore multi-robot cooperative weeding systems, integrating coordination protocols to improve field-level management efficiency;
  • Deeply integrate ADRC and deep learning-based navigation systems to construct intelligent platforms with high stability and dynamic control capabilities.

5. Mainstream Models and Optimization Strategies

5.1. Mainstream Models

With the ongoing advancement of computer vision in agriculture, deep learning models have become the core tools for intelligent soybean production. Based on their network architectures and task adaptability, mainstream models can be categorized into four types: image classification models, object detection models, image segmentation models, and Transformer-based models. Each type exhibits distinct characteristics in terms of processing capabilities, deployment flexibility, and fine-grained recognition performance.
First, image classification models are primarily used to determine the overall health status of soybean leaves and identify disease types within a single image [99]. Representative models include ResNet, which employs residual connections to mitigate degradation in deep networks, allowing for deeper architectures with reduced overfitting risk [100,101]; DenseNet, which enhances feature propagation and reuse through dense connections [102,103], has demonstrated strong performance in classifying diverse soybean leaf disease types [104,105]; and EfficientNet, which utilizes compound scaling to achieve high accuracy with significantly fewer parameters and reduced computational cost [106,107], making it particularly suitable for deployment on edge devices and unmanned platforms. However, these models typically require pre-cropped images focused on a single leaf or region and do not provide information on the precise location of lesions within the image [108].
Second, object detection models are designed to simultaneously localize and classify targets within images, playing a crucial role in intelligent field monitoring and precision spraying. The YOLO (You Only Look Once) family of models, known for their end-to-end structure, fast inference speed, and strong generalization capabilities [88,109,110,111,112,113,114,115,116], are widely used in field-based pest detection. For instance, YOLOv5 enables high-accuracy detection under limited computational resources, while YOLOv8 integrates attention mechanisms to improve the detection of small insects and occluded lesions [110,111,117,118]. In comparison, Faster R-CNN, with its two-stage architecture, achieves a higher detection accuracy [119,120], making it suitable for fine-grained lesion detection in high-resolution images. SSD (Single Shot Detector) [121] offers a balance between speed and accuracy, ideal for medium-scale recognition tasks on vehicle-mounted or UAV platforms.
Third, image segmentation models perform pixel-level delineation of lesion areas and leaf regions, supporting subsequent analyses such as area calculation and severity assessment [70,122]. The widely adopted U-Net architecture [123], originally developed for medical image segmentation, excels in soybean leaf lesion extraction and yield estimation due to its efficiency and strong performance on small datasets [124,125]. Mask R-CNN extends Faster R-CNN with an additional segmentation branch [126], allowing for the simultaneous detection and region delineation of lesions. DeepLabV3+, with atrous convolutions and multi-scale receptive fields, maintains accurate boundary information under complex backgrounds, making it well-suited for large-field disease segmentation tasks [127]. However, segmentation models typically require high-quality annotated data and are computationally intensive, which limits their deployment on low-power devices.
Finally, Transformer-based models have gained prominence in agricultural monitoring and temporal analysis. The Vision Transformer (ViT) divides images into patches and processes them as tokens, enabling effective global modeling for tasks such as analyzing disease progression across soybean growth stages [128,129]. Swin Transformer introduces a shifted window mechanism that balances local and global feature modeling, improving its ability to extract complex visual patterns like lesions and insects [62,130]. DETR, an end-to-end Transformer-based detection framework [131], demonstrates high robustness in localizing pests and diseases in remote sensing imagery. Despite their high performance, Transformer-based models often have large parameter sizes and high computational demands, necessitating further lightweight optimization for practical deployment in agricultural scenarios.
In summary, each model type offers unique advantages and limitations in terms of accuracy, efficiency, and task applicability. Researchers must select appropriate architectures based on specific application contexts. Future research should explore strategies such as model fusion, cross-modal enhancement, and knowledge distillation to develop intelligent recognition systems that balance performance, generalizability, and deployability. Table 1 summarizes the representative models, application tasks, advantages, limitations, and future directions for each of the four model categories.

5.2. Lightweight Architectures, Transfer Learning, and Model Optimization Strategies

To address challenges such as limited computational resources, insufficient annotated data, and complex deployment environments across the soybean industry chain, researchers have proposed a variety of model architecture and training strategy optimizations [132,133]. This section outlines five representative optimization approaches and illustrates their typical applications throughout different stages of soybean production.
Lightweight Model Design: Lightweight architectures are a key solution for overcoming deployment constraints on agricultural edge devices. In tasks such as sowing monitoring, emergence evaluation, and in-field seedling inspection, computational platforms are often vehicle-mounted cameras, low-power embedded boards, or autonomous mobile units. Models such as MobileNet [134], the MobileViT series [135,136,137], and EfficientNet-lite [106] significantly reduce model complexity through techniques like depthwise separable convolutions, group convolutions, and channel pruning. These models enable real-time detection of emergence density, missing seedlings, and seeding uniformity. In harvesting scenarios, lightweight structures can also be embedded into combine harvester vision systems to simultaneously perform pod counting and plant localization, thereby enhancing operational efficiency.
Transfer Learning: Transfer learning [138] has been widely adopted across the soybean industry, particularly in scenarios with limited data or high labeling costs. For instance, during the early growth stage, pretrained models such as ResNet or ViT can be fine-tuned to transfer knowledge from other crops (e.g., cotton) or simulated environments to soybean leaf classification tasks [139], reducing the need for extensive image acquisition. In pod maturity recognition or leaf nutrient diagnosis, transfer learning allows models to achieve acceptable accuracy using only a small number of field-annotated samples, effectively addressing cross-region and cross-variety adaptability issues in agronomic research.
Self- and Semi-Supervised Learning: These learning paradigms offer promising alternatives for addressing label scarcity in field data. Traditional supervised learning struggles to exploit the full potential of unlabeled images. Self-supervised methods such as SimCLR [140], MoCo [141], and MAE [142] can be used to pretrain models on reconstruction or contrastive tasks in applications like sowing zone monitoring or early-stage field variation analysis. When combined with semi-supervised strategies, models can leverage pseudo-labeling to enhance early-stage pest and disease recognition—especially beneficial for emerging diseases or rare soybean cultivars with limited training data.
Model Compression and Deployment Optimization: Compression techniques have broadened the applicability of deep models in real-time tasks such as harvest control and autonomous field inspections. Methods including pruning, quantization, and knowledge distillation enable the compression of large models (e.g., YOLO, ViT, DeepLab) into versions suitable for ARM processors, FPGAs, or edge GPUs. For example, in pod density estimation and harvesting decision-making, quantized YOLOv5 models can be embedded into harvest control systems to achieve low-latency, closed-loop operations encompassing image recognition, motion planning, and control execution—significantly improving precision and responsiveness in automated harvesting equipment.
Multimodal Fusion and Fine-Grained Detection: This is an increasingly important direction for enhancing soybean intelligent recognition systems. Tasks such as quality grading and yield estimation require more than RGB images to accurately reflect intrinsic traits and environmental conditions. Fusing data from infrared imagery, hyperspectral imaging, NDVI vegetation indices, and meteorological sensors allows for deeper model understanding of protein content, pod fullness, and latent diseases. In fine-grained detection tasks—such as identifying early-stage insect pests, soybean inflorescences, or subtle lesion patterns—modules like FPN [143], CBAM [144], and CoordAtt [145] help models perceive minute features more effectively, improving robustness in complex visual backgrounds.
In summary, optimization strategies centered around lightweight design, transfer learning, self-/semi-supervised training, model compression and deployment, and multimodal fusion have demonstrated practical potential across multiple stages of the soybean production chain. These approaches not only enhance model efficiency and environmental adaptability but also lay a solid foundation for advancing agricultural AI toward “end-to-end coordination, low-cost deployment, and high-precision decision-making.”
In addition to remote and UAV-based sensors, field spectroradiometry provides fine-grained, high-resolution spectral data that is particularly valuable for early stress detection and physiological trait estimation in soybean. Instruments such as ASD FieldSpec or other handheld hyperspectral spectrometers can measure reflectance across hundreds of bands, offering rich input features for deep learning models. These datasets have been successfully integrated with CNNs and LSTMs to model nitrogen content, chlorophyll concentration, and drought stress responses at the canopy level. Such field-based data collection bridges the gap between laboratory calibration and large-scale field deployment, improving model robustness and transferability across environmental conditions. Incorporating field spectroradiometry with deep learning represents a promising direction for precision phenotyping in soybean research.

6. Discussion

6.1. Challenges

Despite the considerable potential of deep learning in applications throughout the soybean production phase, from pre-sowing to post-harvest, its real-world implementation still faces several significant challenges.
First, the dual bottleneck of data quality and model generalization continues to hinder performance stability and transferability. Due to the inherent variability of natural agricultural environments, image acquisition is often affected by inconsistent lighting, varying angles, and severe occlusion. Combined with the high cost and low efficiency of manual annotation, these factors have resulted in a scarcity of high-quality soybean datasets for model training. Consequently, models are prone to overfitting in specific scenarios and lack the robustness needed for cross-regional or cross-seasonal deployment [27].
Second, numerous difficulties arise in real-world model deployment. Deep neural networks typically involve large numbers of parameters and high computational complexity, which imposes substantial demands on hardware resources [7]. This is in stark contrast to the resource-constrained conditions of agricultural environments, where edge devices often have limited processing power. In tasks requiring real-time responses—such as weeding and spraying—models often fail to run efficiently on embedded systems, severely limiting their engineering viability.
Moreover, the high complexity of field environments places stringent demands on model perception and robustness. Crops, weeds, and background elements frequently overlap or occlude one another, and natural disturbances are common. As a result, single-modality inputs (e.g., RGB imagery) are often insufficient for robust feature representation [146,147,148]. Although multimodal sensing technologies have been proposed as potential solutions, practical applications face major challenges, including feature heterogeneity, scale mismatches, and temporal–spatial misalignment between modalities. An efficient and unified multimodal fusion framework remains lacking.
More critically, the field of agricultural artificial intelligence lacks a standardized evaluation system. There are no widely accepted benchmarks for task definitions, dataset construction, or model performance assessment. This absence of common standards leads to poor comparability between studies, difficulty in reproducing experiments, and significant barriers to engineering translation. The lack of standardized frameworks directly impedes horizontal benchmarking, vertical iteration, and practical adoption—ultimately hindering the establishment of a robust technological development pathway.

6.2. Future Perspectives

To address the aforementioned challenges, advancing the application of deep learning in the soybean industry requires a dual focus on methodological innovation and system-level integration.
First, the incorporation of reinforcement learning and self-supervised learning offers promising solutions to the issues of data scarcity and decision-making complexity in agricultural settings. Reinforcement learning excels at long-term policy optimization, making it suitable for high-level tasks such as machinery path planning and operational decision-making. In contrast, self-supervised learning constructs pretext tasks to uncover latent structural features from large-scale unlabeled datasets, thereby enhancing the model’s generalization and representational capabilities without relying on manual annotations.
At the perception layer, multisensor data fusion is key to improving field-level recognition accuracy and environmental adaptability. Integrating RGB imagery, infrared thermography, LiDAR, soil moisture, and meteorological sensor data enables the construction of comprehensive phenotypic profiles and facilitates real-time monitoring of microenvironmental changes. This multi-source approach provides a robust data foundation for precision cultivation, early disease warning, and resource regulation [149,150,151].
Simultaneously, improving model interpretability and user accessibility is essential to building user trust and ensuring widespread adoption. Attention visualization, saliency map generation, and other explainability techniques can make model predictions more understandable and verifiable [23,72]. In addition, the development of graphical user interfaces tailored to farmers’ operational habits can lower the technical barrier and significantly improve system usability in real-world agricultural contexts.
On the system integration front, the synergy between deep learning and intelligent agricultural machinery will be central to achieving field-level automation. Embedding vision-based recognition systems into UGVs or UAVs to form a closed-loop workflow—from perception and decision-making to control—will enable coordinated air–ground precision operations and elevate both the intelligence and efficiency of field management.
Moreover, deep learning is playing an increasingly vital role in promoting sustainable agriculture. When combined with precision agriculture strategies, deep models can facilitate quantitative pesticide spraying, water-efficient irrigation, weeds control, and intelligent fertilization, thereby safeguarding yield while reducing resource waste and environmental impact [152,153,154]. Furthermore, coupling crop–environment–management behavior through integrated modeling can enhance the agricultural system’s responsiveness and adaptability to climate change, ultimately improving agroecological resilience.
Finally, the integration of digital twin technology with synthetic simulation data presents a transformative solution for agricultural AI training. By constructing a virtual twin environment for soybean cultivation and generating high-quality synthetic imagery via computer graphics, in conjunction with Sim-to-Real transfer learning frameworks, it becomes possible to significantly reduce dependence on real-world samples. This approach enhances model generalizability and robustness under real-world conditions, accelerating the practical and scalable deployment of deep learning technologies in agricultural production [155].
Among the numerous models and strategies reviewed, certain approaches have demonstrated particularly strong results in enhancing soybean production. For instance, the SoyaTrans model, built on the Swin Transformer architecture, achieved up to 94% accuracy in disease classification, showing high potential for early disease management. MSAFNet, which integrates attention mechanisms and multi-scale feature fusion, delivered high precision in drought stress recognition, contributing to improved irrigation planning. Additionally, YOLOv8-based systems have proven highly effective in real-time pod and pest detection, supporting both mechanical operations and yield estimation. These successes suggest that models emphasizing lightweight design, attention modules, and multimodal integration offer the most practical and accurate solutions in real-world soybean farming environments.

7. Conclusions

This review concludes that deep learning has demonstrated significant potential in addressing various challenges across the soybean production chain, from early-stage perception to end-stage decision-making. While considerable progress has been made in areas such as image-based classification, segmentation, and object detection, practical implementation still faces barriers related to data quality, model generalization, real-time deployment, and reproducibility. Moreover, the complex and dynamic nature of agricultural environments requires further advancements in model robustness and adaptive capability. Future research should prioritize the development of unified evaluation benchmarks, the integration of domain knowledge, and the adoption of emerging paradigms such as reinforcement learning and self-supervised learning. Ultimately, the synergy between methodological innovation and system-level collaboration is expected to accelerate the adoption of deep learning in real-world soybean production and agricultural intelligence at large.

Author Contributions

Conceptualization, P.H. and R.-F.W.; methodology, P.H. and R.-F.W.; validation, H.S., H.-Q.C. and Y.-M.Q.; formal analysis, H.S.; investigation, H.S. and H.-Q.C.; resources, H.S.; data curation, H.S. and Y.-M.Q.; writing—original draft preparation, H.S., H.-Q.C. and Y.-M.Q.; writing—review and editing, P.H. and R.-F.W.; visualization, H.S.; supervision, P.H. and R.-F.W.; project administration, P.H. and R.-F.W.; funding acquisition, H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Open Research Fund of Anhui Province Key Laboratory of Machine Vision Inspection (KLMVI-2024-HIT-12) and the Anhui Province Science and Technology Innovation Breakthrough Plan (No. 202423i08050056).

Data Availability Statement

No data was used in this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, X.; Komatsu, S. Improvement of soybean products through the response mechanism analysis using proteomic technique. Adv. Food Nutr. Res. 2017, 82, 117–148. [Google Scholar]
  2. El-Shemy, H. Soybean and Nutrition; BoD–Books on Demand: London, UK, 2011. [Google Scholar]
  3. Pagano, M.C.; Miransari, M. The importance of soybean production worldwide. In Abiotic and Biotic Stresses in Soybean Production; Elsevier: Amsterdam, The Netherlands, 2016; pp. 1–26. [Google Scholar]
  4. Shea, Z.; Singer, W.M.; Zhang, B. Soybean production, versatility, and improvement. In Legume Crops-Prospects, Production and Uses; IntechOpen: London, UK, 2020; pp. 29–50. [Google Scholar]
  5. Kofsky, J.; Zhang, H.; Song, B.H. The untapped genetic reservoir: The past, current, and future applications of the wild soybean (Glycine soja). Front. Plant Sci. 2018, 9, 949. [Google Scholar] [CrossRef]
  6. Colussi, J.; Schnitkey, G.; Janzen, J.; Paulson, N. The United States, Brazil, and China Soybean Triangle: A 20-Year Analysis. Farmdoc Dly. 2024, 14, 1–4. [Google Scholar]
  7. Yang, Z.Y.; Xia, W.K.; Chu, H.Q.; Su, W.H.; Wang, R.F.; Wang, H. A Comprehensive Review of Deep Learning Applications in Cotton Industry: From Field Monitoring to Smart Processing. Plants 2025, 14, 1481. [Google Scholar] [CrossRef]
  8. Zhang, X.; Li, Q. A Brief Introduction of Main Diseases and Insect Pests in Soybean Production in the Global Top Five Soybean Producing Countries. Plant Dis. Pests 2018, 9, 17–21. [Google Scholar]
  9. Wang, Z.; Wang, R.; Wang, M.; Lai, T.; Zhang, M. Self-supervised transformer-based pre-training method with General Plant Infection dataset. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision (PRCV), Xi’an, China, 25–28 October 2024; pp. 189–202. [Google Scholar]
  10. Yu, S.F.; Wang, C.L.; Hu, Y.F.; Wen, Y.C.; Sun, Z.B. Biocontrol of three severe diseases in soybean. Agriculture 2022, 12, 1391. [Google Scholar] [CrossRef]
  11. Cheng, R.; Mei, R.; Yan, R.; Chen, H.; Miao, D.; Cai, L.; Fan, J.; Li, G.; Xu, R.; Lu, W.; et al. A new distinct geminivirus causes soybean stay-green disease. Mol. Plant 2022, 15, 927–930. [Google Scholar] [CrossRef]
  12. O’Neal, M.E.; Johnson, K.D. Insect pests of soybean and their management. In The Soybean: Botany, Production and Uses; CABI: Wallingford, UK, 2010; pp. 300–324. [Google Scholar]
  13. Hesler, L.S.; Allen, K.C.; Luttrell, R.G.; Sappington, T.W.; Papiernik, S.K. Early-season pests of soybean in the United States and factors that affect their risk of infestation. J. Integr. Pest Manag. 2018, 9, 19. [Google Scholar] [CrossRef]
  14. Gaur, N.; Mogalapu, S. Pests of soybean. In Pests and Their Management; Springer: Singapore, 2018; pp. 137–162. [Google Scholar]
  15. Staniak, M.; Szpunar-Krok, E.; Kocira, A. Responses of soybean to selected abiotic stresses—Photoperiod, temperature and water. Agriculture 2023, 13, 146. [Google Scholar] [CrossRef]
  16. Wang, X.; Wu, Z.; Zhou, Q.; Wang, X.; Song, S.; Dong, S. Physiological response of soybean plants to water deficit. Front. Plant Sci. 2022, 12, 809692. [Google Scholar] [CrossRef]
  17. Destro, D.; Carpentieri-Pípolo, V.; de Souza Kiihl, R.A.; de Almeida, L.A. Photoperiodism and genetic control of the long juvenile period in soybean: A review. Crop Breed. Appl. Biotechnol. 2001, 1, 72–92. [Google Scholar] [CrossRef]
  18. Bagale, S. Nutrient management for soybean crops. Int. J. Agron. 2021, 2021, 3304634. [Google Scholar] [CrossRef]
  19. Yao, M.; Huo, Y.; Ran, Y.; Tian, Q.; Wang, R.; Wang, H. Neural radiance field-based visual rendering: A comprehensive review. arXiv 2024, arXiv:2404.00714. [Google Scholar]
  20. Guan, R.; Liu, T.; Tu, W.; Tang, C.; Luo, W.; Liu, X. Sampling Enhanced Contrastive Multi-View Remote Sensing Data Clustering with Long-Short Range Information Mining. IEEE Trans. Knowl. Data Eng. 2025, 1–15. [Google Scholar] [CrossRef]
  21. Cao, Z.; Lu, Y.; Yuan, J.; Xin, H.; Wang, R.; Nie, F. Tensorized Graph Learning for Spectral Ensemble Clustering. IEEE Trans. Circuits Syst. Video Technol. 2025, 35, 2662–2674. [Google Scholar] [CrossRef]
  22. Saleem, M.H.; Potgieter, J.; Arif, K.M. Automation in agriculture by machine and deep learning techniques: A review of recent developments. Precis. Agric. 2021, 22, 2053–2091. [Google Scholar] [CrossRef]
  23. Qin, Y.M.; Tu, Y.H.; Li, T.; Ni, Y.; Wang, R.F.; Wang, H. Deep Learning for Sustainable Agriculture: A Systematic Review on Applications in Lettuce Cultivation. Sustainability 2025, 17, 3190. [Google Scholar] [CrossRef]
  24. Li, Z.; Sun, C.; Wang, H.; Wang, R.F. Hybrid Optimization of Phase Masks: Integrating Non-Iterative Methods with Simulated Annealing and Validation via Tomographic Measurements. Symmetry 2025, 17, 530. [Google Scholar] [CrossRef]
  25. Guan, R.; Li, Z.; Tu, W.; Wang, J.; Liu, Y.; Li, X.; Tang, C.; Feng, R. Contrastive multiview subspace clustering of hyperspectral images based on graph convolutional networks. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5510514. [Google Scholar] [CrossRef]
  26. Wani, J.A.; Sharma, S.; Muzamil, M.; Ahmed, S.; Sharma, S.; Singh, S. Machine learning and deep learning based computational techniques in automatic agricultural diseases detection: Methodologies, applications, and challenges. Arch. Comput. Methods Eng. 2022, 29, 641–677. [Google Scholar] [CrossRef]
  27. Wang, R.F.; Su, W.H. The application of deep learning in the whole potato production Chain: A Comprehensive review. Agriculture 2024, 14, 1225. [Google Scholar] [CrossRef]
  28. Cui, K.; Shao, Z.; Larsen, G.; Pauca, V.; Alqahtani, S.; Segurado, D.; Pinheiro, J.; Wang, M.; Lutz, D.; Plemmons, R.; et al. PalmProbNet: A Probabilistic Approach to Understanding Palm Distributions in Ecuadorian Tropical Forest via Transfer Learning. In Proceedings of the 2024 ACM Southeast Conference, Tampa, FL, USA, 18–20 April 2024; pp. 272–277. [Google Scholar]
  29. Durai, S.K.S.; Shamili, M.D. Smart farming using machine learning and deep learning techniques. Decis. Anal. J. 2022, 3, 100041. [Google Scholar] [CrossRef]
  30. Zhou, G.; Wang, R.F.; Cui, K. A Local Perspective-based Model for Overlapping Community Detection. arXiv 2025, arXiv:2503.21558. [Google Scholar]
  31. Zhang, W.; Ma, M.; Jiang, Y.; Lian, R.; Wu, Z.; Cui, K.; Ma, X. Center-guided Classifier for Semantic Segmentation of Remote Sensing Images. arXiv 2025, arXiv:2503.16963. [Google Scholar]
  32. Cui, K.; Tang, W.; Zhu, R.; Wang, M.; Larsen, G.D.; Pauca, V.P.; Alqahtani, S.; Yang, F.; Segurado, D.; Fine, P.; et al. Efficient Localization and Spatial Distribution Modeling of Canopy Palms Using UAV Imagery. IEEE Trans. Geosci. Remote. Sens. 2025, 63, 4413815. [Google Scholar] [CrossRef]
  33. Kumari, S.; Dambale, A.S.; Samantara, R.; Jincy, M.; Bains, G. Introduction, history, geographical distribution, importance, and uses of soybean (Glycine max L.). In Soybean Production Technology: Physiology, Production and Processing; Springer: Singapore, 2025; pp. 1–17. [Google Scholar]
  34. Jarecki, W.; Bobrecka-Jamro, D. Effect of sowing date on the yield and seed quality of soybean (Glycine max (L.) Merr.). J. Elem. 2021, 26, 7–18. [Google Scholar] [CrossRef]
  35. Huang, Z.; Wang, R.; Cao, Y.; Zheng, S.; Teng, Y.; Wang, F.; Wang, L.; Du, J. Deep learning based soybean seed classification. Comput. Electron. Agric. 2022, 202, 107393. [Google Scholar] [CrossRef]
  36. Barbedo, J.G.A. Deep learning for soybean monitoring and management. Seeds 2023, 2, 340–356. [Google Scholar] [CrossRef]
  37. Chandel, N.S.; Chakraborty, S.K.; Rajwade, Y.A.; Dubey, K.; Tiwari, M.K.; Jat, D. Identifying crop water stress using deep learning models. Neural Comput. Appl. 2021, 33, 5353–5367. [Google Scholar] [CrossRef]
  38. Tetila, E.C.; Machado, B.B.; Astolfi, G.; de Souza Belete, N.A.; Amorim, W.P.; Roel, A.R.; Pistori, H. Detection and classification of soybean pests using deep learning with UAV images. Comput. Electron. Agric. 2020, 179, 105836. [Google Scholar] [CrossRef]
  39. Razfar, N.; True, J.; Bassiouny, R.; Venkatesh, V.; Kashef, R. Weed detection in soybean crops using custom lightweight deep learning models. J. Agric. Food Res. 2022, 8, 100308. [Google Scholar] [CrossRef]
  40. Attri, I.; Awasthi, L.K.; Sharma, T.P.; Rathee, P. A review of deep learning techniques used in agriculture. Ecol. Inform. 2023, 77, 102217. [Google Scholar] [CrossRef]
  41. Zhang, S.; Feng, H.; Han, S.; Shi, Z.; Xu, H.; Liu, Y.; Feng, H.; Zhou, C.; Yue, J. Monitoring of soybean maturity using UAV remote sensing and deep learning. Agriculture 2022, 13, 110. [Google Scholar] [CrossRef]
  42. Abed, S.H.; Al-Waisy, A.S.; Mohammed, H.J.; Al-Fahdawi, S. A modern deep learning framework in robot vision for automated bean leaves diseases detection. Int. J. Intell. Robot. Appl. 2021, 5, 235–251. [Google Scholar] [CrossRef] [PubMed]
  43. Shaheen, F.; Verma, B.; Asafuddoula, M. Impact of automatic feature extraction in deep learning architecture. In Proceedings of the 2016 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Gold Coast, Australia, 30 November–2 December 2016; pp. 1–8. [Google Scholar]
  44. Zhu, G.; Huang, X.; Peng, X.; Xu, J.; Guo, S.; Zhang, H. A deep learning-based quantitative prediction model for the processing potentials of soybeans as soymilk raw materials. Food Chem. 2024, 453, 139671. [Google Scholar] [CrossRef]
  45. Sharma, R. Artificial intelligence in agriculture: A review. In Proceedings of the 2021 5th International Conference on Intelligent Computing and Control Systems (ICICCS), Madurai, India, 6–8 May 2021; pp. 937–942. [Google Scholar]
  46. Fan, J.; Li, Y.; Yu, S.; Gou, W.; Guo, X.; Zhao, C. Application of internet of things to agriculture—The LQ-FieldPheno platform: A high-throughput platform for obtaining crop phenotypes in field. Research 2023, 6, 0059. [Google Scholar] [CrossRef]
  47. Li, D.; Quan, C.; Song, Z.; Li, X.; Yu, G.; Li, C.; Muhammad, A. High-throughput plant phenotyping platform (HT3P) as a novel tool for estimating agronomic traits from the lab to the field. Front. Bioeng. Biotechnol. 2021, 8, 623705. [Google Scholar] [CrossRef] [PubMed]
  48. Mukherjee, S.; Pal, S.; Pal, A.; Ghosh, D.; Sarkar, S.; Bhand, S.; Sarkar, P.; Bhattacharyya, N. UIISScan 1.1: A Field portable high-throughput platform tool for biomedical and agricultural applications. J. Pharm. Biomed. Anal. 2019, 174, 70–80. [Google Scholar] [CrossRef] [PubMed]
  49. Tian, H.; Wang, T.; Liu, Y.; Qiao, X.; Li, Y. Computer vision technology in agricultural automation—A review. Inf. Process. Agric. 2020, 7, 1–19. [Google Scholar] [CrossRef]
  50. Ghazal, S.; Munir, A.; Qureshi, W.S. Computer vision in smart agriculture and precision farming: Techniques and applications. Artif. Intell. Agric. 2024, 13, 64–83. [Google Scholar] [CrossRef]
  51. Coulibaly, S.; Kamsu-Foguem, B.; Kamissoko, D.; Traore, D. Deep learning for precision agriculture: A bibliometric analysis. Intell. Syst. Appl. 2022, 16, 200102. [Google Scholar] [CrossRef]
  52. Zhou, G.; Wang, R.F. The Heterogeneous Network Community Detection Model Based on Self-Attention. Symmetry 2025, 17, 432. [Google Scholar] [CrossRef]
  53. Shlezinger, N.; Whang, J.; Eldar, Y.C.; Dimakis, A.G. Model-based deep learning. Proc. IEEE 2023, 111, 465–499. [Google Scholar] [CrossRef]
  54. Dong, S.; Wang, P.; Abbas, K. A survey on deep learning and its applications. Comput. Sci. Rev. 2021, 40, 100379. [Google Scholar] [CrossRef]
  55. El Sakka, M.; Ivanovici, M.; Chaari, L.; Mothe, J. A Review of CNN Applications in Smart Agriculture Using Multimodal Data. Sensors 2025, 25, 472. [Google Scholar] [CrossRef]
  56. Gupta, A.; Singh, A. Agri-gnn: A novel genotypic-topological graph neural network framework built on graphsage for optimized yield prediction. arXiv 2023, arXiv:2310.13037. [Google Scholar]
  57. Lu, Y.; Chen, D.; Olaniyi, E.; Huang, Y. Generative adversarial networks (GANs) for image augmentation in agriculture: A systematic review. Comput. Electron. Agric. 2022, 200, 107208. [Google Scholar] [CrossRef]
  58. Jozdani, S.; Chen, D.; Pouliot, D.; Johnson, B.A. A review and meta-analysis of generative adversarial networks and their applications in remote sensing. Int. J. Appl. Earth Obs. Geoinf. 2022, 108, 102734. [Google Scholar] [CrossRef]
  59. Li, P.; Zhou, J.; Sun, H.; Zeng, J. RDRM-YOLO: A High-Accuracy and Lightweight Rice Disease Detection Model for Complex Field Environments Based on Improved YOLOv5. Agriculture 2025, 15, 479. [Google Scholar] [CrossRef]
  60. Park, Y.H.; Choi, S.H.; Kwon, Y.J.; Kwon, S.W.; Kang, Y.J.; Jun, T.H. Detection of soybean insect pest and a forecasting platform using deep learning with unmanned ground vehicles. Agronomy 2023, 13, 477. [Google Scholar] [CrossRef]
  61. Nguyen, C.; Sagan, V.; Maimaitiyiming, M.; Maimaitijiang, M.; Bhadra, S.; Kwasniewski, M.T. Early detection of plant viral disease using hyperspectral imaging and deep learning. Sensors 2021, 21, 742. [Google Scholar] [CrossRef]
  62. Sharma, V.; Tripathi, A.K.; Mittal, H.; Nkenyereye, L. SoyaTrans: A novel transformer model for fine-grained visual classification of soybean leaf disease diagnosis. Expert Syst. Appl. 2025, 260, 125385. [Google Scholar] [CrossRef]
  63. Farah, N.; Drack, N.; Dawel, H.; Buettner, R. A deep learning-based approach for the detection of infested soybean leaves. IEEE Access 2023, 11, 99670–99679. [Google Scholar] [CrossRef]
  64. Omole, O.J.; Rosa, R.L.; Rodriguez, D.Z. Soybean disease detection by deep learning algorithms. In Proceedings of the 2023 International Conference on Software, Telecommunications and Computer Networks (SoftCOM), Split, Croatia, 21–23 September 2023; pp. 1–5. [Google Scholar]
  65. Goshika, S.; Meksem, K.; Ahmed, K.R.; Lakhssassi, N. Deep learning model for classifying and evaluating soybean leaf disease damage. Int. J. Mol. Sci. 2023, 25, 106. [Google Scholar] [CrossRef]
  66. Yu, M.; Ma, X.; Guan, H.; Liu, M.; Zhang, T. A recognition method of soybean leaf diseases based on an improved deep learning model. Front. Plant Sci. 2022, 13, 878834. [Google Scholar] [CrossRef] [PubMed]
  67. Yu, M.; Ma, X.; Guan, H. Recognition method of soybean leaf diseases using residual neural network based on transfer learning. Ecol. Inform. 2023, 76, 102096. [Google Scholar] [CrossRef]
  68. Karlekar, A.; Seal, A. SoyNet: Soybean leaf diseases classification. Comput. Electron. Agric. 2020, 172, 105342. [Google Scholar] [CrossRef]
  69. Jin, Y.; Xia, X.; Gao, Q.; Yue, Y.; Lim, E.G.; Wong, P.; Ding, W.; Zhu, X. Deep learning in produce perception of harvesting robots: A comprehensive review. Appl. Soft Comput. 2025, 174, 112971. [Google Scholar] [CrossRef]
  70. Li, Z.; Xu, R.; Li, C.; Munoz, P.; Takeda, F.; Leme, B. In-field blueberry fruit phenotyping with a MARS-PhenoBot and customized BerryNet. Comput. Electron. Agric. 2025, 232, 110057. [Google Scholar] [CrossRef]
  71. Tan, C.; Sun, J.; Song, H.; Li, C. A customized density map model and segment anything model for cotton boll number, size, and yield prediction in aerial images. Comput. Electron. Agric. 2025, 232, 110065. [Google Scholar] [CrossRef]
  72. Cui, K.; Zhu, R.; Wang, M.; Tang, W.; Larsen, G.D.; Pauca, V.P.; Alqahtani, S.; Yang, F.; Segurado, D.; Lutz, D.; et al. Detection and Geographic Localization of Natural Objects in the Wild: A Case Study on Palms. arXiv 2025, arXiv:2502.13023. [Google Scholar]
  73. Huo, Y.; Wang, R.F.; Zhao, C.T.; Hu, P.; Wang, H. Research on Obtaining Pepper Phenotypic Parameters Based on Improved YOLOX Algorithm. AgriEngineering 2025, 7, 209. [Google Scholar] [CrossRef]
  74. Zhang, Q.Y.; Fan, K.J.; Tian, Z.; Guo, K.; Su, W.H. High-Precision Automated Soybean Phenotypic Feature Extraction Based on Deep Learning and Computer Vision. Plants 2024, 13, 2613. [Google Scholar] [CrossRef]
  75. Cui, K.; Li, R.; Polk, S.L.; Lin, Y.; Zhang, H.; Murphy, J.M.; Plemmons, R.J.; Chan, R.H. Superpixel-based and spatially-regularized diffusion learning for unsupervised hyperspectral image clustering. IEEE Trans. Geosci. Remote. Sens. 2024, 62, 4405818. [Google Scholar] [CrossRef]
  76. Sun, G.; Zhang, Y.; Wang, L.; Zhou, L.; Fei, S.; Han, S.; Xiao, S.; Che, Y.; Yan, L.; Xu, Y.; et al. Bridging the gap between hyperspectral imaging and crop breeding: Soybean yield prediction and lodging classification with prototype contrastive learning. Comput. Electron. Agric. 2025, 230, 109859. [Google Scholar] [CrossRef]
  77. Araus, J.L.; Cairns, J.E. Field high-throughput phenotyping: The new crop breeding frontier. Trends Plant Sci. 2014, 19, 52–61. [Google Scholar] [CrossRef] [PubMed]
  78. Zhou, W.; Chen, Y.; Li, W.; Zhang, C.; Xiong, Y.; Zhan, W.; Huang, L.; Wang, J.; Qiu, L. SPP-extractor: Automatic phenotype extraction for densely grown soybean plants. Crop J. 2023, 11, 1569–1578. [Google Scholar] [CrossRef]
  79. Yang, H.; Fei, L.; Wu, G.; Deng, L.; Han, Z.; Shi, H.; Li, S. A novel deep learning framework for identifying soybean salt stress levels using RGB leaf images. Ind. Crop. Prod. 2025, 228, 120874. [Google Scholar] [CrossRef]
  80. Xu, W.; Ma, X.; Guan, H.; Meng, Y.; Zhang, X. MSAFNet: A multi-scale data fusion-based drought recognition method for three-dimensional images of the soybean plant. Earth Sci. Inform. 2025, 18, 391. [Google Scholar] [CrossRef]
  81. Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef]
  82. Wang, B.; Li, H.; You, J.; Chen, X.; Yuan, X.; Feng, X. Fusing deep learning features of triplet leaf image patterns to boost soybean cultivar identification. Comput. Electron. Agric. 2022, 197, 106914. [Google Scholar] [CrossRef]
  83. Lin, W.; Shu, L.; Zhong, W.; Lu, W.; Ma, D.; Meng, Y. Online classification of soybean seeds based on deep learning. Eng. Appl. Artif. Intell. 2023, 123, 106434. [Google Scholar] [CrossRef]
  84. Li, Y.; Jia, J.; Zhang, L.; Khattak, A.M.; Sun, S.; Gao, W.; Wang, M. Soybean seed counting based on pod image using two-column convolution neural network. IEEE Access 2019, 7, 64177–64185. [Google Scholar] [CrossRef]
  85. Liu, F.; Wang, S.; Pang, S.; Han, Z.; Zhao, L. SmartPod: An Automated Framework for High-Precision Soybean Pod Counting in Field Phenotyping. Agronomy 2025, 15, 791. [Google Scholar] [CrossRef]
  86. Feng, J.; Blair, S.W.; Ayanlade, T.T.; Balu, A.; Ganapathysubramanian, B.; Singh, A.; Sarkar, S.; Singh, A.K. Robust soybean seed yield estimation using high-throughput ground robot videos. Front. Plant Sci. 2025, 16, 1554193. [Google Scholar] [CrossRef]
  87. Wang, R.F.; Tu, Y.H.; Li, X.C.; Chen, Z.Q.; Zhao, C.T.; Yang, C.; Su, W.H. An Intelligent Robot Based on Optimized YOLOv11l for Weed Control in Lettuce. In Proceedings of the 2025 ASABE Annual International Meeting, Toronto, ON, Canada, 13–16 July 2025; p. 1. [Google Scholar]
  88. Wang, R.F.; Tu, Y.H.; Chen, Z.Q.; Zhao, C.T.; Su, W.H. A Lettpoint-Yolov11l Based Intelligent Robot for Precision Intra-Row Weeds Control in Lettuce. Available at SSRN 5162748. 2025. Available online: https://ssrn.com/abstract=5162748 (accessed on 2 June 2025).
  89. Jiang, B.; Zhang, H.Y.; Su, W.H. Automatic localization of soybean seedlings based on crop signaling and multi-view imaging. Sensors 2024, 24, 3066. [Google Scholar] [CrossRef] [PubMed]
  90. Su, W.H.; Sheng, J.; Huang, Q.Y. Development of a three-dimensional plant localization technique for automatic differentiation of soybean from intra-row weeds. Agriculture 2022, 12, 195. [Google Scholar] [CrossRef]
  91. dos Santos Ferreira, A.; Freitas, D.M.; da Silva, G.G.; Pistori, H.; Folhes, M.T. Weed detection in soybean crops using ConvNets. Comput. Electron. Agric. 2017, 143, 314–324. [Google Scholar] [CrossRef]
  92. Fletcher, R.S. Using vegetation indices as input into random forest for soybean and weed classification. Am. J. Plant Sci. 2016, 7, 2186–2198. [Google Scholar] [CrossRef]
  93. Ye, S.; Xue, X.; Si, S.; Xu, Y.; Le, F.; Cui, L.; Jin, Y. Design and testing of an elastic comb reciprocating a soybean plant-to-plant seedling avoidance and weeding device. Agriculture 2023, 13, 2157. [Google Scholar] [CrossRef]
  94. Yao, M.; Huo, Y.; Tian, Q.; Zhao, J.; Liu, X.; Wang, R.; Xue, L.; Wang, H. FMRFT: Fusion mamba and DETR for query time sequence intersection fish tracking. arXiv 2024, arXiv:2409.01148. [Google Scholar]
  95. Bonin-Font, F.; Ortiz, A.; Oliver, G. Visual navigation for mobile robots: A survey. J. Intell. Robot. Syst. 2008, 53, 263–296. [Google Scholar] [CrossRef]
  96. Xu, R.; Li, C. A review of high-throughput field phenotyping systems: Focusing on ground robots. Plant Phenomics 2022, 2022, 9760269. [Google Scholar] [CrossRef]
  97. Kazerouni, I.A.; Fitzgerald, L.; Dooly, G.; Toal, D. A survey of state-of-the-art on visual SLAM. Expert Syst. Appl. 2022, 205, 117734. [Google Scholar] [CrossRef]
  98. Tu, Y.-H.; Wang, R.-F.; Su, W.-H. Active Disturbance Rejection Control—New Trends in Agricultural Cybernetics in the Future: A Comprehensive Review. Machines 2025, 13, 111. [Google Scholar] [CrossRef]
  99. Wu, A.Q.; Li, K.L.; Song, Z.Y.; Lou, X.; Hu, P.; Yang, W.; Wang, R.F. Deep Learning for Sustainable Aquaculture: Opportunities and Challenges. Sustainability 2025, 17, 5084. [Google Scholar] [CrossRef]
  100. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  101. Wu, W.; Huo, L.; Yang, G.; Liu, X.; Li, H. Research into the application of ResNet in soil: A review. Agriculture 2025, 15, 661. [Google Scholar] [CrossRef]
  102. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  103. Hou, Y.; Wu, Z.; Cai, X.; Zhu, T. The application of improved densenet algorithm in accurate image recognition. Sci. Rep. 2024, 14, 8645. [Google Scholar] [CrossRef] [PubMed]
  104. Yogabalajee, V.; Sundaram, K.; Kanagaraj, K. Soybean Leaf Disease Classification using Enhanced Densenet121. In Proceedings of the 2024 10th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 21–23 March 2024; Volume 1, pp. 1515–1520. [Google Scholar]
  105. Bhujade, V.G.; Sambhe, V. Multi-disease classification and severity estimation of cotton and soybean plants using DenseNet. In Proceedings of the International Conference on Advanced Network Technologies and Intelligent Computing, Varanasi, India, 17–19 December 2023; pp. 20–41. [Google Scholar]
  106. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  107. Chawla, S.; Gupta, D.; Pippal, S.K. Review on Architectures of Convolutional Neural Network. In Proceedings of the 2025 3rd International Conference on Disruptive Technologies (ICDT), Greater Noida, India, 7–8 March 2025; pp. 442–448. [Google Scholar]
  108. Wang, Z.; Zhang, H.W.; Dai, Y.Q.; Cui, K.; Wang, H.; Chee, P.W.; Wang, R.F. Resource-Efficient Cotton Network: A Lightweight Deep Learning Framework for Cotton Disease and Pest Classification. Plants 2025, 14, 2082. [Google Scholar] [CrossRef]
  109. Huo, Y.; Yao, M.; Tian, Q.; Wang, T.; Wang, R.; Wang, H. FA-YOLO: Research on Efficient Feature Selection YOLO Improved Algorithm Based on FMDS and AGMF Modules. arXiv 2024, arXiv:2408.16313. [Google Scholar]
  110. Zhao, C.-T.; Wang, R.-F.; Tu, Y.-H.; Pang, X.-X.; Su, W.-H. Automatic lettuce weed detection and classification based on optimized convolutional neural networks for robotic weed control. Agronomy 2024, 14, 2838. [Google Scholar] [CrossRef]
  111. Varghese, R.; Sambath, M. Yolov8: A novel object detection algorithm with enhanced performance and robustness. In Proceedings of the 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS), Chennai, India, 8–10 August 2024; pp. 1–6. [Google Scholar]
  112. Terven, J.; Córdova-Esparza, D.M.; Romero-González, J.A. A comprehensive review of yolo architectures in computer vision: From yolov1 to yolov8 and yolo-nas. Mach. Learn. Knowl. Extr. 2023, 5, 1680–1716. [Google Scholar] [CrossRef]
  113. Wang, C.Y.; Yeh, I.H.; Mark Liao, H.Y. Yolov9: Learning what you want to learn using programmable gradient information. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; pp. 1–21. [Google Scholar]
  114. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J. Yolov10: Real-time end-to-end object detection. Adv. Neural Inf. Process. Syst. 2024, 37, 107984–108011. [Google Scholar]
  115. Di, X.; Cui, K.; Wang, R.F. Toward Efficient UAV-Based Small Object Detection: A Lightweight Network with Enhanced Feature Fusion. Remote Sens. 2025, 17, 2235. [Google Scholar] [CrossRef]
  116. Tian, Y.; Ye, Q.; Doermann, D. Yolov12: Attention-centric real-time object detectors. arXiv 2025, arXiv:2502.12524. [Google Scholar]
  117. Li, X.; Zhang, T.; Yu, M.; Yan, P.; Wang, H.; Dong, X.; Wen, T.; Xie, B. A YOLOv8-based method for detecting tea disease in natural environments. Agron. J. 2025, 117, e70043. [Google Scholar] [CrossRef]
  118. Zhang, L.; Yu, S.; Yang, B.; Zhao, S.; Huang, Z.; Yang, Z.; Yu, H. YOLOv8 forestry pest recognition based on improved re-parametric convolution. Front. Plant Sci. 2025, 16, 1552853. [Google Scholar] [CrossRef] [PubMed]
  119. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef]
  120. Guleria, A.; Varshney, K.; Jindal, S. A systematic review: Object detection. AI Soc. 2025, 1–18. [Google Scholar] [CrossRef]
  121. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In Computer Vision—ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  122. Hemamalini, P.; Chandraprakash, M.; Laxman, R.; Rathinakumari, C.; Senthil Kumaran, G.; Suneetha, K. Thermal canopy segmentation in tomato plants: A novel approach with integration of YOLOv8-C and FastSAM. Smart Agric. Technol. 2025, 10, 100806. [Google Scholar] [CrossRef]
  123. Gao, Y.; Jiang, Y.; Peng, Y.; Yuan, F.; Zhang, X.; Wang, J. Medical Image Segmentation: A Comprehensive Review of Deep Learning-Based Methods. Tomography 2025, 11, 52. [Google Scholar] [CrossRef]
  124. Ingole, V.S.; Kshirsagar, U.A.; Singh, V.; Yadav, M.V.; Krishna, B.; Kumar, R. A Hybrid Model for Soybean Yield Prediction Integrating Convolutional Neural Networks, Recurrent Neural Networks, and Graph Convolutional Networks. Computation 2025, 13, 4. [Google Scholar] [CrossRef]
  125. Zhao, K.; Zhang, Q.; Wan, C.; Pan, Q.; Qin, Y. Visual Mamba UNet fusion multi-scale attention and detail infusion for unsound corn kernels segmentation. Sci. Rep. 2025, 15, 10933. [Google Scholar] [CrossRef] [PubMed]
  126. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  127. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  128. Wang, Y.; Deng, Y.; Zheng, Y.; Chattopadhyay, P.; Wang, L. Vision Transformers for Image Classification: A Comparative Survey. Technologies 2025, 13, 32. [Google Scholar] [CrossRef]
  129. Haruna, Y.; Qin, S.; Chukkol, A.H.A.; Yusuf, A.A.; Bello, I.; Lawan, A. Exploring the synergies of hybrid convolutional neural network and Vision Transformer architectures for computer vision: A survey. Eng. Appl. Artif. Intell. 2025, 144, 110057. [Google Scholar] [CrossRef]
  130. Zhang, J.; Zhou, H.; Liu, K.; Xu, Y. ED-Swin Transformer: A Cassava Disease Classification Model Integrated with UAV Images. Sensors 2025, 25, 2432. [Google Scholar] [CrossRef]
  131. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End-to-end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  132. Jin, C.; Zhou, L.; Pu, Y.; Zhang, C.; Qi, H.; Zhao, Y. Application of deep learning for high-throughput phenotyping of seed: A review. Artif. Intell. Rev. 2025, 58, 76. [Google Scholar] [CrossRef]
  133. Zhou, L.; Han, D.; Sun, G.; Liu, Y.; Yan, X.; Jia, H.; Yan, L.; Feng, P.; Li, Y.; Qiu, L.; et al. Soybean yield estimation and lodging discrimination based on lightweight UAV and point cloud deep learning. Plant Phenomics 2025, 7, 100028. [Google Scholar] [CrossRef]
  134. Howard, A.G. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  135. Mehta, S.; Rastegari, M. Mobilevit: Light-weight, general-purpose, and mobile-friendly vision transformer. arXiv 2021, arXiv:2110.02178. [Google Scholar]
  136. Mehta, S.; Rastegari, M. Separable self-attention for mobile vision transformers. arXiv 2022, arXiv:2206.02680. [Google Scholar]
  137. Wadekar, S.N.; Chaurasia, A. Mobilevitv3: Mobile-friendly vision transformer with simple and effective fusion of local, global and input features. arXiv 2022, arXiv:2209.15159. [Google Scholar]
  138. Jaquier, N.; Welle, M.C.; Gams, A.; Yao, K.; Fichera, B.; Billard, A.; Ude, A.; Asfour, T.; Kragic, D. Transfer learning in robotics: An upcoming breakthrough? A review of promises and challenges. Int. J. Robot. Res. 2025, 44, 465–485. [Google Scholar] [CrossRef]
  139. Liu, D.; Li, Z.; Wu, Z.; Li, C. Digital Twin/MARS-CycleGAN: Enhancing Sim-to-Real Crop/Row Detection for MARS Phenotyping Robot Using Synthetic Images. J. Field Robot. 2025, 42, 625–640. [Google Scholar] [CrossRef]
  140. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A simple framework for contrastive learning of visual representations. In Proceedings of the International Conference on Machine Learning, Virtual, 13–18 July 2020; pp. 1597–1607. [Google Scholar]
  141. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum contrast for unsupervised visual representation learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 9729–9738. [Google Scholar]
  142. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 16000–16009. [Google Scholar]
  143. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  144. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  145. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  146. Jiang, L.; Rodriguez-Sanchez, J.; Snider, J.L.; Chee, P.W.; Fu, L.; Li, C. Mapping of cotton bolls and branches with high-granularity through point cloud segmentation. Plant Methods 2025, 21, 66. [Google Scholar] [CrossRef]
  147. Tan, C.; Sun, J.; Paterson, A.H.; Song, H.; Li, C. Three-view cotton flower counting through multi-object tracking and RGB-D imagery. Biosyst. Eng. 2024, 246, 233–247. [Google Scholar] [CrossRef]
  148. Tan, C.; Li, C.; Sun, J.; Song, H. Three-View Cotton Flower Counting through Multi-Object Tracking and Multi-Modal Imaging. In Proceedings of the 2023 ASABE Annual International Meeting, Omaha, NE, USA, 9–12 July 2023; p. 1. [Google Scholar]
  149. Yang, Z.X.; Li, Y.; Wang, R.F.; Hu, P.; Su, W.H. Deep Learning in Multimodal Fusion for Sustainable Plant Care: A Comprehensive Review. Sustainability 2025, 17, 5255. [Google Scholar] [CrossRef]
  150. Cao, Z.; Xin, H.; Wang, R.; Nie, F. Superpixel-Based Bipartite Graph Clustering Enriched with Spatial Information for Hyperspectral and LiDAR Data. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5505115. [Google Scholar] [CrossRef]
  151. Gupta, D.; Golder, A.; Zhu, R.; Cui, K.; Tang, W.; Yang, F.; Csillik, O.; Alaqahtani, S.; Pauca, V.P. MoSAiC: Multi-Modal Multi-Label Supervision-Aware Contrastive Learning for Remote Sensing. arXiv 2025, arXiv:2507.08683. [Google Scholar]
  152. Wang, H.; Zhu, M.; Li, L.; Wang, L.; Zhao, H.; Mei, S. Regional weed identification method from wheat field based on unmanned aerial vehicle image and shearlets. Trans. Chin. Soc. Agric. Eng. 2017, 33, 99–106. [Google Scholar]
  153. Li, L.; Li, J.; Wang, H.; Georgieva, T.; Ferentinos, K.; Arvanitis, K.; Sygrimis, N. Sustainable energy management of solar greenhouses using open weather data on MACQU platform. Int. J. Agric. Biol. Eng. 2018, 11, 74–82. [Google Scholar] [CrossRef]
  154. Yuan, H.; Li, L.; Wang, J.; Wang, H.; Sigrimis, N.A. Design and test of regulation and control equipment for nutrient solution of water and fertilizer integration in greenhouse. Trans. Chin. Soc. Agric. Eng. 2016, 32, 27–32. [Google Scholar]
  155. Li, Z.; Xu, R.; Li, C.; Fu, L. Simulation of an In-field Phenotyping Robot: System Design, Vision-based Navigation and Field Mapping. In Proceedings of the 2022 ASABE Annual International Meeting, Houston, TX, USA, 17–20 July 2022; p. 1. [Google Scholar]
Figure 1. World Soybeans Usage: Food, Feed, and Processing (1961–2022). [Source: https://ourworldindata.org/grapher/soybean-production-and-use, Last Accessed on 29 May 2025].
Figure 1. World Soybeans Usage: Food, Feed, and Processing (1961–2022). [Source: https://ourworldindata.org/grapher/soybean-production-and-use, Last Accessed on 29 May 2025].
Agronomy 15 01831 g001
Figure 2. PRISMA flow diagram.
Figure 2. PRISMA flow diagram.
Agronomy 15 01831 g002
Figure 3. (a) Architectural outline of the SoyaTrans model; (b) workflow of soybean disease data collection [62].
Figure 3. (a) Architectural outline of the SoyaTrans model; (b) workflow of soybean disease data collection [62].
Agronomy 15 01831 g003
Figure 4. Few sample soybean leaf diseases: (a) bacterial blight, (b) soybean mosaic virus, (c) copper phytotoxicity, (d) charcoal rot, (e) healthy, (f) leaf cercospora, (g) downy mildew, (h) powdery mildew, (i) powdery mildew and rust, (j) rust, (k) rust and target spot. (l) Brown spot, (m) southern blight, and (n) unknown disease [68].
Figure 4. Few sample soybean leaf diseases: (a) bacterial blight, (b) soybean mosaic virus, (c) copper phytotoxicity, (d) charcoal rot, (e) healthy, (f) leaf cercospora, (g) downy mildew, (h) powdery mildew, (i) powdery mildew and rust, (j) rust, (k) rust and target spot. (l) Brown spot, (m) southern blight, and (n) unknown disease [68].
Agronomy 15 01831 g004
Figure 5. (a) Image acquisition process under varying salt stress levels; (b) MSAFNet model architecture [79].
Figure 5. (a) Image acquisition process under varying salt stress levels; (b) MSAFNet model architecture [79].
Agronomy 15 01831 g005
Figure 6. Image acquisition and annotation process [85]. (a) TraitDiscover high-throughput phenotyping platform; (b) example of image annotation.
Figure 6. Image acquisition and annotation process [85]. (a) TraitDiscover high-throughput phenotyping platform; (b) example of image annotation.
Agronomy 15 01831 g006
Figure 7. (a) (A) Data collection for a single plot: Yellow-highlighted cameras recorded the center plot; gray ones recorded others. Three passes were required to fully capture one plot. Python scripts automatically organized videos by plot. Arrows indicate robot movement direction; (B) image sampling: Each row was evenly divided into eight sections using seven splitters. Images from the middle five were used (excluding the first and last). The two rows of each plot were combined into one, yielding 10 images per side, 20 per plot; (b) results of soybean seeds counting [86].
Figure 7. (a) (A) Data collection for a single plot: Yellow-highlighted cameras recorded the center plot; gray ones recorded others. Three passes were required to fully capture one plot. Python scripts automatically organized videos by plot. Arrows indicate robot movement direction; (B) image sampling: Each row was evenly divided into eight sections using seven splitters. Images from the middle five were used (excluding the first and last). The two rows of each plot were combined into one, yielding 10 images per side, 20 per plot; (b) results of soybean seeds counting [86].
Agronomy 15 01831 g007
Figure 8. (a) Schematic diagram of field operation of weeding robot; (b) filed scenario [88].
Figure 8. (a) Schematic diagram of field operation of weeding robot; (b) filed scenario [88].
Agronomy 15 01831 g008
Figure 9. (a) Weed control device installation position; (b) test scenario [93].
Figure 9. (a) Weed control device installation position; (b) test scenario [93].
Agronomy 15 01831 g009
Table 1. Summary of deep learning model types and their applications in soybean production.
Table 1. Summary of deep learning model types and their applications in soybean production.
Model TypeRepresentative ModelsApplication TasksAdvantagesLimitationsFuture Directions
Image ClassificationResNet, DenseNet, EfficientNetDisease classificationHigh accuracy, clear architectureNo lesion localizationMulti-modal and multi-temporal integration
Object DetectionYOLO, Faster R-CNN, SSDLesion and pest detection, seed detection, yield estimationReal-time performance, versatilePoor performance on small targetsMulti-scale detection, small object enhancement
Image SegmentationU-Net, Mask R-CNN, DeepLabV3+Lesion segmentationPixel-level outputLabeling-intensive, high resource demandLightweight architectures, self-supervised learning
Transformer-BasedViT, Swin Transformer, DETRRemote sensing, dynamic modelingStrong global modeling capabilityHigh computational costMulti-modal fusion, lightweight optimization
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, H.; Chu, H.-Q.; Qin, Y.-M.; Hu, P.; Wang, R.-F. Empowering Smart Soybean Farming with Deep Learning: Progress, Challenges, and Future Perspectives. Agronomy 2025, 15, 1831. https://doi.org/10.3390/agronomy15081831

AMA Style

Sun H, Chu H-Q, Qin Y-M, Hu P, Wang R-F. Empowering Smart Soybean Farming with Deep Learning: Progress, Challenges, and Future Perspectives. Agronomy. 2025; 15(8):1831. https://doi.org/10.3390/agronomy15081831

Chicago/Turabian Style

Sun, Huihui, Hao-Qi Chu, Yi-Ming Qin, Pingfan Hu, and Rui-Feng Wang. 2025. "Empowering Smart Soybean Farming with Deep Learning: Progress, Challenges, and Future Perspectives" Agronomy 15, no. 8: 1831. https://doi.org/10.3390/agronomy15081831

APA Style

Sun, H., Chu, H.-Q., Qin, Y.-M., Hu, P., & Wang, R.-F. (2025). Empowering Smart Soybean Farming with Deep Learning: Progress, Challenges, and Future Perspectives. Agronomy, 15(8), 1831. https://doi.org/10.3390/agronomy15081831

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop