Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (272)

Search Parameters:
Keywords = plant image recognition

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
17 pages, 1927 KiB  
Article
ConvTransNet-S: A CNN-Transformer Hybrid Disease Recognition Model for Complex Field Environments
by Shangyun Jia, Guanping Wang, Hongling Li, Yan Liu, Linrong Shi and Sen Yang
Plants 2025, 14(15), 2252; https://doi.org/10.3390/plants14152252 - 22 Jul 2025
Viewed by 192
Abstract
To address the challenges of low recognition accuracy and substantial model complexity in crop disease identification models operating in complex field environments, this study proposed a novel hybrid model named ConvTransNet-S, which integrates Convolutional Neural Networks (CNNs) and transformers for crop disease identification [...] Read more.
To address the challenges of low recognition accuracy and substantial model complexity in crop disease identification models operating in complex field environments, this study proposed a novel hybrid model named ConvTransNet-S, which integrates Convolutional Neural Networks (CNNs) and transformers for crop disease identification tasks. Unlike existing hybrid approaches, ConvTransNet-S uniquely introduces three key innovations: First, a Local Perception Unit (LPU) and Lightweight Multi-Head Self-Attention (LMHSA) modules were introduced to synergistically enhance the extraction of fine-grained plant disease details and model global dependency relationships, respectively. Second, an Inverted Residual Feed-Forward Network (IRFFN) was employed to optimize the feature propagation path, thereby enhancing the model’s robustness against interferences such as lighting variations and leaf occlusions. This novel combination of a LPU, LMHSA, and an IRFFN achieves a dynamic equilibrium between local texture perception and global context modeling—effectively resolving the trade-offs inherent in standalone CNNs or transformers. Finally, through a phased architecture design, efficient fusion of multi-scale disease features is achieved, which enhances feature discriminability while reducing model complexity. The experimental results indicated that ConvTransNet-S achieved a recognition accuracy of 98.85% on the PlantVillage public dataset. This model operates with only 25.14 million parameters, a computational load of 3.762 GFLOPs, and an inference time of 7.56 ms. Testing on a self-built in-field complex scene dataset comprising 10,441 images revealed that ConvTransNet-S achieved an accuracy of 88.53%, which represents improvements of 14.22%, 2.75%, and 0.34% over EfficientNetV2, Vision Transformer, and Swin Transformer, respectively. Furthermore, the ConvTransNet-S model achieved up to 14.22% higher disease recognition accuracy under complex background conditions while reducing the parameter count by 46.8%. This confirms that its unique multi-scale feature mechanism can effectively distinguish disease from background features, providing a novel technical approach for disease diagnosis in complex agricultural scenarios and demonstrating significant application value for intelligent agricultural management. Full article
(This article belongs to the Section Plant Modeling)
Show Figures

Figure 1

15 pages, 5188 KiB  
Article
An Object Detection Algorithm for Orchard Vehicles Based on AGO-PointPillars
by Pengyu Ren, Xuyun Qiu, Qi Gao and Yumin Song
Agriculture 2025, 15(14), 1529; https://doi.org/10.3390/agriculture15141529 - 15 Jul 2025
Viewed by 241
Abstract
With the continuous expansion of the orchard planting area, there is an urgent need for autonomous orchard vehicles that can reduce the labor intensity of fruit farmers and improve the efficiency of operations to assist operators in the process of orchard operations. An [...] Read more.
With the continuous expansion of the orchard planting area, there is an urgent need for autonomous orchard vehicles that can reduce the labor intensity of fruit farmers and improve the efficiency of operations to assist operators in the process of orchard operations. An object detection system that can accurately identify potholes, trees, and other orchard objects is essential to achieve unmanned operation of the orchard vehicle. Aiming to improve upon existing object detection algorithms, which have the problem of low object recognition accuracy in orchard operation scenes, we propose an orchard vehicle object detection algorithm based on Attention-Guided Orchard PointPillars (AGO-PointPillars). Firstly, we use an RGB-D camera as the sensing hardware to collect the orchard road information and convert the depth image data obtained by the RGB-D camera into 3D point cloud data. Then, Efficient Channel Attention (ECA) and Efficient Up-Convolution Block (EUCB) are introduced based on the PointPillars, which can enhance the ability of feature extraction for orchard objects. Finally, we establish an orchard object detection dataset and validate the proposed algorithm. The results show that, compared to the PointPillars, the AGO-PointPillars proposed in this study has an average detection accuracy improvement of 4.64% for typical orchard objects such as potholes and trees, which can prove the reliability of our algorithm. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

14 pages, 7601 KiB  
Article
Evaluation and Optimization of Prediction Models for Crop Yield in Plant Factory
by Yaoqi Peng, Yudong Zheng, Zengwei Zheng and Yong He
Plants 2025, 14(14), 2140; https://doi.org/10.3390/plants14142140 - 10 Jul 2025
Viewed by 319
Abstract
This study focuses on enhancing crop yield prediction in plant factory environments through precise crop canopy image capture and background interference removal. This method achieves highly accurate recognition of the crop canopy projection area (CCPA), with a coefficient of determination (R2) [...] Read more.
This study focuses on enhancing crop yield prediction in plant factory environments through precise crop canopy image capture and background interference removal. This method achieves highly accurate recognition of the crop canopy projection area (CCPA), with a coefficient of determination (R2) of 0.98. A spatial resolution of 0.078 mm/pixel was derived by referencing a scale ruler and processing pixel counts, eliminating outliers in the data. Image post-processing focused on extracting the canopy boundary and calculating the crop canopy area. By incorporating crop yield data, a comparative analysis of 28 prediction models was performed, assessing performance metrics such as MSE, RMSE, MAE, MAPE, R2, prediction speed, training time, and model size. Among them, the Wide Neural Network model emerged as the most optimal. It demonstrated remarkable predictive accuracy with an R2 of 0.95, RMSE of 27.15 g, and MAPE of 11.74%. Furthermore, the model achieved a high prediction speed of 60,234.9 observations per second, and its compact size of 7039 bytes makes it suitable for efficient, real-time deployment in practical applications. This model offers substantial support for managing crop growth, providing a solid foundation for refining cultivation processes and enhancing crop yields. Full article
(This article belongs to the Section Plant Modeling)
Show Figures

Figure 1

18 pages, 6140 KiB  
Article
StomaYOLO: A Lightweight Maize Phenotypic Stomatal Cell Detector Based on Multi-Task Training
by Ziqi Yang, Yiran Liao, Ziao Chen, Zhenzhen Lin, Wenyuan Huang, Yanxi Liu, Yuling Liu, Yamin Fan, Jie Xu, Lijia Xu and Jiong Mu
Plants 2025, 14(13), 2070; https://doi.org/10.3390/plants14132070 - 6 Jul 2025
Viewed by 337
Abstract
Maize (Zea mays L.), a vital global food crop, relies on its stomatal structure for regulating photosynthesis and responding to drought. Conventional manual stomatal detection methods are inefficient, subjective, and inadequate for high-throughput plant phenotyping research. To address this, we curated a [...] Read more.
Maize (Zea mays L.), a vital global food crop, relies on its stomatal structure for regulating photosynthesis and responding to drought. Conventional manual stomatal detection methods are inefficient, subjective, and inadequate for high-throughput plant phenotyping research. To address this, we curated a dataset of over 1500 maize leaf epidermal stomata images and developed a novel lightweight detection model, StomaYOLO, tailored for small stomatal targets and subtle features in microscopic images. Leveraging the YOLOv11 framework, StomaYOLO integrates the Small Object Detection layer P2, the dynamic convolution module, and exploits large-scale epidermal cell features to enhance stomatal recognition through auxiliary training. Our model achieved a remarkable 91.8% mean average precision (mAP) and 98.5% precision, surpassing numerous mainstream detection models while maintaining computational efficiency. Ablation and comparative analyses demonstrated that the Small Object Detection layer, dynamic convolutional module, multi-task training, and knowledge distillation strategies substantially enhanced detection performance. Integrating all four strategies yielded a nearly 9% mAP improvement over the baseline model, with computational complexity under 8.4 GFLOPS. Our findings underscore the superior detection capabilities of StomaYOLO compared to existing methods, offering a cost-effective solution that is suitable for practical implementation. This study presents a valuable tool for maize stomatal phenotyping, supporting crop breeding and smart agriculture advancements. Full article
(This article belongs to the Special Issue Precision Agriculture Technology, Benefits & Application)
Show Figures

Figure 1

16 pages, 3020 KiB  
Article
FA-Unet: A Deep Learning Method with Fusion of Frequency Domain Features for Fruit Leaf Disease Identification
by Xiaowei Li, Wenlin Wu, Fenghua Zhu, Shenhao Guan, Wenliang Zhang and Zheng Li
Horticulturae 2025, 11(7), 783; https://doi.org/10.3390/horticulturae11070783 - 3 Jul 2025
Viewed by 316
Abstract
In the recognition of fruit leaf diseases, image recognition technology based on deep learning has received increasing attention. However, deep learning models often perform poorly in complex backgrounds, and in some cases, they even outperform traditional algorithms. To address this issue, this paper [...] Read more.
In the recognition of fruit leaf diseases, image recognition technology based on deep learning has received increasing attention. However, deep learning models often perform poorly in complex backgrounds, and in some cases, they even outperform traditional algorithms. To address this issue, this paper proposes a Frequency-Adaptive Attention (FA-attention) mechanism that leverages the significance of frequency-domain features in fruit leaf disease regions. By enhancing the processing of frequency domain features, the recognition performance in complex backgrounds is improved. Specifically, FA-attention combines Fourier transform with the attention mechanism to extract frequency domain features as key features. Then, this mechanism is integrated with the Unet model to obtain feature maps strongly related to frequency domain features. These feature maps are fused with multi-scale convolutional feature maps and then used for classification. Experiments were conducted on the Plant Village (PV) dataset and the Plant Pathology (PP) dataset with complex backgrounds. The results indicate that the proposed FA-attention mechanism achieves significant effects in learning frequency domain features. Our model achieves a recognition accuracy of 99.91% on the PV dataset and 89.59% on the PP dataset. At the same time, the convergence speed is significantly improved, reaching 94% accuracy with only 20 epochs, demonstrating the effectiveness of this method. Compared with classical models and state-of-the-art (SOTA) models, our model performs better on complex background datasets, demonstrating strong generalization capabilities. Full article
(This article belongs to the Section Plant Pathology and Disease Management (PPDM))
Show Figures

Figure 1

22 pages, 3237 KiB  
Article
Local Polar Coordinate Feature Representation and Heterogeneous Fusion Framework for Accurate Leaf Image Retrieval
by Mengjie Ye, Yong Cheng, Yongqi Yuan, De Yu and Ge Jin
Symmetry 2025, 17(7), 1049; https://doi.org/10.3390/sym17071049 - 3 Jul 2025
Viewed by 201
Abstract
Leaf shape is a crucial visual cue for plant recognition. However, distinguishing among plants with high inter-class shape similarity remains a significant challenge, especially among cultivars within the same species where shape differences can be extremely subtle. To address this issue, we propose [...] Read more.
Leaf shape is a crucial visual cue for plant recognition. However, distinguishing among plants with high inter-class shape similarity remains a significant challenge, especially among cultivars within the same species where shape differences can be extremely subtle. To address this issue, we propose a novel shape representation and an advanced heterogeneous fusion framework for accurate leaf image retrieval. Specifically, based on the local polar coordinate system, multiscale analysis, and statistical histograms, we first propose local polar coordinate feature representation (LPCFR), which captures spatial distribution from two orthogonal directions while encoding local curvature characteristics. Next, we present heterogeneous feature fusion with exponential weighting and Ranking (HFER), which enhances the compatibility and robustness of fused features by applying exponential weighted normalization and ranking-based encoding within neighborhood distance measures. Extensive experiments on both species-level and cultivar-level leaf datasets demonstrate that the proposed representation effectively captures shape features, and the fusion framework successfully integrates heterogeneous features, outperforming state-of-the-art (SOTA) methods. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

32 pages, 5287 KiB  
Article
UniHSFormer X for Hyperspectral Crop Classification with Prototype-Routed Semantic Structuring
by Zhen Du, Senhao Liu, Yao Liao, Yuanyuan Tang, Yanwen Liu, Huimin Xing, Zhijie Zhang and Donghui Zhang
Agriculture 2025, 15(13), 1427; https://doi.org/10.3390/agriculture15131427 - 2 Jul 2025
Viewed by 316
Abstract
Hyperspectral imaging (HSI) plays a pivotal role in modern agriculture by capturing fine-grained spectral signatures that support crop classification, health assessment, and land-use monitoring. However, the transition from raw spectral data to reliable semantic understanding remains challenging—particularly under fragmented planting patterns, spectral ambiguity, [...] Read more.
Hyperspectral imaging (HSI) plays a pivotal role in modern agriculture by capturing fine-grained spectral signatures that support crop classification, health assessment, and land-use monitoring. However, the transition from raw spectral data to reliable semantic understanding remains challenging—particularly under fragmented planting patterns, spectral ambiguity, and spatial heterogeneity. To address these limitations, we propose UniHSFormer-X, a unified transformer-based framework that reconstructs agricultural semantics through prototype-guided token routing and hierarchical context modeling. Unlike conventional models that treat spectral–spatial features uniformly, UniHSFormer-X dynamically modulates information flow based on class-aware affinities, enabling precise delineation of field boundaries and robust recognition of spectrally entangled crop types. Evaluated on three UAV-based benchmarks—WHU-Hi-LongKou, HanChuan, and HongHu—the model achieves up to 99.80% overall accuracy and 99.28% average accuracy, outperforming state-of-the-art CNN, ViT, and hybrid architectures across both structured and heterogeneous agricultural scenarios. Ablation studies further reveal the critical role of semantic routing and prototype projection in stabilizing model behavior, while parameter surface analysis demonstrates consistent generalization across diverse configurations. Beyond high performance, UniHSFormer-X offers a semantically interpretable architecture that adapts to the spatial logic and compositional nuance of agricultural imagery, representing a forward step toward robust and scalable crop classification. Full article
Show Figures

Figure 1

14 pages, 6074 KiB  
Article
Cross-Modal Data Fusion via Vision-Language Model for Crop Disease Recognition
by Wenjie Liu, Guoqing Wu, Han Wang and Fuji Ren
Sensors 2025, 25(13), 4096; https://doi.org/10.3390/s25134096 - 30 Jun 2025
Viewed by 299
Abstract
Crop diseases pose a significant threat to agricultural productivity and global food security. Timely and accurate disease identification is crucial for improving crop yield and quality. While most existing deep learning-based methods focus primarily on image datasets for disease recognition, they often overlook [...] Read more.
Crop diseases pose a significant threat to agricultural productivity and global food security. Timely and accurate disease identification is crucial for improving crop yield and quality. While most existing deep learning-based methods focus primarily on image datasets for disease recognition, they often overlook the complementary role of textual features in enhancing visual understanding. To address this problem, we proposed a cross-modal data fusion via a vision-language model for crop disease recognition. Our approach leverages the Zhipu.ai multi-model to generate comprehensive textual descriptions of crop leaf diseases, including global description, local lesion description, and color-texture description. These descriptions are encoded into feature vectors, while an image encoder extracts image features. A cross-attention mechanism then iteratively fuses multimodal features across multiple layers, and a classification prediction module generates classification probabilities. Extensive experiments on the Soybean Disease, AI Challenge 2018, and PlantVillage datasets demonstrate that our method outperforms state-of-the-art image-only approaches with higher accuracy and fewer parameters. Specifically, with only 1.14M model parameters, our model achieves a 98.74%, 87.64% and 99.08% recognition accuracy on the three datasets, respectively. The results highlight the effectiveness of cross-modal learning in leveraging both visual and textual cues for precise and efficient disease recognition, offering a scalable solution for crop disease recognition. Full article
(This article belongs to the Section Smart Agriculture)
Show Figures

Figure 1

24 pages, 8787 KiB  
Article
Fine Mapping of QTLs/QTNs and Mining of Genes Associated with Race 7 of the Soybean Cercospora sojina by Combining Linkages and GWAS
by Yanzuo Liu, Bo Hu, Aitong Yu, Yuxi Liu, Pengfei Xu, Yang Wang, Junjie Ding, Shuzhen Zhang, Wen-Xia Li and Hailong Ning
Plants 2025, 14(13), 1988; https://doi.org/10.3390/plants14131988 - 29 Jun 2025
Viewed by 287
Abstract
Soybean frogeye leaf spot (FLS) disease has been reported globally and is caused by the fungus Cercospora sojina, which affects the growth, seed yield, and quality of soybean. Among the 15 physiological microspecies of C. sojina soybean in China, Race 7 is [...] Read more.
Soybean frogeye leaf spot (FLS) disease has been reported globally and is caused by the fungus Cercospora sojina, which affects the growth, seed yield, and quality of soybean. Among the 15 physiological microspecies of C. sojina soybean in China, Race 7 is one of the main pathogenic microspecies. A few genes are involved in resistance to FLS, and they cannot meet the need to design molecular breeding methods for disease resistance. In this study, a soybean recombinant inbred line (RIL3613) population and a germplasm resource (GP) population were planted at two sites, Acheng (AC) and Xiangyang (XY). Phenotypic data on the percentage of leaf area diseased (PLAD) in soybean leaves were obtained via image recognition technology after the inoculation of seven physiological species and full onset at the R3 stage. Quantitative trait loci (QTLs) and quantitative trait nucleotides (QTNs) were mapped via linkage analysis and genome-wide association studies (GWASs), respectively. The resistance genes of FLS were subsequently predicted in the linkage disequilibrium region of the collocated QTN. We identified 114 QTLs and 18 QTNs in the RIL3613 and GP populations, respectively. A total of 14 QTN loci were colocalized in the two populations, six of which presented high phenotypic contributions. Through haplotype–phenotype association analysis and expression quantification, three genes (Glyma.06G300100, Glyma.06G300600, and Glyma.13G172300) located near molecular markers AX-90524088 and AX-90437152 (QTNs) are associated with FLS Chinese Race 7, identifying them as potential candidate resistance genes. These results provide a theoretical basis for the genetic mining of soybean antigray spot No. 7 physiological species. These findings also provide a theoretical basis for understanding the genetic mechanism underlying FLS resistance in soybeans. Full article
Show Figures

Figure 1

29 pages, 4298 KiB  
Article
RGB and Point Cloud-Based Intelligent Grading of Pepper Plug Seedlings
by Fengwei Yuan, Guoning Ma, Qinghao Zeng, Jinghong Liu, Zhang Xiao, Zhenhong Zou and Xiangjiang Wang
Agronomy 2025, 15(7), 1568; https://doi.org/10.3390/agronomy15071568 - 27 Jun 2025
Viewed by 274
Abstract
As an emerging vegetable cultivation technology, plug seedling cultivation significantly improves seedling production efficiency and reduces costs through standardization. Grading and transplanting, as the final step before the sale of plug seedlings, categorizes seedlings into different grades to ensure consistent quality. However, most [...] Read more.
As an emerging vegetable cultivation technology, plug seedling cultivation significantly improves seedling production efficiency and reduces costs through standardization. Grading and transplanting, as the final step before the sale of plug seedlings, categorizes seedlings into different grades to ensure consistent quality. However, most current grading methods can only detect seedling emergence but cannot classify the emerged seedlings. Therefore, this study proposes an intelligent grading method for pepper plug seedlings based on RGB and point cloud images, enabling precise grading using both RGB and 3D point cloud data. The proposed method involves the following steps: First, RGB and point cloud images of the seedlings are acquired using 2D and 3D cameras. The point cloud data is then converted into a 2D representation and aligned with the RGB images. Next, a deep learning-based object detection algorithm identifies the positions of individual seedlings in the RGB images. Using these positions, the seedlings are segmented from both the RGB and 2D point cloud images. Subsequently, a deep learning-based leaf recognition algorithm processes the segmented RGB images to determine leaf count, while another deep learning-based algorithm segments the leaves in the 2D point cloud images to extract their spatial information. Their surface area is measured using 3D reconstruction method to calculate leaf area. Additionally, plant height is derived from the point cloud’s height data. Finally, a classification model is trained using these extracted features to establish a grading system. Experimental results demonstrate that this automated grading method achieves a success rate of 97%, and compared with manual methods, this method has higher production efficiency. Meanwhile, it can grade different tray seedlings by training different models and provide reliable technical support for the quality evaluation of seedlings in industrialized transplanting production. Full article
(This article belongs to the Section Precision and Digital Agriculture)
Show Figures

Figure 1

23 pages, 8232 KiB  
Article
Intelligent Identification of Tea Plant Seedlings Under High-Temperature Conditions via YOLOv11-MEIP Model Based on Chlorophyll Fluorescence Imaging
by Chun Wang, Zejun Wang, Lijiao Chen, Weihao Liu, Xinghua Wang, Zhiyong Cao, Jinyan Zhao, Man Zou, Hongxu Li, Wenxia Yuan and Baijuan Wang
Plants 2025, 14(13), 1965; https://doi.org/10.3390/plants14131965 - 27 Jun 2025
Viewed by 416
Abstract
To achieve an efficient, non-destructive, and intelligent identification of tea plant seedlings under high-temperature stress, this study proposes an improved YOLOv11 model based on chlorophyll fluorescence imaging technology for intelligent identification. Using tea plant seedlings under varying degrees of high temperature as the [...] Read more.
To achieve an efficient, non-destructive, and intelligent identification of tea plant seedlings under high-temperature stress, this study proposes an improved YOLOv11 model based on chlorophyll fluorescence imaging technology for intelligent identification. Using tea plant seedlings under varying degrees of high temperature as the research objects, raw fluorescence images were acquired through a chlorophyll fluorescence image acquisition device. The fluorescence parameters obtained by Spearman correlation analysis were found to be the maximum photochemical efficiency (Fv/Fm), and the fluorescence image of this parameter is used to construct the dataset. The YOLOv11 model was improved in the following ways. First, to reduce the number of network parameters and maintain a low computational cost, the lightweight MobileNetV4 network was introduced into the YOLOv11 model as a new backbone network. Second, to achieve efficient feature upsampling, enhance the efficiency and accuracy of feature extraction, and reduce computational redundancy and memory access volume, the EUCB (Efficient Up Convolution Block), iRMB (Inverted Residual Mobile Block), and PConv (Partial Convolution) modules were introduced into the YOLOv11 model. The research results show that the improved YOLOv11-MEIP model has the best performance, with precision, recall, and mAP50 reaching 99.25%, 99.19%, and 99.46%, respectively. Compared with the YOLOv11 model, the improved YOLOv11-MEIP model achieved increases of 4.05%, 7.86%, and 3.42% in precision, recall, and mAP50, respectively. Additionally, the number of model parameters was reduced by 29.45%. This study provides a new intelligent method for the classification of high-temperature stress levels of tea seedlings, as well as state detection and identification, and provides new theoretical support and technical reference for the monitoring and prevention of tea plants and other crops in tea gardens under high temperatures. Full article
(This article belongs to the Special Issue Practical Applications of Chlorophyll Fluorescence Measurements)
Show Figures

Figure 1

17 pages, 3477 KiB  
Article
Rapid Identification of Mangrove Leaves Based on Improved YOLOv10 Model
by Haitao Sang, Ziming Li, Xiaoxue Shen, Shuwen Wang and Ying Zhang
Forests 2025, 16(7), 1068; https://doi.org/10.3390/f16071068 - 26 Jun 2025
Viewed by 228
Abstract
To address the issue of low recognition accuracy caused by the high morphological similarity of mangrove plant leaves, this study proposes a rapid identification method for mangrove leaves based on the YOLOv10 model, with corresponding improvements made to the baseline model. First, the [...] Read more.
To address the issue of low recognition accuracy caused by the high morphological similarity of mangrove plant leaves, this study proposes a rapid identification method for mangrove leaves based on the YOLOv10 model, with corresponding improvements made to the baseline model. First, the open-source tool LabelImg was employed to annotate leaf images and construct a mangrove leaf species dataset. Subsequently, a PSA attention mechanism was introduced to enhance the extraction of leaf detail features, while the SCDown downsampling method was adopted to preserve key characteristics. Furthermore, a BiFPN architecture incorporating SE modules was implemented to dynamically adjust channel weights for multi-scale feature fusion. Finally, the classification and regression tasks are decoupled by separating the detection head, and the final model is named YOLOv10-MSDet. Experimental results demonstrate that the improved model achieves rapid and accurate identification of various mangrove leaf species, with an average recognition accuracy of 92.4%—a 2.84 percentage point improvement over the baseline model, significantly enhancing the precision of mangrove leaf detection. Full article
Show Figures

Figure 1

23 pages, 6358 KiB  
Article
Optimization of Sorghum Spike Recognition Algorithm and Yield Estimation
by Mengyao Han, Jian Gao, Cuiqing Wu, Qingliang Cui, Xiangyang Yuan and Shujin Qiu
Agronomy 2025, 15(7), 1526; https://doi.org/10.3390/agronomy15071526 - 23 Jun 2025
Viewed by 310
Abstract
In the natural field environment, the high planting density of sorghum and severe occlusion among spikes substantially increases the difficulty of sorghum spike recognition, resulting in frequent false positives and false negatives. The target detection model suitable for this environment requires high computational [...] Read more.
In the natural field environment, the high planting density of sorghum and severe occlusion among spikes substantially increases the difficulty of sorghum spike recognition, resulting in frequent false positives and false negatives. The target detection model suitable for this environment requires high computational power, and it is difficult to realize real-time detection of sorghum spikes on mobile devices. This study proposes a detection-tracking scheme based on improved YOLOv8s-GOLD-LSKA with optimized DeepSort, aiming to enhance yield estimation accuracy in complex agricultural field scenarios. By integrating the GOLD module’s dual-branch multi-scale feature fusion and the LSKA attention mechanism, a lightweight detection model is developed. The improved DeepSort algorithm enhances tracking robustness in occlusion scenarios by optimizing the confidence threshold filtering (0.46), frame-skipping count, and cascading matching strategy (n = 3, max_age = 40). Combined with the five-point sampling method, the average dry weight of sorghum spikes (0.12 kg) was used to enable rapid yield estimation. The results demonstrate that the improved model achieved a mAP of 85.86% (a 6.63% increase over the original YOLOv8), an F1 score of 81.19%, and a model size reduced to 7.48 MB, with a detection speed of 0.0168 s per frame. The optimized tracking system attained a MOTA of 67.96% and ran at 42 FPS. Image- and video-based yield estimation accuracies reached 89–96% and 75–93%, respectively, with single-frame latency as low as 0.047 s. By optimizing the full detection–tracking–yield pipeline, this solution overcomes challenges in small object missed detections, ID switches under occlusion, and real-time processing in complex scenarios. Its lightweight, high-efficiency design is well suited for deployment on UAVs and mobile terminals, providing robust technical support for intelligent sorghum monitoring and precision agriculture management, and thereby playing a crucial role in driving agricultural digital transformation. Full article
Show Figures

Figure 1

44 pages, 8956 KiB  
Article
Gray Brick Wall Surface Damage Detection of Traditional Chinese Buildings in Macau: Damage Quantification and Thermodynamic Analysis Method via YOLOv8 Technology
by Liang Zheng, Jianyi Zheng, Yile Chen, Yuchan Zheng, Wei Lao and Shuaipeng Chen
Appl. Sci. 2025, 15(12), 6665; https://doi.org/10.3390/app15126665 - 13 Jun 2025
Viewed by 510
Abstract
The historical Lingnan gray brick buildings in Macau, a World Heritage Site, are facing severe deterioration due to prolonged disrepair, manifesting as cracks, breakages, moss adhesion, and other types of surface damage. These issues threaten not only the structural stability of the buildings [...] Read more.
The historical Lingnan gray brick buildings in Macau, a World Heritage Site, are facing severe deterioration due to prolonged disrepair, manifesting as cracks, breakages, moss adhesion, and other types of surface damage. These issues threaten not only the structural stability of the buildings but also the conservation of cultural heritage. To address the inefficiencies and low accuracy of traditional manual inspections, this study proposes an automated recognition and quantitative detection method for wall surface damage based on the YOLOv8 deep learning object detection model. A dataset comprising 375 annotated images collected from 162 gray brick historical buildings in Macau was constructed, covering eight damage categories: crack, damage, missing, vandalism, moss, stain, plant, and intact. The model was trained and validated using a stratified sampling approach to maintain a balanced class distribution, and its performance was comprehensively evaluated through metrics such as the mean average precision (mAP), F1 score, and confusion matrices. The results indicate that the best-performing model (Model 3 at the 297th epoch) achieved a mAP of 61.51% and an F1 score up to 0.74 on the test set, with superior detection accuracy and stability. Heatmap analysis demonstrated the model’s ability to accurately focus on damaged regions in close-range images, while damage quantification tests showed high consistency with manual assessments, confirming the model’s practical viability. Furthermore, a portable, integrated device embedding the trained YOLOv8 model was developed and successfully deployed in real-world scenarios, enabling real-time damage detection and reporting. This study highlights the potential of deep learning technology for enhancing the efficiency and reliability of architectural heritage protection and provides a foundation for future research involving larger datasets and more refined classification strategies. Full article
Show Figures

Figure 1

11 pages, 1119 KiB  
Proceeding Paper
Automatic Ladle Tracking with Object Detection and OCR in Steel Melting Shops
by Kabil Murugan, Mahinas Senthilmurugan, Venbha V. Senthilkumar, Harshita Velusamy, Karthiga Sekar, Vasanthan Buvanesan and Manikandan Venugopal
Eng. Proc. 2025, 95(1), 11; https://doi.org/10.3390/engproc2025095011 - 12 Jun 2025
Viewed by 462
Abstract
A ladle tracking system in steel production plants is essential for optimizing the ladle transportation between different processing units. The currently used technologies for ladle tracking, including Radio Frequency Identification (RFID) systems, are not effective due to their high maintenance costs and poor [...] Read more.
A ladle tracking system in steel production plants is essential for optimizing the ladle transportation between different processing units. The currently used technologies for ladle tracking, including Radio Frequency Identification (RFID) systems, are not effective due to their high maintenance costs and poor performance in harsh conditions, leaving a significant gap in developing an automated ladle tracking system. This paper proposes two innovative solutions to address these problems: a computer-vision-based ladle tracking system and an integrated approach of preprocessing techniques with optical character recognition (OCR) algorithms. The first method utilizes a YOLOv8 framework for detecting the two classes from the input images, such as the ladles and their unique numbers. This method achieved a precision of 0.983 and a recall of 0.998 in detecting the classes. The second method involves several preprocessing steps prior to the application of OCR. This is suitable for challenging environments, where the clarity of the images may be compromised. EasyOCR with enhanced preprocessing was able to extract the ladle number with a confidence score of 0.9948. The results demonstrate that vision-based automated ladle tracking is feasible in steel plants, improving operational efficiency, ensuring safety, and minimizing human intervention. Full article
Show Figures

Figure 1

Back to TopTop