Tomato Ripening Detection in Complex Environments Based on Improved BiAttFPN Fusion and YOLOv11-SLBA Modeling
Abstract
:1. Introduction
- (1)
- It is difficult to discern between the transition stage of ripeness in complicated surroundings, and small tomato detection is prone to false negatives.
- (2)
- Some techniques are computationally costly, need a lot of disk space, and take a long time to run, making it impossible to meet the speedy detection requirements of real-world applications.
- (3)
- Accurately identifying tomatoes in the transition stage of ripeness still needs work, and the majority of research relies on color and form features, which are insufficiently reliable for detecting tomatoes in complicated situations.
2. Materials and Methods
2.1. Sources and Acquisition of Images
2.2. Data Preprocessing and Environmental Enhancement
- (1)
- Spray simulation: To replicate the effects of spray cooling operations, droplets of various sizes and colors are sporadically added. Figure 2 illustrates how the transparency mixing (α = 0.6) is used to create 200–500 white line segments (length 10–20 pixels) at random to mimic raindrops. This creates the illusion of water mist in the sprinkler system. The technique is capable of accurately simulating the adhesion and refraction properties of water droplets on the lens surface.
- (2)
- Fog interference: To replicate how fog affects image clarity in the greenhouse, a random fog layer was produced. A non-uniform fog was created using a Gaussian blur (kernel size 101 × 101), and the vision loss brought on by variations in greenhouse humidity was simulated by superimposing a gray mask (transparency α = 0.4). The attenuating effect of fog on visual contrast is seen in Figure 3.
- (3)
- Powerful light interference: To replicate direct sunlight or a powerful light source, a round, bright light was placed at a specific spot in the picture. Weighted fusion (weight 0.2) simulates the halo effect of direct sunlight on the lens by drawing a circular light spot (radius 30–80 pixels) at random. The light spot area displays typical overexposure characteristics, as illustrated in Figure 4.
- (4)
- Shadow masking: Shadow masking was used to remove the image’s leaf shadow effect. To replicate the localized reduction in brightness brought on by leaf projections, black rectangles with random positions and sizes (transparency 0.3) are created. The impact of shadow interference on the fruit recognition region is seen in Figure 5.
2.3. Data Labeling and Segmentation
2.4. YOLOv11 Network Architecture
2.5. YOLOv11-SLBA Network Structure
- Backbone layer improvement: The convolutional kernel sensing field is enlarged and the down-sampling procedure is optimized by substituting the SPPF-LSKA module for the original SPPF module. By capturing a wider range of contextual information while still extracting important image features, this module greatly improves the feature extraction capability for small targets and provides a richer semantic information base for feature fusion and detection tasks later on.
- Neck layer structure optimization: the two-way propagation mechanism is used to accomplish multi-level feature interaction in this research, which replaces the conventional FPN structure with the BiAttFPN feature fusion module. High-level characteristics offer a semantic context, such as the spatial relationship between fruits, branches, and leaves, while low-level features preserve spatial details, such as the texture of the tomato’s surface. By successfully integrating feature maps at various sizes, this dynamically weighted fusion technique significantly improves feature representation and increases the model’s adaptability in complicated scenarios.
- Detection head enhancement: Six sets of feature inputs from the neck layer are received by the detection head architecture with the help of DetectAux. The stability of small target detection is greatly enhanced, detection performance in complex environments is greatly enhanced, and the model’s identification accuracy for tomatoes of varying maturity is effectively improved through the joint optimization of multi-level gradient signals.
2.5.1. Space Pyramid-Large Nucleus Attention SPPF-LSKA
2.5.2. Bidirectional Attention Feature Pyramid Network BiAttFPN
- (1)
- Enhancement of the feature fusion technique: A bidirectional pyramid network for attention features, BiAttFPN, has been enhanced to normalize the fusion features with channel adjustment using the Concat+node_mode combination, which activates denser feature interactions. Multiple Concat operations allow the same layer to receive features from various scales. Simultaneously, redundant information is decreased to guarantee feature reuse efficiency while lowering computation in order to prevent potential feature dilution in BIFPN.
- (2)
- Network structure improvement: a multi-level backtracking connection is used, in which the higher-level features will interact with the bottom-level features multiple times; the number of module repetitions is optimized to increase the fusion path; a hierarchical progressive fusion is used, in which the P4 feature pyramid is built first and then gradually fused upwards. Since P2–P4 progressive fusion preserves more superficially detailed traits, this improvement is more suited for early small target maturity identification in tomatoes. Reducing the number of times particular modules are repeated and preserving performance through more intricate connections are two further ways to maximize computational efficiency.
- (3)
- Improvement of the related formula: the Concat used by BiAttFPN is spliced along the channel dimension as follows:
2.5.3. Auxiliary Detection Head DetectAux
- (1)
- Methodical addition of multi-scale characteristics: At varying stages of ripeness and from different gathering angles, tomatoes show notable variations in size and texture. Small target features and superficial details may be overlooked because traditional detection networks only use a small number of primary detecting heads. In order to improve the model’s perception of multi-scale targets and the overall detection of fine-grained feature expression, this study constructs six sets of DetectAux that are connected to feature maps of various scales.
- (2)
- Increased robustness in complicated environments: The model’s perception of important target areas can be readily impacted by interference variables like obstruction and lighting changes that are frequently present in real-world agricultural contexts. In order to obtain discriminative information from various “receptive field perspectives”, the DetectAux module introduces multiple feature pathways in parallel. This redundant supervision greatly improves the model’s stability and robustness in situations with partial obstruction and complex backgrounds.
- (3)
- Convergence acceleration and supervised reinforcement in the early training phases: DetectAux can also offer extra supervised signals during the training phase, creating a multi-path gradient propagation mechanism that reduces gradient disappearance and boosts model parameter update efficiency. In the early phases of training, the backbone network can be guided by the independent losses produced by many detection heads to collect target features more quickly, reducing training time and enhancing learning quality overall.
- (4)
- Resource management and optimization of computational efficiency: The six DetectAux groups maintain a lightweight structural design despite increasing model complexity. Small convolution kernels and shallow channels are used by each detection head, which requires less parameters than the primary detection head. Even though there are now more parameters overall, it is still appropriate for embedded devices and mid-range GPUs. Furthermore, to improve learning capabilities, all six DetectAux groups are activated during the training phase. Only the primary detection head is kept during the inference phase, and DetectAux output is turned off to prevent any extra computational load during deployment. DetectAux’s modular architecture allows for variable enablement and disabling, making platform switching easy. In addition, DetectAux is inherently compatible with GPU/TPU concurrency optimization and offers parallel computing. Technology such as gradient check-pointing or mixed precision training can be employed to further minimize resource usage in the event of a memory bottleneck.
3. Results and Analysis
3.1. Experimental Environment and Performance Evaluation Index
3.2. Convergence Analysis
3.3. Comparison Experiment
3.3.1. Comparison of Model Performance on Self-Built Datasets
3.3.2. Verification of Generalization Capabilities on Public Datasets
3.3.3. Stability Analysis of Improved Models on Heterogeneous Datasets
3.4. Ablation Experiment
3.5. Evaluation of Testing Effectiveness
3.5.1. Comparison of Model Performance Before and After Improvement
3.5.2. Comparison of Detection Effect of Different Models
4. Discussion
5. Conclusions
- SPPF-LSKA enhances spectral sensitivity, particularly in distinguishing fine color differences (e.g., green–yellow and red ripe stages) under uneven lighting. This is evidenced by a 0.6% mAP50 improvement in complex conditions, reducing the misclassification of transitional ripeness levels.
- BiAttFPN improves feature retention for occluded tomatoes, a common scenario in dense foliage or clustered fruit. The bidirectional attention mechanism increased precision (P) by 3.9% for small targets (<50 × 50 pixels) while maintaining precision (P) above 92%, demonstrating robustness against partial occlusion.
- DetectAux refines ripeness grading accuracy by leveraging multi-scale feature distillation. The six-head architecture reduced false positives between visually similar stages, critical for harvesting robots requiring stage-specific actions.
- High accuracy under variability: Achieved 91.3% mAP50 and 87.5% F1-score despite occlusion/lighting noise, outperforming YOLOv11 by 1.5% mAP50.
- Generalization capability: On tomato-ripeness1, the model attained 93.7% mAP50 and 84.6% F1-score, proving adaptability to diverse environments.
- Real-time readiness: With 10.9 MB memory and less FPS on embedded hardware, YOLOv11-SLBA balances speed and precision for field deployment.
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Nie, H.; Yang, X.; Zheng, S.; Hou, L. Gene-Based Developments in Improving Quality of Tomato: Focus on Firmness, Shelf Life, and Pre- and Post-Harvest Stress Adaptations. Horticulturae 2024, 10, 641. [Google Scholar] [CrossRef]
- Deribe, H.; Beyene, B.; Beyene, B.J.F.S.; Management, Q. Review on pre and post-harvest management on quality tomato (Lycopersicon esculentum Mill.) production. Food Sci. Qual. Manag. 2016, 54, 72–79. [Google Scholar]
- Song, K.; Chen, S.; Wang, G.; Qi, J.; Gao, X.; Xiang, M.; Zhou, Z. Research on High-Precision Target Detection Technology for Tomato-Picking Robots in Sustainable Agriculture. Sustainability 2025, 17, 2885. [Google Scholar] [CrossRef]
- Kabir, H.; Tham, M.-L.; Chang, Y.C. Internet of robotic things for mobile robots: Concepts, technologies, challenges, applications, and future directions. Digit. Commun. Netw. 2023, 9, 1265–1290. [Google Scholar] [CrossRef]
- Sherafati, A.; Mollazade, K.; Saba, M.K.; Vesali, F. TomatoScan: An Android-based application for quality evaluation and ripening determination of tomato fruit. Comput. Electron. Agric. 2022, 200, 107214. [Google Scholar] [CrossRef]
- Halstead, M.; McCool, C.; Denman, S.; Perez, T.; Fookes, C. Fruit quantity and ripeness estimation using a robotic vision system. IEEE Robot. Autom. Lett. 2018, 3, 2995–3002. [Google Scholar] [CrossRef]
- Behera, S.K.; Rath, A.K.; Sethy, P.K. Maturity status classification of papaya fruits based on machine learning and transfer learning approach. Inf. Process. Agric. 2021, 8, 244–250. [Google Scholar] [CrossRef]
- Zu, L.; Zhao, Y.; Liu, J.; Su, F.; Zhang, Y.; Liu, P. Detection and segmentation of mature green tomatoes based on mask R-CNN with automatic image acquisition approach. Sensors 2021, 21, 7842. [Google Scholar] [CrossRef]
- Yu, Y.; Zhang, K.; Yang, L.; Zhang, D. Fruit detection for strawberry harvesting robot in non-structural environment based on Mask-RCNN. Comput. Electron. Agric. 2019, 163, 104846. [Google Scholar] [CrossRef]
- Kang, H.; Zhou, H.; Wang, X.; Chen, C. Real-time fruit recognition and grasping estimation for robotic apple harvesting. Sensors 2020, 20, 5670. [Google Scholar] [CrossRef]
- Zhang, J.; Xie, J.; Zhang, F.; Gao, J.; Yang, C.; Song, C.; Rao, W.; Zhang, Y. Greenhouse tomato detection and pose classification algorithm based on improved YOLOv5. Comput. Electron. Agric. 2024, 216, 108519. [Google Scholar] [CrossRef]
- Wang, C.; Han, Q.; Li, J.; Li, C.; Zou, X. YOLO-BLBE: A Novel Model for Identifying Blueberry Fruits with Different Maturities Using the I-MSRCR Method. Agronomy 2024, 14, 658. [Google Scholar] [CrossRef]
- Zhai, S.; Shang, D.; Wang, S.; Dong, S. DF-SSD: An improved SSD object detection algorithm based on DenseNet and feature fusion. IEEE Access 2020, 8, 24344–24357. [Google Scholar] [CrossRef]
- Bai, Y.; Yu, J.; Yang, S.; Ning, J. An improved YOLO algorithm for detecting flowers and fruits on strawberry seedlings. Biosyst. Eng. 2024, 237, 1–12. [Google Scholar] [CrossRef]
- Seo, D.; Cho, B.-H.; Kim, K.-C. Development of monitoring robot system for tomato fruits in hydroponic greenhouses. Agronomy 2021, 11, 2211. [Google Scholar] [CrossRef]
- Sun, H.; Zheng, Q.; Yao, W.; Wang, J.; Liu, C.; Yu, H.; Chen, C. An Improved YOLOv8 Model for Detecting Four Stages of Tomato Ripening and Its Application Deployment in a Greenhouse Environment. Agriculture 2025, 15, 936. [Google Scholar] [CrossRef]
- Faseeh, M.; Bibi, M.; Khan, M.A.; Kim, D.-H. Deep learning assisted real-time object recognition and depth estimation for enhancing emergency response in adaptive environment. Results Eng. 2024, 24, 103482. [Google Scholar] [CrossRef]
- Wang, L.; Yoon, K.-J. Deep learning for HDR imaging: State-of-the-art and future trends. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 8874–8895. [Google Scholar] [CrossRef]
- Yang, X. Application of High-Speed Optical Measurement Based on Nanoscale Photoelectric Sensing Technology in the Optimization of Football Shooting Mechanics. J. Nanoelectron. Optoelectron. 2023, 18, 1493–1501. [Google Scholar] [CrossRef]
- Ambrus, B.; Teschner, G.; Kovács, A.; Neményi, M.; Helyes, L.; Pék, Z.; Takács, S.; Alahmad, T.; Nyéki, A. Field-grown tomato yield estimation using point cloud segmentation with 3D shaping and RGB pictures from a field robot and digital single lens reflex cameras. Heliyon 2024, 10, e37997. [Google Scholar] [CrossRef]
- GH T1193-2021; Standards for Supply and Distribution Cooperation in the People’s Republic of China. All China Federation of Supply and Marketing Cooperatives (ACFSMC): Beijing, China, 2021.
- Liang, X.; Jia, H.; Wang, H.; Zhang, L.; Li, D.; Wei, Z.; You, H.; Wan, X.; Li, R.; Li, W.; et al. ASE-YOLOv8n: A Method for Cherry Tomato Ripening Detection. Agronomy 2025, 15, 1088. [Google Scholar] [CrossRef]
- Meng, Z.; Du, X.; Xia, J.; Ma, Z.; Zhang, T. Real-time statistical algorithm for cherry tomatoes with different ripeness based on depth information mapping. Comput. Electron. Agric. 2024, 220, 108900. [Google Scholar] [CrossRef]
- Akbar, J.U.M.; Kamarulzaman, S.F.; Muzahid, A.J.M.; Rahman, A.; Uddin, M. A Comprehensive review on deep learning assisted computer vision techniques for smart greenhouse agriculture. IEEE Access 2024, 12, 4485–4522. [Google Scholar] [CrossRef]
- Montesinos López, O.A.; Montesinos López, A.; Crossa, J. Overfitting, Model Tuning, and Evaluation of Prediction Performance. In Multivariate Statistical Machine Learning Methods for Genomic Prediction; Springer: Cham, Switzerland, 2022; pp. 109–139. [Google Scholar]
- He, L.-H.; Zhou, Y.-Z.; Liu, L.; Cao, W.; Ma, J.-H. Research on object detection and recognition in remote sensing images based on YOLOv11. Sci. Rep. 2025, 15, 14032. [Google Scholar] [CrossRef]
- Rahman, M.; Khan, S.I.; Babu, H.M.H. BreastMultiNet: A multi-scale feature fusion method using deep neural network to detect breast cancer. Array 2022, 16, 100256. [Google Scholar] [CrossRef]
- Yang, Z.; Guan, Q.; Zhao, K.; Yang, J.; Xu, X.; Long, H.; Tang, Y. Multi-branch Auxiliary Fusion YOLO with Re-parameterization Heterogeneous Convolutional for Accurate Object Detection. In Pattern Recognition and Computer Vision, Proceedings of the 7th Chinese Conference, PRCV 2024, Urumqi, China, 18–20 October 2024; Springer: Singapore, 2024; pp. 492–505. [Google Scholar]
- Zhang, X.; Wang, Y.; Fang, H. Steel surface defect detection algorithm based on ESI-YOLOv8. Mater. Res. Express 2024, 11, 056509. [Google Scholar] [CrossRef]
- Xu, K.; Zhu, D.; Shi, C.; Zhou, C. YOLO-DBL: A multi-dimensional optimized model for detecting surface defects in steel. J. Membr. Comput. 2025, 1–11. [Google Scholar] [CrossRef]
- Liu, D.; Liang, J.; Geng, T.; Loui, A.; Zhou, T. Tripartite feature enhanced pyramid network for dense prediction. IEEE Trans. Image Process. 2023, 32, 2678–2692. [Google Scholar] [CrossRef]
- Senussi, M.F.; Kang, H.-S. Occlusion Removal in Light-Field Images Using CSPDarknet53 and Bidirectional Feature Pyramid Network: A Multi-Scale Fusion-Based Approach. Appl. Sci. 2024, 14, 9332. [Google Scholar] [CrossRef]
- Thapa, N.; Khanal, R.; Bhattarai, B.; Lee, J. Pine Wilt Disease Segmentation with Deep Metric Learning Species Classification for Early-Stage Disease and Potential False Positive Identification. Electronics 2024, 13, 1951. [Google Scholar] [CrossRef]
- Anbeek, P.; Vincken, K.L.; van Osch, M.J.; Bisschops, R.H.; van der Grond, J. Probabilistic segmentation of white matter lesions in MR imaging. NeuroImage 2004, 21, 1037–1044. [Google Scholar] [CrossRef] [PubMed]
- Li, Z.; Peng, C.; Yu, G.; Zhang, X.; Deng, Y.; Sun, J. Light-head r-cnn: In defense of two-stage object detector. arXiv 2017, arXiv:1711.07264. [Google Scholar]
- Zhang, H.; Tian, Y.; Wang, K.; Zhang, W.; Wang, F.-Y. Mask SSD: An effective single-stage approach to object instance segmentation. IEEE Trans. Image Process. 2019, 29, 2078–2093. [Google Scholar] [CrossRef]
- Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. Detrs beat yolos on real-time object detection. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 17–21 June 2024; pp. 16965–16974. [Google Scholar]
- Du, F.-J.; Jiao, S.-J. Improvement of lightweight convolutional neural network model based on YOLO algorithm and its research in pavement defect detection. Sensors 2022, 22, 3537. [Google Scholar] [CrossRef]
- Tomato Maturity. Tomato Ripeness1 Dataset. 2022. Available online: https://universe.roboflow.com/tomato-maturity/tomato-ripeness1 (accessed on 5 May 2025).
- Du, Z.; Liang, Y. Object detection of remote sensing image based on multi-scale feature fusion and attention mechanism. IEEE Access 2024, 12, 8619–8632. [Google Scholar] [CrossRef]
- Ma, J.; Li, M.; Fan, W.; Liu, J. State-of-the-Art Techniques for Fruit Maturity Detection. Agronomy 2024, 14, 2783. [Google Scholar] [CrossRef]
- Purohit, M.; Singh, M.; Kumar, A.; Kaushik, B.K. Enhancing the surveillance detection range of image sensors using HDR techniques. IEEE Sens. J. 2021, 21, 19516–19528. [Google Scholar] [CrossRef]
- Sattler, C.; Nagel, U.J.; Werner, A.; Zander, P. Integrated assessment of agricultural production practices to enhance sustainable development in agricultural landscapes. Ecol. Indic. 2010, 10, 49–61. [Google Scholar] [CrossRef]
- Liu, J.; Wu, G.; Liu, Z.; Wang, D.; Jiang, Z.; Ma, L.; Zhong, W.; Fan, X.; Liu, R. Infrared and visible image fusion: From data compatibility to task adaption. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 47, 2349–2369. [Google Scholar] [CrossRef]
- Tian, Y.; Pei, S.; Zhang, X.; Zhang, C.; Chawla, N.V. Knowledge distillation on graphs: A survey. ACM Comput. Surv. 2025, 57, 1–16. [Google Scholar] [CrossRef]
- Zhang, K.; Fang, B.; Zhang, Z.; Liu, T.; Liu, K. Exploring future ecosystem service changes and key contributing factors from a “past-future-action” perspective: A case study of the Yellow River Basin. Sci. Total Environ. 2024, 926, 171630. [Google Scholar] [CrossRef] [PubMed]
- Miao, Z.; Yu, X.; Li, N.; Zhang, Z.; He, C.; Li, Z.; Deng, C.; Sun, T. Efficient tomato harvesting robot based on image processing and deep learning. Precis. Agric. 2023, 24, 254–287. [Google Scholar] [CrossRef]
GH/T1193-2021 | Article Category | Color/Size Characteristics | Applicable Scenarios |
---|---|---|---|
Unripe stage + green ripe stage | Flowering and young fruiting | White-green, diameter < 3 cm | Do not pick |
Transition period | Green growth period | Pure green, diameter ≥ 3 cm | Long-term storage |
Early ripening stage | Yellowing and half ripe | The colored area accounts for 30–89% of the total area | Short-haul transportation |
Mid-to-late stage of ripening | Red maturity stage | Red accounts for more than 70% | Instant sales |
Categories | Total Number of Images/Picture | Number of Training Sets/Sheet | Number of Validation Sets/Sheet | Number of Test Sets/Sheet |
---|---|---|---|---|
Flowering and young fruiting | 600 | 422 | 125 | 53 |
Green growth period | 4407 | 3075 | 895 | 437 |
Yellowing and half ripe | 1469 | 1021 | 306 | 142 |
Red maturity stage | 2044 | 1418 | 416 | 210 |
Categories | Total Number of Images/Picture | Number of Training Sets/Sheet | Number of Validation Sets/Sheet | Number of Test Sets/Sheet |
---|---|---|---|---|
Flowering and young fruiting | 1470 | 1107 | 241 | 122 |
Green growth period | 6512 | 4549 | 1298 | 712 |
Yellowing and half ripe | 3899 | 2721 | 784 | 417 |
Red maturity stage | 4041 | 2795 | 838 | 400 |
Datasets | Total Number of Images/Picture | Number of Training Sets/Sheet | Number of Validation Sets/Sheet | Number of Test Sets/Sheet |
---|---|---|---|---|
Pre-expansion dataset | 6493 | 4546 | 1298 | 649 |
Expanded dataset | 10,000 | 7000 | 2000 | 1000 |
Model | Participants /Million | Model Memory/MB | Inference Time/ms | P | R | mAP50 | F1-Score | mAP50-95 |
---|---|---|---|---|---|---|---|---|
Faster R-CNN | 41.12 | 657 | 75 | 60.5% | 88.4% | 87.5% | 71.8% | 61.0% |
SSD | 24.01 | 93 | 22.5 | 89.3% | 78.4% | 85.7% | 83.5% | 59.5% |
RT-DETR | 19.9 | 307.4 | 96.3 | 91.2% | 82.9% | 88.6% | 86.8% | 61.9% |
YOLOv7 | 36.5 | 148 | 5.8 | 89.5% | 84.9% | 90.2% | 87.1% | 63.6% |
YOLOv8 | 3 | 5.8 | 1.2 | 90.1% | 84.5% | 90.4% | 87.2% | 63.9% |
YOLOv11 | 2.8 | 6.9 | 3.6 | 88.6% | 83.4% | 89.8% | 85.9% | 63.8% |
YOLOv11-SLBA | 2.72 | 10.9 | 2.3 | 92.0% | 83.5% | 91.3% | 87.5% | 64.6% |
Model | Participants /Million | Model Memory/MB | Inference Time/ms | P | R | mAP50 | F1-Score |
---|---|---|---|---|---|---|---|
Faster R-CNN | 41 | 657 | 77 | 55.6% | 96.9% | 89.02% | 70.6% |
SSD | 26.28 | 92 | 26 | 70.5% | 91.4% | 87.1% | 79.7% |
RT-DETR | 19.9 | 307.3 | 176.9 | 77.9% | 86.5% | 89.7% | 81.9% |
YOLOv7 | 36.48 | 73 | 6.2 | 78.2% | 91.0% | 88.7% | 84.2% |
YOLOv8 | 3 | 5.8 | 1.7 | 83.0% | 84.0% | 91.5% | 83.5% |
YOLOv11 | 2.58 | 9.8 | 3.2 | 85.9% | 85.3% | 92.7% | 85.6% |
YOLOv11-SLBA | 2.72 | 10.9 | 2.5 | 78.6% | 91.5% | 93.7% | 84.6% |
Datasets | P | R | mAP50 | F1-Score |
---|---|---|---|---|
Public Dataset | 78.6% | 91.5% | 93.7% | 84.6% |
Self-built Datasets | 92.0% | 83.5% | 91.3% | 87.5% |
Methodologies | Model | P | R | mAP50 | F1-Score | ||
---|---|---|---|---|---|---|---|
SPPF-LSKA | BiAttFPN | Aux | |||||
✗ | ✗ | ✗ | YOLOv11 | 88.6% | 83.4% | 89.8% | 85.9% |
✓ | ✗ | ✗ | YOLOv11-SPPF-LSKA | 89.3% | 84% | 90.4% | 86.6% |
✗ | ✓ | ✗ | YOLOv11-BiAttFPN | 92.5% | 81.1% | 90.8% | 86.4% |
✗ | ✗ | ✓ | YOLOv11-Aux | 90.9% | 83.2% | 90.1% | 86.9% |
✓ | ✗ | ✓ | YOLOv11-SPPF-LSKA-Aux | 90.8% | 84.1% | 90.7% | 87.3% |
✓ | ✓ | ✓ | YOLOv11-SLBA | 92% | 83.5% | 91.3% | 87.5% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Hao, Y.; Rao, L.; Fu, X.; Zhou, H.; Li, H. Tomato Ripening Detection in Complex Environments Based on Improved BiAttFPN Fusion and YOLOv11-SLBA Modeling. Agriculture 2025, 15, 1310. https://doi.org/10.3390/agriculture15121310
Hao Y, Rao L, Fu X, Zhou H, Li H. Tomato Ripening Detection in Complex Environments Based on Improved BiAttFPN Fusion and YOLOv11-SLBA Modeling. Agriculture. 2025; 15(12):1310. https://doi.org/10.3390/agriculture15121310
Chicago/Turabian StyleHao, Yan, Lei Rao, Xueliang Fu, Hao Zhou, and Honghui Li. 2025. "Tomato Ripening Detection in Complex Environments Based on Improved BiAttFPN Fusion and YOLOv11-SLBA Modeling" Agriculture 15, no. 12: 1310. https://doi.org/10.3390/agriculture15121310
APA StyleHao, Y., Rao, L., Fu, X., Zhou, H., & Li, H. (2025). Tomato Ripening Detection in Complex Environments Based on Improved BiAttFPN Fusion and YOLOv11-SLBA Modeling. Agriculture, 15(12), 1310. https://doi.org/10.3390/agriculture15121310