Next Article in Journal
Description of Two Promising Walnut (Juglans regia L.) Selections with Lateral Bud Fruitfulness and Large Nuts
Next Article in Special Issue
Artificial Intelligence: A Promising Tool for Application in Phytopathology
Previous Article in Journal
Mineral and Metabolome Analyses Provide Insights into the Cork Spot Disorder on ‘Akizuki’ Pear Fruit
Previous Article in Special Issue
A Comparative Analysis of the Grafting Efficiency of Watermelon with a Grafting Machine
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IMVTS: A Detection Model for Multi-Varieties of Famous Tea Sprouts Based on Deep Learning

1
School of Mechanical Engineering, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
Key Laboratory of Transplanting Equipment and Technology of Zhejiang Province, Hangzhou 310018, China
3
Key Laboratory of Agricultural Equipment for Hilly and Mountainous Areas in Southeastern China (Co-Construction by Ministry and Province), Ministry of Agriculture and Rural Affairs, Hangzhou 310018, China
4
School of Textile Engineering, Zhejiang Sci-Tech University, Hangzhou 310018, China
5
College of Mechanical and Electrical Engineering, Zhongkai University of Agriculture and Engineering, Guangzhou 510225, China
6
Jinhua Polytechnic Zhejiang Key Laboratory of Crop Harvesting Equipment Technology, Jinhua 321017, China
*
Author to whom correspondence should be addressed.
Horticulturae 2023, 9(7), 819; https://doi.org/10.3390/horticulturae9070819
Submission received: 30 April 2023 / Revised: 11 July 2023 / Accepted: 13 July 2023 / Published: 17 July 2023
(This article belongs to the Special Issue Application of Smart Technology and Equipment in Horticulture)

Abstract

:
The recognition of fresh tea leaf sprouts is one of the difficulties in the realization of the automated picking of fresh tea leaves. At present, the research on the detection of fresh tea leaf sprouts is based on a single variety of tea leaves for a specific period or specific place, which has no advantage for the spread, promotion, and application of the methods. To address this problem, an identification of multiple varieties of tea sprouts (IMVTS) model was proposed. First, images of three different varieties of tea (ZhongCha108 (ZC108), ZhongHuangYiHao (ZH), ZiJuan (ZJ)) were obtained, and the multiple varieties of tea (MVT) dataset for training and validating models was created. In addition, the detection effects of adding a convolutional block attention module (CBAM) or efficient channel attention (ECA) module to YOLO v7 were compared. In the detection of the MVT dataset, YOLO v7+ECA and YOLO v7+CBAM showed a higher mean average precision (mAP) than YOLO v7, with 98.82% and 98.80%, respectively. Notably, the IMVTS model had the highest AP for ZC108, ZH, and ZJ compared with the two other models, with 99.87%, 96.97%, and 99.64%, respectively. Therefore, the IMVTS model was proposed on the basic framework of the ECA and YOLO v7. To further illustrate the superiority of the model, this study also conducted a comparison test between the IMVTS model and the mainstream target detection models (YOLO v3, YOLO v5, FASTER-RCNN, and SSD) and the IMVTS model on the VOC dataset, and it is clear from the test results that the mAP of the IMVTS model is ahead of the remaining models. Concisely, the detection accuracy of the IMVTS model can meet the engineering requirements for the automatic harvesting of autumn fresh famous tea leaves, which provides a basis for the future design of detection networks for other varieties of autumn tea sprouts.

1. Introduction

China is known as the home of tea. The country’s planting, output, consumption, and export volume of tea are first in the world. The extensive demand for fresh tea leaves requires effective support [1]. Famous tea is a representative of high-quality tea in China. Its raw materials are one sprout and one leaf of a tea tree. Its harvesting is still mainly manual and labor intensive, and there is a shortage of tea collectors [2]. To achieve the intelligent picking of fresh tea leaf sprouts, researchers use machine vision technology to detect fresh tea leaves. Lei Zhang et al. [3] extracted the R, G, and B components of images of collected fresh tea leaves and then processed the adaptive threshold of the B components. A new component gray diagram was obtained in combination with the G components, and the thresholds were enhanced by segmental transformation, thereby improving the contrast between the tender leaves and the background, and then, the improved watershed function was used to obtain a good division effect. Liang Zhang et al. [4] used Bayesian inference to judge the principles and Bayes methods to build an identification model of the fresh leaf collection status in order to detect purple rose tea tree sprouts in April in real time. Although the above machine learning technology can accurately identify fresh tea leaf sprouts, the identified fresh leaf image background environment is relatively single, and the segmentation accuracy is greatly affected by the characteristics of the fresh tea leaf sprouts. It is often difficult to overcome this shortcoming.
With the rapid development of artificial intelligence technology, the use of deep-learning [5] target detection algorithms, such as Fast R-CNN [6], Faster R-CNN [7], You Only Look Once (YOLO) [8], and Single Shot MultiBox Detector (SSD) [9], etc., are increasingly applied in the agricultural field. In terms of fresh tea leaf sprout recognition, Yatao Li et al. [10] used the YOLO v3 network to detect and identify fresh tea leaf sprouts on an RGB-D image collected with an RGB-D camera and estimated the three-dimensional coordinates of the fresh tea leaf picking position in the corresponding depth chart of the enclosure. It reached 83.18%. Wenkai Xu et al. [11] proposed a detection and classification method for the two-level integration network of the variability domain. This method combined YOLO v3’s fast detection capacity and the high-precision classification ability of DenseNet201 to achieve the accurate detection of fresh tea leaves. Yu-Ting Chen et al. [12] detected the OTTL area in a tea tree image through the Faster R-CNN model and then identified the picking point in the OTTL area with the FCN model to determine the three-dimensional coordinates of the picking point. The YOLO model is currently a widely used detection network. It adopts a one-stage algorithm. The network operates quickly, the memory occupation is small, and it has a fast detection speed [13]. Based on inheriting the advantages of the original YOLO model, YOLO v7 [14] has a higher detection accuracy and faster reasoning speed compared to the previous YOLO series models because of its more complex network structure and training techniques. However, because the YOLO v7 model uses only one main convolutional neural network to predict the category and location of different targets, it also improves the speed and makes the recognition accuracy lower. This may lead to more missed inspections and misunderstandings when using YOLO v7 to detect fresh tea leaf images. At the same time, the applicability of the models proposed in the existing literature to fresh sprouts of different varieties of tea is unknown.
Consequently, this study used the YOLO v7 model as the basic framework to research the identification of multiple varieties of fresh tea leaf sprouts and explored the impact of integrating CBAM and ECA mechanisms on YOLO v7 detection accuracy and generalization effects. According to the results of the research, an IMVTS model was proposed to solve the problem of high-precision detection of multiple tea species. In addition, to demonstrate the superiority of the proposed IMVTS model in detection, comparison tests of the IMVTS model with mainstream target detection models (YOLO v3, YOLO v5, FASTER-RCNN, and SSD) and detection tests of the IMVTS model on VOC datasets were designed.

2. Materials and Methods

2.1. Experimental Data Collection and Preparation

2.1.1. Fresh Tea Leaf Image Collection and MVT Dataset Production

At the experimental base of the Tea Research Institute of the Chinese Academy of Agricultural Sciences (Figure 1), a total of 277 fresh leaf images of 3 tea varieties were collected, of which 90 images were ZhongCha108 (ZC108), 150 were ZhongHuangYiHao (ZH), and 37 were ZiJuan (ZJ). The sampling date was 21 September 2022. The collected fresh tea leaf images were made into a dataset of fresh leaves of multiple tea varieties that could be used to train and verify the network model and were named the MVT dataset. An iQOO Neo3 mobile phone camera was used. The shooting angle was 30~60°, and the shooting distance was 30~50 cm. The images were stored in a JPG format, and the MVT dataset images are shown in Figure 2. The open-source labeling tool Labellmg was used to artificially label the sprout position on the fresh tea leaf images, and the labeled result was saved in the PASCAL VOC format.
The images of ZC108, ZH, and ZJ as the research objects are shown in Figure 2a–c. The leaves of ZC108 are long oval in shape, green in color, slightly rumbled on the leaf surface, and flat on the leaf body, and the sprouts are yellowish-green with less velvet. The leaves of ZH are oval in shape, yellowish green in color, slightly elevated on the leaf surface, slightly inflexed on the leaf body, and the sprouts are slender. ZJ is a large-leaved, medium-sprout species. The leaf blade is upwardly slanting, willow-shaped, with an acuminate tip, and purple in color. Its petiole is purple-red in color. These images can also provide data support for future studies of other tea recognition models.

2.1.2. MVT Dataset Division

The 277 obtained images and corresponding labels were divided into training sets and validation sets according to the proportion of training sets: validation sets (=9:1) so that the model training and validating in the later stage could be performed. To enrich the diversity of samples and improve the robustness of the IMVTS model, the dataset was expanded to 15,000 pieces of the original dataset, with increased or decreased brightness, increased and decreased color saturation, and horizontal flip. The application process is shown in Figure 3. After the enhancement, the number of sprouts of the fresh leaves of the three tea varieties reached 59,337, of which the number of sprouts of ZC108 was 7429, the number of sprouts of ZH was 18,902, and the number of sprouts of ZJ was 33,006.

2.2. Establishment and Training of the IMVTS Model

2.2.1. Experimental Platform and Environmental Configuration

This study used a workstation equipped with the Ubuntu 18.04 operating system for model training. The CPU was Intel i7 12700F (Shenzhen, China) and the GPU was 3080Ti. The development environment was Pytorch 1.11. The development language was Python, and vs. code software was used for debugging.

2.2.2. Establishment of the IMVTS Model

This study used the YOLO v7 model as the basic framework for the recognition of multiple varieties of fresh tea leaf sprouts; the related research has shown that attention mechanisms [15] such as SENet [16], CBAM [17], and ECA [18] improved the network performance. Therefore, to improve the detection accuracy of YOLO v7’s original model on the small target of fresh tea leaves, the effects of adding the CBAM or ECA attention module to YOLO v7 on the detection effect were compared; the final IMVTS network model structure obtained is shown in Figure 4. The IMVTS model mainly includes four parts: input, backbone, head, and output. First, the collected tea leaf images undergo labellmg and random data augmentation, then the tea leaf images are resized into 640 × 640 size RGB images and input to the backbone network The backbone network features the processing of fresh tea leaves with features. Second, the feature information extracted in the backbone network is combined through the characteristics of the head to obtain the characteristics of large, medium, and small sizes. Finally, the characteristics of the characteristic fusion are sent to the REP and Conv module of the head section for detection, and the final result is the output. The workflow of the algorithm is shown in Figure 5. The main network part of the YOLO v7 model is mainly composed of convolution, the extended-ELAN (E-ELAN) module, the MPConv module, and the SPPCSPC module. Among these, the E-ELAN module is based on the original ELAN, changing the calculation block while maintaining the original ELAN’s transition layer structure. The network can learn more features by controlling the shortest and longest gradient path. It can resolve the relatively complicated and variable problems that can be solved for the background environment and targets appearing in the MVT dataset to extract more efficient features. The SPPCSPC module adds parallel MaxPooling operations to a convolution series to obtain different feelings so that the algorithm can adapt to fresh tea leaf images of different resolutions and can make the network speed faster; the accuracy of detecting fresh tea leaf sprouts will also be improved. The main role of the MPConv module is to sample. The experience of the current feature layer is expanded through MaxPooling and then integrated with the normal convolutional fresh tea leaf sprout feature information. It can improve the generalization of the network, suitable for the recognition scene of multiple varieties of tea leaf sprouts [19]. When introducing the attention module, the location information and detail information of the feature information are extracted from the backbone, and there is less semantic information [20]. Feature information is easily lost after processing for small targets such as fresh tea leaf sprouts. To improve the detection accuracy of the YOLO v7 original model for fresh tea leaves, an attention mechanism was added between the backbone and the head. The structure of the attention module is shown in Figure 6. CBAM combines the channel attention mechanism with the spatial attention mechanism. The CBAM module conducts global average pooling and the largest global pooling for the input feature layer, then learns the weight information of each channel through a shared full connection layer and the Sigmoid function. Then, it learns the weight information of each point in the space through a convolutional nucleus of 3 × 3 and the expansion coefficient of 2, and the Sigmoid function. ECA is a channel attention mechanism. The ECA module changes the two full connection layers to one-dimensional convolution. After the Sigmoid function is passed, the channel weight information is obtained. This module avoids the dimension of information and has good cross-channel information acquisition capabilities.

2.2.3. Training Parameter Settings

When training the original model of YOLO v7 and the IMVTS model, only the thawing training was performed. The thawing training set the total training generation epochs to 500; additionally, the batch size was set to 4, the initial learning rate of the model was set to 0.01, the minimum learning rate was set to 0.0001, and the SGD optimizer was used to optimize the model. The momentum parameters of the SGD optimizer were set to 0.937, and the torque annealing function was used to reduce the learning rate. In the same environment, the training loss curves of the YOLO v7 original model, YOLO v7+CBAM model, and IMVTS model are shown in Figure 7.

2.2.4. Model Evaluation Index

To evaluate the detection effect of the IMVTS model, this study used precision (P), recall (R), F1 score, average precision (AP), and mean average precision (mAP) for measurement. Among these, P is the ratio of the number of correctly predicted sprouts to the total number of predicted sprouts in an image, R is the ratio of the number of correctly predicted sprouts to the total number of true sprouts in an image, the F1 score is the equalization of P and R, AP represents the accuracy of each kind of fresh tea leaf sprout recognition, and mAP represents the average value of three kinds of fresh tea leaf sprout recognition. They are calculated as follows (1)–(5):
P = T P T P + F P × 100 %
R = T P T P + F N × 100 %
F 1 = 2 P × R P + R × 100 %
A P = 0 1 P ( R ) d R × 100 %
m A P = 1 n A P 1 + A P 2 + + A P n n
where TP is the number of correctly predicted sprouts; FP is the number of falsely predicted sprouts; FN is the number of true sprouts that were not predicted as sprouts; and n is the number of fresh tea leaves. The values of the above parameters are obtained by counting and tallying the target detection frames in the validation sets.

3. Results and Discussion

3.1. Model Identification Results and Comparative Analysis

3.1.1. Comparison of YOLO v7 Model, YOLO v7+CBAM Model, and IMVTS Model Recognition Results

The YOLO v7 original model, YOLO v7+CBAM model, and IMVTS model were placed in the same environment for detection. The obtained model evaluation indicators are shown in Table 1. It can be seen from Table 1 that after adding an attention mechanism module to the YOLO v7 network, the P, R, F1 score, and mAP were improved. The precision value of the IMVTS model was 99.76%, 0.10% higher than the YOLO v7+CBAM and 0.24% higher than the YOLO v7 model. The recall value of the IMVTS model was 97.03%, 0.10% higher than the YOLO v7+CBAM and 2.46% higher than the YOLO v7 model. The F1 score of the IMVTS model was 0.98, as was that of the YOLO v7+CBAM model, which was 0.01 higher than the YOLO v7 model. The mAP value of the IMVTS model was 98.82%, which was 0.02% higher than the YOLO v7+CBAM and 1.36% higher than the YOLO v7 model. In summary, the IMVTS and YOLO v7+CBAM models were better than the detection effect of the YOLO v7 model, and the IMVTS model had the best detection effect. This is because the characteristic information of fresh tea leaf sprouts is easy to lose after processing. After the ECA module is introduced, it uses a dynamic convolution nucleus to perform 1*1 convolution; it can be used to extract feature information in different areas, avoiding the number of channel dimensions caused by learning channel attention information so that the model can more effectively learn the top-level information, improving the position prediction of the fresh tea sprouts, and at the same time, can reduce the model parameters, thereby increasing the model performance. For the CBAM module, although it combines the channel attention mechanism with the space attention mechanism, there is not enough space for the information-rich feature space. For small targets, such as fresh tea leaf sprouts, the actual effect is worse than the additional ECA module.

3.1.2. Comparison of IMVTS Model with Mainstream Target Detection Models

To verify the advantages of the IMVTS model for multi-species tea fresh leaf shoot detection, the IMVTS model was compared with four mainstream target detection models (YOLO v3, YOLO v5, FASTER-RCNN, and SSD). The comparison experiments of the models were trained and validated under the MVT dataset. The relevant parameters were kept consistent during the experiment, and the detection effect of the model was evaluated by P, R, F1-score, and mAP, and the results are shown in Table 2. The results show that the IMVTS model proposed in this study improves in P, R, F1 values, and mAP relative to the four mainstream target detection models. Among them, IMVTS improved P by 2.38% and 1.79%, R by 7.02% and 6.31%, F1-score by 0.06 and 0.04, and mAP by 4.88% and 4.37%, respectively, relative to YOLO v3 and YOLO v5. IMVTS has higher detection accuracy than YOLO v3 and YOLO v5 because it not only inherits the advantages of the original YOLO model but also achieves higher detection speed with the same computational resources because it uses a faster convolution operation and a smaller model.
IMVTS improved P by 2.92%, R by 26.84%, F1s by 0.07, and mAP by 9.54% relative to FASTER-RCNN. The reason for the improvement in detection accuracy is that FASTER-RCNN uses resnet50 as the backbone, and its feature map only comes from the top-level features, while the MVT dataset has small targets, and there are occlusions and blurring between targets, and only the features of the top-level of the network are used to predict the targets in a single way to extract information, which is not conducive to the localization of target frames. For IMVTS compared to SSD, P improved by 0.44%, R improved by 38.81%, F1score value improved by 0.26, and mAP improved by 12.9%. This is because the SSD model adopts a deep learning network of multi-scale characteristic fusion, but the use of low-level feature information is not enough. At the same time, the resolution of the SSD model is also low, resulting in an insufficient ability to recognize the small target SSD model, which is not conducive to conducting small target detection tasks.

3.1.3. Results and Analysis of Ablation Test

In order to verify the effectiveness of the IMVTS model proposed in this study, different optimization strategies (backbone, model size, and attention) were used in this study. An ablation test was conducted, and the comparative results were shown in Table 3. As can be seen from Table 3, the mAP of the original YOLO v7 model was 97.46%, and after the introduction of the ECA attention mechanism, the mAP increased by 1.36% and reached a peak of 98.82%. Therefore, YOLO v7 and ECA are taken as the basic framework to construct the IMVTS model proposed in this study. In addition, the ablation experiments of four mainstream target detection models (YOLO v3, YOLO v5, FASTER-RCNN, and SSD) were also conducted in this study. It can be seen from Table 3 that the IMVTS model (consisting of YOLO v7 and ECA) has the best detection effect.

3.1.4. IMVTS Model for VOC Dataset Detection Test

To further illustrate the effectiveness of the IMVTS model in practical engineering applications, this study uses the IMVTS model to train and test the VOC dataset, which includes images of birds, boats, buses, etc. The relevant parameters during the experiments were all kept the same as in the previous training of the MVT dataset, and the detection effect of the model was evaluated by P, R, F1-score, and mAP. Furthermore, a comparison test with YOLO v7 was conducted, and the comparison results are shown in Table 4. From Table 4, it can be seen that the IMVTS model has better detection results compared to the original YOLO v7 model.

3.2. Comparison of Model Recognition Effects on the MVT Dataset

To further understand the detection effect of the models for images of fresh leaf sprouts of different varieties of tea. The YOLO v7 and the IMVTS models were used to detect the three different varieties of fresh tea leaves in the selected images. The detection results are shown in Figure 8. It can be seen that the IMVTS model detected more fresh tea leaf sprouts than the YOLO v7 model and had a better detection effect.
Figure 9 shows that when the YOLO v7 original model detects the three varieties of fresh tea leaves, there is a misunderstanding phenomenon. This is because, in the process of generating fresh tea leaf datasets, multiple fresh tea leaves and sprouts often appear in the single picture taken, and these images have problems such as blurry, chaotic positions and overlapping blocks of sprouts. As a result, the selection of the fresh tea leaf sprouts by the frame of artificial labeling is not accurate enough, resulting in error detection of the model. The IMVTS model can effectively reduce the misunderstanding phenomenon, and the specific contrast is as follows.
The results of the original model of YOLO v7, YOLO v7+CBAM, and IMVTS are given in Table 5. In Table 5, it is shown that the AP value of ZC108 can reach up to 99.87%; the AP value of ZJ can reach up to 99.64%; and the AP value of ZH is 96.97%. It can also be seen in Table 2 that compared with the ZC108 and ZJ varieties, the three models have a worse detection effect with ZH. Considering that this study used the same model for all three varieties of fresh tea leaf sprouts, the main reason should be in the images of the fresh tea leaves. Therefore, fresh leaf sprouts and old leaf images of each of the three tea varieties were selected for RGB color gamut analysis. The average and square differences of the RGB of the fresh leaf sprouts and the old leaves of the three tea varieties are shown in Figure 9 and Figure 10. According to Figure 10 and Figure 11, it can be seen that the differences between the three channels of the sprouts and old leaves of ZC108 and the RGB of R, G, and B are 8.33, 30.82, and 31.02, respectively; the differences between the three channels R, G, and B of the square differences of stdRGB are 2.72, 4.79, and 8.04, respectively. The differences between the three channels of the sprouts and old leaves of ZJ and the RGB of R, G, and B are 42.75, 36.52, and 33.76, respectively; the differences between the three channels R, G, and B of the square differences of stdRGB are 15.1, 13.89, and 15.35, respectively. It can be seen that there is a relatively obvious RGB gap between ZC108 and ZJ sprouts and old leaves. It is convenient for the model to distinguish the sprouts and old leaves when the model is detected. The differences between the three channels of the sprouts and old leaves of ZH and the RGB of R, G, and B are 1.34, 1.69, and 5.91, respectively; the differences between the three channels R, G, and B of the square differences of stdRGB are 2.13, 6.11, and 0.14, respectively. It can be seen that the RGB of the fresh leaf sprouts of ZH tea is very low, so the detection accuracy of the model to the fresh leaves of ZH is low.

4. Conclusions

This study proposed an identification of multiple varieties of tea sprouts (IMVTS) model for the detection of fresh leaf sprouts of multiple tea species. The model training and validating dataset of multiple varieties of tea (MVT) consisted of images of three representative tea varieties (ZC108, ZH, and ZJ) in Zhejiang Province. To verify the correctness and advantages of the proposed model, IMVTS was compared with YOLO v7 and YOLO v7+CBAM. The results showed that the IMVTS model had the best detection effect, and the mean average precision (mAP) on the MVT dataset was 98.82%, which improved in comparison with YOLO v7 and YOLO v7+CBAM. In addition, this study also conducted comparison tests of the IMVTS model with mainstream target detection models (YOLO v3, YOLO v5, FASTER-RCNN, and SSD) and the IMVTS model on VOC datasets, and the results also demonstrated the superiority of the proposed IMVTS model. The average precision (AP) values of IMVTS for detecting ZC108, ZH, and ZJ tea leaves were 99.87%, 96.97%, and 99.64%, respectively. Among these, the ZH AP value was lower than those of ZJ and ZC108 because the difference in mean RGB and the difference in variance RGB between sprouts and old leaves in ZH images is smaller, and the colors of the object and background are close, making detection more difficult. In summary, the IMVTS model can improve the detection accuracy of fresh leaf sprouts of three varieties of tea, which can meet the requirements for the automatic picking of fresh leaves of the autumn famous tea and provide a basis for the future detection of fresh leaves of additional varieties of autumn tea. Future research will further focus on the improvement and design of the network structure as the IMVTS model detects objects such as ZH that are not clearly distinguished from the background.

Author Contributions

Writing—original draft preparation, R.Z.; methodology and software, C.L.; formal analysis, T.Y.; investigation, J.C.; data curation, Y.L.; supervision, G.L.; project administration, X.H.; writing—review and editing, Z.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the National Natural Science Foundation of China (Grant No. 52105284), Science and Technology Program of Meizhou, China (Grant No. 2021A0304004), Postdoctoral Science Foundation of China (Grant No. 2022M722819), and supported by China Agriculture Research System of MOF and MARA, and Key Laboratory of Crop Harvesting Equipment Technology of Zhejiang Province (Grant No. 2021KZ04). We also thank the anonymous reviewers for their critical comments and suggestions for improving the manuscript.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data is available on request due to privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, Z. Design and Experiment of Intelligent Picking Robot for Famous Tea. Ph.D. Thesis, Shenyang University of Technology, Shenyang, China, 2020. [Google Scholar]
  2. Zhou, Y.; Wu, Q.; He, L.; Zhao, R.; Jia, J.; Chen, J.; Wu, C. Design and experiment of intelligent picking robot for famous tea. J. Mech. Eng. 2022, 58, 12–23. [Google Scholar]
  3. Zhang, L.; Zou, L.; Wu, C.; Jia, J.; Chen, J. Method of famous tea sprout identification and segmentation based on improved watershed algorithm. Comput. Electron. Agric. 2021, 184, 106108. [Google Scholar] [CrossRef]
  4. Zhang, L.; Zhan, H.; Chen, Y.; Dai, S.; Li, X.; Kenji, I.; Liu, Z.; Li, M. Real-time monitoring of optimum timing for harvesting fresh tea leaves based on machine vision. Int. J. Agric. Biol. Eng. 2019, 12, 6–9. [Google Scholar] [CrossRef]
  5. Yan, L.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar]
  6. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; IEEE: Washington, DC, USA, 2015; pp. 1440–1448. [Google Scholar]
  7. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 4 June 2015; IEEE: Washington, DC, USA, 2015; pp. 91–99. [Google Scholar]
  8. Redmon, J.; Divvala, S.; Girshick, R. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, CA, USA, 8 June 2015; IEEE: Washington, DC, USA, 2016; pp. 779–788. [Google Scholar]
  9. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.-C. SSD: Single Shot Multibox Detector. European Conference on Computer Vision; Springer: Cham, Switzerland, 2016. [Google Scholar]
  10. Li, Y.; He, L.; Jia, J.; Lv, J.; Chen, J.; Qiao, X.; Wu, C. In-field tea shoot detection and 3D localization using an RGB-D camera. Comput. Electron. Agric. 2021, 185, 106149. [Google Scholar] [CrossRef]
  11. Xu, W.; Zhao, L.; Li, J.; Shang, S.; Ding, X.; Wang, T. Detection and classification of tea buds based on deep learning. Comput. Electron. Agric. 2022, 192, 106547. [Google Scholar] [CrossRef]
  12. Chen, Y.-T.; Chen, S.-F. Localizing plucking points of tea leaves using deep convolutional neural networks. Comput. Electron. Agric. 2020, 171, 105298. [Google Scholar] [CrossRef]
  13. Liu, Y.; Cao, X.; Guo, B.; Chen, H.J.; Dai, Z.; Gong, C. Research on detection algorithm about the posture of meat goose in complex scene based on improved YOLO v5. J. Nanjing Agric. Univ. 2022, 32, 1544–1548. [Google Scholar]
  14. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. arXiv 2022, arXiv:2207.02696. [Google Scholar]
  15. Vaswani, A.; Shazier, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhinl, I. Attention is all you need. In Proceedings of the NIPS’17: Proceedings of the 31st International Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  16. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Washington, DC, USA, 2018; pp. 7132–7141. [Google Scholar]
  17. Woo, S.; Park, J.; Lee, J.-Y. CBAM: Convolutional block attention module. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–19. [Google Scholar]
  18. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2022; IEEE: Washington, DC, USA, 2022; pp. 11531–11539. [Google Scholar]
  19. Qi, L.; Gao, J. Small Object Detection Based on Improved YOLOv7. Comput. Eng. 2022, 1000–3428. [Google Scholar] [CrossRef]
  20. Yu, L.; Huang, C.; Tang, J.; Huang, H.; Zhou, Y.; Huang, Y.; Sun, J. Tea Bud Recognition Method Based on Improved YOLOX Model. Guangdong Agric. Sci. 2022, 49, 49–56. [Google Scholar]
Figure 1. Experimental Base of the Tea Research Institute of the Chinese Academy of Agricultural Sciences.
Figure 1. Experimental Base of the Tea Research Institute of the Chinese Academy of Agricultural Sciences.
Horticulturae 09 00819 g001
Figure 2. MVT dataset images; (a) ZC108; (b) ZH; (c) ZJ.
Figure 2. MVT dataset images; (a) ZC108; (b) ZH; (c) ZJ.
Horticulturae 09 00819 g002
Figure 3. MVT dataset enhancement examples. (a,e,i) Image of primitive fresh tea leaves; (b,f,j) Image of fresh tea leaves with increased or decreased brightness; (c,g,k) Image of fresh tea leaves with increased or decreased color saturation; (d,h,l) Image of fresh tea leaves with the horizontal flip.
Figure 3. MVT dataset enhancement examples. (a,e,i) Image of primitive fresh tea leaves; (b,f,j) Image of fresh tea leaves with increased or decreased brightness; (c,g,k) Image of fresh tea leaves with increased or decreased color saturation; (d,h,l) Image of fresh tea leaves with the horizontal flip.
Horticulturae 09 00819 g003aHorticulturae 09 00819 g003b
Figure 4. IMVTS model network structure diagram.
Figure 4. IMVTS model network structure diagram.
Horticulturae 09 00819 g004
Figure 5. Algorithm workflow diagram.
Figure 5. Algorithm workflow diagram.
Horticulturae 09 00819 g005
Figure 6. CBAM and ECA module structure diagram.
Figure 6. CBAM and ECA module structure diagram.
Horticulturae 09 00819 g006
Figure 7. YOLO v7 original model, YOLO v7+CBAM model, and IMVTS model training loss curves.
Figure 7. YOLO v7 original model, YOLO v7+CBAM model, and IMVTS model training loss curves.
Horticulturae 09 00819 g007
Figure 8. Comparison of YOLO v7 and IMVTS detection results; (a) ZC108; (b) ZH; (c) ZJ.
Figure 8. Comparison of YOLO v7 and IMVTS detection results; (a) ZC108; (b) ZH; (c) ZJ.
Horticulturae 09 00819 g008aHorticulturae 09 00819 g008b
Figure 9. Errors in the detection of three kinds of fresh tea leaves; (a) ZC108; (b) ZH; (c) ZJ.
Figure 9. Errors in the detection of three kinds of fresh tea leaves; (a) ZC108; (b) ZH; (c) ZJ.
Horticulturae 09 00819 g009aHorticulturae 09 00819 g009b
Figure 10. Average RGB of sprouts and old leaves of three varieties of tea.
Figure 10. Average RGB of sprouts and old leaves of three varieties of tea.
Horticulturae 09 00819 g010
Figure 11. RGB variance stdRGB of sprouts and old leaves of three varieties of tea.
Figure 11. RGB variance stdRGB of sprouts and old leaves of three varieties of tea.
Horticulturae 09 00819 g011
Table 1. Model performance indicator assessment comparison.
Table 1. Model performance indicator assessment comparison.
ModelP (%)R (%)F1-ScoremAP (%)
YOLO v799.5294.570.9797.46
YOLO v7+CBAM99.6696.930.9898.80
IMVTS99.7697.030.9898.82
Table 2. Results of the model comparison test.
Table 2. Results of the model comparison test.
ModelP (%)R (%)F1-ScoremAP (%)
IMVTS99.7697.030.9898.82
YOLO v397.3890.010.9293.94
YOLO v597.9790.720.9494.45
FASTER-RCNN96.8470.190.8989.28
SSD99.3258.220.7285.92
Table 3. Results of ablation experiments.
Table 3. Results of ablation experiments.
ModelBackboneModel SizeAttentionmAP (%)
ECA98.82
YOLO v7CBAM98.80
NONE97.46
YOLO v7 ECA97.86
YOLO v7_xCBAM97.73
NONE96.58
ECA94.26
YOLO v3 CBAM94.20
NONE93.94
ECA96.03
YOLO v5_lCBAM95.66
NONE94.45
Cspdarknet ECA95.83
YOLO v5_xCBAM95.72
NONE94.51
YOLO v5 ECA94.28
YOLO v5_lCBAM94.23
NONE94.05
Convnext_tiny ECA94.41
YOLO v5_xCBAM94.30
NONE94.12
ECA90.24
Resnet50 CBAM89.76
NONE89.28
FASTER-RCNN ECA87.77
Vgg CBAM87.59
NONE86.33
ECA88.06
Vgg CBAM87.24
NONE85.92
SSD ECA85.42
Mobilenetv2 CBAM84.11
NONE83.31
Table 4. Model comparison results.
Table 4. Model comparison results.
ModelP (%)R (%)F1-ScoremAP (%)
IMVTS87.7972.200.7983.85
YOLO v787.6062.080.7277.90
Table 5. Comparison of the AP results of different varieties of fresh tea leaves.
Table 5. Comparison of the AP results of different varieties of fresh tea leaves.
ModelZC108 (AP%)ZH (AP%)ZJ (AP%)
YOLO v799.7993.3099.30
YOLO v7+CBAM99.8796.8999.64
IMVTS99.8796.9799.64
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, R.; Liao, C.; Yu, T.; Chen, J.; Li, Y.; Lin, G.; Huan, X.; Wang, Z. IMVTS: A Detection Model for Multi-Varieties of Famous Tea Sprouts Based on Deep Learning. Horticulturae 2023, 9, 819. https://doi.org/10.3390/horticulturae9070819

AMA Style

Zhao R, Liao C, Yu T, Chen J, Li Y, Lin G, Huan X, Wang Z. IMVTS: A Detection Model for Multi-Varieties of Famous Tea Sprouts Based on Deep Learning. Horticulturae. 2023; 9(7):819. https://doi.org/10.3390/horticulturae9070819

Chicago/Turabian Style

Zhao, Runmao, Cong Liao, Taojie Yu, Jianneng Chen, Yatao Li, Guichao Lin, Xiaolong Huan, and Zhiming Wang. 2023. "IMVTS: A Detection Model for Multi-Varieties of Famous Tea Sprouts Based on Deep Learning" Horticulturae 9, no. 7: 819. https://doi.org/10.3390/horticulturae9070819

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop