Figure 1.
Flowchart of the training and testing pipeline for tea leaf disease image identification.
Figure 1.
Flowchart of the training and testing pipeline for tea leaf disease image identification.
Figure 2.
Overall architecture of EfficientNetB0, highlighting successive MBConv layers for hierarchical feature extraction in tea leaf disease images.
Figure 2.
Overall architecture of EfficientNetB0, highlighting successive MBConv layers for hierarchical feature extraction in tea leaf disease images.
Figure 3.
Structure of MBConv modules. (a) MBConv1 with depthwise separable convolutions and SE attention; (b) MBConv6 with expanded channels and richer feature extraction.
Figure 3.
Structure of MBConv modules. (a) MBConv1 with depthwise separable convolutions and SE attention; (b) MBConv6 with expanded channels and richer feature extraction.
Figure 4.
SE attention module architecture. It illustrates the channel recalibration mechanism with global average pooling and two fully connected layers.
Figure 4.
SE attention module architecture. It illustrates the channel recalibration mechanism with global average pooling and two fully connected layers.
Figure 5.
Modified EfficientNetB0 (m-EfficientNetB0) architecture. Gray marks indicate the reduction in Stage 7 layers and the removal of Stage 8 to decrease complexity.
Figure 5.
Modified EfficientNetB0 (m-EfficientNetB0) architecture. Gray marks indicate the reduction in Stage 7 layers and the removal of Stage 8 to decrease complexity.
Figure 6.
ECA attention module architecture. It demonstrates lightweight channel attention using a 1D convolutional operation without any dimensionality reduction.
Figure 6.
ECA attention module architecture. It demonstrates lightweight channel attention using a 1D convolutional operation without any dimensionality reduction.
Figure 7.
Sample images from the Tea_Leaf_Disease dataset. The dataset contains five disease categories and one healthy class. Representative samples are shown: (a) Algal Spot, (b) Brown Blight, (c) Gray Blight, (d) Healthy, (e) Helopeltis, and (f) Red Spot. These images highlight the diversity in lesion shapes, colors, and textures.
Figure 7.
Sample images from the Tea_Leaf_Disease dataset. The dataset contains five disease categories and one healthy class. Representative samples are shown: (a) Algal Spot, (b) Brown Blight, (c) Gray Blight, (d) Healthy, (e) Helopeltis, and (f) Red Spot. These images highlight the diversity in lesion shapes, colors, and textures.
Figure 8.
Sample images from the teaLeafBD dataset. This dataset contains six common tea leaf diseases and one healthy class. Representative samples are shown: (a) Algal Spot, (b) Brown Blight, (c) Gray Blight, (d) Helopeltis, (e) Red Spider, (f) Green Mirid Bug, and (g) healthy leaf. The images illustrate the visual diversity of lesions, such as irregular brown patches, concentric blight spots, insect-induced feeding damage, and variations in leaf color and texture.
Figure 8.
Sample images from the teaLeafBD dataset. This dataset contains six common tea leaf diseases and one healthy class. Representative samples are shown: (a) Algal Spot, (b) Brown Blight, (c) Gray Blight, (d) Helopeltis, (e) Red Spider, (f) Green Mirid Bug, and (g) healthy leaf. The images illustrate the visual diversity of lesions, such as irregular brown patches, concentric blight spots, insect-induced feeding damage, and variations in leaf color and texture.
Figure 9.
Confusion matrix for the predictions on the Tea_Leaf_Disease testing set.
Figure 9.
Confusion matrix for the predictions on the Tea_Leaf_Disease testing set.
Figure 10.
Misclassified images from Tea_Leaf_Disease dataset. (a) Correct Gray Blight; (b) Brown Blight misclassified as Gray Blight due to similar dark lesions.
Figure 10.
Misclassified images from Tea_Leaf_Disease dataset. (a) Correct Gray Blight; (b) Brown Blight misclassified as Gray Blight due to similar dark lesions.
Figure 11.
Confusion matrix for the predictions on the teaLeafBD testing set.
Figure 11.
Confusion matrix for the predictions on the teaLeafBD testing set.
Figure 12.
Misclassified teaLeafBD images—Part I. Examples include (a) correctly classified Tea Algal Leaf Spot, (b) Brown Blight misclassified as Tea Algal Leaf Spot, (c) correctly classified Gray Blight, and (d) Brown Blight misclassified as Gray Blight.
Figure 12.
Misclassified teaLeafBD images—Part I. Examples include (a) correctly classified Tea Algal Leaf Spot, (b) Brown Blight misclassified as Tea Algal Leaf Spot, (c) correctly classified Gray Blight, and (d) Brown Blight misclassified as Gray Blight.
Figure 13.
Misclassified teaLeafBD images—Part II. (a) Correct Gray Blight and (b) Red Spider misclassified as Gray Blight due to similar spot morphology.
Figure 13.
Misclassified teaLeafBD images—Part II. (a) Correct Gray Blight and (b) Red Spider misclassified as Gray Blight due to similar spot morphology.
Table 1.
The various hyperparameter settings used in the proposed approach.
Table 1.
The various hyperparameter settings used in the proposed approach.
| Hyperparameter | Value/Method |
|---|
| Epoch | 100 |
| Batch size | 16 |
| Learning rate | 0.001 |
| Optimizer | Adam |
| Loss function | Cross-Entropy |
| Patience in ReduceLROnPlateau | 15 |
| Reduction factor in ReduceLROnPlateau | 0.5 |
Table 2.
Category distribution of tea leaf disease image datasets. The Tea_Leaf_Disease dataset includes six categories, while the teaLeafBD dataset contains seven categories.
Table 2.
Category distribution of tea leaf disease image datasets. The Tea_Leaf_Disease dataset includes six categories, while the teaLeafBD dataset contains seven categories.
| Disease Symptom | Tea_Leaf_Disease | teaLeafBD |
|---|
| Algal Spot | 1000 | 418 |
| Brown Blight | 867 | 506 |
| Gray Blight | 1000 | 1013 |
| Healthy | 1000 | 935 |
| Helopeltis | 1000 | 607 |
| Red Spot | 1000 | NA |
| Red Spider | NA | 515 |
| Green Mirid Bug | NA | 1282 |
| Total | 5867 | 5276 |
Table 3.
Image enhancement techniques used for the training phase.
Table 3.
Image enhancement techniques used for the training phase.
| Image Enhancement | Detailed Description |
|---|
| Random vertical and horizontal flip | Each applied with a 50% probability |
| Random rotation | range |
| Random crop | 80–100% of original area |
| Random affine transformation | Random translation 0–10% of image size |
| Random brightness adjustment | range |
Table 4.
Average 10-fold cross-validation accuracy and standard deviation with/without image enhancement on the Tea_Leaf_Disease dataset during the training phase.
Table 4.
Average 10-fold cross-validation accuracy and standard deviation with/without image enhancement on the Tea_Leaf_Disease dataset during the training phase.
| Image Enhancement | Average Accuracy (%) | Std. Dev. |
|---|
| No | 98.51 | 0.32 |
| Yes | 99.57 | 0.39 |
Table 5.
EfficientNetB0 performance on the Tea_Leaf_Disease testing set with/without image enhancement.
Table 5.
EfficientNetB0 performance on the Tea_Leaf_Disease testing set with/without image enhancement.
| Image Enhancement | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
|---|
| No | 98.21 | 98.20 | 98.15 | 98.17 |
| Yes | 99.23 | 99.21 | 99.24 | 99.22 |
Table 6.
Average 10-fold cross-validation accuracy and standard deviation with/without image enhancement on the teaLeafBD dataset during the training phase.
Table 6.
Average 10-fold cross-validation accuracy and standard deviation with/without image enhancement on the teaLeafBD dataset during the training phase.
|
Image Enhancement | Average Accuracy (%) | Std. Dev. |
|---|
| No | 81.36 | 1.89 |
| Yes | 91.94 | 1.33 |
Table 7.
EfficientNetB0 performance on the teaLeafBD testing set with/without image enhancement.
Table 7.
EfficientNetB0 performance on the teaLeafBD testing set with/without image enhancement.
| Image Enhancement | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
|---|
| No | 77.50 | 75.13 | 74.55 | 74.84 |
| Yes | 88.56 | 87.68 | 85.93 | 86.80 |
Table 8.
Comparison of architectures based on average 10-fold cross-validation accuracy and standard deviation on the Tea_Leaf_Disease dataset during the training phase.
Table 8.
Comparison of architectures based on average 10-fold cross-validation accuracy and standard deviation on the Tea_Leaf_Disease dataset during the training phase.
| Architecture | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| EfficientNetB0 | 99.57 | 0.39 | 49,763 |
| EfficientNetB0 ⊕ Red. S7 | 99.64 | 0.15 | 46,426 |
| EfficientNetB0 − S8 | 99.66 | 0.30 | 48,048 |
| m-EfficientNetB0 | 99.81 | 0.19 | 45,709 |
Table 9.
EfficientNetB0 variant comparison on the Tea_Leaf_Disease testing set.
Table 9.
EfficientNetB0 variant comparison on the Tea_Leaf_Disease testing set.
| Architecture | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| EfficientNetB0 | 99.23 | 99.21 | 99.24 | 99.22 | 0.0170 |
| EfficientNetB0 ⊕ Red. S7 | 99.23 | 99.20 | 99.23 | 99.21 | 0.0119 |
| EfficientNetB0 − S8 | 99.32 | 99.29 | 99.33 | 99.31 | 0.0136 |
| m-EfficientNetB0 | 99.40 | 99.37 | 99.40 | 99.38 | 0.0113 |
Table 10.
Comparison of architectures based on average 10-fold cross-validation accuracy and standard deviation on the teaLeafBD dataset during the training phase.
Table 10.
Comparison of architectures based on average 10-fold cross-validation accuracy and standard deviation on the teaLeafBD dataset during the training phase.
| Architecture | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| EfficientNetB0 | 91.94 | 1.33 | 48,219 |
| EfficientNetB0 ⊕ Red. S7 | 92.34 | 1.09 | 45,302 |
| EfficientNetB0 − S8 | 91.92 | 0.90 | 47,057 |
| m-EfficientNetB0 | 91.87 | 1.40 | 44,476 |
Table 11.
EfficientNetB0 variant comparison on the teaLeafBD testing set.
Table 11.
EfficientNetB0 variant comparison on the teaLeafBD testing set.
| Architecture | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| EfficientNetB0 | 88.56 | 87.68 | 85.93 | 86.80 | 0.0171 |
| EfficientNetB0 ⊕ Red. S7 | 88.37 | 86.67 | 87.55 | 87.11 | 0.0122 |
| EfficientNetB0 − S8 | 88.37 | 86.67 | 86.37 | 86.52 | 0.0136 |
| m-EfficientNetB0 | 89.32 | 87.33 | 86.61 | 87.07 | 0.0114 |
Table 12.
Model complexity of different EfficientNetB0 variants.
Table 12.
Model complexity of different EfficientNetB0 variants.
| Architecture | GFLOPs | Size (MB) | Num. of Parameters |
|---|
| EfficientNetB0 | 0.413874 | 15.60 | 4,016,515 |
| EfficientNetB0 ⊕ Red. S7 | 0.342640 | 8.79 | 2,252,659 |
| EfficientNetB0 − S8 | 0.370807 | 11.82 | 3,032,531 |
| m-EfficientNetB0 | 0.299574 | 5.01 | 1,268,675 |
Table 13.
Average 10-fold cross-validation accuracy and standard deviation of our method with various attention modules on the Tea_Leaf_Disease dataset during the training phase.
Table 13.
Average 10-fold cross-validation accuracy and standard deviation of our method with various attention modules on the Tea_Leaf_Disease dataset during the training phase.
| Method | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| m-EfficientNetB0 | 99.81 | 0.19 | 45,709 |
| m-EfficientNetB0 + CBAM | 98.96 | 0.63 | 51,219 |
| m-EfficientNetB0 + CA | 99.60 | 0.32 | 49,533 |
| m-EfficientNetB0 + MCA | 99.70 | 0.18 | 59,521 |
| m-EfficientNetB0 + SA | 99.57 | 0.30 | 49,330 |
| m-EfficientNetB0 + ECA | 99.81 | 0.21 | 47,353 |
Table 14.
Comparison of m-EfficientNetB0 with different attention modules on the Tea_Leaf_Disease testing set.
Table 14.
Comparison of m-EfficientNetB0 with different attention modules on the Tea_Leaf_Disease testing set.
| Method | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| m-EfficientNetB0 | 99.40 | 99.37 | 99.40 | 99.38 | 0.0113 |
| m-EfficientNetB0 + CBAM | 98.38 | 98.33 | 98.38 | 98.35 | 0.0187 |
| m-EfficientNetB0 + CA | 99.06 | 99.04 | 99.08 | 99.06 | 0.0160 |
| m-EfficientNetB0 + MCA | 98.89 | 98.84 | 98.92 | 98.88 | 0.0317 |
| m-EfficientNetB0 + SA | 98.98 | 98.92 | 98.99 | 98.95 | 0.0163 |
| m-EfficientNetB0 + ECA | 99.49 | 99.49 | 99.48 | 99.48 | 0.0098 |
Table 15.
Average 10-fold cross-validation accuracy and standard deviation of our method with various attention modules on the teaLeafBD dataset during the training phase.
Table 15.
Average 10-fold cross-validation accuracy and standard deviation of our method with various attention modules on the teaLeafBD dataset during the training phase.
| Method | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| m-EfficientNetB0 | 91.87 | 1.40 | 44,476 |
| m-EfficientNetB0 + CBAM | 89.97 | 1.61 | 48,414 |
| m-EfficientNetB0 + CA | 92.25 | 1.09 | 48,864 |
| m-EfficientNetB0 + MCA | 92.20 | 1.02 | 56,191 |
| m-EfficientNetB0 + SA | 92.27 | 1.82 | 47,141 |
| m-EfficientNetB0 + ECA | 92.27 | 1.54 | 45,343 |
Table 16.
Comparison of m-EfficientNetB0 with different attention modules on the teaLeafBD testing set.
Table 16.
Comparison of m-EfficientNetB0 with different attention modules on the teaLeafBD testing set.
| Method | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| m-EfficientNetB0 | 89.32 | 87.33 | 86.61 | 87.07 | 0.0114 |
| m-EfficientNetB0 + CBAM | 88.85 | 88.02 | 86.05 | 87.02 | 0.0182 |
| m-EfficientNetB0 + CA | 88.85 | 87.37 | 86.58 | 86.97 | 0.0142 |
| m-EfficientNetB0 + MCA | 88.75 | 86.77 | 88.36 | 87.56 | 0.0268 |
| m-EfficientNetB0 + SA | 89.51 | 87.88 | 88.51 | 88.19 | 0.0138 |
| m-EfficientNetB0 + ECA | 90.73 | 89.97 | 88.51 | 89.23 | 0.0095 |
Table 17.
Model complexity comparison of m-EfficientNetB0 with different attention modules.
Table 17.
Model complexity comparison of m-EfficientNetB0 with different attention modules.
| Method | GFLOPs | Size (MB) | Num. of Parameters |
|---|
| m-EfficientNetB0 | 0.299574 | 5.01 | 1,268,675 |
| m-EfficientNetB0 + CBAM | 0.300331 | 5.34 | 1,356,127 |
| m-EfficientNetB0 + CA | 0.304910 | 5.89 | 1,494,759 |
| m-EfficientNetB0 + MCA | 0.303574 | 4.30 | 1,079,427 |
| m-EfficientNetB0 + SA | 0.298346 | 4.30 | 1,079,303 |
| m-EfficientNetB0 + ECA | 0.299402 | 4.27 | 1,079,339 |
Table 18.
Average 10-fold cross-validation accuracy and standard deviation with/without ReduceLROnPlateau on the Tea_Leaf_Disease dataset during the training phase.
Table 18.
Average 10-fold cross-validation accuracy and standard deviation with/without ReduceLROnPlateau on the Tea_Leaf_Disease dataset during the training phase.
| ReduceLROnPlateau | Average Accuracy (%) | Std. Dev. |
|---|
| No | 99.47 | 0.27 |
| Yes | 99.81 | 0.19 |
Table 19.
Effect of ReduceLROnPlateau on the Tea_Leaf_Disease testing set.
Table 19.
Effect of ReduceLROnPlateau on the Tea_Leaf_Disease testing set.
| ReduceLROnPlateau | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
|---|
| No | 98.81 | 98.79 | 98.82 | 98.80 |
| Yes | 99.49 | 99.49 | 99.48 | 99.48 |
Table 20.
Average 10-fold cross-validation accuracy and standard deviation with/without ReduceLROnPlateau on the teaLeafBD dataset during the training phase.
Table 20.
Average 10-fold cross-validation accuracy and standard deviation with/without ReduceLROnPlateau on the teaLeafBD dataset during the training phase.
| ReduceLROnPlateau | Average Accuracy (%) | Std. Dev. |
|---|
| No | 91.75 | 1.41 |
| Yes | 92.27 | 1.54 |
Table 21.
Effect of ReduceLROnPlateau on the teaLeafBD testing set.
Table 21.
Effect of ReduceLROnPlateau on the teaLeafBD testing set.
| ReduceLROnPlateau | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) |
|---|
| No | 87.24 | 86.12 | 85.39 | 85.75 |
| Yes | 90.73 | 89.97 | 88.51 | 89.23 |
Table 22.
Average 10-fold cross-validation accuracy and standard deviation of different lightweight models on the Tea_Leaf_Disease dataset during the training phase.
Table 22.
Average 10-fold cross-validation accuracy and standard deviation of different lightweight models on the Tea_Leaf_Disease dataset during the training phase.
| Method | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| SqueezeNet 1.1 | 97.59 | 1.07 | 37,356 |
| MNASNet1_3 | 99.11 | 0.31 | 48,445 |
| ShuffleNetV2_x2_0 | 99.42 | 0.35 | 43,440 |
| MobileNetV4_conv_small | 98.79 | 0.38 | 43,722 |
| RepGhostNet 1.0× | 99.42 | 0.34 | 48,627 |
| Ours | 99.81 | 0.19 | 47,353 |
Table 23.
Performance comparison of different lightweight models on the Tea_Leaf_Disease testing set.
Table 23.
Performance comparison of different lightweight models on the Tea_Leaf_Disease testing set.
| Method | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| SqueezeNet 1.1 | 95.82 | 95.80 | 95.88 | 95.84 | 0.0044 |
| MNASNet1_3 | 99.06 | 99.08 | 99.07 | 99.07 | 0.0099 |
| ShuffleNetV2_x2_0 | 98.98 | 98.94 | 98.99 | 98.96 | 0.0113 |
| MobileNetV4_conv_small | 97.53 | 97.53 | 97.50 | 97.51 | 0.0100 |
| RepGhostNet 1.0× | 99.15 | 99.14 | 99.15 | 99.14 | 0.0186 |
| Ours | 99.49 | 99.49 | 99.48 | 99.48 | 0.0098 |
Table 24.
Average 10-fold cross-validation accuracy and standard deviation of different lightweight models on the teaLeafBD dataset during the training phase.
Table 24.
Average 10-fold cross-validation accuracy and standard deviation of different lightweight models on the teaLeafBD dataset during the training phase.
| Method | Average Accuracy (%) | Std. Dev. | Training Time (s) |
|---|
| SqueezeNet 1.1 | 84.19 | 1.70 | 38,939 |
| MNASNet1_3 | 88.24 | 1.81 | 46,543 |
| ShuffleNetV2_x2_0 | 90.30 | 1.59 | 43,088 |
| MobileNetV4_conv_small | 87.48 | 1.26 | 42,385 |
| RepGhostNet 1.0× | 90.30 | 1.34 | 49,058 |
| Ours | 92.27 | 1.54 | 45,343 |
Table 25.
Performance comparison of different lightweight models on the teaLeafBD testing set.
Table 25.
Performance comparison of different lightweight models on the teaLeafBD testing set.
| Method | Accuracy (%) | Precision (%) | Recall (%) | F1-Score (%) | Inference Time (s) |
|---|
| SqueezeNet 1.1 | 79.49 | 80.82 | 73.74 | 77.12 | 0.0049 |
| MNASNet1_3 | 83.84 | 83.45 | 80.64 | 82.02 | 0.0114 |
| ShuffleNetV2_x2_0 | 88.28 | 86.73 | 86.02 | 86.37 | 0.0147 |
| MobileNetV4_conv_small | 84.97 | 83.60 | 83.06 | 83.33 | 0.0099 |
| RepGhostNet 1.0× | 88.00 | 87.40 | 85.70 | 86.54 | 0.0205 |
| Ours | 90.73 | 89.97 | 88.51 | 89.23 | 0.0095 |
Table 26.
Model complexity comparison of different lightweight models.
Table 26.
Model complexity comparison of different lightweight models.
| Method | GFLOPs | Size (MB) | Num. of Parameters |
|---|
| SqueezeNet 1.1 | 0.263231 | 2.78 | 726,087 |
| MNASNet1_3 | 0.554501 | 19.40 | 5,010,223 |
| ShuffleNetV2_x2_0 | 0.596196 | 20.68 | 5,359,339 |
| MobileNetV4_conv_small | 0.184783 | 9.74 | 2,476,903 |
| RepGhostNet 1.0× | 0.164001 | 11.05 | 2,801,571 |
| Ours | 0.299402 | 4.27 | 1,079,339 |