Next Article in Journal
DIKWP Semantic Judicial Reasoning: A Framework for Semantic Justice in AI and Law
Previous Article in Journal
Leveraging Machine Learning Techniques to Predict Cardiovascular Heart Disease
Previous Article in Special Issue
WTSM-SiameseNet: A Wood-Texture-Similarity-Matching Method Based on Siamese Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases

by
Mst. Tanbin Yasmin Tanny
1,
Tangina Sultana
1,2,
Md. Emran Biswas
1,
Chanchol Kumar Modok
1,
Arjina Akter
1,
Mohammad Shorif Uddin
3 and
Md. Delowar Hossain
2,4,*
1
Department of Electronics and Communication Engineering, Hajee Mohammad Danesh Science and Technology University, Dinajpur 5200, Bangladesh
2
Department of Computer Science and Engineering, Kyung Hee University, Yongin-si 17104, Republic of Korea
3
Department of Computer Science and Engineering, Jahangirnagar University, Dhaka 1342, Bangladesh
4
Department of Computer Science and Engineering, Hajee Mohammad Danesh Science and Technology University, Dinajpur 5200, Bangladesh
*
Author to whom correspondence should be addressed.
Information 2025, 16(8), 638; https://doi.org/10.3390/info16080638
Submission received: 2 June 2025 / Revised: 5 July 2025 / Accepted: 18 July 2025 / Published: 27 July 2025

Abstract

Jute, a vital lignocellulosic fiber crop with substantial industrial and ecological relevance, continues to suffer considerable yield and quality degradation due to pervasive foliar pathologies. Traditional diagnostic modalities reliant on manual field inspections are inherently constrained by subjectivity, diagnostic latency, and inadequate scalability across geographically distributed agrarian systems. To transcend these limitations, we propose DERIENet, a robust and scalable classification approach within a deep ensemble learning framework. It is meticulously engineered by integrating three high-performing convolutional neural networks—ResNet50, InceptionV3, and EfficientNetB0—along with regularization, batch normalization, and dropout strategies, to accurately classify jute leaf diseases such as Cercospora Leaf Spot, Golden Mosaic Virus, and healthy leaves. A key methodological contribution is the design of a novel augmentation pipeline, termed Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), which dynamically modulates photometric and geometric distortions based on image entropy and luminance to synthetically upscale a limited dataset (920 images) into a significantly enriched and diverse dataset of 7800 samples, thereby mitigating overfitting and enhancing domain generalizability. Empirical evaluation, utilizing a comprehensive set of performance metrics—accuracy, precision, recall, F1-score, confusion matrices, and ROC curves—demonstrates that DERIENet achieves a state-of-the-art classification accuracy of 99.89%, with macro-averaged and weighted average precision, recall, and F1-score uniformly at 99.89%, and an AUC of 1.0 across all disease categories. The reliability of the model is validated by the confusion matrix, which shows that 899 out of 900 test images were correctly identified and that there was only one misclassification. Comparative evaluations of the various ensemble baselines, such as DenseNet201, MobileNetV2, and VGG16, and individual base learners demonstrate that DERIENet performs noticeably superior to all baseline models. It provides a highly interpretable, deployment-ready, and computationally efficient architecture that is ideal for integrating into edge or mobile platforms to facilitate in situ, real-time disease diagnostics in precision agriculture.

1. Introduction

Jute is a naturally occurring, environmentally beneficial fiber that finds extensive usage in the textile industry as well as other industrial sectors. The Tiliaceae family includes the plants Corchorus capsularis L. and Corchorus olitorius L., which are the source of this highly economical vegetable fiber, which is second only to cotton [1]. A popular option in a world where environmental preservation is becoming more and more important, jute is highly acclaimed for its affordability as well as its sustainability and biodegradability. Jute is cultivated in Bangladesh, India, Nepal, China, Taiwan, Thailand, Vietnam, and Cambodia, among other countries. At least 87% of the world’s jute production comes from the Indo-Bangladesh region [2]. Despite India being the main producer, Bangladesh remains the world’s leading exporter of jute [3].
Jute is an environmentally degradable polymer that occurs naturally. Polyvinyl chloride, polystyrene, polyethylene, and polypropylene are examples of synthetic polymers that have excellent mechanical strength but are not biodegradable and have a significant negative impact on the environment. Jute, on the other hand, provides an environmentally responsible and sustainable substitute that lowers pollutants and aids in environmental preservation [4,5]. There are other uses for jute leaves outside of just their fiber. They have cultural value and are utilized in many different culinary preparations. Their anti-inflammatory and anti-cancer activities are among their other possible therapeutic qualities [6]. Numerous diseases and pests pose a global danger to jute production. Smallholder farmers, whose livelihoods rely significantly on successful harvests, could suffer serious consequences. The main diseases affecting jute leaves include Cercospora Leaf Spot, Powdery Mildew, and various viral infections such as Yellow Mosaic and Golden Mosaic. These diseases result in brownish lesions, mosaic patterns, yellowing, and defoliation of the leaves. These illnesses impair photosynthesis, lower the quality of fiber, cause growth retardation, and in extreme situations, can kill the plant, all of which have a major impact on yields. The main pests that harm jute are the jute caterpillar, jute weevil, and jute stem borer. These pests lower productivity by harming the plant’s leaves, stems, and fibers. Furthermore, it is crucial to detect and stop these jute pests and leaf diseases as soon as possible because jute is important for both the economy and environmental protection.
The traditional technique of plant disease identification relies on farmers or agricultural experts performing physical inspections; however, this method has major drawbacks. It is time-consuming and resource-intensive, which makes it inappropriate for extensive agricultural operations. A precise diagnosis necessitates expertise, which is sometimes unavailable, especially in rural locations. Additionally, visual assessments are prone to inaccuracies and are subjective, which leads to inconsistent diagnoses [7]. Implementing early disease detection and categorization utilizing the resources and technologies that farmers have access to can significantly reduce all the issues that have come up [8]. Traditional machine-learning techniques for the classification of plant diseases have been used to propose a number of solutions. Furthermore, by eliminating reliance on manually created features [9,10], deep-learning methods in agriculture have opened up new avenues for researchers with remarkable flexibility. Convolutional neural networks (CNNs) have recently featured as a potent tool for any classification problem since they can automatically extract significant characteristics from images without human oversight. Furthermore, a number of CNN architectures [11] have been successfully applied for classifying plant leaf diseases. These include DenseNet [12] for tomato leaf disease detection; AlexNet [13,14], which classified leaf diseases in rice, potato, and tomato; EfficientNet [15,16], demonstrating strong performance in classifying diseases in rice and sugarcane leaves; GoogLeNet [17], employed for rice leaf disease classification; and NASNet [18], used for a variety of plant leaf diseases. In addition, Residual Networks (ResNet) [19], applied to apple leaf diseases; MobileNet [20,21], supported leaf disease classification tasks in apples and beans; SqueezeNet [22], detected maize leaf diseases; Inception [23] was utilized for jute leaf diseases; and Visual Geometric Group (VGG) Networks [24] contributed to multi-crop leaf disease identification.
While significant progress has been made in leaf disease detection for various crops, research on jute leaf diseases remains limited, highlighting a critical gap in addressing jute crop health and productivity [25]. A major obstacle in this discipline is the lack of geographically and environmentally diverse data, which could potentially impact the generalizability of the model to various field conditions. Furthermore, the dependence on simple augmentation approaches and the scarcity of extensive annotated datasets may limit the models’ ability to adjust to the complex variability found in actual agricultural settings. This study aims to address these challenges by developing an effective image-based classification system capable of accurately identifying Cercospora Leaf Spot, Golden Mosaic, and Healthy jute leaves. However, the absence of publicly available datasets makes it difficult to classify jute leaf diseases, largely due to limited research in this field [26]. To mitigate this constraint, we utilized the “Jute Leaf Disease Detection” dataset available on Kaggle (https://www.kaggle.com/datasets/mdsaimunalam/jute-leaf-disease-detection, accessed on 3 March 2025) and a proposed data-augmentation technique, Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), to expand the size of the dataset and improve the model’s adaptability and resilience. To develop a strong classification scheme, we applied transfer learning (TL) and enhanced performance by introducing a Deep Ensemble of ResNet, Inception, and EfficientNet (DERIENet), our proposed ensemble model that integrates various deep-learning architectures. Utilizing criteria including accuracy, precision, recall, F1-score, ROC curve, and a confusion matrix, the model’s efficacy was evaluated. The DERIENet model’s superior performance over baseline models demonstrated its efficacy in identifying jute leaf diseases, according to the data.
According to our research, the DERIENet model enhances crop management and lowers losses by improving the accuracy of jute leaf disease diagnosis and giving agricultural professionals a dependable tool for early identification. Below is a summary of the main conclusions drawn from this study:
  • We introduced DERIENet, a collaborative ensemble deep-learning architecture designed to accurately classify jute leaf diseases. DERIENet achieves a classification accuracy of 99.89% by integrating the predictive capabilities of several refined deep-learning models—EfficientNetB0, ResNet50, and InceptionV3—into a unified framework enhanced with regularization and dropout layers, significantly improving resilience and robustness against noise and variability in agricultural images.
  • We proposed a novel data-augmentation pipeline, Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), which dynamically perturbs image geometry and photometry based on entropy-aware heuristics. This critical innovation enabled the synthetic expansion of a limited dataset from 920 to 7800 images, dramatically enhancing generalization capability and mitigating overfitting in data-constrained scenarios.
  • We designed a custom ensemble feature fusion strategy using global average pooling and dense layers with L2 regularization and batch normalization, which not only preserved discriminative features from individual CNN backbones but also facilitated a compact, low-variance representation, enabling higher classification fidelity with minimal computational overhead.
  • We conducted a comprehensive comparative evaluation against multiple state-of-the-art baseline models, including MobileNetV2, DenseNet201, InceptionV3, VGG16, ResNet50, and EfficientNetB0, conclusively demonstrating DERIENet’s superior performance across all standard classification metrics and its statistical significance through ablation and ensemble analysis. Moreover, we demonstrated the model’s practical feasibility for real-time deployment in precision agriculture, validating its efficacy through precision, recall, and F1-score metrics, supported by interpretability tools such as ROC curves and confusion matrices, underscoring its potential for deployment on mobile or edge computing devices for in-field jute disease detection.
This document’s structure is as follows: A comprehensive overview of recent studies on jute leaf disease prediction is provided in Section 2, which also presents the pertinent papers. In Section 3, the methodology is described, along with the modeling techniques used in our study and image preprocessing steps. The results and discussion are provided in Section 4, where we compare our proposed model with baseline methods and assess its efficacy. In Section 5, the results and their consequences are interpreted in detail. Section 6 concludes the work by summarizing our findings and offering suggestions for more research.

2. Literature Review

Due to the wide range of symptoms and the need for personal inspection, jute leaf disease detection is extremely difficult. To address these issues, however, recent advancements in machine-learning (ML) and deep-learning (DL) approaches have shown great promise. With an emphasis on the use of ML, DL, and federated learning (FL) techniques, this section examines a number of important works on jute disease and pest detection.
Rahman et al. [27] suggested using the Self-Generated Mask (SGM) technique to identify diseases by identifying the Region of Interest (ROI), to extract an image’s Region of Interest, in jute leaves and stems. A 1772-image, 10-class dataset was used for SGM-enabled CNN training, which reduced annotation time while producing results similar to human annotation. A popular method for improving model accuracy is ensemble learning. Using a real-time dataset, Daphal et al. [28] produced five types of sugarcane leaves: rust, red rot, foliar, yellow leaf, and healthy, with a total of 2569 images. Transfer-learning and ensemble approaches were used in a piece of comparative research to categorize the dataset for illnesses of the sugarcane leaves. With a spatially tuned deep CNN and a sequential CNN, the ensemble model achieved an accuracy of 86.53%. Patil et al. [29] presented "Rice Fusion," an inventive multisensory data integration architecture that diagnoses rice diseases by combining a CNN with a multilayer perceptron (MLP). The model was trained and evaluated using 3200 manually collected rice samples from four categories, and it demonstrated excellent outcomes with a test accuracy of 95.31%. Yigit et al. [30] indicated an SVM classifier for the straightforward binary categorization of sugarcane leaves into healthy and damaged categories. By analyzing the color and texture properties of the photos, the researchers achieved an accuracy of 92.91%. A CNN-based technique for identifying yellow mosaic and chlorosis in jute leaves was presented by Karim et al. [31]. It achieved 96% accuracy on a small dataset of 600 images without feature extraction or preprocessing, underscoring the need for more study and bigger datasets to improve model performance. YOLO-JD, a deep-learning model that incorporates advanced modules such as SPPM and DSCFEM, was proposed by Li et al. [32]. It demonstrated the effectiveness of advanced deep-learning architectures for precise disease identification by achieving a state-of-the-art mean average precision (mAP) of 96.63% when trained on a large dataset covering ten disease types. JuLeDI, a four-layer CNN model trained on 4740 images across three disease categories, was also developed by Uddin et al. [7]. It outperformed GPDCNN and SVM and achieved 96% accuracy, highlighting the usefulness of lightweight CNN models for plant disease diagnosis and their potential for practical applications. A deep-learning approach for jute leaf disease identification was developed by Akhand et al. [33], using a dataset of 1820 images. Using SVM, Random Forest, XGBoost, and custom CNN classifiers with DenseNet121, ResNet152, Xception, and VGG19 extractors, the model’s accuracy of 98.41% and AUC of 0.9988 showed how effective deep learning is in diagnosing diseases. RH Hridoy et al. [34] proposed JuteNet, a hybrid deep-learning model utilizing multi-scale feature fusion of Xception, InceptionResNetV2, and InceptionV3, to detect jute diseases and pests. Trained on jute leaf and stem images, JuteNet achieved 99.47% accuracy on a six-class test set of 2803 images, outperforming the individual models, which achieved 91.83%, 96.11%, and 98.86%, respectively. Kaushik et al. [35] used a dataset of 1820 annotated images from Kaggle to diagnose jute leaf diseases using a deep-learning technique based on the ResNet50 architecture. The dataset underwent preprocessing procedures like as rescaling, normalization, and data augmentation in order to reduce overfitting and improve model resilience. After completing these procedures, the improved model showed minimal misclassification rates, balanced precision and recall metrics, and an accuracy of 94%.
The detection of jute pests has been transformed by developments in DL and ML, which allow for accurate and timely identification to improve pest control. Pests have a major influence on the production of jute. Therefore, DL and ML-powered image-based detection systems assist farmers in promptly identifying infestations, minimizing the need for manual inspections, and increasing productivity. Using the VGG19 CNN architecture with transfer learning on the ImageNet database, Sourav et al. [36] detected four common jute pests with 95.86% accuracy and integrated the model into iOS and Android apps for real-world applications. On a dataset of 2200 pictures, Wang et al. [37] created PestDetector, a deep CNN model that achieved 99.18% training accuracy and 99.00% validation accuracy, and 98% test accuracy; nevertheless, the small sample size hampered its generalizability. A transfer-learning-based model called JutePestDetect was presented by Talukder et al. [38]. It was tested on 17 pest classes, each with about 380 pictures, and DenseNet201 had the best accuracy of 99%. The research, however, only examined early-stage detection and did not analyze potential dataset biases. A lightweight CNN model that exceeded current techniques with 99.43% accuracy was proposed by Rana et al. [39]. It was a workable solution for farmers, utilizing SHAP visualizations to understand predictions and Gradio for real-time accessibility. Ali et al. [40] developed transfer-learning-based models using a VCI-validated dataset of 7235 pictures from 17 pest classes; DenseNet201 had the highest accuracy of 97%.
Federated learning (FL), a solution to data privacy problems, has become a viable method for identifying jute diseases. Using FedAvg, Vats et al. [26] created a FL-based CNN model that achieved a macro-average precision of 73.41–90.72%; nevertheless, real-world deployment requires improvements. FL’s potential for decentralized learning was demonstrated by Bansal et al. [41], who used data from numerous users to categorize five jute leaf diseases. They achieved macro-average precision between 79.92% and 83.67%. Rajput et al. [42] improved this method by combining FL with CNN in several ecological zones, demonstrating its suitability for extensive crop health monitoring with a 98% accuracy rate on client OB5. In a similar study, [43] achieved 98% accuracy using federated averaging with CNNs in decentralized data sets, while improving precision, recall, and F1-score, confirming the efficacy of FL in the classification of diseases while preserving data security and scalability. Conventional methods and resistance breeding have been investigated as alternatives to ML and DL procedures in the fight against jute diseases. For resistance to stem rot (Macrophomina phaseolina), a significant jute disease, Rijal et al. [44] assessed six Corchorus olitorius jute germplasm lines. According to disease rating scales of 0.73, 0.68, and 0.63, JRO-524 has continuously shown strong resistance over a three-year period (2019–2021), making it a viable option for resistance breeding. Likewise, Reza et al. [45] created an Android-based system for jute stem disease identification that uses hue-based segmentation and a Multi-SVM classifier. Although it produces findings in less than three seconds, the accuracy of the system is dependent on a pre-defined database, which could reduce its usefulness if the database is not diverse or indicative of real-world circumstances. Table 1 provides a thorough overview of the research findings in the field of plant leaf disease identification, with a focus on various architectural approaches.
Our suggested DERIENet model employs an ensemble approach that combines InceptionV3, ResNet50, and EfficientNetB0 to detect jute leaf diseases. Through the positive traits of each architecture, the models are combined to enhance classification performance. We begin with 920 high-quality images from the Kaggle jute leaf dataset and use an advanced data-augmentation technique (GLOAR) to reach 7800 images and overcome the problem of limited data. Better robustness and flexibility across a range of environmental variables are ensured by the larger dataset size. The designation of jute leaf diseases with a high accuracy of 99.89% is achieved by the DERIENet model, which has been refined by transfer learning.

3. Methodology

The DERIENet architecture utilizes jute leaf images as input to accurately predict their respective class labels. The procedure begins with collecting a dataset of jute plant images, then proceeds to preprocessing stages such as data augmentation and scaling to a uniform size. These actions improve the size, diversity, and quality of the dataset. After processing, the dataset is split into subsets for testing, validation, and training purposes. A number of baseline models, including MobileNetV2, DenseNet201, InceptionV3, VGG16, ResNet50, and EfficientNetB0, are refined through the integration of transfer learning. To enhance classification performance, the DERIENet model is introduced, incorporating models such as InceptionV3, ResNet50, and EfficientNetB0. According to a comparative study, the suggested ensemble technique is superior to the baseline models and achieves the highest classification accuracy, demonstrating its efficacy in detecting jute leaf diseases. Figure 1 has an illustration of the general structure of the suggested method.

3.1. Data Description

The dataset employed in this research, titled “Jute Leaf Disease Detection,” can be accessed at https://www.kaggle.com/datasets/mdsaimunalam/jute-leaf-disease-detection, accessed on 5 March 2025. It comprises 920 high-resolution images that are divided into three classes (depicted in Figure 2): Cercospora Leaf Spot, Golden Mosaic, and Healthy Leaf. With the full consent of local agricultural officials, these images were first collected during an extensive field trip in Dinajpur and Brahmanbaria, Bangladesh. Using the proposed data-augmentation technique (GLOAR), the total number of images was increased to 7800 to improve model transferability and enhance the dataset’s diversity.
In particular, the dataset includes 2604 images (33.385%) for the Golden Mosaic and Healthy leaf classes, and 2592 images (33.230%) for Cercospora Leaf Spot. The dataset’s distribution of categories after augmentation is presented in Figure 3.

3.2. Data Preprocessing

A crucial phase in obtaining raw data for ML application is data preparation; ensuring that the dataset is organized, consistent, and free from noise enhances the model’s capability for learning and generalizing it [46,47]. Data augmentation and resizing were two important preprocessing methods used in this work to enhance the dataset’s usability.

3.2.1. Data Resizing

To make sure that all input images have consistent dimensions and are compatible with the neural network model, image resizing is a crucial preprocessing step. Every image included in this study was reduced to 224 × 224 pixels, which was the fixed resolution. By ensuring uniformity throughout the dataset, this scaling procedure helps the model handle photos efficiently for both training and testing. The complete instructions for resizing images are provided by Algorithm 1.
Algorithm 1 Image resizing algorithm
Require: List of image files I, target dimensions w × h (e.g., 224 × 224 pixels)
Ensure: Resized images saved in appropriate directories
1: for each image i m g I  do
2:        Open the image file
3:         if image is not in RGB format then
4:             Convert the image to RGB
5:         end if
6:        Resize the image to w × h
7:        Save the resized image to the output directory
8:end for

3.2.2. Data Augmentation

An innovative augmentation technique called Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), described in Algorithm 2, is deployed to enhance the efficacy of the model under a variety of input situations. By creating a controlled cascade of transformations, this technique greatly increases the dataset’s size and produces a huge amount of visual variation. Context-aware transformation is made possible by the GLOAR algorithm, which dynamically modifies the augmentation strength according to the global brightness and entropy of each image. Among the geometric augmentations are translations (up to 15 % ), perspective warping, and random affine transformations (rotations up to 30 ° , scaling between 0.8 and 1.2 , shear up to 15 ° ). Furthermore, by cropping small patches, modifying them (for example, by pixel shuffling, rotation, or blurring), and then re-inserting them, a micro-patch jittering step creates local perturbations that resemble localized distortion or occlusion. Adaptive brightness adjustment within ( 1 0.2 α , 1 + 0.2 α ) and contrast variation inside ( 1 0.3 α , 1 + 0.3 α ) are examples of entropy-aware photometric transformations. α is an intensity factor that is obtained from image statistics. To improve spectral diversity, hue and saturation are also slightly altered in the HSV color space. All images have been cropped to 224 × 224 pixels after augmentation, and their pixel values are scaled to the [ 0 , 1 ] range for normalization. Any missing pixels from transformations are filled in using nearest-neighbor filling. Figure 4 displays examples of the augmented images.
Algorithm 2 Geometric Localized Occlusion and Adaptive Rescaling (GLOAR)
Require: Raw input image I
Ensure: Augmented image I
1: Compute global brightness B and entropy H of I
2: Set augmentation intensity parameter α f ( B , H ) , where f is a learnt or heuristic function
3: Step 1: Geometric Transformations (weighted by α )
4: Apply random affine transformation: rotation ( ± 30 ), scaling ( 0.8 1.2 ), shear ( ± 15 )
5: Apply random translation ( ± 15 % of image height/width)
6: Apply perspective warping with probability p = α
7: Step 2: Localized Micro-Patch Jittering
8: for i = 1 to n random patches do
9:        Crop a patch P i of size 8 % 15 % at a random location
10:        Apply perturbation: small rotation, pixel shuffle, or blur on P i
11:        Reinsert P i into its original location in I
12: end for
13: Step 3: Photometric Transformations (entropy-aware)
14: Adjust brightness in range ( 1 0.2 α , 1 + 0.2 α )
15: Adjust contrast in range ( 1 0.3 α , 1 + 0.3 α )
16: Convert to HSV color space
17: Perturb hue by ± 10
18: Scale saturation in range ( 0.9 , 1.1 )
19: return Augmented image I

3.3. Train, Validation and Test Split

To improve model generalization and robustness, the original dataset, which included 920 images of excellent quality taken in Dinajpur and Brahmanbaria, Bangladesh, was augmented to 7800 images. A total of 6000 images were assigned for training, with 900 images allotted for both testing and validation. The three classes in the dataset, namely Golden Mosaic, Cercospora Leaf Spot, and Healthy, are evenly distributed among all splits. With the training set aiding in feature learning, the validation set aiding in hyperparameter tuning to avoid overfitting, and the test set offering an objective performance evaluation, this hierarchical partitioning guarantees efficient model training.

3.4. Proposed DERIENet Model

For the purpose of detecting jute leaf diseases, our proposed DERIENet model integrates EfficientNetB0, InceptionV3, and ResNet50. As previously mentioned, the Jute Leaf Disease dataset was initially obtained from Kaggle. In order to improve the quality of the dataset and its ability to adapt to new data, a data-augmentation technique (GLOAR) was applied, ensuring variability in image features. As described in Section 3, each image was then resized to a consistent 224 × 224 pixel dimension. To enhance feature extraction and mitigate overfitting, three TL models were applied by incorporating additional dense layers with L2 regularization, batch normalization, ReLU activation, and dropout layers. These models were trained using the Adam optimizer with the categorical cross-entropy loss function, ensuring efficient learning and convergence. After evaluating their individual performances, the three top-performing TL models were selected and integrated into a custom ensemble framework. This ensemble approach leverages the advantages of each model, resulting in enhanced classification precision and adaptability in detecting jute leaf diseases.

DERIENet Architecture

The DERIENet architecture is specifically designed to handle RGB images of size 224 × 224 × 3 , ensuring compatibility with most deep-learning frameworks. Pre-trained models are utilized as feature extractors, with each model initialized using ImageNet weights. The configuration of the suggested model for identifying jute leaf diseases is depicted in Figure 5.
Let X represent the input image, where X R 224 × 224 × 3 . To extract feature maps, the models M EfficientNetB 0 , M InceptionV 3 , M ResNet 50 independently process the input image X. Compact feature vectors are generated by passing these feature maps through global average pooling (GAP) layers. Specifically, for each model:
E pool = GAP ( M EfficientNetB 0 ( X ) )
I pool = GAP ( M InceptionV 3 ( X ) )
R pool = GAP ( M ResNet 50 ( X ) )
Subsequently, these pooled feature vectors E pool , I pool , R pool are concatenated to form a single feature vector C :
C = concat ( E pool , I pool , R pool )
The concatenated vector C is processed further by passing it through a sequence of dense, fully linked layers. The final classification decision and the refinement of the characteristics retrieved by the basic models depend heavily on these layers. Figure 6 illustrates that the complex set structure of these fully interconnected layers is displayed. The ReLU (Rectified Linear Unit) function activates each of the 512 neurons in the first fully linked layer to add non-linearity:
D 1 = ReLU ( W 1 · C + b 1 )
This layer’s weights are subjected to L2 regularization in order to avoid overfitting. The first dense layer’s L2 penalty term is as follows:
L 2 Regularization Term for D 1 : λ i W 1 i 2
The strength of regularization is denoted by λ . The model is better able to generalize and prevent overfitting when the value of λ applies a stronger penalty to the weights. Following L2 regularization, the output of the dense layer is subject to batch normalization, which stabilizes training by lowering internal covariate shift:
D 1 norm = BatchNorm ( D 1 )
After batch normalization, a 50% dropout layer is also added. By randomly deactivating half of the neurons during training, dropout helps avoid overfitting and strengthens the network by requiring it to acquire more generic properties that are independent of particular neurons.
D 1 drop = Dropout ( 0.5 ) ( D 1 norm )
The second fully connected layer consists of 256 neurons and follows a similar structure. Additionally, the dropout layer mitigates overfitting by introducing noise during training. Similar to the previous layer, the weights in this layer undergo L2 regularization, followed by batch normalization and another dropout layer.
D 2 = ReLU ( W 2 · ( D 1 drop ) + b 2 )
D 2 drop = Dropout ( 0.5 ) ( BatchNorm ( D 2 ) )
The model’s output layer, which comprises three neurons, represents each class of jute leaf disease. The model’s predictions are probabilistic and comprehensible as confidence scores for each class when the SoftMax activation function converts the dense layers’ output into a probability distribution across the classes. The SoftMax activation function is used by the output layer to give each class a probability score:
O = SoftMax ( W 3 · D 2 drop + b 3 )
The Adam optimizer is employed to train the model. It has a learning rate of 1 × 10 4 and uses categorical cross-entropy as the loss function. Early stopping is used to avoid needless overfitting by terminating training after 10 consecutive epochs without improving validation loss. The ReduceLROnPlateau technique reduces the learning rate by 50% if the validation loss fails to improve after five epochs. Later on in the training process, this technique aids in fine-tuning the model’s weights, particularly when achieving further gains at a faster learning rate becomes challenging. At last, the completely trained model, represented by M c , is prepared for comparison and implementation.

3.5. Baseline Models

We utilized a variety of TL models for comparison that have demonstrated strong performance in picture classification. These models, including EfficientNetB0, ResNet50, InceptionV3, MobileNetV2, DenseNet201, and VGG16, are useful for a range of image identification applications because they each have unique architectural benefits.

3.5.1. EfficientNetB0

EfficientNetB0 [48] adaptively recalibrates channel-wise feature responses using Mobile Inverted Bottleneck Convolution (MBConv) blocks with swish activation and Squeeze-and-Excitation (SE) blocks. These components are characterized as follows:
Output = BN ( Conv 1 × 1 ( BN ( DWConv 3 × 3 ( BN ( Conv 1 × 1 ( x ) ) ) ) ) )
Swish ( x ) = x · σ ( x ) = x 1 + e x
SE ( x ) = x · σ ( W 2 · δ ( W 1 · GlobalAvgPool ( x ) ) )
where Conv 1 × 1 : 1 × 1 pointwise convolution, DWConv 3 × 3 : 3 × 3 depthwise convolution, BN: Batch Normalization, W 1 , W 2 : weight matrices for fully connected layers, δ : ReLU activation and σ : Sigmoid activation. After the MBConv blocks, a final pooling and dense layer is applied:
Output = SoftMax ( W · GlobalAvgPool ( x ) + b )
where W: weight matrix, b: bias, GlobalAvgPool ( x ) : Averages each channel before classification.

3.5.2. ResNet50

The ResNet-50 model presents a residual learning architecture in which the residual block serves as the fundamental building block [49]. The residual block’s output, y , can be written as follows:
y = F ( x , { W i } ) + x
where the input feature map is x and the residual function, F ( x , { W i } ) , represents the learnt transformation of the input x . Three layers make up the bottleneck design of the residual block in ResNet-50:
F ( x ) = W 3 · σ ( W 2 · σ ( W 1 · x ) )
where W 1 represents a 1 × 1 convolution used for channel reduction, W 2 is a 3 × 3 convolution responsible for spatial processing, and W 3 is another 1 × 1 convolution that restores the channels. Additionally, σ denotes the ReLU activation function.
To match the dimensions if the input and output are different, a projection shortcut is utilized:
y = F ( x ) + W s x
where W s is a 1 × 1 convolutional layer.

3.5.3. InceptionV3

Inception-v3 uses methods such as factorized convolutions, Inception modules, label smoothing regularization, and grid size reduction to increase computational efficiency and model performance [50]. The factorized convolutions approach substitutes smaller convolutions for larger ones, such as 5 × 5 ). For instance, two 3 × 3 convolutions are produced by factorizing a 5 × 5 convolution.
y = W 3 × 3 ( 2 ) · σ ( W 3 × 3 ( 1 ) · x )
The Inception module utilizes parallel convolutional paths to process the input at several scales. These pathways’ results are concatenated.
y = Concat [ W 1 × 1 · x , W 3 × 3 · σ ( W 1 × 1 · x ) , W 3 × 1 · σ ( W 1 × 3 · σ ( W 1 × 1 · x ) ) , MaxPool ( x ) ]
Regularization of label smoothing softens the target labels, preventing overfitting. Regularization of label smoothing softens the target labels, preventing overfitting. The below formula is employed to estimate the adjusted probabilities:
q ( k | x ) = ( 1 ϵ ) δ k , y + ϵ K
where ϵ = 0.1 is the smoothing factor, K is the number of classes, and δ k , y is the Dirac delta function, which equals 1 for the true label y and 0 otherwise. The spatial dimensions of the feature maps are essentially decreased using strided convolutions and max-pooling for grid size reduction.
y = Concat Conv 3 × 3 , stride = 2 ( x ) , MaxPool 3 × 3 , stride = 2 ( x )

3.5.4. MobileNetV2

MobileNetV2 reduces memory utilization and computational complexity using bottleneck blocks and depthwise separable convolutions [51]. The computational cost of a conventional convolution is:
C std = h i · w i · d i · d j · k 2
The cost of a depthwise separable convolution is:
C dw = h i · w i · d i k 2 + d j
The cost for a bottleneck block with expansion factor t, input channels d , output channels d , and kernel size k is:
C bottleneck = h · w · d · t d + k 2 + d
The peak memory required for a computational graph G is bounded by:
M ( G ) = max o p G A op in | A | + B op out | B | + | op |
Let F ( x ) = [ A N B ] x be a bottleneck block, where A and B are linear transformations, and N = ReLU 6 dwise ReLU 6 .
F ( x ) = [ A N B ] x

3.5.5. DenseNet201

Let L represent the total count of layers in the network. The outcome of the -th layer ( { 1 , 2 , . . . , L } ) is determined by applying a function to the concatenated feature maps from all preceding layers [52].
x = H [ x 0 , x 1 , , x 1 ]
where x represents the resulting feature maps of the -th layer, [ x 0 , x 1 , , x 1 ] indicates the combination of maps of features from all previous layers along the channel dimension, and H refers to the composite function applied at the -th layer, which includes the required transformations.

3.5.6. VGG16

The VGG-16 model comprises multiple convolutional layers succeeded by a fully connected layer [53]. The operations in each layer are as follows: The output of the -th convolutional layer ( { 1 , 2 , , L } ) is calculated by applying a convolution operation with a 3 × 3 filter and a ReLU activation:
x = ReLU ( Conv 3 × 3 ( x 1 ) )
where x 1 is the input map of features from the preceding layer. Max-pooling utilizing a 2 × 2 window and an interval of 2 is used subsequent to convolutional layers, diminishing the spatial dimensions of the feature maps:
x pool = MaxPool 2 × 2 ( x )
The result of the fully connected layers is computed as:
y = SoftMax ( W 3 · ReLU ( W 2 · ReLU ( W 1 · x flattened + b 1 ) + b 2 ) + b 3 )
where x flattened denotes the flattened output from the final pooling layer, whereas W 1 , W 2 , W 3 represent the weight matrices for the fully connected layers, accompanied with the associated bias terms b 1 , b 2 , b 3 .

3.6. Model Performance Metrics

Model assessment is a crucial stage in the machine-learning process, which establishes the model’s capacity to generalize to unseen data. This study used many methodologies, including the confusion matrix, receiver operating characteristic (ROC) curve, and several key performance measures, to assess the model’s efficacy. Accuracy, recall, F1-score, and precision are metrics that provide important information about the overall efficacy of the model [54].

3.6.1. Accuracy

The proportion of properly estimated instances to all predictions is known as accuracy [55], and it is a key evaluation parameter in machine learning. It offers an overall assessment of a model’s performance on unseen data. Accuracy is computed using samples from the test set to assess the generalizability of a model.
The accuracy is defined as:
Accuracy = M N × 100 %
where N is the overall number of instances in the test set, and M denotes the total quantity of samples for which the model accurately predicted the class labels.

3.6.2. Precision

The accuracy of positive forecasts is measured by a parameter called precision. It calculates how many accurately identified positive instances there were out of all the instances that were projected to be positive [56]. In a multiclass problem, the one-vs-all technique is used to calculate precision for each class. Additional names for precision include Positive Predictive Value (PPV).
The precision for each class c is given by:
Precision c = T P c T P c + F P c
where T P c denotes the quantity of true positive samples accurately categorized as class c, and F P c signifies the quantity of false positive samples inaccurately classified as class c.

3.6.3. Recall

The completeness of positive predictions is assessed using a statistic called recall [57], sometimes referred to as sensitivity. The percentage of real positive cases that the model accurately detected is measured. When dealing with a multiclass situation, the one-vs-all technique is used to determine recall for each class.
The recall for each class c is given by:
Recall c = T P c T P c + F N c
where T P c is the number of true positive samples correctly classified as class c, and F N c is the number of false negative samples that should have been classified as class c but were misclassified as other classes.

3.6.4. F1-Score

The F1-score offers a singular metric that equilibrates accuracy and recall by calculating their harmonic mean. [58]. This approach is highly beneficial for addressing unbalanced datasets as it considers both False Positives and False Negatives. An elevated F1-score indicates that the model is proficient in identifying affirmative examples, with recall and precision in equilibrium.
The F1-score for each class c is given by:
F 1 -score c = 2 × Precision c × Recall c Precision c + Recall c
where Precision c and Recall c are the precision and recall values for class c, respectively.

4. Results and Discussion

Several baseline models, such as EfficientNetB0, ResNet50, InceptionV3, VGG-16, DenseNet201, and MobileNetV2 were studied for jute leaf disease identification. The best-performing models were chosen for additional development. Our suggested DERIENet model achieved an accuracy of 99.89%, outperforming all baseline models. Ablation research validated the contributions of each component, and an analysis of per-class precision, recall, and F1-scores successfully solved the problem of class imbalance. Our method’s superiority was further illustrated by a comparison with the most advanced models. With its exceptional accuracy, robustness, and usefulness in actual agricultural settings, the suggested ensemble model sets a new benchmark for jute leaf disease detection.

4.1. Results of Baseline Models

This section utilizes ROC curve analysis and confusion matrix analysis to analyze the effectiveness of six baseline models. These evaluation methods aid in comprehending each model’s classification accuracy as well as how well it can differentiate between various categories.

4.1.1. Confusion Matrix Analysis

The jute leaf disease classification results of the baseline models, shown in Figure 7, provide useful information based on confusion matrices generated from 900 test cases. As the image shows, the models’ performance varied, with some demonstrating strengths in particular areas. While EfficientNetB0 performed well in distinguishing between Healthy Leaf and Cercospora Leaf Spot, it misclassified three instances of Healthy Leaf as Cercospora Leaf Spot and three instances of Cercospora Leaf Spot as Golden Mosaic. ResNet50 effectively classified both Golden Mosaic and Cercospora Leaf Spot. However, it misclassified two cases of Cercospora Leaf Spot as Healthy Leaf and 13 cases of Golden Mosaic. InceptionV3 correctly detected Healthy Leaf and Cercospora Leaf Spot. However, it misclassified five Golden Mosaic cases as Cercospora Leaf Spot and nine Cercospora Leaf Spot cases as Golden Mosaic, with a few minor misclassifications. MobileNetV2 exhibited noticeable misclassifications, particularly in Golden Mosaic, where 4 cases were incorrectly classified as Healthy Leaf and 9 as Cercospora Leaf Spot. One instance of Cercospora Leaf Spot was also mistakenly identified as Healthy Leaf, and 13 cases as Golden Mosaic. DenseNet201 successfully recognized all 300 Golden Mosaic samples. Ten cases of Cercospora Leaf Spot, however, were mistakenly identified as Golden Mosaic. Healthy Leaf had 298 accurate predictions and only two incorrect classifications: Cercospora Leaf Spot and Golden Mosaic. VGG16 effectively identified 267 Cercospora Leaf Spot cases; however, it incorrectly categorized 14 as Golden Mosaic and 19 as Healthy Leaf. Furthermore, it misidentified three Healthy Leaf cases as Cercospora Leaf Spot and 49 Golden Mosaic cases as such.

4.1.2. ROC Curve Analysis

AUC scores for the evaluation of ROC curves for six develops of the baseline model ranged from 93.38% to an almost flawless 99.99%, indicating a remarkable classification ability in the diagnosis of jute leaf disease (Figure 8).
The most accurate model was DenseNet201, which obtained remarkable AUC scores of 99.97% for Golden Mosaic detection and 99.99% for Cercospora Leaf Spot and Healthy Leaf categories. Across all disease categories, the ResNet50 design consistently produced excellent results (AUC: 99.48–99.82%). With AUC values above 99.87% for every disease class, InceptionV3 and EfficientNetB0 also continued to perform at a high level. Despite being lightweight, MobileNetV2 outperformed some more complicated architectures and produced impressive results (AUC: 99.58–99.97%). Visual examination of the ROC plots revealed that the best-performing models had curves that were primarily centered in the upper-left quadrant, indicating a low percentage of false positives and a high proportion of real positives. However, VGG16 was an anomaly, showing comparatively poorer performance, especially when it came to identifying the Cercospora Leaf Spot (AUC = 93.38%). This ROC-based evaluation ensures that, prior to deployment, the models meet the exacting accuracy requirements needed for precision agriculture.

4.2. Results of Proposed DERIENet Model

The DERIENet model produced exceptional results with a 99.89% accuracy, precision, recall, and F1-score. This outstanding performance made DERIENet the top model in this study, exceeding all baseline models. The training and validation graphs (Figure 9) further illustrate the model’s exceptional performance.
Over 50 epochs, the accuracy and loss curves showed a consistent convergence, with training and validation accuracy hitting 99.89% and the loss dropping off drastically to about 0.1. Strong learning and generalization are indicated by these patterns, indicating that the model successfully avoids overfitting. Figure 10, which displays the training and test confusion matrices, provides additional evidence of the DERIENet model’s performance. As demonstrated, the model correctly identified 899 out of 900 test-set images of jute leaf disease, with just one misclassification. The model’s excellent accuracy and low errors are demonstrated by the thorough confusion matrix analysis, which also shows how well it detects and categorizes jute leaf diseases.
Figure 11 displays the receiver operating characteristic (ROC) curve for the DERIENet model, illustrating its remarkable performance. Each of the three categories has an area under the curve (AUC) of 1, which illustrates the model’s outstanding class distinctiveness and flawless classification performance.
Accuracy alone is not a trustworthy way to evaluate a model’s performance since high accuracy might conceal poor performance on underrepresented classes, which is a prevalent problem in publicly available datasets with class imbalance [55]. We tackled this by concentrating on macro-averaged and weighted-averaged precision, recall, and F1-scores, which guarantee that all classes receive the same treatment regardless of sample size. Regarding F1-score, recall, and precision, our suggested model produced remarkable results of 99.89%. While correctly recognizing True Positives, the model successfully reduces False Positives and False Negatives, based on the elevated recall and precision metrics.
Furthermore, the precision, recall, and F1-scores for every class are shown in Table 2. The classification report of the suggested model makes it abundantly evident how well it can identify the states of jute leaves in both diseased and healthy groups. The model’s 100% recall, 99.67% precision, and 99.83% F1-score for the Cercospora Leaf Spot class demonstrate its exceptional ability to differentiate Cercospora from other classes with almost no false negatives. The Golden Mosaic model yielded an F1-score of 99.83% with 100% precision and 99.67% recall overall. These findings highlight how the model has a very low risk of misclassification and is highly reliable in detecting even minor disease instances. The model’s total accuracy in detecting disease-free samples is demonstrated by its perfect classification for the Healthy Leaf class, which attained 100% precision, recall, and F1-score. This steady and excellent performance in every class confirms the model’s resilience and applicability for practical use in automated agricultural diagnostic systems.

4.3. Model Performance Comparison

The suggested DERIENet model performs noticeably better in testing and training than separate baseline models. Its test accuracy of 99.89% is excellent, and it consistently earns superior categorization results. DERIENet performs better in classification than EfficientNetB0 (98.56%), ResNet50 (96.78%), and InceptionV3 (97.89%). The accuracy of the ensemble overcomes even DenseNet201, which achieves a strong test accuracy of 98.67%. Compared to the suggested ensemble, conventional models like MobileNetV2, which attains 96.44%, and VGG16, which only reaches 90.22%, are less accurate and resilient. Furthermore, DERIENet maintains consistently high precision (99.89%), recall (99.89%), and F1-score (99.89%) throughout testing. The test findings show that it performs better than baseline models on all important performance parameters, highlighting its remarkable versatility. Regarding concerns about potential overfitting, several techniques beyond early stopping were employed during training to mitigate overfitting and improve generalization. These include the use of both L1 and L2 regularization (with L2 providing the maximum accuracy), dropout layers with a rate of 0.5, and significant data augmentation (GLOAR). Furthermore, early stopping was implemented according to validation loss, and the ReduceLROnPlateau callback was used to adaptively modify the learning rate. All these techniques improved the model’s capacity to generalize within the dataset and successfully decreased overfitting.
Table 3 compiles the performance of several models, with particular attention to test accuracy, precision, recall, and F1-score. Using the same measures, Figure 12 compares the performance of these models and offers a visual representation of the data.

4.4. Ablation Study

The notable performance gains at each step of the model’s construction and augmentation process are demonstrated by the ablation research in Table 4. Two models combined resulted in more improvements, with EfficientNetB0 and InceptionV3 attaining 99.44% and EfficientNetB0 and ResNet50 reaching 99.67%. The combination of ResNet50 and InceptionV3 achieved an accuracy of 98.89%, which, while commendable, was lower in comparison to the other two-model combinations. However, evaluating the combined model of EfficientNetB0, ResNet50, and InceptionV3 without data augmentation (GLOAR) resulted in a significant drop in accuracy to 85.71% due to overfitting, highlighting the crucial importance of data augmentation.The accuracy increased to 99.22% when L1 regularization was used on the same three-model ensemble with augmentation (GLOAR). Finally, the proposed DERIENet model, which integrates all three architectures with a proposed data-augmentation technique (GLOAR) and L2 regularization, attained the maximum accuracy of 99.89%. This illustrates the strong benefits of model ensemble and augmentation in boosting performance and minimizing overfitting.

5. Discussion

One of the key variables in this work was the original dataset’s small size, which necessitated the use of calculated techniques to guarantee dependable model performance. A computer vision-based jute leaf disease classification system was developed using the dataset used in this work, which included high-resolution photos from three categories: Healthy Leaf, Golden Mosaic, and Cercospora Leaf Spot. Nevertheless, there are significant obstacles in developing strong deep-learning models on comparatively limited datasets. In order to boost dataset diversity and reduce the possibility of overfitting, we used GLOAR, a proposed data-augmentation technique. The augmentation methods decrease the possibility of overfitting while increasing the dataset [59]. A number of baseline models were first assessed for the classification of jute leaf disease, including ResNet50, InceptionV3, VGG16, DenseNet201, MobileNetV2, and EfficientNetB0. The three designs that performed the best were chosen for additional improvement using regularization approaches to lessen overfitting, batch normalization, and dense layers. The suggested DERIENet model was then created by combining these improved models using an ensemble technique.
This study evaluated the proposed DERIENet model through comprehensive experiments, where it consistently outperformed baseline models across various key performance indicators. As illustrated in Figure 7, the confusion matrices of every baseline model were examined in order to determine which models were most suited for the final ensemble. EfficientNetB0 properly classified 297 cases, demonstrating the best accuracy in detecting Cercospora Leaf Spot. With no misclassifications, InceptionV3 successfully classified Healthy Leaf images. Although DenseNet201 accurately classified all 300 Golden Mosaic samples, its significantly higher runtime made it less suitable for inclusion. ResNet50 was therefore selected for the ensemble because of its effectiveness and well-rounded performance. EfficientNetB0, InceptionV3, and ResNet50 were ultimately chosen for the ensemble due to their accuracy and lightweight design, which made them ideal for the DERIENet framework.
DERIENet’s performance evaluation further demonstrates how effective the selected ensemble architecture is. It outscored all baseline models in classification across all important evaluation metrics. As shown in Figure 10, DERIENet successfully classified 299 out of 300 test samples, with only one misclassification. Table 3 emphasizes the model’s impressive precision, recall, F1-score, and accuracy, all of which were recorded at 99.89%. Furthermore, as seen in Figure 13, the model showed a high degree of accuracy in classifying it.
As shown in Table 5, all models reported in prior research were evaluated using the Jute Leaf Disease dataset from this study. The results show that the suggested DERIENet model performs noticeably better than these current methods in terms of classification accuracy. In particular, when tested on the Jute Leaf Disease dataset, traditional architectures such as four-layer CNN [7], CNN [31], DCNN [36], and Lightweight CNN [39] obtained accuracies ranging from 95.43% to 96.33%. Conversely, transfer-learning models such as ResNet50 [35] and DenseNet201 [38] showed improved performance, 96.50% and 98.21%, respectively, but still failed to surpass our proposed model. More advanced methods, such as federated learning (FL) + CNN [42] and ResNet152 + Custom CNN [33], obtained accuracies of 97.67% and 98.98%, respectively. With a classification accuracy of 99.89%, our suggested DERIENet model outperformed all other approaches noticeably.
This study exclusively focused on three categories in jute leaf disease classification: Cercospora Leaf Spot, Golden Mosaic, and Healthy Leaf. The model was developed based on these diseases, but it has yet to be tested on additional jute diseases in other contexts. Incorporating a wider variety of jute diseases into the model could enhance its relevance in practical situations. The model would be more flexible and perform better in a range of agricultural settings if the dataset were expanded to include more varied disease kinds.

6. Conclusions

This study introduces DERIENet, a high-precision ensemble deep neural network model for automatically classifying diseases of jute leaves. The model incorporates InceptionV3, ResNet50, and EfficientNetB0 to use their combined capabilities and enhance classification performance. In order to provide resilience across a variety of disease patterns, the suggested data-augmentation technique (GLOAR) was used to enhance inference and avoid overfitting. In order to preserve model dependability while better representing disease variability, the dataset was enlarged. After evaluating six transfer-learning models in comparison, the top three models were chosen and fused using an ensemble technique to maximize performance. This led to DERIENet exceeding all baseline models in precision, recall, and F1-score, achieving an outstanding classification accuracy of 99.89%. The scalable and effective early jute disease detection solution offered by this high-performance model allows for prompt intervention and helps farmers reduce crop losses. A more sustainable and effective agricultural system could result from DERIENet’s ability to modernize jute farming, increase crop yields, and improve fiber quality by enabling quick and accurate disease identification.
Future studies should focus on enhancing DERIENet’s capacity to identify a greater variety of jute leaf diseases and evaluating how well it detects different jute plant infections to expand the model’s reach and applicability. Additionally, DERIENet’s incorporation of lightweight neural networks enhances its processing efficiency, making it appropriate for use in mobile apps. In instances with limited resources, this would enable real-time disease diagnosis, giving farmers a diagnostic option that is both accessible and reasonably priced. A mobile or web-based application would also make it simpler to engage with the model, enabling farmers and agricultural specialists to use DERIENet for effective disease management.

Author Contributions

M.T.Y.T.: Conceptualization, Software, Writing—Original Draft, Writing-Review, and Editing. T.S.: Project administration, Supervision, Conceptualization, Writing—Original Draft, Writing—review and editing. M.E.B.: Conceptualization, Software, Writing—Original Draft, and Editing. C.K.M.: Writing—review and editing. A.A.: Writing—review and editing. M.S.U.: Project administration, Supervision, Writing—Review and Editing. M.D.H.: Project administration, Supervision, Conceptualization, Writing—Original Draft, Writing—review and editing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partly supported by the Institute of Information & Communications Technology Planning & Evaluation (IITP)-ITRC (Information Technology Research Center) grant funded by the Korea government (MSIT) (IITP-2024-RS-2024-00438239) and the Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT) (No. RS-2024-00509257, Global AI Frontier Lab). This was also supported by the Institute of Research and Training (IRT) of Hajee Mohammad Danesh Science and Technology University, Bangladesh.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ahmed, Z.; Sarkar, S. Review on jute leaf: A powerful biological tool. Int. J. Sci. Res. Updat. 2022, 4, 64–85. [Google Scholar] [CrossRef]
  2. Islam, M.M. Advanced Production Technology and Processing of Jute. In Agronomic Crops: Volume 1: Production Technologies; Springer: Singapore, 2019; pp. 387–440. [Google Scholar] [CrossRef]
  3. Sharna, S.C.; Kamruzzaman, M. Comparative advantages of jute export in Bangladesh, China and India. Res. Agric. Livest. Fish. 2020, 7, 183–190. [Google Scholar] [CrossRef]
  4. Miah, M.J.; Khan, M.A.; Khan, R.A. Fabrication and characterization of jute fiber reinforced low density polyethylene based composites: Effects of chemical treatment. J. Sci. Res. 2011, 3, 249–259. [Google Scholar] [CrossRef]
  5. Akter, S.; Sadekin, M.N.; Islam, N. Jute and jute products of Bangladesh: Contributions and challenges. Asian Bus. Rev. 2020, 10, 143–152. [Google Scholar] [CrossRef]
  6. Islam, M.M. Biochemistry, Medicinal and Food values of Jute (Corchorus capsularis L. and C. olitorius L.) leaf: A Review. Int. J. Enhanc. Res. Sci. Technol. Eng. 2013, 2, 35–44. [Google Scholar]
  7. Uddin, M.S.; Munsi, M.Y. Juledi: Jute leaf disease identification using convolutional neural network. In Proceedings of the International Conference on Information, Communication and Computing Technology, New Delhi, India, 27 May 2023; Springer: Berlin/Heidelberg, Germany, 2023; pp. 661–677. [Google Scholar]
  8. Li, L.; Zhang, S.; Wang, B. Plant Disease Detection and Classification by Deep Learning—A Review. IEEE Access 2021, 9, 56683–56698. [Google Scholar] [CrossRef]
  9. Biswas, M.E.; Sultana, T.; Mandal, A.K.; Morshed, M.G.; Hossain, M.D. Spatio-temporal feature engineering and selection-based flight arrival delay prediction using deep feedforward regression network. Electronics 2024, 13, 4910. [Google Scholar] [CrossRef]
  10. Biswas, M.E.; Hossain, M.D.; Huh, E.-N.; Sultana, T. Spatiotemporal Data-Driven Hourly Bike-Sharing Demand Prediction Using ApexBoost Regression. Int. J. Data Sci. Anal. 2025, 1–24. [Google Scholar] [CrossRef]
  11. Lu, J.; Tan, L.; Jiang, H. Review on convolutional neural network (CNN) applied to plant leaf disease classification. Agriculture 2021, 11, 707. [Google Scholar] [CrossRef]
  12. Albahli, S.; Nawaz, M. DCNet: DenseNet-77-based CornerNet model for the tomato plant leaf disease detection and classification. Front. Plant Sci. 2022, 13, 957961. [Google Scholar] [CrossRef]
  13. Matin, M.M.H.; Khatun, A.; Moazzam, M.G.; Uddin, M.S. An efficient disease detection technique of rice leaf using AlexNet. J. Comput. Commun. 2020, 8, 49–57. [Google Scholar] [CrossRef]
  14. Arya, S.; Singh, R. A Comparative Study of CNN and AlexNet for Detection of Disease in Potato and Mango leaf. In Proceedings of the 2019 International Conference on Issues and Challenges in Intelligent Computing Techniques (ICICT), Ghaziabad, India, 27–28 September 2019; IEEE: Piscataway, NJ, USA, 2019; Volume 1, pp. 1–6. [Google Scholar]
  15. Tanny, M.T.Y.; Biswas, M.E.; Akter, A.; Modok, C.K.; Nipu, M.H.N.; Hasan, M.; Hossain, M.D.; Mandal, A.K.; Hossain, M.S.; Sultana, T. Customized EfficientNet: A Transfer Learning Approach for Rice Leaf Disease Classification with Advanced Resizing and Augmentation. In Proceedings of the 2025 International Conference on Electrical, Computer and Communication Engineering (ECCE), Chittagong, Bangladesh, 13–15 February 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 1–6. [Google Scholar]
  16. Kunduracıoğlu, İ.; Paçal, İ. Deep Learning-Based Disease Detection in Sugarcane Leaves: Evaluating EfficientNet Models. J. Oper. Intell. 2024, 2, 321-235. [Google Scholar] [CrossRef]
  17. Yang, L.; Yu, X.; Zhang, S.; Long, H.; Zhang, H.; Xu, S.; Liao, Y. GoogLeNet based on residual network and attention mechanism identification of rice leaf diseases. Comput. Electron. Agric. 2023, 204, 107543. [Google Scholar] [CrossRef]
  18. Adedoja, A.; Owolawi, P.A.; Mapayi, T. Deep learning based on nasnet for plant disease recognition using leave images. In Proceedings of the 2019 International Conference on Advances in Big Data, Computing and Data Communication Systems (icABCD), Winterton, South Africa, 5–6 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–5. [Google Scholar]
  19. Li, X.; Rai, L. Apple leaf disease identification and classification using resnet models. In Proceedings of the 2020 IEEE 3rd International Conference on Electronic Information and Communication Technology (ICEICT), Shenzhen, China, 13–15 November 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 738–742. [Google Scholar]
  20. Bi, C.; Wang, J.; Duan, Y.; Fu, B.; Kang, J.R.; Shi, Y. MobileNet based apple leaf diseases identification. Mob. Netw. Appl. 2022, 27, 172–180. [Google Scholar] [CrossRef]
  21. Elfatimi, E.; Eryigit, R.; Elfatimi, L. Beans leaf diseases classification using mobilenet models. IEEE Access 2022, 10, 9471–9482. [Google Scholar] [CrossRef]
  22. Theerthagiri, P.; Ruby, A.U.; Chandran, J.G.C.; Sardar, T.H.; Shafeeq BM, A. Deep SqueezeNet learning model for diagnosis and prediction of maize leaf diseases. J. Big Data 2024, 11, 112. [Google Scholar] [CrossRef]
  23. Haque, R.; Miah, M.M.; Sultana, S.; Fardin, H.; Noman, A.A.; Al-Sakib, A.; Hasan, M.K.; Rafy, A.; Shihabur, R.M.; Rahman, S. Advancements in Jute Leaf Disease Detection: A Comprehensive Study Utilizing Machine Learning and Deep Learning Techniques. In Proceedings of the 2024 IEEE International Conference on Power, Electrical, Electronics and Industrial Applications (PEEIACON), Rajshahi, Bangladesh, 12–13 September 2024; pp. 248–253. [Google Scholar] [CrossRef]
  24. Paymode, A.S.; Malode, V.B. Transfer learning for multi-crop leaf disease image classification using convolutional neural network VGG. Artif. Intell. Agric. 2022, 6, 23–33. [Google Scholar] [CrossRef]
  25. Sreedevi, A.; Srinivas, K. Deep learning methodologies towards leaf disease detection: A review. In Impending Inquisitions in Humanities and Sciences; CRC Press: Boca Raton, FL, USA, 2024; pp. 270–280. [Google Scholar]
  26. Vats, S.; Kukreja, V.; Mehta, S. A New Era in AgriTech: Federated Learning CNN for Jute Leaf Disease Identification. In Proceedings of the 2024 IEEE International Conference on Interdisciplinary Approaches in Technology and Management for Social Innovation (IATMSI), Gwalior, India, 14–16 March 2024; Volume 2, pp. 1–6. [Google Scholar] [CrossRef]
  27. Rahman, M.I.; Nazim Uddin, M. Jute Plant Disease Detection by obtaining RoI with SGM using Image Processing Technique and CNN. In Proceedings of the 2023 26th International Conference on Computer and Information Technology (ICCIT), Cox’s Bazar, Bangladesh, 13–15 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
  28. Daphal, S.D.; Koli, S.M. Enhancing sugarcane disease classification with ensemble deep learning: A comparative study with transfer learning techniques. Heliyon 2023, 9, e18261. [Google Scholar] [CrossRef] [PubMed]
  29. Patil, R.R.; Kumar, S. Rice-fusion: A multimodality data fusion framework for rice disease diagnosis. IEEE Access 2022, 10, 5207–5222. [Google Scholar] [CrossRef]
  30. Yigit, E.; Sabanci, K.; Toktas, A.; Kayabasi, A. A study on visual features of leaves in plant identification using artificial intelligence techniques. Comput. Electron. Agric. 2019, 156, 369–377. [Google Scholar] [CrossRef]
  31. Hasan, M.Z.; Ahamed, M.S.; Rakshit, A.; Hasan, K.M.Z. Recognition of Jute Diseases by Leaf Image Classification using Convolutional Neural Network. In Proceedings of the 2019 10th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kanpur, India, 6–8 July 2019; pp. 1–5. [Google Scholar] [CrossRef]
  32. Li, D.; Ahmed, F.; Wu, N.; Sethi, A.I. Yolo-JD: A Deep Learning Network for jute diseases and pests detection from images. Plants 2022, 11, 937. [Google Scholar] [CrossRef]
  33. Akhand, M.N. Deep Learning-Based Jute Leaf Disease Identification. In Proceedings of the Undergraduate Conference on Intelligent Computing and Systems (UCICS 2025), Varendra University, Rajshahi, Bangladesh, 26–27 February 2025. [Google Scholar]
  34. Hridoy, R.H.; Yeasmin, T.; Mahfuzullah, M. A deep multi-scale feature fusion approach for early recognition of jute diseases and pests. In Inventive Systems and Control: Proceedings of ICISC 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 553–567. [Google Scholar]
  35. Kaushik, P.; Khurana, S. Jute Leaf Disease Detection using ResNet50: A Deep Learning Approach for Precision Agriculture. In Proceedings of the 2025 International Conference on Electronics and Renewable Systems (ICEARS), Tuticorin, India, 11–13 February 2025; pp. 1901–1906. [Google Scholar] [CrossRef]
  36. Sourav, M.S.U.; Wang, H. Intelligent identification of jute pests based on transfer learning and deep convolutional neural networks. Neural Process. Lett. 2023, 55, 2193–2210. [Google Scholar] [CrossRef]
  37. Karim, D.Z.; Bushra, T.A.; Saif, M.M. PestDetector: A Deep Convolutional Neural Network to Detect Jute Pests. In Proceedings of the 2022 4th International Conference on Sustainable Technologies for Industry 4.0 (STI), Los Alamitos, CA, USA, 17–18 December 2022; pp. 1–6. [Google Scholar] [CrossRef]
  38. Talukder, M.S.H.; Chowdhury, M.R.; Sourav, M.S.U.; Rakin, A.A.; Shuvo, S.A.; Sulaiman, R.B.; Nipun, M.S.; Islam, M.; Islam, M.R.; Islam, M.A.; et al. JutePestDetect: An intelligent approach for jute pest identification using fine-tuned transfer learning. Smart Agric. Technol. 2023, 5, 100279. [Google Scholar] [CrossRef]
  39. Rana, S.; Samad, M.F. End-to-end jute-pest detection by explainable lightweight cnn. In Proceedings of the 2024 6th International Conference on Electrical Engineering and Information & Communication Technology (ICEEICT), Dhaka, Bangladesh, 2–4 May 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 230–235. [Google Scholar]
  40. Islam, M.T.; Rahman, M.S. An Efficient Deep Learning Approach for Jute Pest Classification Using Transfer Learning. In Proceedings of the 2024 6th International Conference on Electrical Engineering and Information & Communication Technology (ICEEICT), Dhaka, Bangladesh, 2–4 May 2024; pp. 1473–1478. [Google Scholar] [CrossRef]
  41. Bansal, A.; Vats, S.; Prasad, C.; Kukreja, V.; Mehta, S. Collaborative Computing: Federated Learning CNNs for Jute Leaf Disease Analysis. In Proceedings of the 2024 3rd International Conference for Innovation in Technology (INOCON), Bangalore, India, 1–3 March 2024; pp. 1–6. [Google Scholar] [CrossRef]
  42. Rajput, K.; Garg, A.; Bhati, J.P.; Kukreja, V.; Mehta, S. Agricultural AI Insights: Deciphering Jute Leaf Diseases Through Federated Learning CNNs. In Proceedings of the 2024 10th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 14–15 March 2024; Volume 1, pp. 1589–1595. [Google Scholar] [CrossRef]
  43. Rajput, K.; Garg, A.; Kukreja, V.; Mehta, S. Agricultural Intelligence: Federated Learning CNN’s Models for Jute Leaf Disease Analysis. In Proceedings of the 2024 11th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 14–15 March 2024; pp. 1–6. [Google Scholar] [CrossRef]
  44. Rijal, T.R.; Karki, S. Evaluation of promising jute (Corchorus olitorius) germplasm against stem rot caused by (Macrophomina phaseolina)(Tassi) Goid. Int. J. Appl. Agric. Sci. 2022, 8, 192–195. [Google Scholar]
  45. Reza, Z.N.; Nuzhat, F.; Mahsa, N.A.; Ali, M.H. Detecting jute plant disease using image processing and machine learning. In Proceedings of the 2016 3rd International Conference on Electrical Engineering and Information Communication Technology (ICEEICT), Dhaka, Bangladesh, 22–24 September 2016; pp. 1–6. [Google Scholar] [CrossRef]
  46. Biswas, M.E.; Shahzamal, M.; Haque, M.D. Machine Learning Approach to Estimate Requirements for Target Productivity of Garments Employees. In Proceedings of the 2024 6th International Conference on Electrical Engineering and Information & Communication Technology (ICEEICT), Dhaka, Bangladesh, 2–4 May 2024; IEEE: New York, NY, USA, 2024; pp. 921–926. [Google Scholar]
  47. Biswas, M.E.; Ahamed, M.S.; Rafi, M.A.H.; Haque, M.D. An Adaptive Machine Learning Approach for Electrical Fault Detection and Minimization. In Proceedings of the 2024 2nd International Conference on Information and Communication Technology (ICICT), Dhaka, Bangladesh, 21–22 October 2024; IEEE: New York, NY, USA, 2024; pp. 274–278. [Google Scholar]
  48. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; PMLR: Breckenridge, CO, USA, 2019; pp. 6105–6114. [Google Scholar]
  49. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  50. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  51. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4510–4520. [Google Scholar]
  52. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  53. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  54. Grandini, M.; Bagli, E.; Visani, G. Metrics for multi-class classification: An overview. arXiv 2020, arXiv:2008.05756. [Google Scholar] [CrossRef]
  55. Ahmed, S.; Hasan, M.B.; Ahmed, T.; Sony, M.R.K.; Kabir, M.H. Less is more: Lighter and faster deep neural architecture for tomato leaf disease classification. IEEE Access 2022, 10, 68868–68884. [Google Scholar] [CrossRef]
  56. Barud Demilie, W. Plant disease detection and classification techniques: A comparative study of the performances. J. Big Data 2024, 11, 5. [Google Scholar] [CrossRef]
  57. Joseph, D.S.; Pawar, P.M.; Chakradeo, K. Real-Time Plant Disease Dataset Development and Detection of Plant Disease Using Deep Learning. IEEE Access 2024, 12, 16310–16333. [Google Scholar] [CrossRef]
  58. Pandian, J.A.; Kumar, V.D.; Geman, O.; Hnatiuc, M.; Arif, M.; Kanchanadevi, K. Plant disease detection using deep convolutional neural network. Appl. Sci. 2022, 12, 6982. [Google Scholar] [CrossRef]
  59. Bansal, P.; Kumar, R.; Kumar, S. Disease detection in apple leaves using deep convolutional neural network. Agriculture 2021, 11, 617. [Google Scholar] [CrossRef]
Figure 1. Block diagram of the proposed methodology.
Figure 1. Block diagram of the proposed methodology.
Information 16 00638 g001
Figure 2. Studied classes of jute leaves used for analysis.
Figure 2. Studied classes of jute leaves used for analysis.
Information 16 00638 g002
Figure 3. Class Distribution of jute leaves.
Figure 3. Class Distribution of jute leaves.
Information 16 00638 g003
Figure 4. Some examples of augmented images.
Figure 4. Some examples of augmented images.
Information 16 00638 g004
Figure 5. Architecture of the proposed DERIENet model.
Figure 5. Architecture of the proposed DERIENet model.
Information 16 00638 g005
Figure 6. Schematic illustration of the layers in the suggested model.
Figure 6. Schematic illustration of the layers in the suggested model.
Information 16 00638 g006
Figure 7. Test Confusion Matrix for Baseline Models.
Figure 7. Test Confusion Matrix for Baseline Models.
Information 16 00638 g007
Figure 8. ROC curve for baseline models.
Figure 8. ROC curve for baseline models.
Information 16 00638 g008
Figure 9. Training and validation performance (accuracy and loss) of the DERIENet model.
Figure 9. Training and validation performance (accuracy and loss) of the DERIENet model.
Information 16 00638 g009
Figure 10. Train and test confusion matrices of DERIENet model.
Figure 10. Train and test confusion matrices of DERIENet model.
Information 16 00638 g010
Figure 11. ROC curve of DERIENet model.
Figure 11. ROC curve of DERIENet model.
Information 16 00638 g011
Figure 12. Performance comparison of proposed DERIENet and baseline models.
Figure 12. Performance comparison of proposed DERIENet and baseline models.
Information 16 00638 g012
Figure 13. Visualization of true and predicted outputs obtained from the model.
Figure 13. Visualization of true and predicted outputs obtained from the model.
Information 16 00638 g013
Table 1. Overview of previous work on leaf disease detection methods.
Table 1. Overview of previous work on leaf disease detection methods.
Ref.Model(s) UsedDataset#Images#ClassesTransfer
Learning
Ensemble
Learning
Data
Augmentation
Accuracy
[7]Four-layer CNNJute leaf(own)47403NoNoYes96%
[28]CNN, Xception, ResNet50, MobileNetV2, VGG19, and EfficientNetB7Sugarcane (own)25695YesYesNo86.53%
[29]CNN, MLPRice(own)32004NoYesNo95.31%
[30]SVMMulti-plant (Folio)63732NoNoNo92.91%
[31]CNNJute leaf6002NoNoNo96%
[32]YOLO-JDJute leaf and pest441810NoNoNo96.63%
[33]ResNet152 + Custom CNNJute leaf(Kaggle)18203YesNoNo98.41%
[35]ResNet50Jute leaf(Kaggle)1820-YesNoYes94%
[36]DCNNJute pest15354YesNoNo95.86%
[38]DenseNet201Jute pest380 per class17YesNoYes99%
[43]FL-based CNNJute leaf-5NoNoNo98%
Table 2. Classification report of the proposed model (DERIENet).
Table 2. Classification report of the proposed model (DERIENet).
ClassPrecision (%)Recall (%)F1-Score (%)Support
Cercospora Leaf Spot99.6710099.83300
Golden Mosaic10099.6799.83300
Healthy Leaf100100100300
Accuracy 99.89900
Macro Average99.8999.8999.89900
Weighted Average99.8999.8999.89900
Table 3. Comparative evaluation of the models’ performance.
Table 3. Comparative evaluation of the models’ performance.
Model NameTrain ResultsTest ResultsTrain AccuracyTest Accuracy
PrecisionRecallF1-ScorePrecisionRecallF1-Score
EfficientNetB099.80%99.80%99.80%98.56%98.56%98.56%99.83%98.56%
ResNet5096.54%96.47%96.44%96.81%96.78%96.78%97.17%96.78%
InceptionV398.91%98.90%98.90%97.89%97.89%97.88%99.02%97.89%
MobileNetV298.01%97.97%97.97%96.44%96.44%96.44%98.25%96.44%
DenseNet20199.07%99.05%99.05%98.71%98.67%98.67%99.27%98.67%
VGG1692.47%92.45%92.41%90.49%90.22%90.16%92.70%90.22%
DERIENet (Proposed)99.98%99.98%99.98%99.89%99.89%99.89%99.95%99.89%
Table 4. Ablation analysis of DERIENet’s components in jute leaf disease detection.
Table 4. Ablation analysis of DERIENet’s components in jute leaf disease detection.
ModelTest Classification Accuracy
EfficientNetB0 + InceptionV399.44%
EfficientNetB0 + ResNet5099.67%
ResNet50 + InceptionV398.89%
EfficientNetB0 + ResNet50 + InceptionV385.71%
(Without GLOAR)
EfficientNetB0 + ResNet50 + InceptionV3 99.22%
(With L1 Regularization)
DERIENet (Proposed)99.89%
Table 5. Performance comparison between the model suggested in this paper and earlier research.
Table 5. Performance comparison between the model suggested in this paper and earlier research.
No.ModelRef.Accuracy
1 four-layer CNN Uddin et al. [7] 96.21%
2 CNN Karim et al. [31] 95.45%
3 ResNet152 + Custom CNN Akand et al. [33] 98.98%
4 ResNet50 Kaushik et al. [35] 96.50%
5 DCNN Sourav et al. [36] 96.33%
6 DenseNet201 Talukder et al. [38] 98.21%
7 Lightweight CNN Rana et al. [39] 95.43%
8 FL + CNN Rajput et al. [42] 97.67%
9DERIENetOur Proposed Method99.89%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tanny, M.T.Y.; Sultana, T.; Biswas, M.E.; Modok, C.K.; Akter, A.; Uddin, M.S.; Hossain, M.D. DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases. Information 2025, 16, 638. https://doi.org/10.3390/info16080638

AMA Style

Tanny MTY, Sultana T, Biswas ME, Modok CK, Akter A, Uddin MS, Hossain MD. DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases. Information. 2025; 16(8):638. https://doi.org/10.3390/info16080638

Chicago/Turabian Style

Tanny, Mst. Tanbin Yasmin, Tangina Sultana, Md. Emran Biswas, Chanchol Kumar Modok, Arjina Akter, Mohammad Shorif Uddin, and Md. Delowar Hossain. 2025. "DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases" Information 16, no. 8: 638. https://doi.org/10.3390/info16080638

APA Style

Tanny, M. T. Y., Sultana, T., Biswas, M. E., Modok, C. K., Akter, A., Uddin, M. S., & Hossain, M. D. (2025). DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases. Information, 16(8), 638. https://doi.org/10.3390/info16080638

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop