Next Article in Journal
Application of Microwave Transmission Sensors for Water Cut Metering under Varying Salinity Conditions: Device, Algorithm and Uncertainty Analysis
Next Article in Special Issue
A Survey on Low-Latency DNN-Based Speech Enhancement
Previous Article in Journal
High-Sensitivity Slot-Loaded Microstrip Patch Antenna for Sensing Microliter-Volume Liquid Chemicals with High Relative Permittivity and High Loss Tangent
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Pest Detection Framework with a Medium-Scale Benchmark to Increase the Agricultural Productivity

by
Suliman Aladhadh
1,*,
Shabana Habib
1,
Muhammad Islam
2,
Mohammed Aloraini
3,
Mohammed Aladhadh
4 and
Hazim Saleh Al-Rawashdeh
5
1
Department of Information Technology, College of Computer, Qassim University, Buraydah 51452, Saudi Arabia
2
Department of Electrical Engineering, College of Engineering and Information Technology, Onaizah Colleges, Onaizah 56447, Saudi Arabia
3
Department of Electrical Engineering, College of Engineering, Qassim University, Unaizah 56452, Saudi Arabia
4
Department of Food Science and Human Nutrition, College of Agriculture and Veterinary Medicine, Qassim University, Buraydah 51452, Saudi Arabia
5
Department of Cyber Security, College of Engineering and Information Technology, Onaizah Colleges, Onaizah 56447, Saudi Arabia
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(24), 9749; https://doi.org/10.3390/s22249749
Submission received: 28 October 2022 / Revised: 18 November 2022 / Accepted: 21 November 2022 / Published: 12 December 2022
(This article belongs to the Special Issue Audio, Image, and Multimodal Sensing Techniques)

Abstract

:
Insect pests and crop diseases are considered the major problems for agricultural production, due to the severity and extent of their occurrence causing significant crop losses. To increase agricultural production, it is significant to protect the crop from harmful pests which is possible via soft computing techniques. The soft computing techniques are based on traditional machine and deep learning-based approaches. However, in the traditional methods, the selection of manual feature extraction mechanisms is ineffective, inefficient, and time-consuming, while deep learning techniques are computationally expensive and require a large amount of training data. In this paper, we propose an efficient pest detection method that accurately localized the pests and classify them according to their desired class label. In the proposed work, we modify the YOLOv5s model in several ways such as extending the cross stage partial network (CSP) module, improving the select kernel (SK) in the attention module, and modifying the multiscale feature extraction mechanism, which plays a significant role in the detection and classification of small and large sizes of pest in an image. To validate the model performance, we develop a medium-scale pest detection dataset that includes the five most harmful pests for agriculture products that are ants, grasshopper, palm weevils, shield bugs, and wasps. To check the model’s effectiveness, we compare the results of the proposed model with several variations of the YOLOv5 model, where the proposed model achieved the best results in the experiments. Thus, the proposed model has the potential to be applied in real-world applications and further motivate research on pest detection to increase agriculture production.

1. Introduction

In agricultural production, pest detection has always been a serious problem, which is responsible for 20% of annual crop losses globally [1,2]. In 2021, the affected areas caused by pests and major diseases in China had reached almost 400 million hectares. Therefore, timely detection of crop diseases and pests is crucial to agricultural production, which has a significant impact on grain production, agricultural development, and farmers’ income increase [3]. Building an artificially intelligent model based on agricultural image processing is one of the more effective ways to detect pests and classify them according to the class label [4], which can give an efficient response and intervention to agricultural production, increase the effectiveness of pest detection model, reduce the losses in the agricultural production.
Therefore, researchers used traditional machine learning (ML) [5,6], and deep learning-based models for an efficient pest detection system. The traditional methods of detecting insect pests based on morphological features are limited by the need for trained taxonomists to make accurate identifications [7]. It is important to note that traditional methods for pest detection come with many limitations. Recently, several methods for automatic pest detection using traditional machine learning are proposed [8]. For example, Faithpraise et al. [9], proposed K-means clustering algorithms for pests detection. Manual feature extraction and relative filter were used to recognize various species of pests, which is time-consuming when the dataset is large. Rumpf et al. [10], proposed a support vector machine and spectral vegetation-based sugar beet diseases recognition. These methods are capable of pest detection; however, several limitations restrict the traditional ML-based model from real-world implementation such as when multiple features have to be extracted manually, traditional ML algorithms are often inefficient. Secondly, in the traditional ML-based methods, the manual features extraction and classification is time-consuming, tedious, error-prone and requires computer experts.
The concept of deep learning refers to ML that uses multilevel neural networks for learning and extracting automatic end-to-end deep features [11,12,13,14,15,16,17]. This strategy improves detection performance while reducing the time and effort of manual feature extraction [18,19,20]. A DL-based method for pests and disease recognition in tomato leaves was proposed by Shijie et al. [21]., which obtained an average accuracy of 89%. However, this approach is only applicable to recognize pests in a simple background, which restrict the system from real-world implementation. Gandi et al. [22]., employed a Generative Adversarial Network to augment the dataset, and then the augmented dataset was fed to a pre-trained CNN model for plant disease classification. Another approach, used a DL-based method for fruit fly identification and obtained 95.68% accuracy [23]. DL-based methods to recognize ten different pest species were proposed by Dawei et al. [24]., and achieved 93.84% accuracy. Analysis of previous work shows that DL methods significantly improve pest classification performance. However, several factors hinder the existing of DL-based methods from real-world implementation such as lack of suitability for mobile devices, deficiency of robustness, lower accuracy, and a high equipment cost. Therefore, we propose a novel method for efficient pest detection based on a modified YOLOv5 model. The YOLOv5 model has several advantages over state-of-the-art object detection model, such as fast inference speed, higher mean average precision (mAP), strong customization, and lower computational complexity, which ensures the detection accuracy. We further improve the YOLOv5 model with several modifications to achieve higher mAP with lower computational cost. In the designing of the proposed model, we firstly modified the cross stage partial network (CSP) to focus more on the shallow feature extraction while the feature extraction module iterates to get more details. Furthermore, the modified select kernel (SK) in the attention module has been proposed in residual blocks, where the channel dimension is reweighted and fused. In the detection head, the multi-scale features detection is improved to detect weak and small objects. To the end of this, the major contribution of the proposed work is as follows:
  • We develop a medium-scale pest (insects) detection dataset that includes diverse images, captured in a challenging environment, where the object has high visual similarity with the background. The dataset consists of five different classes that allow a network to efficiently detect and recognize the pest species.
  • We propose YOLOv5s models with several modifications such as extending the CSP module, improving the SK Attention module, and modifying the multiscale feature extraction mechanism to efficiently detect pest, and reduce computational cost.
  • We perform experiments with various versions of YOLOv5 using a self-created dataset, where the proposed model achieves the best results in terms of model accuracy and time complexity analysis.

2. Related Work

The agriculture field plays a vital role to hike the economy of the countries, so it is essential to find-out harmful pests in natural environments. Therefore, several researchers have been working to develop an automatic system [25], for the recognition of insects in the agriculture field. Cheeti et al. [26], utilized a Convolution Neural Network (CNN) [27] and You Look Only Once (YOLO) model to classify and detect pests in the agriculture fields. They used their own manually created dataset from the internet through which they obtained promising performance in terms of the testing results. As follow-up research by Mique et al. [28], developed a technique for detecting rice pests and diseases with the help of CNN [29], and image processing. They trained their model on collected images from the internet and achieved 90.9% training results in terms of accuracy. In addition, they deployed their proposed model on the mobile application for public use. However, their technique is quite expensive and needs further improvement in terms of accuracy. Nam et al. [30], proposed a pre-trained Single Shot MultiBox Detector (SSD) model with some fine-tuning strategies for the accurate detection and classification of trap insects. They employed Deep CNN (DCNN) [31], and achieved 84% and 86% of testing accuracies using the custom dataset. However, their proposed model had an enormous performance as compared to the state-of-the-art. Li et al. [32], employed the DCNN networks (namely Faster-RCNN, Mask-RCNN, and Yolov5) for the effective detection of insects in agriculture fields using the IPI02 dataset. They achieved a promising performance and stated that the Faster-RCNN and Mask-RCNN have better results than Yolov5, which leads to 99% while Yolov5 has 97% performance in terms of accuracy, but the real-time testing speed for pest detection and localization of Yolov5 was faster than the Faster-RCNN and Mask-RCNN. Alsanea et al. [33], proposed an effective region-based CNN to detect and classify red palm weevil (RPW), their model achieved optimal performance in terms of evaluation matrices using the RPW dataset, however, the inferencing speed and model complexity restricted their method from a real-time implementation. Kouba et al. [34], utilized a sensor-based technique for monitoring agriculture. They used a customs dataset, which was created with the help of accelerometer sensors. In addition, their method is deployed on a mobile application for public use, which early detects the RPW based on their movement. Hu et al. [35], used a near-infrared imaging technology-based method and YOLOv5 for the accurate classification and detection of the pest in the agriculture fields. They obtained promising performance which was 99.7% of mAP using their custom dataset. Burhan et al. [36], compared the performance of four pre-trained deep learning models (namely VGG16, VGG19, ResNet50, and ResNet50V2) with some fine-tuning strategies for the detection of pests and identification of rice field diseases in agriculture fields. They achieved comparatively promising performance with an accuracy of 86.799%. However, their proposed model needs further improvement to enhance the performance in terms of evaluation matrices. Svenning et al., [37], proposed a pretrained CNN-based model with some fine tuning techniques for the classification of carabid beetle species. They achieved an average classification of 74.6% while it classified 51.9% of test images accurately to species level. The speed of their proposed model in the testing phase bounded the model from a real-time implementation. Chen et al. [38], proposed an AI mobile-based model for the detection of pests in the agriculture fields using a custom dataset. They focused on different types of pretrained deep learning (DL) models named faster region-based convolutional neural networks (R-CNNs), single-shot detectors (SSDs), and YOLOv4 for correct identification. They stated that the YOLOv4 achieved comparatively better performance in terms of F1-score, i.e., 100% f1-score in mealybugs, 89% in Coccide, and 97% in Diaspididae.
Liu et al. [39], proposed an end-to-end region-based DL model named PostNet for multi-class classification and identification of pests using the MPD2018 database. Their article is focused on three tiers of processing, in the first tier, they utilized the concept of Channel-Spatial Attention (CSA) to enhance the performance of the model. The second tier is based on region proposal network while the third is focused on the replacement of the fully connected (FC) layers. The experimental results show that the PestNet achieved 75.46% mAP, which is comparatively better than state-of-the-art methods. However, their model has enormous performance and needs more enhancement in terms of evaluation matrices. As a follow of research Liu et al. [40], implemented a DL-based YOLOv3 model to detect pest and tomato diseases in a natural environment in the agriculture fields using a custom dataset. Similarly, Legaspi et al. [41], implemented a DL-based YOLOv3 model for the classification and detection of pests especially whiteflies and fruit flies using a custom dataset. Their research is focused on hardware implementation namely Raspberry Pi, desktop, and web applications for public use. Their experimental result demonstrates that the model obtained 83.07% performance in terms of accuracy for the classification and detection of pests. However, their technique needs further enhancement for accurate prediction. Lim et al. [42], proposed a pre-trained convolution-based AlexNet model with several fine-tuning strategy techniques for the accurate classification of insects in real environments. Karar et al. [43], presented a DL technique based on the mobile application to recognize pests in the agriculture fields. Their experiment results are focused on Faster-RCNN for the accurate identification of pests in a real environment. The experiments of the proposed model show that the model achieved 99.0% accuracy in the testing phase. Likewise, their proposed model is comparatively better than other state-of-the-art DL architectures such as SSD and traditional back propagation (BP) neural networks. G.M.Esgario et al. [44], proposed a CNN model for automatic biotic stress detection in coffee leaves. They also developed a mobile app to assist coffee farmers. Furthermore, Habib et al. [45] proposed a traditional machine learning-based model to automatically recognize and classify the brown- and yellow-rusted diseases in wheat crops. In [46], the researchers proposed a novel DeepPestNet model for pest recognition, which consists of eight convolutional layers and three fully connected layers, and achieved higher performance. However, their method is computationally expensive.
Based on the literature review, traditional machine methods learning need two basic steps such as features extraction and classification, where the selection of an effective features extractor and classifier is a major concern, time-consuming, and requires experts in the fields. The deep learning model leverages these issues by adopting an end-to-end features extraction mechanism, however, in the literature some of the methods produced limited results and the other requires large computational resources, which cannot be deployable over resource constraints devices for real-time pest recognition.
To cope with this, we propose a DL-based YOLOv5 pipeline for the real-time recognition of pests in the natural environment. We conduct our experiment on a custom dataset and achieved promising performance in terms of accuracy. In addition, our proposed model is also focused on reducing the false prediction rate and outperforming other state-of-the-art methods. In the coming section, we provide details information on the proposed method of our article.

3. The Proposed Method

In this section, we provide a detailed description of the proposed DL-based object detection model. For object detection, one-stage DCNN such as YOLOv3, YOLOv4, and YOLOv5 have obtained remarkable performance in terms of inference speed, model size, and accuracy. In these versions, the YOLOv5 is the most recent version, which utilizes various network structures and two different varieties of CSP modules to increase the YOLOv4 performance. Therefore, this work presents a method based on YOLOv5 for pest detection and recognition by modifying the network structure. As a result, the proposed model obtained good performance to detect weak and smaller objects efficiently and effectively in the tested images. The proposed framework is demonstrated in Figure 1, which consists of three parts that are backbone, neck, and head. The dimension of the input images changes from 512 × 512 × 3 to 256 × 256 × 4 after the focus module operation. Then, we use an extended CSP module in the backbone to extract promising features from shallow and deep feature maps after the operation of the focus module. We also introduce an attention mechanism in the CSP, which focuses more on small objects in an image. In the SPP module, a concatenation operation is used for fusing the acquired results from channel dimension using four pooling layers that efficiently solve the problem in anchors and feature maps. In the neck part, a path aggregation network (PANet) is used for feature pyramid generation and the bottom-up and top-down approach is used for multiscale features fusion obtained from the backbone network and increase the object detection performance using different scales. Finally, in the head part, we use four different sets of feature maps for object detection at different scales with recognition of class labels and score predictions.

3.1. Extended CSP Module

In the DL-based model the hidden layers are gradually increasing, which extracts the semantic information of high-level features more precisely, with a reduced number of dimensions. In contrast, the dimensionality of the shallow layer is higher, which extracts the low-level feature in the network. For weak and small object detection with a few features, deep CNN may fail to achieve promising results. To increase the feature extraction capability for the weak and smaller objects in the input image, it is essential to make full use of the high-level features of the CNN in the shallow layer. Therefore, in the features extraction part, we extended the thickness of the CSP module in the shallow layer for feature extraction. This strategy is a follow-up in the upcoming layer to extract the multi-scale object features from shallow to deeper layers. Furthermore, when deepening the CSP in the entire network for feature extraction by controlling the depth and width factors, we expand the CSP module thickness to extract shallow features. The backbone part of the proposed model is shown in Figure 2. This strategy increases the model size and complexity but also increases the capability to extract prominent features from the shallow layers which are beneficial to detect small and weak objects in images. Moreover, the CSP part splits the features maps into two branches for feature extraction and then fuses them, which can obtain a richer gradient combination with a reduced number of calculations.
In the proposed model, we stack the convolution and CSP modules three times after the focus module. In this regard, the shallow layer is expended to the similar size of CSP module feedback as the deeper layer, and the feature maps with different sizes are acquired step-wise. Then, we obtain full fine-grained features of shallow and deep semantic information as shown in Figure 3, where convolution represents three basic operations that are 2D convolution, batch normalization, and activation function. By the concatenation operation, the feature maps that have two convolution branches and the attention (SK Layer) are fused. Finally, a 128 × 128 features vector is extracted in the shallow layer. When comparing the proposed model with the YOLOv5m model that added 108 layers to extend the shallow layer of the CSP module, we only added 18 layers to the network without compromising the network performance.

3.2. Modified SK Attention Module

The visualization part of any system focuses on a piece of information that helps to evaluate the image and ignores unnecessary information. In the DL-based model, an attention module can be used in the residual blocks of the shallow layer in the feature extraction stage for prominent feature selection and assigned more weights to weak and small objects to enhance feature extraction capabilities for accurate object recognition/detection [47,48]. The SK Net model adaptively modifies the receptive field size according to the multi-scales of the input information. Thus, we propose an improved version of the SK attention mechanism in each CSP and utilizes two convolutional operations with different filter sizes for channel weight learning. The output feature maps continue to perform 1 × 1 convolution as shown in Figure 4.
The improved version of the SK attention mechanism is directly employed in the residual blocks, which is divided into three parts such as (1) Split: separates the input vector to perform convolution operation with two different filters size i.e., 3 × 3 and 5 × 5 to achieve the output vectors U 1 and U 2 and acquired U after the addition process. (2) Fuse: utilizes the global average pooling denoted by ( F g p ) for the matrix compression to 1 × 1 × C, and employs a channel descriptor for the representation of each channel information. Thus, the dependency between channels is determined, which is mathematically formulated in Equations (1) and (2). The fully connected layers ( F f c ) established the relationship between the channels flexibility and nonlinearity. In the proposed work, we use two F f c layers to add nonlinearity, fit the complex correlation between channels, decrease the training parameters and computations as much as possible, and obtain the weight value, as given in the following Equation:
F g p U = 1 W × H i = 1 W j 1 H U i , j
where W represents the weight, H height, i ith row, and j jth column of the given image, respectively.
F f c   F g p ,   ϖ = δ β F g p ,   ϖ
The ϖ is used for weight, δ is the Relu activation, and β is the batch normalization.
(3) Scale: a simple weighting operation, where the calculated weight values in the fusion stage are multiplied with the original matrix to achieve the outcome of the SK blocks. This strategy improves the feature extraction process for weak and small object detection. Then the matrix columns are fused to utilize the shallow and deep layer features. We use a fully connected layer, where a sigmoid function is directly multiplied by the vector U to achieve the vector V as given in Equation (3).
F s c a l e   U ,   F f c = V 1 + V 2 = U 1 F f c + U 2 F f c
The F s c a l e   U , F f c is channel-wise multiplication to multiply the feature maps of U with the obtained weight of F f c stage, and outputting the weighted feature. In the proposed model SK is an efficient and effective module that can be directly used in the network. The SK module has a strong model generalization capability by obtaining different receptive field features and an adaptive adjustment structure that significantly detect and recognize small and even large pest in the tested images.

3.3. Multiscale Feature Detection

In the YOLOv5 object detection model, three kinds of output feature maps are used for object detection with different sizes, which utilizes 8 different downsampling output features maps for small object detection. The object in the proposed dataset for pest detection and recognition is weak and small. Therefore, we employ a feature scale to focus more on the smaller objects. When the feature maps are upsampled to 64 × 64 size, we continue to upsample the feature maps to acquire 4 downsampling feature maps. In the meantime, the extended 128 × 128 feature maps are combined with the similar size feature maps of the second layer in the backbone part of the network to make full use of the deep and shallow layer features. After the fusion of multi-scale features, the four different scales of these features are 18 × 18, 32 × 32, 64 × 64, and 128 × 128. The YOLOv5 adaptively computes suitable anchors according to different datasets, which makes the convergence capability of the model simple and detects the object with different scales. The first step is the input image selection, which will be used for the prediction. In the second step, we use four different layers for detection i.e., P2, P3, P4, and P5, which predict the values of central point tx, ty, the height th, width tw, and the confidence score. Afterward, a loss function between the ground truth and the model prediction is calculated for each detection layer. Through this, our model gradually optimizes the error rate and increases the generalization capability. The loss function for each detection layer is similar, which is obtained by computing the sum of the class loss, confidence loss, and bounding box regression loss. The mathematical expression of the P2 detection layer loss is given in Equation (4).
l o s s p 2 = l o s s c l a s s + l o s s o b j e c t + l o s s b o u n d i n _ b o x
where the bounding box loss utilizes CIoU, the class loss is computed according to the categorical cross entropy, and the confidence loss is obtained by categorical cross entropy with logistics loss to fit numerical stability.

3.4. Psuedo Code Algorithms

According to our model structure, the pseudocode is designed, which explains the training procedure of our model step-by-step as presented in Algorithm 1.
Algorithm 1: Psuedocode of the proposed model
Input: Dataset samples S = {(X1, Y1),(X2, Y2),…,(Xn, Yn)}. The S is categorized into a training set (TrainX, TrainY), a validation Set (valx, Valy), and a testing set (testx, testy), where x is the number of pest images, and y is the corresponding image labels.
T denoted the number of training epochs.
Output: converge model
Load the (TrainX, TrainY), and (valx, Valy);
Augment the (TrainX, TrainY);
Begin:
Initialize weights and biases.
For m = 1, 2, 3, …, T:
Features extraction using CSP
Input the feature SK Attention Module
Generate the attention map using SK Attention Module
Fed the extraction features from SK Attention Module to Multiscale Feature Detection
Weight the multiscale feature maps, and calculate the output of the Multiscale Feature Detection.
Model fit (Optimizer, (TrainX, TrainY)) → (M(m))
Model evaluate (M(m), (ValX, ValY)) → mAP(m).
End For
Save the optimal model which has max mAP in T epochs.
End
Load the testing set;
Load the optimal model in terms of object detection performances.

4. Dataset Collection

Dataset collection is a major part of model training in the field of artificial intelligence. In this article, we develop a new dataset that includes diverse images, captured in a challenging environment, which consists of five different classes: ants, grasshoppers, palm_weevil, shield_bug, and wasps. The ants class includes 392 frames, the grasshopper class has 315 images, palm_weevil class has 148 images, shield_bug class has 392 images, and wasps class has 318 images as tabulated in Table 1.
We annotate the dataset according to the object detection model using labeling tool which is publicly available on GitHub (accessed on 23 June 2022). The labeling is written in Python programing language and uses Qt cross-platform Graphical User Interface (GUI) toolkit (Qt GUI 5.15.11). The YOLO-based model required .txt annotated file for data labeling, therefore we annotate the dataset according to the YOLO formats.
All images have the same size, i.e., 512 × 512 and represent the three channels (Red, Green, and Blue). In addition, the datasets are separated into three portions thar are training, validation, and testing, where the training sets consist of 70% data, a validation sets consist of 20%, and a testing sets consist of 10%. Likewise, we implement a pre-trained YOLOv5s-based model with some fine-tuning strategies using the proposed dataset for the detection of pests in the agriculture fields. Our proposed model achieves promising performance in terms of evaluation matrices as compared to state-of-the-art methods. Sample images of each class of the dataset is presented in Figure 5.

5. Experiments and Results

This section provides detailed information about the dataset and the implementation of the pre-trained YOLOv5 architecture. In Section 5.1, we discuss the experimental setup of the paper; followed by the convergence of the proposed model, and then the experimental evaluations are presented.

5.1. Experimental Setup

The experimental results of this article for insect detection are conducted in Pytorch with CUDA support. All the experiments were performed on the windows operating system, equipped with a Core i7-9700KF CPU, and graphics NVIDIA Corporation TU104 (GeForce RTX 3070 Super GPU) with 8 GB of RAM. To assess the testing performance of the YOLOv5 architecture, it is difficult to use the existing evaluation metrics for each problem. The precision, recall, and mAP are utilized in our experiments as the evaluation matrices and the detailed description of these metrics are available in [49,50,51,52].
Precision is a measurement matric that is based on a confusion matrix, used to check model performance in the field of machine learning and deep learning. In addition, it is referred to as true positive samples divided by true positive and false positive samples, as formulated in Equation (5).
P = T P T P + F P
Recall sometimes referred to as ‘sensitivity’ [53]. It is only dependent on the positive samples of the data and does not care about negative samples. In addition, it is calculated as the number of true positive samples divided by the number of true positive and false negative samples, as given in Equation (6).
R = T P T P + F N
Mean average precision (mAP) is used to evaluate the object detection models, which measures the ground-truth bounding box to the localized box and gives a score. The highest score leads the model toward accurate detection, as formulated in Equation (7).
m A P = 1 N i 1 N A P i
whereas T P describes the number of positive detected samples, F P indicates the number of negative detected samples, and F N represents the number of positive samples that are not accurately classified.

5.2. Convergence Results of the Proposed Model

First, we extract the images from open-source directories and a manual labeling technique is utilized to label each pest by its name to train the YOLOv5-based architectures for detection to protect the agriculture fields. In the experiments, we use 200 epochs with a batch size of 10 to train the model. To obtain better performance, the Stochastic Gradient Descent (SGD) algorithm is utilized to train the model better and optimize the network during training. Furthermore, we store the optimal trained weights of the model after completing the training process. The result of the model is evaluated using the validation and test images. In this paper, the training and validation sets are fed to the model as input for training. The loss graphs of the training and validation is determined after 200 epochs as mentioned in Figure 6, which contains detection frame, object loss, classification loss, precision, recall, and mean average precision (mAP).
The loss graph represents the model performance that how our model is accurately predict the object. The model can achieve the target task when the loss function reaches a smaller value. The object loss function is focused on measuring the probability of the target task based on the area of interest. Higher accuracy depends on the smaller value of the loss function. The classification loss function can accurately classify the object category. The accurate classification of the object depends upon the lower loss as shown in Figure 6. Similarly, precision and recall are the model performance measuring terms in ML and DL. The higher precision and recall are the evidence of model accuracy as presented in Figure 6.
In Figure 6, the value of the loss function leads to declining during training, and the model parameters and their weights are continuously updated based on the SGD algorithm. After a few epochs, the model is capable of constantly reducing the loss function value while it is able to rapidly enhance the accuracy, recall rate, and precision as shown in Figure 6. Our proposed model achieves optimal performance in terms of evaluation matrices such as precision, recall, mAP, detection frame, object loss, and classification loss. The loss function value of the model in training and validation sets leads approximately to downward trends while the precision, recall, and mAP were on the peak on 200 epochs.
The performance of the proposed model is further investigated by the confusion matrix using the test dataset as shown in Figure 7. In the confusion matrix, we have five different categories of pests including one extra category named background FN. The basic purpose of the background FN category is to highlight the none detected object in the image. To further investigate the confusion matrix of our model, the correct prediction of ants, grasshoppers, palm-weevil, shield_bug, and wasps’ classes is 0.72, 0.87, 0.89, 0.92, and 0.84, respectively.

5.3. Comparing the Proposed Model with the Various Versions of YOLOv5 Models and the Current State-of-the-Art Models

In this section, we compare the performance of our model with nine different versions of state-of-the-art models. The performance of the proposed model is comparatively higher than the other models for pest detection such as ants, grasshoppers, palm-weevil, shield-bug, and wasps as described in Table 2. In the experiments, the Faster-RCNN achieves the second best performance as compared to other models. However, our model surpasses the Faster-RCNN by achieving higher values of precision, recall, and mAP, i.e., 0.018, 0.015, and 0.011, respectively. During the experiments, the lower performance is achieved by YOLOv3 and YOLOv4 models.
The YOLOv5n achieves an average value of 0.87, 0.878, and 0.895 for precision, recall, and mAP, respectively, which is the lowest in all YOLOv5 versions. The YOLOv5s achieves 0.906% precision, 0.835% recall, and 0.901% mAP for pests detection while the YOLOv5m obtains 0.936% precision, 0.845% recall, and 0.907% mAP. The YOLOv5l and YOLOv5x achieve good results in terms of precision, recall, and mAP as presented in Table 2, where the YOLOv5x outperformes the YOLOv5l by achieving a higher value for precision, recall, and mAP. In the Table 2, it can be observed that the proposed model achieves higher performance than other versions of YOLOv5, which surpasses the YOLOv5x by obtaining higher precision, recall, and mAP values.

5.4. Model Complexity Analysis

The detailed feasibility analysis of the proposed model in terms of parameters, model size, and frame per second (FPS) using CPU is described in Table 3. To approximate the inferencing time, we determine the Giga Floating Point Operations per Second (GFLOPs), model size, and Frame Per Second (FPS) of each model and compare it with the proposed model as presented in Table 3. The higher GFLOPs, model size, and lower inferencing speed are associated with the YOLOV5x, which restricts the system from real-world applications. In Table 3, it can be seen that the proposed model is a suitable choice for pest detection due to the higher inference speed and lower GFLOPs and model size, which increase the potential of our model to be implemented in real-time.

5.5. Visual Result of the Proposed Model

The proposed model is utilized to detect five kinds of pests in the natural environment and check the classification and identification performance of the model. The visual results of our model are shown in Figure 8. In this figure, the proposed model achieves better performance in terms of detection and classification, where the proposed model draws an accurate bounding box around the object and assigns a correct class label to the object. Thus, the visualized results of our model show real-time applicability.

6. Conclusions

In this research, we have experimented with nine different object detection models including the proposed model, which is the most efficient and accurate model as proven by the experimental section. All the experiments were performed using the manually collected dataset, which consists of five different classes i.e., ants, grasshopper, palm weevils, shield bugs, and wasps. Our model surpassed the state-of-the-art models by achieving higher values for precision, recall, and mAP, i.e., 0.018, 0.015, and 0.011, respectively. Furthermore, the proposed models can be deployed in real-time due to the higher inferencing speed with lower GFLOPs calculations and model size. As a result, the proposed model is highly capable of detecting and recognizing different types of species in real time.
In the future, we aim to increase the model performance with reduced model size and GFLOPs calculations by introducing novel mechanisms in the backbone architecture and using vision transformer for pests detection. Further, we plan to increase the number of training and pest species in the dataset to increase the model’s robustness.

Author Contributions

Conceptualization, M.A. (Mohammed Aloraini) and S.A.; methodology, S.H.; software, S.H.; validation, H.S.A.-R., M.A. (Mohammed Aladhadh) and S.A.; formal analysis, M.I.; investigation, M.A. (Mohammed Aloraini); resources, H.S.A.-R.; data curation, H.S.A.-R.; writing—original draft preparation, S.H.; writing—review and editing, M.A. (Mohammed Aloraini) and H.S.A.-R.; visualization, M.A. (Mohammed Aladhadh); supervision, S.H.; project administration, S.A.; funding acquisition, M.I. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The authors extend their appreciation to the Deputyship for Research & Innovation, Ministry of Education, Saudi Arabia for funding this research work through the project number (QU-IF-4-4-3-30507). The authors also thank to Qassim University for technical support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Amiri, A.N.; Bakhsh, A. An effective pest management approach in potato to combat insect pests and herbicide. 3 Biotech 2019, 9, 16. [Google Scholar] [CrossRef]
  2. Fernández, R.M.; Petek, M.; Gerasymenko, I.; Juteršek, M.; Baebler, S.; Kallam, K.; Giménez, E.M.; Gondolf, J.; Nordmann, A.; Gruden, K.; et al. Insect pest management in the age of synthetic biology. Plant Biotechnol. J. 2022, 20, 25–36. [Google Scholar] [CrossRef]
  3. Habib, S.; Khan, I.; Aladhadh, S.; Islam, M.; Khan, S. External Features-Based Approach to Date Grading and Analysis with Image Processing. Emerg. Sci. J. 2022, 6, 694–704. [Google Scholar] [CrossRef]
  4. Zhou, J.; Li, J.; Wang, C.; Wu, H.; Zhao, C.; Teng, G. Crop disease identification and interpretation method based on multimodal deep learning. Comput. Electron. Agric. 2021, 189, 106408. [Google Scholar] [CrossRef]
  5. Khan, Z.A.; Ullah, W.; Ullah, A.; Rho, S.; Lee, M.Y.; Baik, S.W. An Adaptive Filtering Technique for Segmentation of Tuberculosis in Microscopic Images; IEEE: New York, NY, USA, 2020; pp. 184–187. [Google Scholar]
  6. Ullah, R.; Hayat, H.; Siddiqui, A.A.; Siddiqui, U.A.; Khan, J.; Ullah, F.; Hassan, S.; Hasan, L.; Albattah, W.; Islam, M.; et al. A Real-Time Framework for Human Face Detection and Recognition in CCTV Images. Math. Probl. Eng. 2022, 2022, 3276704. [Google Scholar] [CrossRef]
  7. Al-Hiary, H.; Bani-Ahmad, S.; Reyalat, M.; Braik, M.; Alrahamneh, Z. Fast and accurate detection and classification of plant diseases. Int. J. Comput. Appl. 2011, 17, 31–38. [Google Scholar] [CrossRef]
  8. Nguyen, T.N.; Lee, S.; Nguyen-Xuan, H.; Lee, J. A novel analysis-prediction approach for geometrically nonlinear problems using group method of data handling. Comput. Methods Appl. Mech. Eng. 2019, 354, 506–526. [Google Scholar] [CrossRef]
  9. Faithpraise, F.; Birch, P.; Young, R.; Obu, J.; Faithpraise, B.; Chatwin, C. Automatic plant pest detection and recognition using k-means clustering algorithm and correspondence filters. Int. J. Adv. Biotechnol. Res. 2013, 4, 189–199. [Google Scholar]
  10. Rumpf, T.; Mahlein, A.-K.; Steiner, U.; Oerke, E.-C.; Dehne, H.-W.; Plümer, L. Early detection and classification of plant diseases with support vector machines based on hyperspectral reflectance. Comput. Electron. Agric. 2010, 74, 91–99. [Google Scholar] [CrossRef]
  11. Khan, Z.A.; Hussain, T.; Haq, I.U.; Ullah, F.U.M.; Baik, S.W. Towards efficient and effective renewable energy prediction via deep learning. Energy Rep. 2022, 8, 10230–10243. [Google Scholar] [CrossRef]
  12. Ullah, W.; Hussain, T.; Khan, Z.A.; Haroon, U.; Baik, S.W. Intelligent dual stream CNN and echo state network for anomaly detection. Knowl.-Based Syst. 2022, 253, 109456. [Google Scholar] [CrossRef]
  13. Khan, Z.A.; Hussain, T.; Baik, S.W. Boosting energy harvesting via deep learning-based renewable power generation prediction. J. King Saud Univ.-Sci. 2022, 34, 101815. [Google Scholar] [CrossRef]
  14. Yar, H.; Imran, A.S.; Khan, Z.A.; Sajjad, M.; Kastrati, Z. Towards smart home automation using IoT-enabled edge-computing paradigm. Sensors 2021, 21, 4932. [Google Scholar] [CrossRef] [PubMed]
  15. Ali, H.; Farman, H.; Yar, H.; Khan, Z.; Habib, S.; Ammar, A. Deep learning-based election results prediction using Twitter activity. Soft Comput. 2022, 26, 7535–7543. [Google Scholar] [CrossRef]
  16. Rehman, A.; Saba, T.; Kashif, M.; Fati, S.M.; Bahaj, S.A.; Chaudhry, H. A revisit of internet of things technologies for monitoring and control strategies in smart agriculture. Agronomy 2022, 12, 127. [Google Scholar] [CrossRef]
  17. Malik, M.A.; Awan, M.J.; Saleem, M.R.; Rehman, A.; Alyami, J. A Novel Method for Lung Segmentation of Chest with Convolutional Neural Network. In Prognostic Models in Healthcare: AI and Statistical Approaches; Saba, T., Rehman, A., Roy, S., Eds.; Springer: Singapore, 2022; pp. 239–260. [Google Scholar]
  18. Sajjad, M.; Khan, Z.A.; Ullah, A.; Hussain, T.; Ullah, W.; Lee, M.Y.; Baik, S.W. A novel CNN-GRU-based hybrid approach for short-term residential load forecasting. IEEE Access 2020, 8, 143759–143768. [Google Scholar] [CrossRef]
  19. Khan, Z.A.; Hussain, T.; Ullah, A.; Rho, S.; Lee, M.; Baik, S.W. Towards efficient electricity forecasting in residential and commercial buildings: A novel hybrid CNN with a LSTM-AE based framework. Sensors 2020, 20, 1399. [Google Scholar] [CrossRef] [Green Version]
  20. Shoukat, A.; Akbar, S.; Hassan, S.A.E.; Rehman, A.; Ayesha, N. An Automated Deep Learning Approach to Diagnose Glaucoma using Retinal Fundus Images. In Proceedings of the 2021 International Conference on Frontiers of Information Technology (FIT), Islamabad, Pakistan, 13–14 December 2021; IEEE: New York, NY, USA, 2021; pp. 120–125. [Google Scholar]
  21. Shijie, J.; Peiyi, J.; Siping, H. Automatic detection of tomato diseases and pests based on leaf images. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017; IEEE: New York, NY, USA, 2017; pp. 2510–2537. [Google Scholar]
  22. Gandhi, R.; Nimbalkar, S.; Yelamanchili, N.; Ponkshe, S. Plant disease detection using CNNs and GANs as an augmentative approach. In Proceedings of the 2018 IEEE International Conference on Innovative Research and Development (ICIRD), Bangkok, Thailand, 11 May 2018; IEEE: New York, NY, USA, 2018; pp. 1–5. [Google Scholar]
  23. Leonardo, M.M.; Carvalho, T.J.; Rezende, E.; Zucchi, R.; Faria, F.A. Deep feature-based classifiers for fruit fly identification (Diptera: Tephritidae). In Proceedings of the 2018 31st SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI), Paraná, Brazil, 29 October–1 November 2018; IEEE: New York, NY, USA, 2018; pp. 41–47. [Google Scholar]
  24. Dawei, W.; Limiao, D.; Jiangong, N.; Jiyue, G.; Hongfei, Z.; Zhongzhi, H. Recognition pest by image-based transfer learning. J. Sci. Food Agric. 2019, 99, 4524–4531. [Google Scholar] [CrossRef] [PubMed]
  25. Khan, Z.A.; Ullah, A.; Ullah, W.; Rho, S.; Lee, M.; Baik, S.W. Electrical energy prediction in residential buildings for short-term horizons using hybrid deep learning strategy. Appl. Sci. 2020, 10, 8634. [Google Scholar] [CrossRef]
  26. Cheeti, S.; Kumar, G.S.; Priyanka, J.S.; Firdous, G.; Ranjeeva, P.R. Pest Detection and Classification Using YOLO and CNN. Ann. Rom. Soc. Cell Biol. 2021, 25, 15295–15300. [Google Scholar]
  27. Albattah, W.; Kaka Khel, M.H.; Habib, S.; Islam, M.; Khan, S.; Kadir, K.A. Hajj Crowd Management Using CNN-Based Approach. Comput. Mater. Contin. 2021, 66, 2183–2197. [Google Scholar] [CrossRef]
  28. Mique, E.L., Jr.; Palaoag, T.D. Rice pest and disease detection using convolutional neural network. In Proceedings of the 2018 International Conference on Information Science and System, San Francisco, CA, USA, 13–16 December 2018; pp. 147–151. [Google Scholar]
  29. Habib, S.; Alsanea, M.; Aloraini, M.; Al-Rawashdeh, H.S.; Islam, M.; Khan, S. An Efficient and Effective Deep Learning-Based Model for Real-Time Face Mask Detection. Sensors 2022, 22, 2602. [Google Scholar] [CrossRef]
  30. Nam, N.T.; Hung, P.D. Pest detection on traps using deep convolutional neural networks. In Proceedings of the 2018 International Conference on Control and Computer Vision, Singapore, 15–18 June 2018; pp. 33–38. [Google Scholar]
  31. Khan, Z.A.; Ullah, A.; Haq, I.U.; Hamdy, M.; Hamdy, M.; Maurod, G.M.; Muhammad, K.; Hijji, M.; Baik, S.W. Efficient short-term electricity load forecasting for effective energy management. Sustain. Energy Technol. Assess. 2022, 53, 102337. [Google Scholar] [CrossRef]
  32. Li, W.; Zhu, T.; Li, X.; Dong, J.; Liu, J. Recommending Advanced Deep Learning Models for Efficient Insect Pest Detection. Agriculture 2022, 12, 1065. [Google Scholar] [CrossRef]
  33. Alsanea, M.; Habib, S.; Khan, N.F.; Alsharekh, M.F.; Islam, M.; Khan, S. A Deep-Learning Model for Real-Time Red Palm Weevil Detection and Localization. J. Imaging 2022, 8, 170. [Google Scholar] [CrossRef]
  34. Koubaa, A.; Aldawood, A.; Saeed, B.; Hadid, A.; Ahmed, M.; Saad, A.; Alkhouja, H.; Ammar, A.; Alkanhal, M. Smart Palm: An IoT framework for red palm weevil early detection. Agronomy 2020, 10, 987. [Google Scholar] [CrossRef]
  35. Hu, Z.; Xiang, Y.; Li, Y.; Long, Z.; Liu, A.; Dai, X.; Lei, X.; Tang, Z. Research on Identification Technology of Field Pests with Protective Color Characteristics. Appl. Sci. 2022, 12, 3810. [Google Scholar] [CrossRef]
  36. Burhan, S.A.; Minhas, S.; Tariq, A.; Hassan, M.N. Comparative study of deep learning algorithms for disease and pest detection in rice crops. In Proceedings of the 2020 12th International Conference on Electronics, Computers and Artificial Intelligence (ECAI), Bucharest, Romania, 25–27 June 2020; IEEE: New York, NY, USA, 2020; pp. 1–5. [Google Scholar]
  37. Hansen, O.L.P.; Svenning, J.-C.; Olsen, K.; Dupont, S.; Garner, B.H.; Iosifidis, A.; Price, B.W.; Høye, T.T. Species-level image classification with convolutional neural network enables insect identification from habitus images. Ecol. Evol. 2020, 10, 737–747. [Google Scholar] [CrossRef]
  38. Chen, J.-W.; Lin, W.-J.; Cheng, H.-J.; Hung, C.-L.; Lin, C.-Y.; Chen, S.-P. A smartphone-based application for scale pest detection using multiple-object detection methods. Electronics 2021, 10, 372. [Google Scholar] [CrossRef]
  39. Liu, L.; Wang, R.; Xie, C.; Yang, P.; Wang, F.; Sudirman, S.; Liu, W. PestNet: An end-to-end deep learning approach for large-scale multi-class pest detection and classification. IEEE Access 2019, 7, 45301–45312. [Google Scholar] [CrossRef]
  40. Liu, J.; Wang, X. Tomato diseases and pests detection based on improved Yolo V3 convolutional neural network. Front. Plant Sci. 2020, 11, 898. [Google Scholar] [CrossRef]
  41. Legaspi, K.R.B.; Sison, N.W.S.; Villaverde, J.F. Detection and Classification of Whiteflies and Fruit Flies Using YOLO. In Proceedings of the 2021 13th International Conference on Computer and Automation Engineering (ICCAE), Melbourne, Australia, 20–22 March 2021; IEEE: New York, NY, USA, 2021; pp. 1–4. [Google Scholar]
  42. Lim, S.; Kim, S.; Kim, D. Performance effect analysis for insect classification using convolutional neural network. In Proceedings of the 2017 7th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 24–26 November 2017; IEEE: New York, NY, USA, 2017; pp. 210–215. [Google Scholar]
  43. Karar, M.E.; Alsunaydi, F.; Albusaymi, S.; Alotaibi, S. A new mobile application of agricultural pests recognition using deep learning in cloud computing system. Alex. Eng. J. 2021, 60, 4423–4432. [Google Scholar] [CrossRef]
  44. Esgario, J.G.; de Castro, P.B.; Tassis, L.M.; Krohling, R.A. An app to assist farmers in the identification of diseases and pests of coffee leaves using deep learning. Inf. Process. Agric. 2022, 9, 38–47. [Google Scholar] [CrossRef]
  45. Habib, S.; Khan, I.; Islam, M.; Albattah, W.; Alyahya, S.M.; Khan, S.; Hassan, M.K. Wavelet frequency transformation for specific weeds recognition. In Proceedings of the 2021 1st International Conference on Artificial Intelligence and Data Analytics (CAIDA), Riyadh, Saudi Arabia, 6–7 April 2021; IEEE Prince Sultan University: Riyadh, Saudi Arabia, 2021; pp. 97–100. [Google Scholar]
  46. Ullah, N.; Khan, J.A.; Alharbi, L.A.; Raza, A.; Khan, W.; Ahmad, I. An Efficient Approach for Crops Pests Recognition and Classification Based on Novel DeepPestNet Deep Learning Model. IEEE Access 2022, 10, 73019–73032. [Google Scholar] [CrossRef]
  47. Yar, H.; Hussain, T.; Khan, Z.A.; Lee, M.Y.; Baik, S.W. Fire Detection via Effective Vision Transformers. J. Korean Inst. Next Gener. Comput. 2021, 17, 21–30. [Google Scholar]
  48. Yar, H.; Hussain, T.; Agarwal, M.; Khan, Z.A.; Gupta, S.K.; Baik, S.W. Optimized Dual Fire Attention Network and Medium-Scale Fire Classification Benchmark. IEEE Trans. Image Process. 2022, 31, 6331–6343. [Google Scholar] [CrossRef]
  49. Habib, S.; Hussain, A.; Albattah, W.; Islam, M.; Khan, S.; Khan, R.U.; Khan, K. Abnormal Activity Recognition from Surveillance Videos Using Convolutional Neural Network. Sensors 2021, 21, 8291. [Google Scholar] [CrossRef]
  50. Jan, H.; Yar, H.; Iqbal, J.; Farman, H.; Khan, Z.; Koubaa, A. Raspberry pi assisted safety system for elderly people: An application of smart home. In Proceedings of the 2020 First International Conference of Smart Systems and Emerging Technologies (SMARTTECH), Riyadh, Saudi Arabia, 3–5 November 2020; IEEE: New York, NY, USA, 2020; pp. 155–160. [Google Scholar]
  51. Yar, H.; Hussain, T.; Khan, Z.A.; Koundal, D.; Lee, M.Y.; Baik, S.W. Vision sensor-based real-time fire detection in resource-constrained IoT environments. Comput. Intell. Neurosci. 2021, 2021, 5195508. [Google Scholar] [CrossRef]
  52. Li, D.; Ahmed, F.; Wu, N.; Sethi, A.I. YOLO-JD: A Deep Learning Network for Jute Diseases and Pests Detection from Images. Plants 2022, 11, 937. [Google Scholar] [CrossRef]
  53. Khan, Z.A.; Hussain, T.; Ullah, F.U.M.; Gupta, S.K.; Lee, M.Y.; Baik, S.W. Randomly Initialized CNN with Densely Connected Stacked Autoencoder for Efficient Fire Detection. Eng. Appl. Artif. Intell. 2022, 116, 105403. [Google Scholar] [CrossRef]
Figure 1. Illustrated the overall framework of the proposed model for pest detection.
Figure 1. Illustrated the overall framework of the proposed model for pest detection.
Sensors 22 09749 g001
Figure 2. The Backbone part of the proposed work with the extended CSP module.
Figure 2. The Backbone part of the proposed work with the extended CSP module.
Sensors 22 09749 g002
Figure 3. The extended CSP module of the proposed work with modified SK attention module.
Figure 3. The extended CSP module of the proposed work with modified SK attention module.
Sensors 22 09749 g003
Figure 4. Represents the internal architecture of the modified SK attention module of the proposed work.
Figure 4. Represents the internal architecture of the modified SK attention module of the proposed work.
Sensors 22 09749 g004
Figure 5. Sample images of each class of the self-created dataset for early pest detection.
Figure 5. Sample images of each class of the self-created dataset for early pest detection.
Sensors 22 09749 g005
Figure 6. Illustrates the model effectiveness using different evaluation metrics, the X-axis represents the number of epochs and Y-axis are the corresponding score of each evaluation matrix.
Figure 6. Illustrates the model effectiveness using different evaluation metrics, the X-axis represents the number of epochs and Y-axis are the corresponding score of each evaluation matrix.
Sensors 22 09749 g006
Figure 7. Represents the confusion matrix of the proposed model using a self-created dataset.
Figure 7. Represents the confusion matrix of the proposed model using a self-created dataset.
Sensors 22 09749 g007
Figure 8. Represents the visual results of the proposed model which show the model effective analysis.
Figure 8. Represents the visual results of the proposed model which show the model effective analysis.
Sensors 22 09749 g008
Table 1. Tabulated form of the self-created dataset.
Table 1. Tabulated form of the self-created dataset.
ClassNumber of Images
Ants392
Grasshopper315
Palm_weevil148
Shield_bug392
Wasps318
Table 2. Comparative analysis of the proposed model with various versions of the YOLOv5 models.
Table 2. Comparative analysis of the proposed model with various versions of the YOLOv5 models.
ModelsClassesPrecisionRecallmAP
Faster RCNNAll0.920.890.924
Ants0.730.740.76
Grasshopper0.980.991
Palm_weevil0.990.880.98
Shield_bug0.960.971
Wasps0.940.860.91
YoloV3All0.820.870.86
Ants0.590.750.64
Grasshopper0.860.840.91
Palm_weevil0.910.930.95
Shield_bug0.880.910.93
Wasps0.870.90.88
YoloV4All0.850.870.89
Ants0.650.760.71
Grasshopper0.870.830.93
Palm_weevil0.930.920.96
Shield_bug0.90.940.95
Wasps0.880.910.89
Yolov5nAll0.870.8780.895
Ants0.5730.740.677
Grasshopper0.9230.8750.944
Palm_weevil10.9830.995
Shield_bug0.9330.9720.978
Wasps0.9220.8210.881
Yolov5sAll0.9060.8350.901
Ants0.7970.6790.781
Grasshopper0.9040.8750.88
Palm_weevil9.96510.995
Shield_bug0.9730.9150.977
Wasps0.8880.7060.871
YoloV5mAll0.9360.8450.907
Ants0.8610.6550.731
Grasshopper0.94810.995
Palm_weevil10.8160.995
Shield_bug0.9430.9340.97
Wasps0.930.8210.846
Yolov5lAll0.8490.890.917
Ants0.7230.7219.756
Grasshopper10.9330.995
Palm_weevil0.75810.984
Shield_bug0.8850.9580.965
Wasps0.8810.8390.886
Yolo5xAll0.9120.8820.921
Ants0.6970.7240.764
Grasshopper0.96910.995
Palm_weevil10.8660.975
Shield_bug0.9720.9780.992
Wasps0.9220.8410.89
Our modelAll0.9380.8960.934
Ants0.790.760.80
Grasshopper0.9810.996
Palm_weevil10.8860.98
Shield_bug0.9750.9780.993
Wasps0.9470.8560.9
Table 3. Comparing the model complexity analysis of the proposed model with five different versions of YOLOv5.
Table 3. Comparing the model complexity analysis of the proposed model with five different versions of YOLOv5.
ModelGFLOPsModel sizeFPS (CPU)
YOLOV5n4.23.6531.02
YOLOV5s1614.121.25
YOLOV5m48.340.210.16
YOLOV5l108.388.56.62
YOLOV5x204.71653.90
The proposed model4.81328.60
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Aladhadh, S.; Habib, S.; Islam, M.; Aloraini, M.; Aladhadh, M.; Al-Rawashdeh, H.S. An Efficient Pest Detection Framework with a Medium-Scale Benchmark to Increase the Agricultural Productivity. Sensors 2022, 22, 9749. https://doi.org/10.3390/s22249749

AMA Style

Aladhadh S, Habib S, Islam M, Aloraini M, Aladhadh M, Al-Rawashdeh HS. An Efficient Pest Detection Framework with a Medium-Scale Benchmark to Increase the Agricultural Productivity. Sensors. 2022; 22(24):9749. https://doi.org/10.3390/s22249749

Chicago/Turabian Style

Aladhadh, Suliman, Shabana Habib, Muhammad Islam, Mohammed Aloraini, Mohammed Aladhadh, and Hazim Saleh Al-Rawashdeh. 2022. "An Efficient Pest Detection Framework with a Medium-Scale Benchmark to Increase the Agricultural Productivity" Sensors 22, no. 24: 9749. https://doi.org/10.3390/s22249749

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop