Next Article in Journal
Using Deep Neural Networks to Evaluate Leafminer Fly Attacks on Tomato Plants
Previous Article in Journal
Socioeconomic and Environmental Impact Assessment of Different Power-Sourced Drip Irrigation Systems in Punjab, Pakistan
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

IOT-Enabled Model for Weed Seedling Classification: An Application for Smart Agriculture

1
School of Computer Science, University of Petroleum & Energy Studies, Dehradun 248007, India
2
Department of Data Science and Engineering, Manipal University Jaipur, Jaipur 303007, India
3
Department of Information Technology, Manipal University Jaipur, Jaipur 303007, India
4
Department of Computer Science and Engineering, Institute of Technology & Management, Gwalior 475001, India
5
Department of Electrical Power Engineering, Faculty of Electrical Engineering and Computer Science, VSB-Technical University of Ostrava, 708-00 Ostrava, Czech Republic
6
Department of Electrical Engineering Fundamentals, Faculty of Electrical Engineering, Wroclaw University of Science and Technology, 50-370 Wroclaw, Poland
*
Authors to whom correspondence should be addressed.
AgriEngineering 2023, 5(1), 257-272; https://doi.org/10.3390/agriengineering5010017
Submission received: 6 January 2023 / Revised: 14 January 2023 / Accepted: 23 January 2023 / Published: 29 January 2023

Abstract

:
Smart agriculture is a concept that refers to a revolution in the agriculture industry that promotes the monitoring of activities necessary to transform agricultural methods to ensure food security in an ever-changing environment. These days, the role of technology is increasing rapidly in every sector. Smart agriculture is one of these sectors, where technology is playing a significant role. The key aim of smart farming is to use the technologies to increase the quality and quantity of agricultural products. IOT and digital image processing are two commonly utilized technologies, which have a wide range of applications in agriculture. IOT is an abbreviation for the Internet of things, i.e., devices to execute different functions. Image processing offers various types of imaging sensors and processing that could lead to numerous kinds of IOT-ready applications. In this work, an integrated application of IOT and digital image processing for weed plant detection is explored using the Weed-ConvNet model to provide a detailed architecture of these technologies in the agriculture domain. Additionally, the regularized Weed-ConvNet is designed for classification with grayscale and color segmented weed images. The accuracy of the Weed-ConvNet model with color segmented weed images is 0.978, which is better than 0.942 of the Weed-ConvNet model with grayscale segmented weed images.

1. Introduction

These days, the world has more IOT-connected devices than humans. The Internet of things (IOT) is a theory which defines the notion of interrelated computing nodes, devices, or things being linked to the Internet and being able to link themselves to other computing devices [1]. It mainly comprises networks, sensor equipment, and wireless communications. The Internet of things includes electronic appliances, linked security systems, vehicles, thermostats, vending machines, ATMs, speaker systems, buildings, alarm clocks, and a wide range of other applications. The Internet of things is used to advance intelligence by offering a high level of interaction between the environment and humans. It improves reliability, flexibility, and proficiency by focusing on time saving, cost reduction, and resource usage [2]
Visual information is the most significant sort of information that can be observed, extracted, and understood by humans. Visual information processing consists of almost one third of the cortical area of the human brain. Image processing is an area that performs some processes on an acquired image generally to enhance the quality of the image or to analyze it to obtain some information [3]. Currently, image processing and analysis is among the speedily rising applied technologies. It has diverse uses such as medical image processing, remote sensing, machine vision, robotics, video processing, surveillances, self-driving cars, gesture recognitions, and many more [4].
Smart agriculture is a mechanism for agriculture farm supervision that practices information technology to ensure that the crops and soil obtain precisely what is essential for the finest productivity. The aim of smart agriculture is to ensure productivity, sustainability, and safety [5]. Smart agriculture heavily depends on specific equipment and information technology. The smart agriculture methods include real-time monitoring of crops, soil, and environment. During the last decades, smart agriculture was restricted to operations that could assist in information technology infrastructure. Additionally, related computational resources needed to be fully implemented and benefited from the profits of precision agriculture. Nowadays, mobile computing applications, smart sensors, cloud computing, and drones promote smart agriculture as feasible for farming at any scale [6,7].
The numerous roles of Internet of things technologies in smart agriculture are crop observation, water management, soil management, and pesticide control. The enhanced responsiveness of operations is one of the advantages of employing IOT in agriculture. Farmers can swiftly react to any substantial change in air, weather, humidity, quality, or the condition of each crop or soil in the field using real-time surveillance and forecast systems. Moreover, the Internet of things, with its real-time, precise, and joint characteristics can offer great modifications to the agricultural supply chain management and deliver a critical technology for forming an even flow of agricultural logistics [8,9]. Systems, drones, remotes, remote sensors, and computer imaging, combined with IOT are utilized in smart agriculture to supervise surveys, crops, and map fields, as well as deliver data to farmers for rational farm management plans to save both money and time.
There are several uses of digital image processing in smart farming. Digital Image processing has been demonstrated to be an effective mechanism for processing analysis in multiple applications of smart agriculture and farming [10]. Various kinds of imaging methods such as photometric feature-based imaging, fluorescence imaging, thermal imaging, and hyper-spectral imaging have considerably contributed to the last decades [11]. The accessibility of wireless communication networks along with image processing methods can transform the condition of receiving effective expert guidance within a short period and at reasonable price. It has an extensive range of uses such as crop management, fruit grading, plant health monitoring, nutrient deficiency recognition, weed recognition and plant height/size analysis. The image processing-based analysis of the parameters has been demonstrated to be precise and less time consuming compared with conventional approaches [12].
Image processing and Internet of things have so far been used for several applications individually. Various separate applications of these tools are available in agriculture and have attained degrees of success; however, very few applications are offered with the integration of both these technologies. By integration of digital image processing and IOT, better results can be obtained in the smart agriculture domain [13,14]. Some studies such as “Implementation of IoT and Image processing in smart agriculture” have offered the integration of these technologies. This work has offered a method to integrate IOT and digital image processing to determine environmental factors or man-made factors such as pesticides and fertilizers that are explicitly obstructing the progress of a plant [15]. The rest of the work is distributed into Section 2, Section 3 and Section 4. Section 2 provides the details of the proposed framework for an IOT-enabled framework for smart agriculture, which includes an example of weed seedling classification. Section 3 discusses the experiment and results and Section 4 presents the conclusions. This work offers the following contributions:
  • An IOT-enabled framework for smart agriculture;
  • Design of a well regularized CNN framework for weed seedling classification;
  • Analysis of the performance of weed seedling classification based on grayscale and color channel information.

2. Methods

A weed plant is undesirable and in most situations it is “a plant in the wrong place”. Weed control is important in smart agriculture. Weed plant control is the botanical section of pest control that aims to end weeds, particularly harmful weeds, from contending with wanted flora and fauna comprising domesticated plants and livestock, and in natural situations stopping non-native species that compete with native species. Approaches for weed control comprise chemical attack with herbicides, hand cultivation with hoes, powered cultivation with cultivators, stifling with mulch or soil solarization, lethal wilting with extraordinary heat, or burning [16]. However, detection of the weed plant is necessary for weed control by herbicides [17]. This section gives an overall design for an example of a setup for weed plant classification application in the smart agriculture domain. The IOT and image analytics on the cloud model for weed plant seedling detection is presented in Figure 1. In this model, the IOT device is used to capture and transfer the image of the weed plant to the cloud using a wireless network. The image analytics is performed on the cloud and the message is communicated to the farmer.
Figure 1 shows the assembly for an IOT-enabled smart capturing system and the crop sensory system. The diagram is showing the framework for the proposed approach. The layered architecture is also shown in the figure where the analytics part ran onto the cloud environment. The actual flow sequences for the IOT-enabled framework is provided below.
AI is a subdivision of computer science that deals with how to make a computer or machine behave like a human. In Figure 2 flow graph is shown. Intelligence is an intangible concept that is composed of learning, reasoning, problem solving, perception and linguistic intelligence [18,19]. Learning is the ability to improve behavior with experience. In machine learning, we explore different algorithms to build models from the available set of data. Deep learning is a significant domain behind automated vehicles, self-navigating drones and self-driving cars, permitting them to recognize a stop sign or discriminate between a lamppost, a vehicle, and a pedestrian. DL is based on unsupervised and supervised learning methods that use artificial neural networks [20,21,22,23,24,25,26,27,28,29]. DL is a subgroup of ML that has networks proficient in learning from data that are unlabeled or unstructured. The term deep, indicates the extent of hidden layers in the neural network (NN). A normal NN has 2–3 hidden layers whereas, in a deep neural network (DNN), it can have more. To train a deep learning model, it uses a large quantity of labelled data and a neural network design. Based on the connectivity of neurons in the neural network, we can have many types of DNN, e.g., multi-layer perceptron (MLP), convolutional neural network (CNN), generative adversarial networks (GAN), recurrent neural network (RNN); these and others are changing the way we interact with the world [30,31].

2.1. Regularized Convolution Neural Network

CNN is a supervised learning-based model that is useful in the vision domain. The input layer and the intermediate hidden layer are convolution layers whereas the output layer is a fully connected network. Convolution layers consist of different filters to learn the different features. Convolution kernels perform the convolution operation. A convolution is the basic process of applying a filter to an input to produce an activation. The convolution of continuous functions f and h is given below [21].
( f * g ) ( t ) = f ( τ ) g ( t τ ) d τ = f ( t τ ) g ( τ ) d τ  
The corresponding convolution operation for discrete functions F and G is defined as below.
( F * G ) ( n ) = m = F ( m ) G ( n m ) = m = F ( n m ) G ( m )  
Equivalently, the above 1-D convolution operation for the 2-D convolution case is performed as below.
( F * G ) ( r , c ) = m = M M n = N N F ( r n , c m ) G ( n , m )  
The function H is referred to as a filter since it is utilized to convolve over the image function F. There are pooling layers between two consecutive convolutions layers, which is used to reduce the likelihood of overfitting. The activation function ReLU became a popular choice in deep learning and continues to produce excellent results today. It was created to overcome the previously described vanishing gradient problem. A neuron’s activation function determines whether it should be activated. It indicates that throughout the prediction step, it will employ simplified mathematical procedures to decide whether the neuron’s input to the network is essential or not. This function is defined as below [12,22].
f ( x ) = { 0   i f   x < 0 x   i f   x 0
It has the gradient as provided below.
d d x f ( x ) = { 0   i f   x < 0 1   i f   x 0
Figure 3 displays the architecture of Weed-ConvNet model. The ConvNet algorithm 1 used in this work is presented below.
Algorithm 1. Regularized Deep Weed-ConvNet Model.
Input: A total of 12 weed plant species are represented by an image dataset of distinct plants.
Output: The plant species class represents one of the 12 possible categories for the input weed images.
1. Transform image I ( x , y ) to HSV (hue, saturation, and value) color space.
2. Resize the images 128 × 128 size for dimension reduction.
3. Perform data augmentations rotation, scaling, and flipping.
4. Learning-Forward Pass:
 For each convolution filter apply convolution on the image matrix.
  Generate feature map:
   feature_map = y   =   0 c o l u m n s ( x   =   0 r o w s f i l t e r   ( x a ,   y b ) ,   I ( x , y ) )
    Univariate vector = max (feature_map(0,y))
  end for
5. To build a single feature vector, combine all univariate features.
6. Apply the SoftMax operation to the feature vector attained in step 5 as follows:
f j ( z ) = e z j k e z k
7. Argmax (softmax_outputs)
Learning-Back Propagation:
8. Loss function (categorical cross-entropy):
L C E = 1 N i   =   1 N l o g e W y i T x i + b y i j   =   1 n e W j T x i + b j
9. Weight update:
w i = w i α E w i + β w i γ α w i
where α is learning rate, β is momentum, and is γ weight decay.
During model learning, overfitting is the most common problem. Data generalization would be weak if a simple model is used. On training data, a complex model can perform well, but on test data, it may perform poorly. The model’s efficiency would be greatly improved if overfitting is minimized. The concept of regularization is used to eliminate overfitting and the selection of a suitable complexity model issue. Under regularization, we do certain adjustments/changes in the model so that it performs well on training as well as test data. We tune the function by incorporating the penalty term with the error function. There are multiple regularization methods. The following are some of the most used regularization methods [23,24]:
L1 Regularization (Lasso penalization): The absolute value of the coefficients is added in this approach, and the sum is applied as a penalty to the error function. As a result, some parameters are reduced to zero. For obvious reasons, L1 regularization is more robust than L2 regularization. Because L2 regularization requires the square of the weights, the cost of outliers in the data grows exponentially. Because L1 regularization uses the absolute values of the weights, the cost only grows linearly.
L2 Regularization (Ridge penalization): The most common type of regularization technique is L2 regularization, which is also known as weight decay or ridge regression. Ridge regression adds a penalty term to the loss function that is the “squared magnitude” of the coefficient. L2 regularization reduces all weights to small values, preventing the model from learning any complex concept in relation to any specific node/feature and thus avoiding overfitting.
Elastic Net: It is a mixture of L1 and L2 regularization. The sum of the square of coefficients and the sum of the absolute value of coefficients is added as a penalty to the error function.
Dropout: Dropout is a learning process in which neurons are overlooked at random. They “disappear” at random intervals. It means that during the forward pass, their support for downstream neuron activation is temporarily halted, and any weight changes are unsuccessful for the neuron during the backward pass.
Data Augmentation: Data augmentation is the simplest way to reduce overfitting. Its aim is to expand the size of the training data. In this case, you can enlarge the training image data by rotating it at a certain angle, vertically or horizontally flipping it, shrinking, scaling, moving it, and so on.
Early stopping: This is a type of cross-validation scheme where the dataset is separated into the training set, testing set, and validation set. When the performance on the validation set is found poorer, the training of the model is immediately discontinued.
Add Random Noise: Adding Gaussian (random) noise with a zero mean and defined standard deviation improves the generalization error and the structure of the mapping problem in this form.
Batch Normalization: Batch normalization normalizes the yield of a preceding activation layer by deducting the batch mean and dividing by the batch standard deviation. It is a method of increasing the speed and stability of neural networks by adding extra layers to a DNN. The new layer performs standardizing and normalizing operations on the input of a previous layer.
The convolution neural network-based classification model is an effective tool for weed plant classification [25]. The Weed-ConvNet model used in weed plant classification work employs data augmentation, batch normalization, and Gaussian dropout to develop a robust regularized weed plant classification model.

2.2. Image Data Set

A public image dataset of weed images of approximately 4234 distinct plants representing 12 weed plant species [26] is used in this work. The Aarhus University Signal Processing group collected these 12 weed plant seedling species. The sample weed images from each class of species are shown in Figure 4.
All the weed images are segmented using color image processing in the HSV space followed by morphological operations. The sample segmented weed images are shown in Figure 5.

3. Experiments and Results Discussion

To implement Weed-ConvNet, the most effective architecture applies one or more stacks of convolution + pooling layers with suitable activation function, followed by a flatten layer and then finally one or two dense layers. The convolution neural network model with five convolution layers is designed for this weed plant classification problem. All these layers are followed by activation layers.
Batch normalization layers are used after activation layers. These normalization layers at each batch normalize the activations of the previous layer, i.e., converts the mean activation close to 0 and the activation standard deviation close to 1. Gaussian dropout layers are also applied after the max pooling layers. These dropout layers apply multiplicative 1-centered Gaussian noise. As it is a regularization layer, it remains active only at training time. The last layer is the dense layer with the SoftMax activation function. Table 1 gives the details of each layer of the Weed-ConvNet model including output shape and parameters. Total trainable and non-trainable parameters are 17,329,804 and 3264, respectively.
To reduce the computation cost, the segmented images are cropped to size 64 × 64 from center. Data augmentation was performed over the cropped images. Over the cropped weed images, five main types of data augmentation techniques are used: image shifts with 1-pixel width and height shift range parameters, image flips with horizontal and vertical flip cases, image rotations with a 10-degree rotation range parameter, image brightness, and image zoom. The image dataset consists of total 5659 weed images. The dataset is divided into a 70% training dataset and a 30% testing dataset, i.e., 3961 and 1698 weed images in each category, respectively.
The learning parameters as provided in Table 2 are applied to compile the model. The word metric in this table refers to how the model’s efficiency is calculated by classification accuracy. A loss function is a function that is used to calculate a loss value that the training method then attempts to reduce by adjusting the weights. The word optimizer refers to a method for determining how the network weights will be updated based on the loss function’s performance. The Adam optimization algorithm is one of the best optimization procedures available, and it works well with a variety of deep learning models.
The first step in the actual learning of the designed model is the fitting of the training data. Depending on the scale of the dataset, preparation takes up most of the time. An epoch is a complete transit of the training weed data through the learning algorithm. The algorithm’s epoch number is a significant hyperparameter. It defines the number of epochs or full passes through the algorithm’s learning phase for the training set. We partition the epoch into numerous smaller batches since one epoch is too large to provide to the computer all at once. The number of samples from the training set considered during each training iteration is referred to as the batch size. The weights are adjusted and the loss is calculated based on the batch size. The training process is completed with the total number of instances once after one epoch is finished. The validation dataset is used to evaluate the model’s learning performance. The training is repeated for the specified number of epochs. Stopping early and establishing the validation accuracy stabilization between consecutive epochs can improve this iteration. The learning hyperparameters that are used in this study are listed in Table 3.
Two separate experiments were performed to measure the classification performance. In the first experiment, the grayscale segmented weed images are used, while in the second experiment, color segmented weed images are used. The same Weed-ConvNet model is utilized in both experiments. The training is repeated for 50 epochs in both cases. Figure 6 and Figure 7 display the training and testing accuracy and loss curves for weed plant classification with grayscale segmented weed images and color segmented weed images, respectively.
From the above curves, it is marked that there is a narrow difference between the training and the validation loss. This specifies that the network has learned the training data properly. These trained models neither overfitted nor underfitted means well regularized. Table 4 presents the test results in terms of a confusion matrix for Weed-ConvNet-based classification with grayscale segmented weed images. A confusion matrix is a metric designed to assess the quality of a classification model. It is a square matrix whose dimensions the number of classes in a classifier determines. In this N × N matrix, rows denote to true classes and the columns signify the categorized class by the model.
Table 5 presents the results in terms of a confusion matrix for Weed-ConvNet-based classification with color segmented weed images. By comparing the confusion matrices, it can be observed that the color segmented weed images give low false classifications.
Output metrics are measured using the above confusion matrices to further test the qualified models. Precision, recall, and the F1 score are all used as performance indicators. The proportion of accurate predictions to total predictions is defined as classification accuracy. Precision is described as the proportion of correct predictions to the total number of correct predictions expected. This metric assesses the classifier’s ability to predict positive outcomes. The ratio of correct positive predictions to total positive predictions is referred to as sensitivity or recall. The F1 score combines precision and recall in relation to a specific positive class. The F1 score is a weighted average of recall and precision, with one signifying the best and zero representing the worst [27]. These metrics are specified below.
Precision = T r u e + T r u e + + F a l s e +
Sensitivity = T r u e + T r u e + + F a l s e
F 1 Score = 2 * Precision * Sensitivity Precision + Sensitivity
Accuracy = T r u e + + T r u e T r u e + + T r u e + F a l s e + F a l s e +
where:
  • T r u e + and T r u e are the truly labelled positive and negative weed samples, respectively.
  • F a l s e + represents the no. of negative weed samples labelled incorrectly as positive.
  • F a l s e denotes the no. of positive weed samples labelled incorrectly as negative.
Table 6 consists of these measures. These three performance measures are higher in color segmented weed images which are 0.86, 0.87, and 0.86 respectively. The test accuracy of the Weed-ConvNet model with the color segmented weed images is 0.978, which is better than 0.942 of the Weed-ConvNet model with the grayscale segmented weed images.
Figure 8 displays the ROC curves to affirm the robustness of these weed plant classification problems. ROC analysis is a graphical method for analyzing a classifier’s performance. It characterizes a classifier’s performance using two statistics: true positive rate (TPR) and false positive rate (FPR). The x-axis of a ROC curve normally shows the FPR, while the y-axis displays the TPR. This means that the plot’s top left corner is the “ideal” point, with a TPR of one and a FPR of zero. This implies that having a larger area under the curve (AUC) is generally better. These ROC plots show that the Weed-ConvNet model with color segmented weed images outperforms greyscale segmented weed images, as indicated by the area under curves of these curves.
ROC plots are commonly used in binary classification to investigate a classifier’s performance. Binarizing the output is needed to extend the ROC plot and AUC to multi-label or multi-class problems. Micro-averaging is a binarized plot in which the ROC is plotted by treating each member of the label matrix as a binary classification. Tang et al. developed an alternative multi-class label evaluation metric that delivers the same weight to each class label (2014). As shown in these ROC curves, for the first Weed-ConvNet model, the micro-average area and macro-average area are 09.4 and 0.96, respectively. With the color segmented weed images, the metrics micro-average area and macro-average area improve to 0.99 in each case.

4. Conclusions

IOT technology-based smart farming can empower cultivators and farmers to decrease waste and improve production. With the help of an IOT-based system for monitoring the crop field, crops can be improved with the help of humidity, light, image, temperature, and moisture sensors. Out of these sensors, an image-acquisition-enabled IOT device plays a vital role. This work discussed a case study where image processing and IOT played an important role for weed plant seedling detection and classification. The early detection of weed seedling can enable farmers to use herbicides or other control mechanisms on time. The regularized Weed-ConvNet is made to classify weed images that have been colored and gray scaled. The Weed-ConvNet model’s accuracy with color segmented weed images is 0.978, which is higher than the model’s accuracy with grayscale segmented weed images of 0.942. In future, it will be vital to determine that on which crops can an IOT-enabled model be applied to, since weeds spread amongst different crops. It might not be able to distinguish some weeds from crops since they resemble them so closely. Additionally, the density of plant seeding, stage of plant development, and other aspects will have a big impact on the IOT-enabled model’s ability to identify weeds.

Author Contributions

Conceptualization, S.T. and M.J.; methodology, S.T.; software, A.K.S.; validation, S.T. and A.K.S.; formal analysis, A.J.; investigation, R.G., D.G. and Z.L.; resources, M.J., R.G. and A.J.; writing—original draft preparation, S.T.; writing—review and editing, A.K.S., M.J., R.G. and M.G.; visualization, Z.L. and M.G.; supervision, D.G. All authors have read and agreed to the published version of the manuscript.

Funding

SGS Grant from VSB—Technical University of Ostrava under grant number SP2023/005.

Informed Consent Statement

Not Applicable.

Data Availability Statement

There are no data available for this research.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gubbi, J.; Buyya, R.; Marusic, S.; Palaniswami, M. Internet of Things (IoT): A vision, architectural elements, and future directions. Future Gener. Comput. Syst. 2013, 29, 1645–1660. [Google Scholar] [CrossRef] [Green Version]
  2. Lee, I.; Lee, K. The Internet of Things (IoT): Applications, investments, and challenges for enterprises. Bus. Horiz. 2015, 58, 431–440. [Google Scholar] [CrossRef]
  3. Goenka, N.; Tiwari, S. AlzVNet: A volumetric convolutional neural network for multiclass classification of Alzheimer’s disease through multiple neuroimaging computational approaches. Biomed. Signal Process. Control 2022, 74, 103500. [Google Scholar] [CrossRef]
  4. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Lipper, L.; Thornton, P.; Campbell, B.M.; Baedeker, T.; Braimoh, A.; Bwalya, M.; Hottle, R. Climate-smart agriculture for food security. Nat. Clim. Change 2014, 4, 1068. [Google Scholar] [CrossRef]
  6. Sinha, B.B.; Dhanalakshmi, R. Recent advancements and challenges of Internet of Things in smart agriculture: A survey. Future Gener. Comput. Syst. 2022, 126, 169–184. [Google Scholar] [CrossRef]
  7. Adamides, G.; Edan, Y. Human–robot collaboration systems in agricultural tasks: A review and roadmap. Comput. Electron. Agric. 2023, 204, 107541. [Google Scholar] [CrossRef]
  8. Gondchawar, N.; Kawitkar, R.S. IoT based smart agriculture. Int. J. Adv. Res. Comput. Commun. Eng. 2016, 5, 838–842. [Google Scholar]
  9. TongKe, F. Smart agriculture based on cloud computing and IOT. J. Converg. Inf. Technol. 2013, 8, 210–216. [Google Scholar]
  10. Jhuria, M.; Kumar, A.; Borse, R. Image processing for smart farming: Detection of disease and fruit grading. In Proceedings of the 2013 IEEE Second International Conference on Image Information Processing (ICIIP-2013), Shimla, India, 9–11 December 2013; pp. 521–526. [Google Scholar]
  11. Bhange, M.; Hingoliwala, H.A. Smart farming: Pomegranate disease detection using image processing. Procedia Comput. Sci. 2015, 58, 280–288. [Google Scholar] [CrossRef] [Green Version]
  12. Hamuda, E.; Glavin, M.; Jones, E. A survey of image processing techniques for plant extraction and segmentation in the field. Comput. Electron. Agric. 2016, 125, 184–199. [Google Scholar] [CrossRef]
  13. Mekala, M.S.; Viswanathan, P. A Survey: Smart agriculture IoT with cloud computing. In Proceedings of the 2017 International Conference on Microelectronic Devices, Circuits and Systems (ICMDCS), Vellore, India, 10–12 August 2017; pp. 1–7. [Google Scholar]
  14. Suma, N.; Samson, S.R.; Saranya, S.; Shanmugapriya, G.; Subhashri, R. IOT based smart agriculture monitoring system. Int. J. Recent Innov. Trends Comput. Commun. 2017, 5, 177–181. [Google Scholar]
  15. Kapoor, A.; Bhat, S.I.; Shidnal, S.; Mehra, A. Implementation of IoT (Internet of Things) and Image processing in smart agriculture. In Proceedings of the 2016 International Conference on Computation System and Information Technology for Sustainable Solutions (CSITSS), Bengaluru, India, 6–8 October 2016; pp. 21–26. [Google Scholar]
  16. Lottes, P.; Khanna, R.; Pfeifer, J.; Siegwart, R.; Stachniss, C. UAV-based crop and weed classification for smart farming. In Proceedings of the 2017 IEEE International Conference on Robotics and Automation (ICRA), Singapore, 29 May–3 June 2017; pp. 3024–3031. [Google Scholar]
  17. Sadjadi, F.A.; Farmer, M.E. Smart Weed Recognition/Classification System. U.S. Patent No. 5,606,821, 4 March 1997. [Google Scholar]
  18. Dang, M.P.; Le, H.G.; Le Chau, N.; Dao, T.P. An Optimized Design of New XYθ Mobile Positioning Microrobotic Platform for Polishing Robot Application Using Artificial Neural Network and Teaching-Learning Based Optimization. Complexity 2022, 2022, 2132005. [Google Scholar] [CrossRef]
  19. Dang, M.P.; Le, H.G.; Nguyen, N.P.; Le Chau, N.; Dao, T.P. Optimization for a New XY Positioning Mechanism by Artificial Neural Network-Based Metaheuristic Algorithms. Comput. Intell. Neurosci. 2022, 2022, 9151146. [Google Scholar] [CrossRef]
  20. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436. [Google Scholar] [CrossRef] [PubMed]
  21. Sharma, A.K.; Tiwari, S.; Aggarwal, G.; Goenka, N.; Kumar, A.; Chakrabarti, P.; Jasiński, M. Dermatologist-Level Classification of Skin Cancer Using Cascaded Ensembling of Convolutional Neural Network and Handcrafted Features Based Deep Neural Network. IEEE Access 2022, 10, 17920–17932. [Google Scholar] [CrossRef]
  22. Tiwari, S. An analysis in tissue classification for colorectal cancer histology using convolution neural network and colour models. Int. J. Inf. Syst. Model. Des. 2018, 9, 1–19. [Google Scholar] [CrossRef]
  23. Tripathi, N.; Jadeja, A. A survey of regularization methods for deep neural network. Int. J. Comput. Sci. Mob. Comput. 2014, 3, 429–436. [Google Scholar]
  24. Mikołajczyk, A.; Grochowski, M. Data augmentation for improving deep learning in image classification problem. In Proceedings of the 2018 International Interdisciplinary PhD Workshop (IIPhDW), Swinoujście, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar]
  25. Alimboyong, C.R.; Hernandez, A.A. An Improved Deep Neural Network for Classification of Plant Seedling Images. In Proceedings of the 2019 IEEE 15th International Colloquium on Signal Processing & Its Applications (CSPA), Penang, Malaysia, 8–9 March 2019; pp. 217–222. [Google Scholar]
  26. Giselsson, T.M.; Jørgensen, R.N.; Jensen, P.K.; Dyrmann, M.; Midtiby, H.S. A public image database for benchmark of plant seedling classification algorithms. arXiv 2017, arXiv:1711.05458. [Google Scholar]
  27. Tiwari, S. A Pattern Classification Based approach for Blur Classification. Indones. J. Electr. Eng. Inform. 2017, 5, 162–173. [Google Scholar]
  28. Ramani, P.; Pradhan, N.; Sharma, A.K. Classification Algorithms to Predict Heart Diseases—A Survey. In Computer Vision and Machine Intelligence in Medical Image Analysis. Advances in Intelligent Systems and Computing; Gupta, M., Konar, D., Bhattacharyya, S., Biswas, S., Eds.; Springer: Singapore, 2020; Volume 992. [Google Scholar]
  29. Sharma, N.; Litoriya, R.; Sharma, A. Application and Analysis of K-Means Algorithms on a Decision Support Framework for Municipal Solid Waste Management. In Advanced Machine Learning Technologies and Applications. AMLTA 2020. Advances in Intelligent Systems and Computing; Hassanien, A., Bhatnagar, R., Darwish, A., Eds.; Springer: Singapore, 2021; Volume 1141. [Google Scholar]
  30. Tang, K.; Wang, R.; Chen, T. Towards maximizing the area under the ROC curve for multi-class classification problems. In Proceedings of the Twenty-Fifth AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 7–11 August 2011. [Google Scholar]
  31. Le Chau, N.; Tran, N.T.; Dao, T.P. Topology and size optimization for a flexure hinge using an integration of SIMP, deep artificial neural network, and water cycle algorithm. Appl. Soft Comput. 2021, 113, 108031. [Google Scholar] [CrossRef]
Figure 1. (a): A typical model of IOT-based image analytics for smart agriculture [13]. (b): Layered architecture for IOT-enabled smart agriculture consisting of a crop collection layer, sensory layer, analytical/cloud layer, and a prediction layer.
Figure 1. (a): A typical model of IOT-based image analytics for smart agriculture [13]. (b): Layered architecture for IOT-enabled smart agriculture consisting of a crop collection layer, sensory layer, analytical/cloud layer, and a prediction layer.
Agriengineering 05 00017 g001
Figure 2. Flow graph of the entire system.
Figure 2. Flow graph of the entire system.
Agriengineering 05 00017 g002
Figure 3. Architecture of the Weed-ConvNet model.
Figure 3. Architecture of the Weed-ConvNet model.
Agriengineering 05 00017 g003
Figure 4. Weed plant image dataset samples.
Figure 4. Weed plant image dataset samples.
Agriengineering 05 00017 g004
Figure 5. Segmented weed plant image dataset samples.
Figure 5. Segmented weed plant image dataset samples.
Agriengineering 05 00017 g005
Figure 6. Training and testing accuracy and loss curves for weed plant classification with grayscale segmented weed images.
Figure 6. Training and testing accuracy and loss curves for weed plant classification with grayscale segmented weed images.
Agriengineering 05 00017 g006
Figure 7. Training and testing accuracy and loss curves for weed plant classification with color segmented weed images.
Figure 7. Training and testing accuracy and loss curves for weed plant classification with color segmented weed images.
Agriengineering 05 00017 g007
Figure 8. Receiver operating characteristic curves: (a) experiment with grayscale segmented weed images; (b) experiment with grayscale segmented weed images.
Figure 8. Receiver operating characteristic curves: (a) experiment with grayscale segmented weed images; (b) experiment with grayscale segmented weed images.
Agriengineering 05 00017 g008
Table 1. Layers of regularized Weed-ConvNet model.
Table 1. Layers of regularized Weed-ConvNet model.
Layer (Type)Output ShapeParameters
c o n v 2 d _ 1   ( C o n v 2 D ) ( N o n e ,   64 ,   64 ,   32 )   896
a c t i v a t i o n _ 1   ( R e L U ) ( N o n e ,   64 ,   64 ,   32 ) 0
b a t c h _ n o r m a l i z a t i o n _ 1 ( N o n e ,   64 ,   64 ,   32 ) 128
m a x _ p o o l i n g 2 d _ 1 ( N o n e ,   32 ,   32 ,   32 )   0
g a u s s i a n _ d r o p o u t _ 1 ( N o n e ,   32 ,   32 ,   32 ) 0
c o n v 2 d _ 2   ( C o n v 2 D ) ( N o n e ,   32 ,   32 ,   64 ) 18496
a c t i v a t i o n _ 2   ( R e L U ) ( N o n e ,   32 ,   32 ,   64 ) 0
b a t c h _ n o r m a l i z a t i o n _ 2 ( N o n e ,   32 ,   32 ,   64 ) 256
m a x _ p o o l i n g 2 d _ 2   ( N o n e ,   16 ,   16 ,   64 ) 0
g a u s s i a n _ d r o p o u t _ 2 ( N o n e ,   16 ,   16 ,   64 ) 0
c o n v 2 d _ 3   ( C o n v 2 D ) ( N o n e ,   16 ,   16 ,   128 ) 73856
a c t i v a t i o n _ 3   ( R e L U ) ( N o n e ,   16 ,   16 ,   128 ) 0
b a t c h _ n o r m a l i z a t i o n _ 3 ( N o n e ,   16 ,   16 ,   128 ) 512
m a x _ p o o l i n g 2 d _ 3 ( N o n e ,   8 ,   8 ,   128 ) 0
g a u s s i a n _ d r o p o u t _ 3 ( N o n e ,   8 ,   8 ,   128 )   0
c o n v 2 d _ 4   ( C o n v 2 D ) ( N o n e ,   8 ,   8 ,   128 )       147584
a c t i v a t i o n _ 4   ( R e L U ) ( N o n e ,   8 ,   8 ,   128 ) 0
b a t c h _ n o r m a l i z a t i o n _ 4 ( N o n e ,   8 ,   8 ,   128 ) 512
c o n v 2 d _ 5   ( C o n v 2 D ) ( N o n e ,   8 ,   8 ,   256 )   295168
b a t c h _ n o r m a l i z a t i o n _ 5 ( N o n e ,   8 ,   8 ,   256 ) 1024
g a u s s i a n _ d r o p o u t _ 4 ( N o n e ,   8 ,   8 ,   256 )       0
f l a t t e n _ 1   ( F l a t t e n ) ( N o n e ,   16384 )   0
d e n s e _ 1   ( D e n s e )     ( N o n e ,   1024 ) 16778240
b a t c h _ n o r m a l i z a t i o n _ 6 ( N o n e ,   1024 ) 4096
d e n s e _ 2   ( D e n s e ) ( N o n e ,   12 ) 12300
a c t i v a t i o n _ 5   ( S o f t m a x ) ( N o n e ,   12 ) 0
T r a i n a b l e   p a r a m e t e r s :   17 , 329 , 804
N o n t r a i n a b l e   p a r a m e t e r s :   3 , 264
Table 2. Learning parameters of Weed-ConvNet model.
Table 2. Learning parameters of Weed-ConvNet model.
Learning ParameterMetric
Metric accuracy
Loss function categorical   cross   entropy
Optimizer adam
Table 3. Training hyperparameter optimization for Weed-ConvNet model.
Table 3. Training hyperparameter optimization for Weed-ConvNet model.
Training HyperparameterSearch SpaceSelected Value
Learning rate[0.1, 0.01, 0.001, 0.0001, 0.00001] 0.001
Epochs[10, 20, 30, 40, 50, 60, 80, 100] 50
Batch size[8, 16, 32, 64] 32
Early stopping p a r a m e t e r :   v a l l o s s ,   p a t i e n c e   [ 1 , 2.3.4.5 ] p a t i e n c e :   1
Conv2D layer 1 channels[8, 16, 32, 48][16]
Conv2D layer 2 channels[8, 16, 32, 48][16]
Conv2D layer 3 channels[16, 32, 48, 64, 128][32]
Conv2D layer 4 channels[16, 32, 48, 64, 128][64]
Conv2D layer 5 channels[16, 32, 48, 64, 128, 256][128]
Kernel size for layers [1,2,3,4,5,6,7,8][8]
Padding[0,1,2,3,4][0]
Conv2D stride[1,2,3,4][2]
Dropout rate layer 1,2[0.1, 0.2, 0.3, 0.4][0.2]
Dropout rate layer 1,2[0.1, 0.2, 0.3, 0.4][0.4]
Table 4. Confusion matrix for Weed-ConvNet-based classification with grayscale segmented weed images.
Table 4. Confusion matrix for Weed-ConvNet-based classification with grayscale segmented weed images.
Class B l a c k g r a s s C h a r l o c k C l e a v e r s C o m m o n   C h i c k w e e d C o m m o n   w h e a t F a t   H e n L o o s e   S i l k y b e n t M a i z e S c e n t l e s s   M a y w e e d S h e p h e r d s   P u r s e S m a l l f l o w e r e d   C r a n e s b i l l S u g a r   b e e t
B l a c k g r a s s 140104004210007
C h a r l o c k 01234032000006
C l e a v e r s 07740125000844
C o m m o n   C h i c k w e e d 020101021501335160
C o m m o n   w h e a t 0020621211004
F a t   H e n 022061341200211
L o o s e   S i l k y b e n t 2002061115060005
M a i z e 0152210650003
S c e n t l e s s   M a y w e e d 81263043015841015
S h e p h e r d s   P u r s e 01924310042503
S m a l l f l o w e r e d   C r a n e s b i l l 0913032230013420
S u g a r   b e e t 030061010001118
Table 5. Confusion matrix for Weed-ConvNet-based classification with color segmented weed images.
Table 5. Confusion matrix for Weed-ConvNet-based classification with color segmented weed images.
Class B l a c k g r a s s C h a r l o c k C l e a v e r s C o m m o n   C h i c k w e e d C o m m o n   w h e a t F a t   H e n L o o s e   S i l k y b e n t M a i z e S c e n t l e s s   M a y w e e d S h e p h e r d s   P u r s e S m a l l f l o w e r e d   C r a n e s b i l l S u g a r   b e e t
B l a c k g r a s s 46000444400025
C h a r l o c k 11302012000002
C l e a v e r s 0699013000005
C o m m o n   C h i c k w e e d 01021021030012
C o m m o n   w h e a t 2011535102008
F a t   H e n 00210154000021
L o o s e   S i l k y b e n t 351105918510035
M a i z e 0000100770001
S c e n t l e s s   M a y w e e d 33170102155105
S h e p h e r d s   P u r s e 0302010015202
S m a l l f l w .   C r a n e s b i l l 02201100201762
S u g a r   b e e t 01000102003132
Table 6. Performance metrics for regularized Weed-ConvNet-based classification with grayscale and color segmented weed images.
Table 6. Performance metrics for regularized Weed-ConvNet-based classification with grayscale and color segmented weed images.
ClassPerformance Measures with Grayscale Segmented Weed ImagesPerformance Measures with Color Segmented Weed Images
PrecisionRecallF1 ScorePrecisionRecallF1 Score
Black-grass0.330.130.190.530.440.48
Charlock0.630.890.740.880.940.91
Cleavers0.610.650.630.920.870.89
Common chickweed0.740.460.570.950.950.95
Common wheat0.310.850.450.780.730.75
Fat hen0.860.840.850.850.960.9
Loose silky-bent0.760.610.680.80.760.78
Maize0.560.820.670.910.970.94
Scentless mayweed0.910.470.620.970.870.92
Shepherd’s purse0.640.410.50.980.850.91
Small-flowered cranesbill0.940.720.820.940.950.94
Sugar beet0.460.850.60.780.950.85
Average0.70.640.640.860.870.86
Test Loss0.1690.059
Test Accuracy0.9420.978
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tiwari, S.; Sharma, A.K.; Jain, A.; Gupta, D.; Gono, M.; Gono, R.; Leonowicz, Z.; Jasiński, M. IOT-Enabled Model for Weed Seedling Classification: An Application for Smart Agriculture. AgriEngineering 2023, 5, 257-272. https://doi.org/10.3390/agriengineering5010017

AMA Style

Tiwari S, Sharma AK, Jain A, Gupta D, Gono M, Gono R, Leonowicz Z, Jasiński M. IOT-Enabled Model for Weed Seedling Classification: An Application for Smart Agriculture. AgriEngineering. 2023; 5(1):257-272. https://doi.org/10.3390/agriengineering5010017

Chicago/Turabian Style

Tiwari, Shamik, Akhilesh Kumar Sharma, Ashish Jain, Deepak Gupta, Miroslava Gono, Radomir Gono, Zbigniew Leonowicz, and Michał Jasiński. 2023. "IOT-Enabled Model for Weed Seedling Classification: An Application for Smart Agriculture" AgriEngineering 5, no. 1: 257-272. https://doi.org/10.3390/agriengineering5010017

APA Style

Tiwari, S., Sharma, A. K., Jain, A., Gupta, D., Gono, M., Gono, R., Leonowicz, Z., & Jasiński, M. (2023). IOT-Enabled Model for Weed Seedling Classification: An Application for Smart Agriculture. AgriEngineering, 5(1), 257-272. https://doi.org/10.3390/agriengineering5010017

Article Metrics

Back to TopTop