A Review of Plant Disease Detection Systems for Farming Applications

: The globe and more particularly the economically developed regions of the world are currently in the era of the Fourth Industrial Revolution (4IR). Conversely, the economically developing regions in the world (and more particularly the African continent) have not yet even fully passed through the Third Industrial Revolution (3IR) wave, and Africa’s economy is still heavily dependent on the agricultural ﬁeld. On the other hand, the state of global food insecurity is worsening on an annual basis thanks to the exponential growth in the global human population, which continuously heightens the food demand in both quantity and quality. This justiﬁes the signiﬁcance of the focus on digitizing agricultural practices to improve the farm yield to meet the steep food demand and stabilize the economies of the African continent and countries such as India that are dependent on the agricultural sector to some extent. Technological advances in precision agriculture are already improving farm yields, although several opportunities for further improvement still exist. This study evaluated plant disease detection models (in particular, those over the past two decades) while aiming to gauge the status of the research in this area and identify the opportunities for further research. This study realized that little literature has discussed the real-time monitoring of the onset signs of diseases before they spread throughout the whole plant. There was also substantially less focus on real-time mitigation measures such as actuation operations, spraying pesticides, spraying fertilizers, etc., once a disease was identiﬁed. Very little research has focused on the combination of monitoring and phenotyping functions into one model capable of multiple tasks. Hence, this study highlighted a few opportunities for further focus.


Introduction
Over the last two decades, we have seen a significant increase in the discussion of the Fourth Industrial Revolution (4IR) among academics and policymakers in both developing and industrialized countries [1].The 4IR critique is marked by the merging of the real and virtual worlds and the collapse of almost all industries [1,2].For others, the assembling of cyber-physical systems, cloud technology, the Internet of Things, and the Internet of Services and their integration while interacting with humans in real time to maximize the generation of value is known as the Fourth Industrial Revolution [3].Some thinkers assert that some old jobs will vanish because of the alleged revolutionary power of 4IR, opening the door for a new array of jobs and markets that will necessitate the creation of new areas of expertise [1][2][3].The word "fourth" typically implies that there have been three revolutions before the Industrial Revolution 4.0 [3].Through mechanization and steam engines, the First Industrial Revolution greatly increased the productivity of manufacturing methods.Because there was more electrical power available during the second, assembly lines and mass production became a reality [4,5].The Third Industrial Revolution saw the widespread adoption of computing and digitalization [6].The 4IR is currently where we are, and this era is dominated using cyber-physical systems to improve life-sustaining processes such as production works (refer to Figure 1).Growth in automation marked each shift from one revolution to the next [3].Productivity rose by approximately 50-fold with each revolution, even if many jobs from the previous industrial age were rendered obsolete [7].All revolutions by their very nature are disruptive, and the preceding three revolutions brought about significant modifications to the economic and social landscape [6,7].
Appl.Sci.2023, 13, x FOR PEER REVIEW 2 of 29 assembly lines and mass production became a reality [4,5].The Third Industrial Revolution saw the widespread adoption of computing and digitalization [6].The 4IR is currently where we are, and this era is dominated using cyber-physical systems to improve lifesustaining processes such as production works (refer to Figure 1).Growth in automation marked each shift from one revolution to the next [3].Productivity rose by approximately 50-fold with each revolution, even if many jobs from the previous industrial age were rendered obsolete [7].All revolutions by their very nature are disruptive, and the preceding three revolutions brought about significant modifications to the economic and social landscape [6,7].In the 1970s, it was believed that automating repetitive tasks would liberate people, resulting in more free time and less working time [1].Despite advancements in technology, this promise remained mostly unmet [1,2].Now, the Fourth Industrial Revolution, which builds on digitalization and information and communication technologies (ICT), is thought to be revolutionizing everything [6].It is projected that new technologies including artificial intelligence (AI), biotechnology, the Internet of Things (IoT), quantum computing, and nanotechnology will alter how we interact with one another, perform our jobs, run our economies, and even "the mere meaning of being a human being" [7].It should be noted that the definition of the Fourth Industrial Revolution employed in this paper brightens a technology-centric understanding of 4IR; however, one should bear in mind the other important factors including the implications for society, politics, law, and ethics.Even though 4IR has been the topic of discussion on many international forums, there have not been many systematic attempts to analyze the state of the art of this new industrial revolution wave [6].This situation may be more apparent in Africa, where the Third Industrial Revolution has mostly not even fully begun [1][2][3].Therefore, African academics have expressed skepticism and caution regarding the alleged advantages of information and communications technology (ICT) in African environments.Swaminathan [8] stated the following: "Such a dream of transforming an agro-based economy into an information society must either be a flight of fancy or thinking hardly informed by the industrial economic background of developed economies that are in transition to informational economies.For an economy with about half of its adult population engaged in the food production sector, In the 1970s, it was believed that automating repetitive tasks would liberate people, resulting in more free time and less working time [1].Despite advancements in technology, this promise remained mostly unmet [1,2].Now, the Fourth Industrial Revolution, which builds on digitalization and information and communication technologies (ICT), is thought to be revolutionizing everything [6].It is projected that new technologies including artificial intelligence (AI), biotechnology, the Internet of Things (IoT), quantum computing, and nanotechnology will alter how we interact with one another, perform our jobs, run our economies, and even "the mere meaning of being a human being" [7].It should be noted that the definition of the Fourth Industrial Revolution employed in this paper brightens a technology-centric understanding of 4IR; however, one should bear in mind the other important factors including the implications for society, politics, law, and ethics.Even though 4IR has been the topic of discussion on many international forums, there have not been many systematic attempts to analyze the state of the art of this new industrial revolution wave [6].This situation may be more apparent in Africa, where the Third Industrial Revolution has mostly not even fully begun [1][2][3].Therefore, African academics have expressed skepticism and caution regarding the alleged advantages of information and communications technology (ICT) in African environments.Swaminathan [8] stated the following: "Such a dream of transforming an agro-based economy into an information society must either be a flight of fancy or thinking hardly informed by the industrial economic background of developed economies that are in transition to informational economies.For an economy with about half of its adult population engaged in the food production sector, and about 70% of its development budget sourced from donor support, any talk of transition into an information society sounds like a far-fetched dream [8]".
Monzurul [9] argued that one cannot leap into the information age.Although African leaders and officials have spoken out in support of 4IR's goals, most of the continent's nations continue to be heavily dependent on an agrarian economy [10].Pachade [5] stated that critics frequently advocate that some community ICT projects have been unsuccessful partly because of the technology/reality divide.Africa has previously been described as a technological and digital wilderness [3,10,11].It is evident that Africa still lags the rest of the international community regarding the Fourth Industrial Revolution.This is due to several factors such as poor infrastructure and over-reliance on the primary sector-agriculture [6].
Agriculture remains the backbone of the African continent; it is a crucial part of the global economy and plays an important role in providing food for the rapidly growing population and hence its heightened food demand [8,10].According to the United Nations, the world's population is anticipated to reach over 10 billion people by 2050, virtually doubling global food consumption [3].Therefore, global agricultural productivity will need to rise by 1.75% each year to meet the resulting food demand [3,11].The Global Harvest Initiative (GHI) estimated that productivity is currently increasing at a rate of 1.63% annually since the farmers are already being assisted by precision agriculture and advanced technologies such as automation, machine learning, computer vision, and artificial intelligence in keeping up with the food demand [5].Global navigation satellite systems (GNSSs) are playing a particularly significant role as enablers in the transformation of the agricultural sector through precision agriculture.
Prashar [12] defined precision agriculture as a smart form of farm governance using digital systems, sensors, microcontrollers, actuators, robotics, and communication systems to achieve the goals of sustainability, revenue, and environmental conservation.Swaminathan [8] defined it as the integration of different computer tools into conventional agricultural methods to maximize the farm harvest and achieve self-sufficiency in farming operations.Precision agriculture (also known as digital farming or intelligent agriculture) includes (but is not limited to) the following: pest detection, weed detection, plant disease detection, morphology, irrigation monitoring and control, soil monitoring, air monitoring, humidity monitoring, and harvesting [4,[6][7][8]12].This paper aimed to study in detail the recent research trends in precision agriculture-particularly in the disease/pest/weed detection area-to comprehend the artificial intelligence (AI) tools and scientific background required to implement these machine learning (ML)-based precision agriculture systems.
The disease/pest/weed detection system was chosen because it possesses a multipurpose architecture that can be applied in several diverse applications on a farm with only amendments to the software and limited changes to the hardware; for example, a disease, weed, pest, nutrient deficiency, or morphological feature.Detection systems all have similar working principles in which a high-quality picture is acquired from a farm specimen, and an ML algorithm is then fed that picture after processing to classify what it detected in the given picture.Therefore, these systems can have similar prototypic architectures, and a farmer can have one universal robotic system that has a few changeable parts (such as cameras and sensors) and different software that are specific to different activities.This paper aimed to present and summarize the recent research trends in precision agriculture-particularly in the disease/pest/weed detection area-to identify the opportunities for further research.Its general architecture can be seen in Figure 2. The following research questions were addressed in this study:

•
What are the recent precision agriculture research developments, particularly for disease/pest/weed detection systems?

•
What are the found limitations and gaps in the literature review?

•
What are the arising opportunities for further research?

•
Lastly, what topological amendments can be made to the traditional precision agricultural systems to make them more economical to employ in rural farms and make them more accessible?

•
Lastly, what topological amendments can be made to the traditional precision agricultural systems to make them more economical to employ in rural farms and make them more accessible?
Figure 2. The general structure of this review paper.

Literature Review: Precision Agriculture Research Developments
Monitoring and early identification of diseases, pests, and weeds are imperative in an effective farming operation [1].In conventional agricultural practices, farmers rely upon visual observations of specimens to identify diseased leaves, fruits, roots, and other parts of crops [4,6].However, this method is faced with several challenges that include the need for continuous checking and observation of specimens, which is tedious and expensive for large farms but most importantly, much less accurate [1,2,11].Badage [1] asserted that agriculturalists often consult experts for the identification of infections of their crops, which incurs even more costs and results in longer turnaround times.The earlierstated limitations of classical farming methods coupled with the pressure to keep up with an exponentially growing demand for food both in quantity and quality have served as the push factors for researchers to devise new strategies and tools to digitize the agricultural field with the prime objective of increasing the farm yields and produce [13].The following subsection discusses the general plant disease detection system; one should note that the same general topology can be used to monitor pests, weeds, morphological features, and similar factors.

Plant Disease/Pest/Weed Detection System Basic Principles
Detection of diseases, pests, or weeds is achieved by utilizing machine learning (ML) [3,5,6,11,12].Shruthi [3] defined ML as an intelligent technique in which a machine is capacitated to recognize a pattern, recall historical information, and train itself without being commanded to do so.Both supervised and unsupervised training strategies can be utilized for machine training [8].While there are distinct training and assessment datasets for supervised training, there is no such distinction for unsupervised training datasets

Literature Review: Precision Agriculture Research Developments
Monitoring and early identification of diseases, pests, and weeds are imperative in an effective farming operation [1].In conventional agricultural practices, farmers rely upon visual observations of specimens to identify diseased leaves, fruits, roots, and other parts of crops [4,6].However, this method is faced with several challenges that include the need for continuous checking and observation of specimens, which is tedious and expensive for large farms but most importantly, much less accurate [1,2,11].Badage [1] asserted that agriculturalists often consult experts for the identification of infections of their crops, which incurs even more costs and results in longer turnaround times.The earlier-stated limitations of classical farming methods coupled with the pressure to keep up with an exponentially growing demand for food both in quantity and quality have served as the push factors for researchers to devise new strategies and tools to digitize the agricultural field with the prime objective of increasing the farm yields and produce [13].The following subsection discusses the general plant disease detection system; one should note that the same general topology can be used to monitor pests, weeds, morphological features, and similar factors.

Plant Disease/Pest/Weed Detection System Basic Principles
Detection of diseases, pests, or weeds is achieved by utilizing machine learning (ML) [3,5,6,11,12].Shruthi [3] defined ML as an intelligent technique in which a machine is capacitated to recognize a pattern, recall historical information, and train itself without being commanded to do so.Both supervised and unsupervised training strategies can be utilized for machine training [8].While there are distinct training and assessment datasets for supervised training, there is no such distinction for unsupervised training datasets [12].Prashar [12] further stated that since ML is an evolving procedure, the machine's performance becomes better with time.As soon as the machine has finished learning or training, it may classify the data, make predictions, and even generate fresh test data from which its re-trains itself, and the process goes on and on [8].Adekar [4] defined ML as a decision-making tool capable of visualizing the potentially complicated interrelationships between important parameters on a farm and making educated predictions and/or decisions.
The author further provided an illustration of an ML application in precision agriculture as seen in Figure 3.In the three-level precision agricultural layout shown, the first level, which is the physical layer, represents all the field equipment such as sensors, trackers, actuators, and probes that are in physical contact with the farm environment and are collecting data for further processing [4].In the second level, the edge layer is where the processing of the data collected in Level 1 is taking place to convert the raw data into useful information that is used to inform the decision making.The decision making takes place at this level through computational tools such as computers, microcontrollers, microprocessors, and similar ones [4].In the third level (the cloud layer), the storage of data for iterative training of the machine takes place [4].Therefore, the plant disease detection system is made up of two main subsystems, viz. the image-processing system and the classification system.The image processing is further subdivided into four steps.The four most cited different classification protocols are summarized in Table 1.[12].Prashar [12] further stated that since ML is an evolving procedure, the machine's performance becomes better with time.As soon as the machine has finished learning or training, it may classify the data, make predictions, and even generate fresh test data from which its re-trains itself, and the process goes on and on [8].Adekar [4] defined ML as a decision-making tool capable of visualizing the potentially complicated inter-relationships between important parameters on a farm and making educated predictions and/or decisions.
The author further provided an illustration of an ML application in precision agriculture as seen in Figure 3.In the three-level precision agricultural layout shown, the first level, which is the physical layer, represents all the field equipment such as sensors, trackers, actuators, and probes that are in physical contact with the farm environment and are collecting data for further processing [4].In the second level, the edge layer is where the processing of the data collected in Level 1 is taking place to convert the raw data into useful information that is used to inform the decision making.The decision making takes place at this level through computational tools such as computers, microcontrollers, microprocessors, and similar ones [4].In the third level (the cloud layer), the storage of data for iterative training of the machine takes place [4].Therefore, the plant disease detection system is made up of two main subsystems, viz. the image-processing system and the classification system.The image processing is further subdivided into four steps.The four most cited different classification protocols are summarized in Table 1.The latest studies of phenomics and high-throughput picture-data gathering are available; however, most of the research on image interpretation and processing can be found in textbooks that dive into extensive detail into the methodologies [14].Figure 4 summarizes the techniques for image acquisition and processing generally utilized for plant disease detection systems.The latest studies of phenomics and high-throughput picture-data gathering are available; however, most of the research on image interpretation and processing can be found in textbooks that dive into extensive detail into the methodologies [14].Figure 4 summarizes the techniques for image acquisition and processing generally utilized for plant disease detection systems.

Image Processing Image Acquisition
Image collection is the first step in a system for detecting plant diseases [6,8,12].Image sensors, scanners, and unmanned aerial vehicles (UAVs) can all be used to capture photos of plants [3].The commonly utilized image-acquisition tools are a charge-coupled device (CCD) and a complementary metal-oxide-semiconductor (CMOS) [15].Both of these camera technologies convert light signals and protons to digital data, which is then further transformed into a picture [15,16].However, their methods of turning the light signals into image data vary [16].In a CCD camera, the light signals are transferred through a series of adjacent pixels before being amplified and converted into image data at the end of these pixel strings [17,18].This enables CCD cameras to possess minimal degradation during the image-acquisition process [19].CCD cameras generate sharp pictures with reduced distortion [18].Contrarily, in CMOS cameras the light signals are collected, amplified, and converted at each pixel of the image sensor [15].This enables the CMOS devices to generate images faster than CCD devices since each pixel can convert light signals into an image locally [17].CMOS devices are normally preferred in projects with a low budget since they are cheap compared to CCD devices, have a lower power consumption, and can acquire high-quality images faster than their CCD counterparts [17][18][19].Figure 5 shows the serial versus localized pixel image conversion of CCD and

Image Processing Image Acquisition
Image collection is the first step in a system for detecting plant diseases [6,8,12].Image sensors, scanners, and unmanned aerial vehicles (UAVs) can all be used to capture photos of plants [3].The commonly utilized image-acquisition tools are a charge-coupled device (CCD) and a complementary metal-oxide-semiconductor (CMOS) [15].Both of these camera technologies convert light signals and protons to digital data, which is then further transformed into a picture [15,16].However, their methods of turning the light signals into image data vary [16].In a CCD camera, the light signals are transferred through a series of adjacent pixels before being amplified and converted into image data at the end of these pixel strings [17,18].This enables CCD cameras to possess minimal degradation during the image-acquisition process [19].CCD cameras generate sharp pictures with reduced distortion [18].Contrarily, in CMOS cameras the light signals are collected, amplified, and converted at each pixel of the image sensor [15].This enables the CMOS devices to generate images faster than CCD devices since each pixel can convert light signals into an image locally [17].CMOS devices are normally preferred in projects with a low budget since they are cheap compared to CCD devices, have a lower power consumption, and can acquire high-quality images faster than their CCD counterparts [17][18][19].Figure 5 shows the serial versus localized pixel image conversion of CCD and CMOS image sensors, respectively.An imaging acquisition tactic known as time delay and integration (TDI) can be combined with either CCD or CMOS technology to drastically improve their image-acquisition capabilities [20].Applications involving fast-moving objects and requiring high precision and the capacity to function in extremely dim lighting environments use TDI [20,21].Refer to Figure 6 for an example of a high-speed application of TDI technology in which a high-velocity train was captured with a normal and a TDI-featured camera in the left and right pictures, respectively.When the camera was operated in normal mode, the image of the train was a blur due to its high velocity and dim lighting conditions; however, the incorporation of a TDI mode countered these challenges and produced a clear detailed picture of the train.An imaging acquisition tactic known as time delay and integration (TDI) can be combined with either CCD or CMOS technology to drastically improve their image-acquisition capabilities [20].Applications involving fast-moving objects and requiring high precision and the capacity to function in extremely dim lighting environments use TDI [20,21].Refer to Figure 6 for an example of a high-speed application of TDI technology in which a high-velocity train was captured with a normal and a TDI-featured camera in the left and right pictures, respectively.When the camera was operated in normal mode, the image of the train was a blur due to its high velocity and dim lighting conditions; however, the incorporation of a TDI mode countered these challenges and produced a clear detailed picture of the train.
After an image has been captured with a CCD or CMOS device with or without TDI technology incorporated, the captured image should proceed to the following step of the image processing, which is normally image segmentation [3,5,11,12,16].The segmentation of an image is a process in which the features of interest are extracted from the rest of the image and irrelevant features are masked [10].The features of interest are referred to as the foreground, while the irrelevant ones are referred to as the background [16].The creation of the foreground versus background is dependent on picture properties such as color, spectrum brightness, edge detection, and neighbor resemblance [17].However, image pre-processing may occasionally be necessary before an effective image segmentation can take place [3,8,11,22].
Image Pre-Processing This is a crucial step in an ML-based disease detection system [14].Pre-processing of an image deals with the correct setting of image contrast and filtration of interference signals resulting in noise and hence blurry images [18,19].This procedure can greatly enhance the precision of feature extraction and the correct disease detection in general [15].Preprocessing typically involves straightforward treatments such as image cutting, clipping, cropping, filtering, trimming, and deblurring [3].Wang [23] explained that a typical image pre-processing procedure that is generally employed in image-based detection systems comprises image acquisition, gray scaling, filtering, binarization, and edge filtering.
which a high-velocity train was captured with a normal and a TDI-featured camera in the left and right pictures, respectively.When the camera was operated in normal mode, the image of the train was a blur due to its high velocity and dim lighting conditions; however, the incorporation of a TDI mode countered these challenges and produced a clear detailed picture of the train.After an image has been captured with a CCD or CMOS device with or without TDI technology incorporated, the captured image should proceed to the following step of the image processing, which is normally image segmentation [3,5,11,12,16].The segmentation of an image is a process in which the features of interest are extracted from the rest of the image and irrelevant features are masked [10].The features of interest are referred to as the foreground, while the irrelevant ones are referred to as the background [16].The creation of the foreground versus background is dependent on picture properties such as color, spectrum brightness, edge detection, and neighbor resemblance [17].However, The first step in the procedure illustrated in Figure 7 involves the transformation of a colored image into a gray image [23].This conversion stage into a gray image may be omitted in applications in which color features are of relevance; otherwise, this step is crucial because it is much simpler and faster to process an image in a gray color format [17].The second stage involves the denoising of a specimen image because in most cases, images are not without interference with the noise signal, which affects the visibility of the features in the specimen images [23].The third step then includes image segmentation, which will be explained more broadly in the coming Section.The last step involves the forming of an outline image, which can be achieved by masking the leafstalk as well as holes while keeping the outer connected region [15,23].Wakhare [24] proposed a similar procedure to that illustrated in Figure 7 for plant-leaf feature identification applications under reallife varying lighting conditions.This procedure involves the conversion of a specimen image into grayscale, noise suppression as well as smoothing, and formation of the image outline through edge filtering.In a comparative study conducted by Ekka [25], a histogram equalization method was proven to be the most effective form of image enhancement of the gray images that were originally color images.Conversely, Kolhalkar [26] found that red-green-blue (RGB) camera images offer more valuable image enhancement compared to those converted to grayscale in the context of identifying diseases on the plant leaves.
Therefore, we could not conclude which image pre-processing technique is better than the other, rather the application in which the image is used, and thing kind of image involved in that application shall be considered in the selection of an appropriate preprocessing technique.
under real-life varying lighting conditions.This procedure involves the conversion of a specimen image into grayscale, noise suppression as well as smoothing, and formation of the image outline through edge filtering.In a comparative study conducted by Ekka [25], a histogram equalization method was proven to be the most effective form of image enhancement of the gray images that were originally color images.Conversely, Kolhalkar [26] found that red-green-blue (RGB) camera images offer more valuable image enhancement compared to those converted to grayscale in the context of identifying diseases on the plant leaves.Therefore, we could not conclude which image pre-processing technique is better than the other, rather the application in which the image is used, and thing kind of image involved in that application shall be considered in the selection of an appropriate preprocessing technique.

Image Segmentation
Image segmentation is a pivotal part of image-based plant feature identification and phenotyping systems [23].Segmentation of an image involves the separation between the foreground and the background [15]; that is, the isolation of the feature of interest and masking of the irrelevant part of the image [24][25][26].The features of interest are normally identified by comparing adjacent pixels for similarity by looking at the three main parameters, viz. the texture, color, and shape [15,17].Table 2 shows a list of free data libraries available to the public for use in the image segmentation process.

CA-RET
Offers a wide range of tools for creating forecasting analytics utilizing R's extensive model library.It contains techniques for the pre-processing learning algorithm, determining the relevance of parameters, and presenting networks [30].

Rweka
Data pre-processing, categorization, analysis, grouping, clustering algorithms, and image-processing methods for all Java-based machine learning methods [31].

ROCR
A tool for assessing and displaying the accuracy of rating classifiers [32].

Earth
Utilizes the methods from Friedman's publications "Multivariate Adaptive Regression Splines" and "Fast MARS" to create a prediction model [34].

TREE
A library containing functions designated to work with trees [35].

R, C
Igraph Contains functions for manipulating large graphs and displaying [34].

Python, R Scikit-learn
Offers a standardized interface for putting the machine into practicing the learning of algorithms.
It comprises various auxiliary tasks such as data pre-processing operations, information resampling methods, assessment criteria, and search portals for adjusting and performance optimization of methods [36].

Tensorflow
Toolkit for quick computation of numbers in artificial intelligence and machine learning [40].https://www.tensorflow.org/(accessed on 18 February 2023) PyBrain A versatile, powerful, and user-friendly machine learning library that offers algorithms that may be used for a range of machine learning tasks [41].

Pylearn2
A specially created library for machine learning to make learning much easier for developers.It is quick and provides a researcher with a great deal of versatility [42].
http://deeplearning.net/software/ pylearn2 (accessed on 18 February 2023) Java Java-ML A collection of machine learning and data mining techniques that aim to offer a simple-to-use and extendable API.Algorithms rigorously adhere to their respective interfaces, which are maintained as basic for each type of algorithm's interface [43].
http://java-ml.sourceforge.net/(accessed on 17 February 2023) ELKI A data mining software that intends to make it possible to create and evaluate sophisticated data mining algorithms and study how they interact with database search architecture [44].
http://elki.dbs.ifi.lmu.de/(accessed on 16 February 2023) JSAT A library designed to fill the need for a general purpose, reasonably high-efficiency, and versatile library in the Java ecosystem that is not sufficiently satisfied by Weka and Java-ML [45]

MLC++
Supervised machine learning methods and functions in a C++ ecosystem [52].

Accord
Fully C#-written machine learning platform with audio and picture analysis libraries [54].
http://accord-framework.net/ (accessed on 13 February 2023) A very popular example of an image segmentation technique is thresholding [55].Threshold segmentation is a process of converting a color or grayscale image into a binary image (as shown in Figure 8) with the sole purpose of making feature classification easier [55,56].The output binary images consist of black and white colored pixels that correspond to the background and foreground, respectively, or vice versa [26,55,56].

MLC++
Supervised machine learning methods and functions in a C++ ecosystem [52].
mlc/source.html (accessed on 13 February 2023) Accord Fully C#-written machine learning platform with audio and picture analysis libraries [54].
http://accord-framework.net/ (accessed on 13 February 2023) A very popular example of an image segmentation technique is thresholding [55].Threshold segmentation is a process of converting a color or grayscale image into a binary image (as shown in Figure 8) with the sole purpose of making feature classification easier [55,56].The output binary images consist of black and white colored pixels that correspond to the background and foreground, respectively, or vice versa [26,55,56].Threshold segmentation is mathematically defined as follows, where T refers to a certain threshold intensity, g is the black or white pixel of a binary image, and f is the gray level of the input picture [56]: Threshold segmentation is subdivided into global, local, and adaptive thresholding [15,57].Global thresholding is applied in scenarios where there is enough distribution between the intensity distribution of the foreground compared to the background [15].Hence, a single threshold value is selected and used to distinguish between the features of significance and the background [15,55].Local thresholding is applied in cases where there is no distinct difference in intensity distribution between the background and the foreground and hence it is not conducive to selecting a single threshold value [55].In such a case, an image is partitioned into smaller images, and different threshold values for each partitioned picture are selected [15].Adaptive thresholding is also appropriate for images with uneven intensity distribution because a threshold value is calculated for each pixel Threshold segmentation is mathematically defined as follows, where T refers to a certain threshold intensity, g is the black or white pixel of a binary image, and f is the gray level of the input picture [56]: Threshold segmentation is subdivided into global, local, and adaptive thresholding [15,57].Global thresholding is applied in scenarios where there is enough distribution between the intensity distribution of the foreground compared to the background [15].Hence, a single threshold value is selected and used to distinguish between the features of significance and the background [15,55].Local thresholding is applied in cases where there is no distinct difference in intensity distribution between the background and the foreground and hence it is not conducive to selecting a single threshold value [55].In such a case, an image is partitioned into smaller images, and different threshold values for each partitioned picture are selected [15].Adaptive thresholding is also appropriate for images with uneven intensity distribution because a threshold value is calculated for each pixel [57].The Otsu thresholding method is another thresholding technique used for image segmentation [15].In this technique, a measure of spread for the pixel intensity levels on either side of the threshold is listed by looping through all the reasonable threshold values [58].The intent is to decide the threshold value for which the summation of foreground and background escalates is at its minimum [15,58].The fundamental characteristic of the Otsu thresholding method is the fact that it implements the threshold values automatically instead of it being preselected by the user [58]; (2) below shows the mathematical definition for the thresholding in the Otsu method.
Another segmentation method applied in image processing is watershed transformation [59].A grayscale image undergoes a transition called a watershed [59,60].In a metaphorical sense, the name alludes to a geologic catchment or drainage split that divides parallel catchments [59].The watershed conversion locates the lines that follow the tops of ridges by treating the image it operates upon as a topographic map; the luminosity of each pixel denotes its elevation [60].Figure 9 is an example of a watershed-segmented image in which the black pixels denote the background, the gray pixels denote the features to be extracted, and the white pixels correspond to watershed lines [61].
(, ) = 0, 0 ℎ Another segmentation method applied in image processing is watershed transformation [59].A grayscale image undergoes a transition called a watershed [59,60].In a metaphorical sense, the name alludes to a geologic catchment or drainage split that divides parallel catchments [59].The watershed conversion locates the lines that follow the tops of ridges by treating the image it operates upon as a topographic map; the luminosity of each pixel denotes its elevation [60].Figure 9 is an example of a watershed-segmented image in which the black pixels denote the background, the gray pixels denote the features to be extracted, and the white pixels correspond to watershed lines [61].On the other hand, Grabcut is a very popular and innovative segmentation technique that takes into consideration the textural and boundary conditions of an image [62].This segmentation method is based on the iterative graph-cut method in which a mathematical function is derived to implement the background as well as the foreground [63].Each pixel in an image is then assessed to decide whether it falls in the background or the foreground [62,63].The Grabcut segmentation method is preferred in most applications because of minimal user interference in the operation of this technique; however, it is not without its drawbacks [62].The Grabcut sequence cycles take a long time to implement because of the complexity of the thresholding equation [63].The segmentation is also poor in scenarios where the background is complex and there is minimal distinction between the features of interest and the background [64].Several distinct segmentation methods and algorithms exist in the literature.The suitability of a particular method is based on a On the other hand, Grabcut is a very popular and innovative segmentation technique that takes into consideration the textural and boundary conditions of an image [62].This segmentation method is based on the iterative graph-cut method in which a mathematical function is derived to implement the background as well as the foreground [63].Each pixel in an image is then assessed to decide whether it falls in the background or the foreground [62,63].The Grabcut segmentation method is preferred in most applications because of minimal user interference in the operation of this technique; however, it is not without its drawbacks [62].The Grabcut sequence cycles take a long time to implement because of the complexity of the thresholding equation [63].The segmentation is also poor in scenarios where the background is complex and there is minimal distinction between the features of interest and the background [64].Several distinct segmentation methods and algorithms exist in the literature.The suitability of a particular method is based on a particular application, and hence this study was not able to rule out certain segmentation methods or determine which ones outperform the others.

Feature Extraction
One of the foundational elements of computer-vision-based image recognition is the extraction of features [65].A feature is data that are utilized to solve a particular computer vision problem and is a constituting part of a raw image [64].The feature vectors include the features that have been retrieved from an image [66].An extensive range of techniques is used to identify the items in an image while creating feature vectors [62].Edges, image pixel intensity, geometry, texture, and image modifications such as Fourier, Wavelet, or permutations of pixels from various color images are the primary features [46,66].Use as a set of classifiers and machine learning algorithms is feature extraction's ultimately purpose [66].The feature extraction in plant leaf disease-monitoring systems is subdivided into three spheres that include texture, color, and shape [20,21,46,65].

Shape Features
The shape is a basic characteristic of a leaf used in feature extraction of leaf images during image processing [66].The primary shape parameters include the length (L), which is the displacement between the two points in the longest axis; the width (W), which denotes the displacement between the shortest axis; the diameter (D), which denotes the maximum distance between the points; the area (A), which denotes the surface area of all the pixels found within the margin of a leaf picture; and the perimeter (P), which denotes the accumulative length of the pixels around the margin of a leaf picture [55,58,62,64].From the 5 defined primary characteristics of shape features, 11 distinct secondary features are formed by mathematical definitions involving 2 or more primary variables [59].These 11 features are called the morphological features of a plant.The morphological features are as follows: • Circularity (C)-a feature defining the degree to which a leaf conforms to a perfect circle.It is defined as [60]: • Rectangularity (R)-a feature defining the degree to which a leaf conforms to a rectangle.It is defined as [55]: • Aspect ratio (AS)-ratio of width to length of a leaf.It is defined as [55]: • Smooth factor (SF)-ratio of leaf picture area when 5 × 5 and 2 × 2 regular smoothing filters have been used [58].

•
Perimeter-to-diameter ratio (PDr)-ratio of the perimeter to the diameter of a leaf.It is defined as [64]: • Perimeter to length plus width ratio (PLWr)-ratio of the perimeter to length plus width of a leaf.It is defined as [64]: • Narrow factor (NFr)-ratio of diameter to length of a leaf [60]: • Area convexity (ACr)-area ratio between the area of a leaf and the area of its convex hull [59].

•
Perimeter convexity (ACr)-the ratio between the perimeter of a leaf to that of its convex hull [60].

•
Eccentricity (Ar)-the degree to which a leaf shape is a centroid [64].

•
Irregularity (Ir)-ratio of the diameters of an inscribed to the circumscribed circles on the image of a leaf [59].

Color Features
Other researchers and scholars chose to implement the color features as the pivotal features during the extraction process [67].The color features normally cited in the literature on leaf feature extraction include the following:

•
Color standard deviation (σ)-a measure of how much the different colors found in an image match one another or are rather different from one another [60].If an image is differentiated into an array of its basic building blocks (the pixels), then i is a pointer moving across the rows of pixels in an array from the origin to the very last row M, while j is a pointer moving across the columns of pixels in an array from the origin to the very last column N. At any point, a pixel color intensity is defined by p(i, j), where i and j denote the coordinate position of a pixel in an image array.Therefore, the color standard deviation is mathematically defined as follows: • Color mean (µ)-a measure to identify a dominant color in a leaf image.This feature is normally used to identify the leaf type [63].It is mathematically defined as follows: • Color skewness (ϕ)-a measure to identify a color symmetry in a leaf image [21,46]: • Color kurtosis (ϕ)-a measure to identify a color shape dispersion in a leaf image [65]: Texture Features There are also several textural features referenced by authors such as Singh [68], Martsepp [69], and Ponce [70].Using the same assumption of an image partitioned into pixels in the above Section, the following are the textural features used for feature extraction in plant leaves:

•
Entropy (Entr)-this is a measure of the complexity and uniformity of a texture of a leaf image [68]: • Contrast (Con)-this is a measure of how clear the features are in a leaf image; it is also referred to as the moment of inertia [69,70]: • Energy (En)-this is a measure of the degree of uniformity of a gray image.It is also called the second moment [69]: • Correlation (Cor)-this is a measure of whether there is a similar element in a sample picture that corresponds to the re-occurrence of a similar matrix within a large array of pixels [68]. where: • Difference moment inverse (DMI)-this is a measure of the degree of homogeneity in an image [69]: Other textural features include the maximum probability, which is the highest response to correlation; the standard deviation and/or variance, which is the aggregate texture observed in a leaf picture; and the average illuminance, which is the average light distribution across the leaf when an image was captured [66,[68][69][70].The selection of a particular color, shape, or textural feature strictly depends on the application of the system being designed.

Feature Classification
The classification techniques are machine learning algorithms that are used to categorize input sample data into different classes or groups of belonging or membership [3,5,11,56].These classifiers may employ supervised learning, unsupervised learning, and reinforcement learning methods during their training [39].Supervised learning occurs when a person is a trainer of the model and may use pre-formed datasets to conduct the training [39,53].Unsupervised learning occurs when there is no training data available and the algorithm must train itself and improve its classification efficiency by iteratively adjusting itself [5,39,53].Reinforcement learning occurs when the algorithm makes classification rulings based on the feedback applied by the environment to it [12,39].In the case of vision-based plant disease-monitoring systems, the most cited classification algorithms include support vector machines (SVM), artificial neural networks, k-nearest neighbors' machines, and fuzzy machines.The following subsections discuss these classification techniques.

SVM Classifier
The support vector machine, sometimes known as SVM, is a predictive model used to solve both regression and classification tasks [3].It is a supervised learning model that works well for numerous practical problems and can solve both linear and non-linear tasks [3,71].The SVM concept is relatively simple; a vector or a hyperplane that splits the data into groups is generated by this technique [72].
In Figure 10, the optimal hyperplane is used to separate the two classes of data (the blue squares and green circles).The two planes (dashed lines) parallel to the optimal hyperplane are called the positive and negative imaginary planes, which are the planes passing through the closest data points to either side of an optimal hyperplane [72].These closest points to the optimal hyperplane are called the support vectors and are used to determine the exact position of an optimal hyperplane [73].There might be several possible hyperplanes, but the optimal hyperplane is the one with the maximum marginal distance, which is the distance between the two marginal planes [72,73].The maximized margin results in a more generalized solution compared to smaller margins; should the training data change, the algorithm with a smaller margin will have accuracy challenges [73].In some cases, data classes are not always easily separable with a straight line or place as in the case of Figure 10.Therefore, when data classes show a property of non-linearity, transforming a space in which these data classes occur from a low-dimension (often two-dimensional) into a high-dimension space (often three-dimensional) space using the kernel method.The kernel method is a computation of a dot product of the dimensions in the new high-dimension space [72][73][74]; (17) below gives the general solution of a hyperplane, where → x is any data point or support vector, → ω is the weight vector that applies the bias of the support vectors, and ω 0 is the constant [74].
and g(

ANN Classifier
An ANN is a supervised learning model that is a collection of interlinked inp output nodes in which each link has an associated bias value called a weight [75].A input layer, one or perhaps more intermediate layers that are normally called hidde ers, and one or more output layers make up the structure of an ANN [75,76].The w of each connection is modulated as the network operates to facilitate neural network ing [76].The performance of the network is enhanced by adjusting the weight co ously [75].ANN can be divided into two groups based on connection types: feed-fo networks and recurrent networks [33].In contrast to recurrent neural networks, fee ward neural networks do not have cycle-forming connections between units [76].T chitecture, transfer function, and learning rule all have an impact on how a neural ne behaves [49,76].The weighted total of input triggers the activation of neural networ rons [75].Figure 11 shows a generalized model of an ANN model with the input the hidden intermediate layer (purple layer), and the output layer.

ANN Classifier
An ANN is a supervised learning model that is a collection of interlinked input and output nodes in which each link has an associated bias value called a weight [75].A single input layer, one or perhaps more intermediate layers that are normally called hidden layers, and one or more output layers make up the structure of an ANN [75,76].The weight of each connection is modulated as the network operates to facilitate neural network learning [76].The performance of the network is enhanced by adjusting the weight continuously [75].ANN can be divided into two groups based on connection types: feed-forward networks and recurrent networks [33].In contrast to recurrent neural networks, feed-forward neural networks do not have cycle-forming connections between units [76].The architecture, transfer function, and learning rule all have an impact on how a neural network behaves [49,76].The weighted total of input triggers the activation of neural network neurons [75].Figure 11 shows a generalized model of an ANN model with the input layer, the hidden intermediate layer (purple layer), and the output layer.
ously [75].ANN can be divided into two groups based on connection types: networks and recurrent networks [33].In contrast to recurrent neural netw ward neural networks do not have cycle-forming connections between unit chitecture, transfer function, and learning rule all have an impact on how a n behaves [49,76].The weighted total of input triggers the activation of neural rons [75].Figure 11 shows a generalized model of an ANN model with th the hidden intermediate layer (purple layer), and the output layer.

k-NN Classifier
The k-nearest neighbors algorithm, sometimes known as k-NN, is the most straightforward machine learning technique [78].It is a non-parametric technique used for problems involving regression and classification [74,78].Non-parametric implies that no dataset for initial training is necessary [78].Therefore, k-NN does not require the use of any presumptions [79].The k-closest training examples in the feature space provide the input for classification and regression tasks, respectively [78].Whether k-NN is applied for classification or regression determines the results [79].The outcome of the k-NN classifier is a class of belonging [74,78,79].Based on the predominant kind of its neighborhood, the given data point is classed [79].The input point is awarded to the category that has the highest frequency among its k-closest neighbors [78].In most cases, k is a small positive integer such as 1.The result of a k-NN regression is just a value of the property for the attribute.The aggregate of the variables of the k-closest neighbors constitutes this number [79].
Figure 12 shows a space with numerous data points or vectors that can be classified into two classes: the red class and the green class.Now, assume there exists a data point at any location in the space shown in Figure 12 that is unknown whether it belongs to either the red or green class.The k-NN will then proceed through the following computational steps to assign that point a class of belonging:

•
Take the uncategorized data point as input to a model.• Measure the spatial distance between this unclassified point to all the other already classified points.The distance can be computed via Euclidean, Minkowski, or Manhattan formulae [80].

•
Check the points with the shortest displacement from the unknown data point to be classified for a certain K value (K is defined by the supervisor of the algorithm) and separate these points by the class of belonging [80].

•
Select the correct class of membership as the one with the most frequent vectors as the neighbors of the unknown data point [80].
• Measure the spatial distance between this unclassified point to all the other already classified points.The distance can be computed via Euclidean, Minkowski, or Manhattan formulae [80].

•
Check the points with the shortest displacement from the unknown data point to be classified for a certain K value (K is defined by the supervisor of the algorithm) and separate these points by the class of belonging [80].

•
Select the correct class of membership as the one with the most frequent vectors as the neighbors of the unknown data point [80].The most cited method of computing the spatial distance between the data point p to be classified and its neighbors qn is the Euclidean Formulae (18) [74,80]: The most cited method of computing the spatial distance between the data point p to be classified and its neighbors qn is the Euclidean Formulae (18) [74,80]: (18)   Fuzzy Classifier The fuzzy classifier system is a supervised learning model that enables computational variables, outputs, and inputs to assume a spectrum of values over predetermined bands [81].By developing fuzzy rules that connect the values of the input variables to internal or output variables, the fuzzy classifier system is trained [82].It has mechanisms for credit assignment and conflict resolution that combine elements of typical fuzzy classifier systems [81].A genetic algorithm is used by the fuzzy classifier system to develop suitable fuzzy rules [83].
As shown in Figure 13, fuzzy sets display a continuous membership, and a data point membership classification can be ruled as the extent (µ) to which it belongs to a certain fuzzy set.For example, 690 mm in Figure 13 has a degree of membership µ(960) on the close fuzzy set that is 0.7.It can also be seen in Figure 9 that a data point can belong to multiple fuzzy sets, and the degrees of membership to each set may or may not (in the intersection points) differ since some fuzzy sets overlap with each other.Table 3 summarizes the advantages and disadvantages of all the classification techniques discussed in this section.

Fuzzy Classifier
The fuzzy classifier system is a supervised learning model that enables computational variables, outputs, and inputs to assume a spectrum of values over predetermined bands [81].By developing fuzzy rules that connect the values of the input variables to internal or output variables, the fuzzy classifier system is trained [82].It has mechanisms for credit assignment and conflict resolution that combine elements of typical fuzzy classifier systems [81].A genetic algorithm is used by the fuzzy classifier system to develop suitable fuzzy rules [83].
As shown in Figure 13, fuzzy sets display a continuous membership, and a data point membership classification can be ruled as the extent (µ) to which it belongs to a certain fuzzy set.For example, 690 mm in Figure 13 has a degree of membership µ(960) on the close fuzzy set that is 0.7.It can also be seen in Figure 9 that a data point can belong to multiple fuzzy sets, and the degrees of membership to each set may or may not (in the intersection points) differ since some fuzzy sets overlap with each other.Table 3 summarizes the advantages and disadvantages of all the classification techniques discussed in this section.• Works very accurately when there is a clear formation of a hyperplane [74].
• Accuracy difficulties with a large amount of training data [71].
• Works more accurately in high-dimension spaces such as 3D and 4D [51].
• Susceptibility to noise and overlapping data classes [75].

•
The number of characteristics for a single dataset must not exceed the number of data points in the training set [74].

•
The machine is learning continuously, and the accuracy is improving and iterable [50].
• Accuracy is data-dependent; more training data translate to a more accurate classification and vice versa [75].
• Accuracy difficulties with a large amount of training data [79].
• Simple to add new data to the model to extend its scope [80].
• Relatively easy to implement with only the two parameters to work out: the k value and the geometric distance between the points [78].

Fuzzy Classifier
• Unclear, distorted, degraded, or vague input data is accommodated by the model [81].
• More flexibility and ease to change the rules [83].
• Require excessive supervision in a form of testing and validation [82].

•
Robust in applications with no exact input format [82].

•
The is no universal approach to implementing fuzzy classification models, which adds to their inaccuracy [83].

Literature Survey: Plant Disease/Nutrient Deficiency Monitoring Systems
Many authors in the literature have proposed plant disease/pest/weed detection systems that employ the above-described general format.Literature shows that this technology of plant disease detection models has been developing at a faster rate in the last two decades and achieving high success in terms of classification accuracy and efficiency.

Tabulated Summary of Plant Disease/Nutrient Deficiency Monitoring Systems publications
Table 4 summarizes a literature survey on these systems.Several publications have been consulted for this research study.A few aspects have been noted for each publication such as the type of crop investigated, the number of crop disease covered in the study and the classification results achieved.

•
Very little research discussed the combination of these monitoring and phenotyping tasks into one system to reduce costs and improve technology availability to farmers and add convenience.

•
Little research discussed the post-harvest benefits of disease/nutrient deficiency detection or similar systems.

•
Most research papers on plant disease detection models processed two-dimensional images captured from plant samples.In cases where samples were in the form of fruits, single-input cameras or a two-dimensional view may pose a challenge because of the spherical or cylindrical nature of most fruits.The authors noticed that the fruit disease symptoms or any types of defects are not always evenly distributed across the surface area of a sample fruit; Figure 14 shows an example.Therefore, in high-throughput and high-speed applications, a sample fruit might be oriented such that the diseased part is masked or hidden from the camera's line of sight, so an incorrect classification is highly probable.Most research papers on plant disease detection models processed two-dimensional images captured from plant samples.In cases where samples were in the form of fruits, single-input cameras or a two-dimensional view may pose a challenge because of the spherical or cylindrical nature of most fruits.The authors noticed that the fruit disease symptoms or any types of defects are not always evenly distributed across the surface area of a sample fruit; Figure 14 shows an example.Therefore, in highthroughput and high-speed applications, a sample fruit might be oriented such that the diseased part is masked or hidden from the camera's line of sight, so an incorrect classification is highly probable.• Few studies discussed the importance of optimum optical or lighting conditions in the successful operation of an image-based plant disease detection model and their relationship to classification accuracy and efficiency.
Hence, this study took advantage of the second-to-last opportunity outlined above and proposes two conceptual ideas as a mitigation measure.The purpose of these two propositions is to give a classification model a virtual three-dimensional view of a sample fruit so that a classification model "sees" the total surface of a sample fruit so as not to miss any important details before making a final classification.The two proposed ideas are:

•
A multicamera-input fruit disease detection model • A dynamic-input fruit disease detection model A multicamera-input fruit disease detection model has an improved input system that features multiple-input camera sensors specially arranged in a circular setup and equidistant from each other with a sample fruit at the central point.These cameras capture the surface of a sample fruit at different angles such that all the fruit surface is captured (refer to Figure 15).
The classification model should classify each input image from each camera independently and consolidate all the results to make a final classification.The final classification should be decided as follows: • If at least one input image is classified as a diseased sample, set the final classification to a "diseased sample".

•
Otherwise, set the final classification to a "healthy sample".
ppl.Sci.2023, 13, x FOR PEER REVIEW Hence, this study took advantage of the second-to-last opportunity outlin proposes two conceptual ideas as a mitigation measure.The purpose of the sitions is to give a classification model a virtual three-dimensional view of so that a classification model "sees" the total surface of a sample fruit so as n important details before making a final classification.The two proposed ide  The classification model should classify each input image from each pendently and consolidate all the results to make a final classification.The fi tion should be decided as follows: • If at least one input image is classified as a diseased sample, set the fina to a "diseased sample".

•
Otherwise, set the final classification to a "healthy sample".
A dynamic input fruit disease detection model, on the other hand, main input camera but instead features a revolving sample stand that rotates in determined angle ϴ while an input image is captured per rotation until the ence of a sample fruit has been captured (refer to Figure 16).A dynamic input fruit disease detection model, on the other hand, maintains a singleinput camera but instead features a revolving sample stand that rotates in steps of a predetermined angle θ while an input image captured per rotation until the full circumference of a sample fruit has been captured (refer to Figure 16).All the capture samples are processed similarly as in the multicamera input disease detection mode.The authors foresee that these two may have different pros and cons such as the classification cycle time per each sample; however, this still needs to be examined in more detail.

Conclusions
This paper has presented the background on the research developments in plant disease detection models for agricultural applications.Substantial progress has been achieved in this research area, several crops have been considered, and several disease or nutrient detection models have been proposed that are capable of classifying each with no less than 75% accuracy as presented in the literature survey section of this paper.This study has found image processing and machine learning to be the widely used tools amongst a large proportion of researchers to implement plant disease or nutrient detection models.
This study also presented a few opportunities that the authors believe are worth further research (presented in Section 2.2.2) and has proposed two separate improvements that can be made to the classical disease classification models to improve the classification accuracy and efficiency.Much more can still be done to further improve the accuracy levels of some monitoring systems presented in Table 4 such as improving the training data.This study is already serving as a foundation for a Doctor of Philosophy Research Project that seeks to explore some of the research opportunities presented in this paper.All the capture samples are processed similarly as in the multicamera input disease detection mode.The authors foresee that these two may have different pros and cons such as the classification cycle time per each sample; however, this still needs to be examined in more detail.

Conclusions
This paper has presented the background on the research developments in plant disease detection models for agricultural applications.Substantial progress has been achieved in this research area, several crops have been considered, and several disease or nutrient detection models have been proposed that are capable of classifying each with no less than 75% accuracy as presented in the literature survey section of this paper.This study has found image processing and machine learning to be the widely used tools amongst a large proportion of researchers to implement plant disease or nutrient detection models.
This study also presented a few opportunities that the authors believe are worth further research (presented in Section 2.2.2) and has proposed two separate improvements that can be made to the classical disease classification models to improve the classification accuracy and efficiency.Much more can still be done to further improve the accuracy levels of some monitoring systems presented in Table 4 such as improving the training data.This study is already serving as a foundation for a Doctor of Philosophy Research Project that seeks to explore some of the research opportunities presented in this paper.

Figure 1 .
Figure 1.The evolution of industrial revolutions over time (DUT Inaugural Lecture).

Figure 1 .
Figure 1.The evolution of industrial revolutions over time (DUT Inaugural Lecture).

Figure 2 .
Figure 2. The general structure of this review paper.

Figure 6 .
Figure6.Impact of TDI incorporation in CMOS and CCD sensors[5]

Figure 6 .
Figure6.Impact of TDI incorporation in CMOS and CCD sensors[5]

•
During the literature survey presented in earlier sections, the following opportunities that the authors of this paper believe have seen little interest from researchers are as follows: • Little discussion of the real-time monitoring of the onset signs of diseases before they spread throughout the whole plant.• Few papers discussed real-time monitoring and real-time mitigation measures such as actuation operations, spraying pesticides, and spraying fertilizers, to name a few examples.• Very little research discussed the combination of these monitoring and phenotyping tasks into one system to reduce costs and improve technology availability to farmers and add convenience.• Little research discussed the post-harvest benefits of disease/nutrient deficiency detection or similar systems.•

Figure 14 .
Figure 14.A sample fruit with uneven distribution of the disease-infected surface area.•Fewstudies discussed the importance of optimum optical or lighting conditions in the successful operation of an image-based plant disease detection model and their relationship to classification accuracy and efficiency.

Figure 14 .
Figure 14.A sample fruit with uneven distribution of the disease-infected surface area.

•
A multicamera-input fruit disease detection model •A dynamic-input fruit disease detection model A multicamera-input fruit disease detection model has an improved inp features multiple-input camera sensors specially arranged in a circular setu tant from each other with a sample fruit at the central point.These camer surface of a sample fruit at different angles such that all the fruit surface is c to Figure15).

Figure 16 .
Figure 16.A dynamic input design for the fruit disease detection system.

Figure 16 .
Figure 16.A dynamic input design for the fruit disease detection system.

Table 1 .
Summary of image-processing steps and different classification techniques in plant disease detection. A

Table 1 .
Summary of image-processing steps and different classification techniques in plant disease detection.

Table 2 .
Table showing a list of image segmentation ML libraries.

Table 3 .
Table showing pros and cons of different classification methods.

Table 3 .
Table showing pros and cons of different classification methods.

Table 4 .
Summary of a literature survey on plant disease/pest/weed detection systems.
•During the literature survey presented in earlier sections, the following opportunities that the authors of this paper believe have seen little interest from researchers are as follows:•Little discussion of the real-time monitoring of the onset signs of diseases before they spread throughout the whole plant.•Fewpapers discussed real-time monitoring and real-time mitigation measures such as actuation operations, spraying pesticides, and spraying fertilizers, to name a few examples.