Next Article in Journal
Evaluation of a System to Assess Herbicide Movement in Straw under Dry and Wet Conditions
Next Article in Special Issue
Performance of Neural Networks in the Prediction of Nitrogen Nutrition in Strawberry Plants
Previous Article in Journal
An Effective and Affordable Internet of Things (IoT) Scale System to Measure Crop Water Use
Previous Article in Special Issue
Enhanced Deep Learning Architecture for Rapid and Accurate Tomato Plant Disease Diagnosis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Performance Comparison of CNN Models for Bean Phenology Classification Using Transfer Learning Techniques

by
Teodoro Ibarra-Pérez
1,
Ramón Jaramillo-Martínez
1,*,
Hans C. Correa-Aguado
1,
Christophe Ndjatchi
1,
Ma. del Rosario Martínez-Blanco
2,
Héctor A. Guerrero-Osuna
3,
Flabio D. Mirelez-Delgado
1,
José I. Casas-Flores
4,
Rafael Reveles-Martínez
1 and
Umanel A. Hernández-González
1
1
Instituto Politécnico Nacional, Unidad Profesional Interdisciplinaria de Ingeniería Campus Zacatecas (UPIIZ), Zacatecas 98160, Mexico
2
Laboratorio de Inteligencia Artificial Avanzada (LIAA), Universidad Autónoma de Zacatecas, Zacatecas 98000, Mexico
3
Posgrado en Ingeniería y Tecnología Aplicada, Unidad Académica de Ingeniería Eléctrica, Universidad Autónoma de Zacatecas, Zacatecas 98000, Mexico
4
Instituto Nacional de Investigaciones Forestales, Agrícolas y Pecuarias, Campo Experimental Zacatecas (INIFAP), Zacatecas 98500, Mexico
*
Author to whom correspondence should be addressed.
AgriEngineering 2024, 6(1), 841-857; https://doi.org/10.3390/agriengineering6010048
Submission received: 4 January 2024 / Revised: 1 March 2024 / Accepted: 1 March 2024 / Published: 18 March 2024
(This article belongs to the Special Issue Application of Artificial Neural Network in Agriculture)

Abstract

:
The early and precise identification of the different phenological stages of the bean (Phaseolus vulgaris L.) allows for the determination of critical and timely moments for the implementation of certain agricultural activities that contribute in a significant manner to the output and quality of the harvest, as well as the necessary actions to prevent and control possible damage caused by plagues and diseases. Overall, the standard procedure for phenological identification is conducted by the farmer. This can lead to the possibility of overlooking important findings during the phenological development of the plant, which could result in the appearance of plagues and diseases. In recent years, deep learning (DL) methods have been used to analyze crop behavior and minimize risk in agricultural decision making. One of the most used DL methods in image processing is the convolutional neural network (CNN) due to its high capacity for learning relevant features and recognizing objects in images. In this article, a transfer learning approach and a data augmentation method were applied. A station equipped with RGB cameras was used to gather data from images during the complete phenological cycle of the bean. The information gathered was used to create a set of data to evaluate the performance of each of the four proposed network models: AlexNet, VGG19, SqueezeNet, and GoogleNet. The metrics used were accuracy, precision, sensitivity, specificity, and F1-Score. The results of the best architecture obtained in the validation were those of GoogleNet, which obtained 96.71% accuracy, 96.81% precision, 95.77% sensitivity, 98.73% specificity, and 96.25% F1-Score.

1. Introduction

According to the Food and Agriculture Organization of the United Nations (FAO), plant plagues are among the main causes of the loss of over 40 percent of food crops worldwide, exceeding losses of up to USD 220 billion each year [1]. In Mexico, bean production contributes MXN 5927 million to the annual income. However, in 2021, the registered loss was more than MXN 222 million, mainly due to diseases caused by viruses transmitted in seeds, aphids, white flies, and other similar insects [2]. Several factors interfere with food security, such as climate change [3,4,5], the lack of pollinators [6,7], plagues and plant diseases [8], the result of the COVID-19 pandemic, and the present war between Russia and Ukraine [9], among others.
Insect plagues, diseases, and other organisms significantly affect the quality and production of crops. These organisms feed off plants and transmit diseases that can cause severe disruption in the growth and development of plants, causing a major impact on food security, the economy, and the environment, thereby decreasing the availability of food, increasing production costs, and affecting the growth of rural areas and developing countries [10].
It is important to mention the strategies used to mitigate the effects produced by plagues and diseases around the world, such as the selection of resistant varieties, crop rotation, the use of natural enemies of plagues, and the rational use of chemical products, among others. More efforts need to be made to implement mechanisms and innovative strategies to reduce loss in food crops and sustainably contribute to food security [9,11,12].
In recent years, the use of artificial intelligence (AI) in applications has increased exponentially. Proof of this is the appearance of works related to image recognition, especially in the field of agriculture, where various approaches to using deep learning (DL) methods to classify the phenology of different food crops around the world have been presented. This allows us to have knowledge of the record of critical moments in the life cycle of the plant to program treatments, effectively and timely apply pesticides or fungicides, and prevent and control plagues and diseases; this offers great advantages in precision agriculture in a nonharmful manner, and it helps minimize damage to crops [13,14,15,16,17].
DL methods are used to identify the different phenological stages of crops, and there exists a diversity of approaches to address the classification of topics related to agricultural decision making that mainly influence the estimation of agricultural production. In this regard, the present work proposes a comparative study of the performance of four models of convolutional neural networks (CNNs), AlexNet, VGG19, SqueezeNet, and GoogleNet, in classifying the phenological stages of bean crops; the performance of each of the models is compared through the following metrics: accuracy, precision, sensitivity, specificity, and F-1 Score. The results are used to choose the architecture that best models the classification problem in bean phenology.
The goal of analyzing the different CNN architectures is to identify the best-performing one and, in the future, to embed networks in compact systems so that farmers can identify the phenological stages of plants, allowing them to take preventive measures.
The organization of the present work is structured as follows: Section 2 describes the most relevant works on transfer learning, related concepts, and a description of the CNN models used in the present work. Section 3 describes the methodology of the investigation work. Section 4 contains the obtained results and their discussion. Section 5 presents the conclusions, and finally, future work is described.

2. Related Work

DL has been used to obtain high-quality maps. In this regard, Ge et al. [18] mapped crops from different regions in the period from planting to vigorous growth and compared the maps obtained by using conventional methods with those obtained with DL, where the latter reached 87% accuracy.
On the other hand, Yang et al. [17] proposed the identification of the different phenological phases of rice from RGB images captured by a drone with a CNN model that incorporated techniques such as spatial pyramidal sampling, knowledge transfer, and external data, which are essential for timely estimation and output, in comparison to previous approaches where data on the vegetation index in temporal series and diverse methods based on thresholds were used. The obtained results show that the approach has high precision in the identification of the phenology of rice, with 83.9% precision and a mean absolute error (MAE) of 0.18.
The approach of learning by transfer is used in various applications, such as the prediction of the performance of numerous crops worldwide, where Wang et al. [19] used remote sensors with satellite images to estimate the outcome of soy crops through algorithms and deep learning, offering an inexpensive and efficient alternative in comparison to conventional techniques that are generally expensive and difficult to expand in regions with limited access to data.
Identifying the phenology in diverse crops allows for determining critical moments for timely agricultural activities. In this regard, Reeb et al. [20] implemented a pre-trained CNN model named ResNet18 to classify the phenology of A. petiolata, comparing the results obtained to the classification conducted by a group of non-expert humans. During the validation stage of the model, 86.4% of the results obtained from a set of 2448 images were classified correctly by the proposed model. Subsequently, an evaluation of the precision of the model was made and compared to human precision, where the model correctly classified 81.7% of a total of 241 images. In contrast, the non-expert humans achieved a precision of 78.6%.
On the other hand, Datt et al. [15] used the CNN to recognize eight phenological stages in apple crops. The knowledge transfer model used was Inception-v3, which trained a set of images from the Srinagar region in India. The number of images captured in the area was 1200, extending to 7000 images through data augmentation techniques, obtaining results in comparison to other models such as Xception, Xception-v3, ResNet50, VGG16, and VGG19 where the metric used was for F1-Score and achieved values of 0.97, 0.96, 0.66, 0.96, and 0.95, respectively.
Through the use of characteristic descriptors, Yalcin et al. [16] proposed the implementation of automatic learning algorithms to compare learning algorithms based on a CNN to recognize and classify phenological stages in various types of crops such as wheat, corn, barley, lentil, cotton, and pepper through images collected by cameras at farming stations located in parts of the territory of Turkey. The AlexNet model used for the classification of phenology significantly exceeded automatic learning algorithms during the performance evaluation.
The combination of conventional techniques with DL methods can offer alternatives to the solution of classification problems, such as the case of the application of hybrid methods, to get to know the estimation of agricultural production in the work carried out by Zhao et al. [21], in which they used the knowledge transfer technique to learn from an existing model based on the combination of biomass algorithms of wheat crops and the model of simple performance. The results show a precise estimation of the wheat harvest with both models since they reveal a good correlation of R 2 = 0.83 and a root mean squared error (RMSE) of 1.91 t ha−1.
The use of temporal series and other techniques for phenological classification is usually relevant in the work carried out by Taylor et al. [22], who propose a model for creating temporal series of the phenological cycle. They use a hidden Markov post-process model to address the temporal correlation between images, reaching F1-Scores of 0.86 to 0.91. The results show the temporal progression of the crops from the emergency to the harvest, providing the daily phenological stages on a temporal scale.
DL techniques are used to classify the different phenological stages in different types of crops, including bean crops. The diversity of approaches and techniques used to classify images varies the obtained results compared to techniques and strategies used in diverse studies during the last decade. Thereupon, CNN produces trusted results, has a grand capacity for generalization in the classification of images, and has a high capacity for extracting features related to the phenology in different crops.

2.1. Convolutional Neural Networks (CNNs)

CNNs basically consist of three blocks: the first in a layer of convolution that allows the extraction of features of an image; the second is a block that consists of a layer of maximum grouping to execute a subsample of pixels and reduce the dimensionality, allowing the reduction of computational costs; and finally, the third block involves fully connected layers to provide the network with the capacity of classification [23,24,25]. The general description of the CNN architecture is shown in Figure 1, where features of the images are identified, extracted, and classified.

2.2. Transfer Learning

In general, CNN results are better if trained by more extensive data sets than small ones. However, many applications do not have large data sets, and transfer learning can be helpful in those applications where the data set is smaller in ImageNet [26]. For this reason, a re-trained model from large data sets can be used to learn new features from a comparatively smaller data set [27]. Figure 2 describes a block diagram of the transfer learning approach used in this study.
Recently, the scientific community has taken a particular interest in the transfer learning approach in diverse fields, such as medicine and agriculture, among others [28,29,30,31]. This approach allows previously acquired knowledge and avoids training with large quantities of data when training new deep architecture models [28,32,33].

2.3. Re-Trained Neural Networks

Using a re-trained CNN model has significant advantages in comparison to the design of models from zero, which require large sets of data and training that can take considerable time, including weeks, translating into high computational costs. On the other hand, a re-trained model can have a high capacity for generalization and accelerate convergence [34].
In this study, four models of re-trained CNN were used to evaluate the performance in the image classification identifying phenological phases in bean crops: AlexNet, VGG19, SqueezeNet, and GoogleNet. A brief introduction of each of the re-trained models is included below.

2.3.1. AlexNet Model

The algorithms for object detection and image classification were evaluated with the AlexNet model developed by Krizhevsky et al. [35] and the ImageNet Large-Scale Visual Recognition Challenge (LSVRC) for the model training [36]. The architecture is characterized by using a new activating function, the Rectified Linear Unit (ReLU), to add non-linearity, solve the gradient evanescent problem, and accelerate network training. CNN consists of eight layers in total: the first five layers are of convolution, some of which are followed by max-pooling layers, and the following three layers are fully connected, followed by an exit layer of 1000 neurons for the SoftMax for multiclass classification. The AlexNet model was trained with over one million images from the ImageNet database created by Deng et al. [26]. The entry size of the images is 227 × 227, a total of 60 million parameters and 650 thousand neurons.

2.3.2. VGG19

The VGG19 network is a nineteen-layer deep convolutional neural network developed by Simonyan et al. [37]. This model uses small filters of 3 × 3 in each of the sixteen convolution layers. Next, it uses three fully connected layers to classify images into 1000 categories. The ImageNet database developed by Deng et al. [26] was used to train the VGG19 model. The layers used for the extraction of features are separated into five groups where a layer of max-pooling follows each group, and it is required to insert and image the size of 224 × 224 to generate the label corresponding to the exit.

2.3.3. SqueezeNet

The SqueezeNet architecture uses compression techniques to reduce the model size without compromising its performance with a fire module that, instead of using convolutional layers followed by grouping layers, uses a combination of filters that combine convolutions of 1 × 1 and 3 × 3 to reduce the number of parameters. The SqueezeNet model proposed by Iandola et al. [38] contains fifty times fewer parameters than the AlexNet model, has eighteen layers of profundity, and requires a size of 227 × 227.

2.3.4. GoogleNet

This model, also known as Inception v1, was developed by Szegedy et al. [32]. It consists of twenty-two layers of profundity with entry images the size of 224 × 224. GoogleNet uses average pooling after the last convolutional layer instead of fully connected layers. The convolution modules called “Inception modules” are composed of multiple convolutions of different sizes (1 × 1, 3 × 3, and 5 × 5), allowing the network to capture the features on different spatial layers, facilitating the representation of fine details and complex patterns.
Table 1 summarizes the main features in terms of the parameters used, profundity, and size of the different architectures of CNN networks proposed in this study.

3. Materials and Methods

The methodology used in this study consists of three phases, as shown in Figure 3. The first phase describes the data acquisition procedure and the construction and general features of the obtained images. The second phase describes the transfer learning of CNN architectures used in this study and the configuration of hyperparameters, such as the learning rate, the lot size by iteration, the number of epochs, and the optimizer. The third phase describes the evaluation of the proposed models to measure the performance by applying different metrics.

3.1. Acquisition of Data

The two selected bean parcels are located in the municipality of Calera de Víctor Rosales in the state of Zacatecas, Mexico (22°54′14.6″ N 102°39′32.5″ W). The variety of beans used was pinto Saltillo, and the data were collected between 12 May and 15 August of the year 2023. The camera model used was HC-801Pro, with 4G technology, a range of optical vision of 120°, IP65 protection, and a resolution of 30 megapixels to acquire high-quality images.
Two cameras were installed to capture the images, as shown in Figure 4. To determine the number of images for the training and testing data set, according to Tylor et al. [22], the average time for bean harvest after its emergence is from 65 to 85 days approximately, which is why an average of eight to ten images were captured per day since the emergence of the plant.
The shooting method used was for two samples per sequence for intervals of time between 8:00, 10:00, 12:00, 16:00, and 18:00 h, obtaining a total of 814 images, allowing the experimental data to include the bean growing cycle in the vegetative phase and the reproductive phase, from the germination phase of the plant (V0) through the emergence phase of the plant (V1), primary leaves (V2), the first trifoliate leaf (V3), the third trifoliate leaf (V4), prefloration (R5), floration (R6), pod formation (R8), pod filling (R8), and maturation (R9).
Generally, the bean’s phenology is classified into ten classes and divided into two main categories: the vegetative and the production phases. However, for this investigation, only four classes were selected according to the most significant number of examples per class since they tend to be the most representative, according to Etemadi et al. [39].
An example of each class of the obtained data set can be observed in Figure 5, labeling the vegetative phase in the stage of primary leaves, first and third trifoliate leaves (V2–V4), reproductive phase in the stage of prefloration and floration (R5–R6), reproductive phase in the stage of formation and filling of pods (R7–R8), and reproductive phase in the stage of maturation (R9).
The training data set and tests used per class can be observed in Figure 6. Most images have a resolution of 5120 × 3840 pixels. However, the images were re-dimensioned to adjust the size according to the entry specifications for each proposed model [34].

3.2. Data Augmentation

The data augmentation contributes to avoiding overfitting the network and memorizing the exact details of the images during training. This is a common problem when the CNN model is exposed to small data sets where the learned patterns are not generalized into new data [40,41].
Currently, there is a tendency in deep learning training algorithms to allow for the increase of the initial data set through data augmentation techniques, obtaining results that can improve the precision performance in deep learning algorithms [42]. A series of aleatory transformations increased data to exploit the few examples of images and increase the precision of the proposed CNN models. The strategies of data augmentation used were rotation, translation, reflection, and scaling. Examples are shown in Figure 7.

3.3. Training of the Models

For the training data set and tests, the images were divided randomly in a partition of 70% of the training data set and 30% for the test set. Table 2 shows the configuration of the experimental equipment used in this investigation. The four models previously selected were trained by the ImageNet database, which contains more than 15 million images [26].
The hyperparameters selected in this study from the revised literature and the previously mentioned hardware capacity are described in Table 3. The selection of hyperparameters significantly affects the performance of CNN models, which is why a good selection is crucial. The hyperparameters were standardized for each model to compare the performance of the proposed models [31,43,44].
The optimizer used is the Stochastic Gradient Descent with Momentum (SGDM) method, which combines stochastic gradient descent and momentum techniques. Each iteration calculates the gradient using a random sample from the training set. Then, the weight is updated considering the previous update, allowing convergence acceleration and keeping it at a local minimum.
Momentum is employed to improve the precision and velocity of the training by adding a fraction from the previous step to the present step in the weight update. This allows it to overcome the obstacle of local minimums and maintain a constant impulse in the direction of the gradient.
The epochs refer to the number of iterations carried out regarding the correlation of forward and reverse propagation to reduce loss. The size of the epochs describes the number of examples used in each iteration of the training algorithm. The learning rate is defined by the velocity size in which the optimization function performs a search to converge [45].

3.4. Performance Evaluation

At present, an extended variety of metrics are used to evaluate the performance of CNN models, where information is given about the aspects and characteristics that allow the evaluation of the performance of the models. A number of true positives (TP), true negatives (TN), false positives (FP), and false negatives (FN) are needed to calculate the performance of the models. These cases represent the combinations of true and predicted classes in classification problems. Therefore, TP + TN + FP + FN equals the total number of samples and is described in a confusion matrix [46].
When trained by transfer learning, the biased distributions appear naturally, producing an intrinsic unbalance. This is why it is necessary to employ metrics that evaluate the global performance of each model. In this regard, the metrics employed [16,20,22,46,47] were used to compare the performance of the models without setting aside the different characteristics of the training and validation data used in this study.
TP is the true positive, which means the prediction is positive. FP is the false positive, which means the prediction is negative. However, the prediction is positive. FN is the false negative, which means a positive prediction, but the result is negative. TN is the true negative, which means a negative result prediction.
The confusion matrix is a tool that allows the visualization of a model’s performance when classifying and containing the previously defined elements. The rows in the matrix represent the true class, and the columns, the predicted class, and the primary diagonal cells describe the correctly classified observations. In contrast, the lateral diagonals correspond to the incorrectly classified observations.
In this study, five metrics were used to evaluate the performance of the proposed models: accuracy, precision, sensitivity, specificity, and F1-Score [27,48]. Accuracy is the relation between the number of correct predictions and the total number of made predictions, as calculated by Equation (1).
A c c u r a c y = T P + T N T P + T N + F P + F N
Precision measures the proportion of correct predictions made by the model, in other words, the number of correctly classified elements as positives out of a total of elements identified as positive. The mathematical representation is described in Equation (2).
P r e c i s i o n = T P T P + F P
Sensitivity is also known as recall; it calculates the proportion of correctly identified cases as positive from a total of true positives, as described in Equation (3).
S e n s i t i v i t y = T P T P + F N
Specificity is the opposite of sensitivity or recall and calculates the portion of cases identified as negatives. It is calculated by Equation (4).
S p e c i f i c i t y = T N T N + F P
F1-Score allows the combination of precision and sensitivity or recall, where the value of one indicates a good balance between precision and sensitivity in the classification model. Its mathematical representation is described in Equation (5).
F 1 - s c o r e = 2 × p r e c i s i o n × r e c a l l p r e c i s i o n + r e c a l l

4. Results and Discussion

Table 4 provides a detailed calculation for each of the different metrics of AlexNet’s architecture. A sensitivity of 100% was obtained for the prediction of classes R9 and V2–V4, which correspond to the reproductive phenological phase in the stage of maturation and the vegetative phase in the stage of primary leaves and first and third trifoliate leaves, respectively.
The averages obtained for the accuracy, precision, sensitivity, specificity, and F1-Score were 95.8%, 94.1%, 97.2%, 98.6%, and 95.5% of the predicted classes during validation, respectively.
Table 5 shows the different metrics in the VGG19 model. In classes R5-R6, corresponding to the reproductive phenological phase in the stages of prefloration and floration, a precision of 100% was reached.
A 95% precision average for all classes can be observed. On the other hand, maximum sensitivity was also obtained for classes R9 and V2–V4, corresponding to the phenological phase of reproduction in the maturation stage and vegetative phase of primary leaves and first and third trifoliate leaves, respectively. An average of 97.4% of sensitivity is observed in each class. In addition, averages reached for specificity are 98.8% and 96% F1-Score in all classes, achieving the best scores compared to the other architectures.
A detailed calculation for each metric used in the architecture SqueezeNet can be shown in Table 6. For the classes V2–V4, which corresponds to the vegetative phenological phase in the stage of primary leaves and first and third trifoliate leaves, a sensitivity of 100% was obtained. On the other hand, averages in accuracy, precision, sensitivity, and F1-Score of 95.8%, 93.4%, 95.9%, 98.6%, and 94.4%, respectively, are observed for all predicted classes during validation.
The different metrics of the GoogleNet model can be observed in Table 7, where the average obtained for precision is 96.8% in all predicted classes, and there is a maximum sensitivity in the prediction of classes V2–V4 that corresponds to the vegetative phenological phase in the stage of primary leaves and first and third trifoliate leaves. The averages observed in the metrics of accuracy, precision, sensitivity, specificity, and F1-Score are 96.7%, 96.8%, 95.7%, 98.7%, and 96.2%, respectively, for all predicted classes during validation, which concur with the metric of precision obtained for the VGG19 model.
Table 8 shows the results obtained in each metric, with the best values obtained during the architecture’s evaluation highlighted in bold. It shows that the architecture of VGG19 and GoogleNet obtained the best performance, and both concur in accuracy. On the other hand, the architecture with the lowest performance observed is SqueezeNet due to the values obtained that are generally lower than those obtained in other architectures. However, SqueezeNet required the least training time in comparison to the others.
A comparison of the accuracy obtained in each of the models during the training and validation stages is shown in Figure 8, where the architecture AlexNet reaches the highest accuracy percentage compared to the other models during the training stage. However, in the validation stage, VGG19 reached the highest accuracy percentage. This stage shows that the GoogleNet model obtained the lowest accuracy percentage compared to the other models. However, it is observed that this model reached the highest accuracy percentage during the validation stage, just like VGG19; in other words, they obtained a greater capacity.
The architecture performance summary with measurements obtained in each metric calculated during the validation stage can be observed in Figure 9; the accuracy metric obtained during the training stage is included for each model. It is observed that the GoogleNet architecture maintained the best balance in the projection of all metrics. On the other hand, it is observed that the architecture SqueezeNet obtained the best performance compared to the rest of the architectures, except in precision, where the result is the same as the AlexNet architecture, and sensitivity is less than the obtained with the GoogleNet architecture.
Based on the obtained results, the GoogleNet architecture has a higher performance during the validation compared to its performance during the training process; the cause of this behavior could be due to the limited amount of training data, the adequate selection of hyperparameters, and possibly an over-adjustment. However, the difference between the precision of training and validation is 1.4% compared to the models AlexNet, VGG19, and SqueezeNet, which present a difference of 1.7%, 3.3%, and 1.7%, respectively—considering that increased training data will give an outcome with a tendency to decrease the performance during the validation of each architecture.
According to the behavior during the validation, the architectures AlexNet and SqueezeNet presented a low-balanced tendency in the metrics, obtaining low results for precision and F1-Score. On the other hand, the architecture VGG19 registered the same level of performance as GoogleNet but with lower precision and F1-Score, giving a reason to consider the architecture GoogleNet as having the best global performance.
Figure 10 shows the confusion matrix of the four proposed CNN models. It also provides a detailed analysis of instance numbers correctly classified by each proposed architecture. Compared to other architectures, the AlexNet architecture presented problems in correctly classifying the class R9, which corresponds to the reproductive phenological phase in the stage of maturation, achieving the classification of only 85.2% of instances.
The confusion matrix of the VGG19 architecture shows its high capacity to classify instances correctly. The diagonal shows the correctly classified instances; however, class R9 presented the most difficulty. The SqueezeNet architecture, like the previous architectures, shows more difficulties in correctly classifying the class R9; however, for the GoogleNet architecture, the class presents minimal difficulties.
Table 9 describes a comparison summary of the performance results obtained by other authors concerning the techniques and metrics used for this work. It breaks down the values obtained for accuracy, precision, sensitivity, and F1-Score for each proposed architecture.

5. Conclusions

The proposed methodology shows that the proposed CNN models allow the correct classification of more than 90% of the samples, even when working with an unbalanced and relatively minor data set. In addition, each analyzed architecture has different characteristics, such as the number of layers and used filters. However, it is crucial to highlight a suitable selection of metrics to discriminate one architecture from the other.
Evaluating different CNN topologies is significant for future work since the architectures can present bias due to being trained with numerous images from which many are not part of the final classification. In this regard, evaluating the performance by transfer with new data lays the foundation for new work, such as identifying nutrients or plagues for this species. The joint evaluation of the metrics accuracy, precision specificity, sensitivity, and F1-Score allows the obtention of a multifaceted analysis, resulting in a higher performance GoogleNet architecture. Even though the global performance of each model is acceptable, data augmentation can modify the performance of all architectures.
One of the main limitations in the implementation of CNN models is the lack of data for certain classes, for which the main contribution of this study is to be able to distinguish the performance obtained through a reduced data set, where the application of data augmentation techniques other than reducing the overfitting in training helps improve the capacity of generalization in the network in comparison to other studies where augmentation techniques were not applied in the same way as the performance results of the models in Table 9.
On the other hand, through a methodological analysis, the performance was compared and evaluated by applying five metrics to four CNN models. The GoogleNet architecture obtained the best performance, showing the best results in most metrics, obtaining 96.71% accuracy, 96.81% precision, 95.77% sensitivity, 98.73% specificity, and 96.25% F1-Score.

6. Future Work

This study will open other alternatives that could be applied using the same transfer-by-learning approach for the controlled prevention of plagues and diseases in bean crops through timely intervention and automated computerized image classification.

Author Contributions

Conceptualization, T.I.-P. and U.A.H.-G.; methodology, R.J.-M., M.d.R.M.-B. and J.I.C.-F.; software, U.A.H.-G. and H.A.G.-O.; validation, R.R.-M. and T.I.-P.; formal analysis, H.C.C.-A., U.A.H.-G. and R.R.-M.; investigation, C.N., H.C.C.-A. and R.J.-M.; resources, U.A.H.-G. and H.A.G.-O.; data curation, C.N., M.d.R.M.-B. and J.I.C.-F.; writing—original draft preparation, T.I.-P. and J.I.C.-F.; writing—review and editing, T.I.-P. and R.J.-M.; visualization, R.R.-M. and H.A.G.-O.; supervision, C.N. and M.d.R.M.-B.; project administration, T.I.-P. and F.D.M.-D.; funding acquisition, H.C.C.-A., F.D.M.-D. and T.I.-P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Instituto Politécnico Nacional (IPN) under grant number SIP/20230388.

Data Availability Statement

The data presented in this study are available upon request from the corresponding author.

Acknowledgments

We want to deeply thank Instituto Nacional de Investigaciones Forestales, Agrícolas y Pecuarias Campo Experimental Zacatecas (INIFAP) for providing us with the experimental field for this research and the Consejo Zacatecano de Ciencia, Tecnología e Innovación (COZCyT). We sincerely thank the people who provided support and advice for this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. FAO. Food and Agriculture Organization of the United Nations International Year of Plant Health. Available online: https://www.fao.org/plant-health-2020/about/en/ (accessed on 11 December 2023).
  2. Velia, A.; Garay, A.; Alberto, J.; Gallegos, A.; Muro, L.R. El Cultivo Del Frijol Presente y Futuro Para México; INIFAP: Celaya, Gto., México, 2021; Volume 1, ISBN 978-607-37-1318-4. [Google Scholar]
  3. Gregory, P.J.; Ingram, J.S.I.; Brklacich, M. Climate Change and Food Security. Philos. Trans. R. Soc. B Biol. Sci. 2005, 360, 2139–2148. [Google Scholar] [CrossRef]
  4. Chakraborty, S.; Newton, A.C. Climate Change, Plant Diseases and Food Security: An Overview. Plant Pathol. 2011, 60, 2–14. [Google Scholar] [CrossRef]
  5. Mutengwa, C.S.; Mnkeni, P.; Kondwakwenda, A. Climate-Smart Agriculture and Food Security in Southern Africa: A Review of the Vulnerability of Smallholder Agriculture and Food Security to Climate Change. Sustainability 2023, 15, 2882. [Google Scholar] [CrossRef]
  6. Bailes, E.J.; Ollerton, J.; Pattrick, J.G.; Glover, B.J. How Can an Understanding of Plant-Pollinator Interactions Contribute to Global Food Security? Curr. Opin. Plant Biol. 2015, 26, 72–79. [Google Scholar] [CrossRef]
  7. Saha, H.; Chatterjee, S.; Paul, A. Role of Pollinators in Plant Reproduction and Food Security: A Concise Review. Res. J. Agric. Sci. 2023, 14, 72–79. [Google Scholar]
  8. Trebicki, P.; Finlay, K. Pests and Diseases under Climate Change; Its Threat to Food Security; John Wiley & Sons Ltd.: Chichester, UK, 2019; Volume 1, ISBN 9781119180654. [Google Scholar]
  9. Alam, F.B.; Tushar, S.R.; Zaman, S.M.; Gonzalez, E.D.S.; Bari, A.M.; Karmaker, C.L. Analysis of the Drivers of Agriculture 4.0 Implementation in the Emerging Economies: Implications towards Sustainability and Food Security. Green Technol. Sustain. 2023, 1, 100021. [Google Scholar] [CrossRef]
  10. Mcbeath, J.H.; Mcbeath, J. Environmental Change and Food Security in China; Beniston, M., Ed.; Springer: Fairbanks, AK, USA, 2010; Volume 35, ISBN 978-1-4020-9179-7. [Google Scholar]
  11. Calicioglu, O.; Flammini, A.; Bracco, S.; Bellù, L.; Sims, R. The Future Challenges of Food and Agriculture: An Integrated Analysis of Trends and Solutions. Sustainability 2019, 11, 222. [Google Scholar] [CrossRef]
  12. He, J.; Chen, K.; Pan, X.; Zhai, J.; Lin, X. Advanced Biosensing Technologies for Monitoring of Agriculture Pests and Diseases: A Review. J. Semicond. 2023, 44, 23104. [Google Scholar] [CrossRef]
  13. Mallick, M.D.T.; Biswas, S.; Das, A.K.; Saha, H.N.; Chakrabarti, A.; Deb, N. Deep Learning Based Automated Disease Detection and Pest Classification in Indian Mung Bean. Multimed. Tools Appl. 2023, 82, 12017–12041. [Google Scholar] [CrossRef]
  14. Hadipour-Rokni, R.; Asli-Ardeh, E.A.; Jahanbakhshi, A.; Paeen-Afrakoti, I.E.; Sabzi, S. Intelligent Detection of Citrus Fruit Pests Using Machine Vision System and Convolutional Neural Network through Transfer Learning Technique. Comput. Biol. Med. 2023, 155, 106611. [Google Scholar] [CrossRef]
  15. Datt, R.M.; Kukreja, V. Phenological Stage Recognition Model for Apple Crops Using Transfer Learning. In Proceedings of the 2022 2nd International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Greater Noida, India, 28–29 April 2022; pp. 1537–1542. [Google Scholar]
  16. Yalcin, H. Phenology Recognition Using Deep Learning. In Proceedings of the 2018 Electric Electronics, Computer Science, Biomedical Engineerings’ Meeting (EBBT), Istanbul, Turkey, 18–19 April 2018; pp. 1–5. [Google Scholar]
  17. Yang, Q.; Shi, L.; Han, J.; Yu, J.; Huang, K. A near Real-Time Deep Learning Approach for Detecting Rice Phenology Based on UAV Images. Agric. For. Meteorol. 2020, 287, 107938. [Google Scholar] [CrossRef]
  18. Ge, S.; Zhang, J.; Pan, Y.; Yang, Z.; Zhu, S. Transferable Deep Learning Model Based on the Phenological Matching Principle for Mapping Crop Extent. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102451. [Google Scholar] [CrossRef]
  19. Wang, A.X.; Tran, C.; Desai, N.; Lobell, D.; Ermon, S. Deep Transfer Learning for Crop Yield Prediction with Remote Sensing Data. In Proceedings of the 1st ACM SIGCAS Conference on Computing and Sustainable Societies, COMPASS, Menlo Park and San Jose, CA, USA, 20–22 June 2018. [Google Scholar]
  20. Reeb, R.A.; Aziz, N.; Lapp, S.M.; Kitzes, J.; Heberling, J.M.; Kuebbing, S.E. Using Convolutional Neural Networks to Efficiently Extract Immense Phenological Data From Community Science Images. Front. Plant Sci. 2022, 12, 787407. [Google Scholar] [CrossRef]
  21. Zhao, Y.; Han, S.; Meng, Y.; Feng, H.; Li, Z.; Chen, J.; Song, X.; Zhu, Y.; Yang, G. Transfer-Learning-Based Approach for Yield Prediction of Winter Wheat from Planet Data and SAFY Model. Remote Sens. 2022, 14, 5474. [Google Scholar] [CrossRef]
  22. Taylor, S.D.; Browning, D.M. Classification of Daily Crop Phenology in PhenoCams Using Deep Learning and Hidden Markov Models. Remote Sens. 2022, 14, 286. [Google Scholar] [CrossRef]
  23. Bailer, C.; Habtegebrial, T.; Varanasi, K.; Stricker, D. Fast Feature Extraction with CNNs with Pooling Layers. arXiv 2018, arXiv:1805.03096. [Google Scholar] [CrossRef]
  24. Paymode, A.S.; Malode, V.B. Transfer Learning for Multi-Crop Leaf Disease Image Classification Using Convolutional Neural Network VGG. Artif. Intell. Agric. 2022, 6, 23–33. [Google Scholar] [CrossRef]
  25. Bosilj, P.; Aptoula, E.; Duckett, T.; Cielniak, G. Transfer Learning between Crop Types for Semantic Segmentation of Crops versus Weeds in Precision Agriculture. J. Field Robot. 2020, 37, 7–19. [Google Scholar] [CrossRef]
  26. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A Large-Scale Hierarchical Image Database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  27. Rahman, T.; Chowdhury, M.E.H.; Khandakar, A.; Islam, K.R.; Islam, K.F.; Mahbub, Z.B.; Kadir, M.A.; Kashem, S. Transfer Learning with Deep Convolutional Neural Network (CNN) for Pneumonia Detection Using Chest X-ray. Appl. Sci. 2020, 10, 3233. [Google Scholar] [CrossRef]
  28. Barbhuiya, A.A.; Karsh, R.K.; Jain, R. CNN Based Feature Extraction and Classification for Sign Language. Multimed Tools Appl. 2021, 80, 3051–3069. [Google Scholar] [CrossRef]
  29. Salehi, A.W.; Khan, S.; Gupta, G.; Alabduallah, B.I.; Almjally, A.; Alsolai, H.; Siddiqui, T.; Mellit, A. A Study of CNN and Transfer Learning in Medical Imaging: Advantages, Challenges, Future Scope. Sustainability 2023, 15, 5930. [Google Scholar] [CrossRef]
  30. Kim, H.E.; Cosa-Linan, A.; Santhanam, N.; Jannesari, M.; Maros, M.E.; Ganslandt, T. Comparison of Three Dimensional Reconstruction and Conventional Computer Tomography Angiography in Patients Undergoing Zero-Ischemia Laparoscopic Partial Nephrectomy. BMC Med. Imaging 2022, 22, 47. [Google Scholar] [CrossRef]
  31. Narvekar, C.; Rao, M. Flower Classification Using CNN and Transfer Learning in CNN-Agriculture Perspective. In Proceedings of the 2020 3rd International Conference on Intelligent Sustainable Systems (ICISS), Thoothukudi, India, 3–5 December 2020; pp. 660–664. [Google Scholar]
  32. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  33. Jogin, M.; Mohana, M.; Madhulika, M.; Divya, G.; Meghana, R.; Apoorva, S. Feature Extraction Using Convolution Neural Networks (CNN) and Deep Learning. In Proceedings of the 2018 3rd IEEE International Conference on Recent Trends in Electronics, Information & Communication Technology (RTEICT), Bangalore, India, 18–19 May 2018; pp. 2319–2323. [Google Scholar]
  34. Cuevas-Rodriguez, E.O.; Galvan-Tejada, C.E.; Maeda-Gutiérrez, V.; Moreno-Chávez, G.; Galván-Tejada, J.I.; Gamboa-Rosales, H.; Luna-García, H.; Moreno-Baez, A.; Celaya-Padilla, J.M. Comparative Study of Convolutional Neural Network Architectures for Gastrointestinal Lesions Classification. PeerJ 2023, 11, e14806. [Google Scholar] [CrossRef]
  35. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  36. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  37. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50× Fewer Parameters and <0.5 MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  39. Etemadi, F.; Hashemi, M.; Zandvakili, O.R.; Mangan, F.X. Phenology, Yield and Growth Pattern of Faba Bean Varieties. Int. J. Plant Prod. 2018, 12, 243–250. [Google Scholar] [CrossRef]
  40. Kolar, Z.; Chen, H.; Luo, X. Transfer Learning and Deep Convolutional Neural Networks for Safety Guardrail Detection in 2D Images. Autom. Constr. 2018, 89, 58–70. [Google Scholar] [CrossRef]
  41. Lopez, A.; Giro-I-Nieto, X.; Burdick, J.; Marques, O. Skin Lesion Classification from Dermoscopic Images Using Deep Learning Techniques. In Proceedings of the 2017 13th IASTED International Conference on Biomedical Engineering (BioMed), Innsbruck, Austria, 20–21 February 2017; pp. 49–54. [Google Scholar]
  42. Kurek, J.; Antoniuk, I.; Górski, J.; Jegorowa, A.; Świderski, B.; Kruk, M.; Wieczorek, G.; Pach, J.; Orłowski, A.; Aleksiejuk-Gawron, J. Data Augmentation Techniques for Transfer Learning Improvement in Drill Wear Classification Using Convolutional Neural Network. Mach. Graph. Vis. 2019, 28, 3–12. [Google Scholar] [CrossRef]
  43. Hassan, S.M.; Maji, A.K.; Jasiński, M.; Leonowicz, Z.; Jasińska, E. Identification of Plant-Leaf Diseases Using Cnn and Transfer-Learning Approach. Electronics 2021, 10, 1388. [Google Scholar] [CrossRef]
  44. Thenmozhi, K.; Reddy, U.S. Crop Pest Classification Based on Deep Convolutional Neural Network and Transfer Learning. Comput. Electron. Agric. 2019, 164, 104906. [Google Scholar] [CrossRef]
  45. Alzubaidi, L.; Zhang, J.; Humaidi, A.J.; Al-Dujaili, A.; Duan, Y.; Al-Shamma, O.; Santamaría, J.; Fadhel, M.A.; Al-Amidie, M.; Farhan, L. Review of Deep Learning: Concepts, CNN Architectures, Challenges, Applications, Future Directions. J. Big Data 2021, 8, 53. [Google Scholar] [CrossRef] [PubMed]
  46. Qin, J.; Hu, T.; Yuan, J.; Liu, Q.; Wang, W.; Liu, J.; Guo, L.; Song, G. Deep-Learning-Based Rice Phenological Stage Recognition. Remote Sens. 2023, 15, 2891. [Google Scholar] [CrossRef]
  47. Han, J.; Shi, L.; Yang, Q.; Huang, K.; Zha, Y.; Yu, J. Real-Time Detection of Rice Phenology through Convolutional Neural Network Using Handheld Camera Images. Precis. Agric. 2021, 22, 154–178. [Google Scholar] [CrossRef]
  48. Johnson, J.M.; Khoshgoftaar, T.M. Survey on Deep Learning with Class Imbalance. J. Big Data 2019, 6, 27. [Google Scholar] [CrossRef]
Figure 1. The architecture of convolutional network.
Figure 1. The architecture of convolutional network.
Agriengineering 06 00048 g001
Figure 2. Block diagram of the concept of transfer learning.
Figure 2. Block diagram of the concept of transfer learning.
Agriengineering 06 00048 g002
Figure 3. Diagram of proposed methodology.
Figure 3. Diagram of proposed methodology.
Agriengineering 06 00048 g003
Figure 4. Installation of the GSM camera station in the open field for the capture of images: (a) camera station for the capture of images; (b) schematic diagram for the acquisition of images.
Figure 4. Installation of the GSM camera station in the open field for the capture of images: (a) camera station for the capture of images; (b) schematic diagram for the acquisition of images.
Agriengineering 06 00048 g004
Figure 5. Descriptive stages of the phenology of the bean: (a) vegetative phase in primary leaves, first and third trifoliate leaves; (b) reproductive phase in prefloration and floration; (c) reproductive stage in the formation and filling of pods; (d) reproductive phase in maturation.
Figure 5. Descriptive stages of the phenology of the bean: (a) vegetative phase in primary leaves, first and third trifoliate leaves; (b) reproductive phase in prefloration and floration; (c) reproductive stage in the formation and filling of pods; (d) reproductive phase in maturation.
Agriengineering 06 00048 g005
Figure 6. Images for training and tests per class.
Figure 6. Images for training and tests per class.
Agriengineering 06 00048 g006
Figure 7. Data augmentation in an image of the phenology of the bean: (a) original image without data increase; (b) image after rotation; (c) image after translation; (d) image after reflection; (e) image after scaling.
Figure 7. Data augmentation in an image of the phenology of the bean: (a) original image without data increase; (b) image after rotation; (c) image after translation; (d) image after reflection; (e) image after scaling.
Agriengineering 06 00048 g007
Figure 8. Accuracy of models during training and validation.
Figure 8. Accuracy of models during training and validation.
Agriengineering 06 00048 g008
Figure 9. Summary of model performance: (a) AlexNet model; (b) VGG19 model; (c) SqueezeNet model; (d) GoogleNet model.
Figure 9. Summary of model performance: (a) AlexNet model; (b) VGG19 model; (c) SqueezeNet model; (d) GoogleNet model.
Agriengineering 06 00048 g009
Figure 10. Confusion matrix of CNN models: (a) AlexNet architecture; (b) VGG19 architecture; (c) SqueezeNet architecture; (d) GoogleNet architecture.
Figure 10. Confusion matrix of CNN models: (a) AlexNet architecture; (b) VGG19 architecture; (c) SqueezeNet architecture; (d) GoogleNet architecture.
Agriengineering 06 00048 g010aAgriengineering 06 00048 g010b
Table 1. Summary of presented architectures.
Table 1. Summary of presented architectures.
CNN ArchitectureYearDeveloperProfundity (Number of Layers)Size (MB)Number of Parameters
AlexNet2012Krizhevsky et al. [35]824060 million
VGG192014Simonyan et al. [37]19550138 million
SqueezeNet2016Iandola et al. [38]1851.2 million
GoogleNet2014Google22504 million
Table 2. Configuration of experimental equipment.
Table 2. Configuration of experimental equipment.
Name of ConfigurationParameters
ProcessorAMD Ryzen 7, 5700U, 8 cores at 1.8 GHz
Graphic CardRadeon Graphics up to 4.3 GHz
RAM16 GB
Operative systemWindows 11
Programing platformMATLAB (2022a)
Table 3. Training hyperparameters of pre-trained models.
Table 3. Training hyperparameters of pre-trained models.
Re-Trained CNN ModelImage SizeOptimizerMomentumEpochsBatch SizeLearning Rate
AlexNet227 × 227Stochastic Gradient Descent with Momentum0.9151280.001
VGG19224 × 224
SqueezeNet227 × 227
GoogleNet224 × 224
Table 4. Calculation of metric for the AlexNet model.
Table 4. Calculation of metric for the AlexNet model.
NameClassesAveraging
True Positive6293235558.25
False Positive32412.5
False Negative28002.5
True Negative176140216187179.75
Accuracy0.9588470.9588470.9588470.9588470.958847
Precision0.9538460.9789470.8518510.9821420.941697
Sensitivity0.9687500.920792110.972385
Specificity0.9832400.9859150.9818180.9946800.986413
F1-Score0.9612400.9489790.9200000.9909900.955302
Table 5. Calculation of metrics for VGG19 model.
Table 5. Calculation of metrics for VGG19 model.
NameClassesAveraging
True Positive6097235558.75
False Positive03412
False Negative44002
True Negative179139216187180.25
Accuracy0.9670780.9670780.9670780.9670780.967078
Precision10.9700000.8518510.9821420.950998
Sensitivity0.9375000.960396110.974474
Specificity10.9788730.9818180.9946800.988843
F1-Score0.9677410.9651740.9200000.9909900.960976
Table 6. Calculation of metrics for SqueezeNet model.
Table 6. Calculation of metrics for SqueezeNet model.
NameClassesAveraging
True Positive5997225558.25
False Positive03612.50
False Negative54102.50
True Negative179139214187179.75
Accuracy0.9588470.9588470.9588470.9588470.958847
Precision10.9700000.7857140.9821420.934464
Sensitivity0.9218750.9603960.95652110.959698
Specificity10.9788730.9727270.9946800.986570
F1-Score0.9593490.9651740.8627450.9909900.944564
Table 7. Calculation of metrics for GoogleNet model.
Table 7. Calculation of metrics for GoogleNet model.
NameClassesAveraging
True Positive6099215558.75
False Positive15112
False Negative42202
True Negative178137219187180.25
Accuracy0.9670780.9670780.9670780.9670780.967078
Precision0.9836060.9519230.9545450.9821420.968054
Sensitivity0.9375000.9801980.91304310.957685
Specificity0.9944130.9647880.9954540.9946800.987334
F1-Score0.9600000.9658530.9333330.9909900.962544
Table 8. Comparison of results for each of the architectures.
Table 8. Comparison of results for each of the architectures.
MetricsAlexNetVGG19SqueezeNetGoogleNet
Accuracy0.95880.96710.95880.9671
Precision0.94170.95100.93450.9681
Sensitivity0.97240.97450.95970.9577
Specificity0.98640.98880.98660.9873
F1-Score0.95530.96100.94460.9625
Time [minutes]17.31168.7620.1925.48
Table 9. Comparison of similar work recently published.
Table 9. Comparison of similar work recently published.
AuthorPhenologyTechniqueData AugmentationAccuracy [%]Precision [%]Sensitivity [%]F1-Score [%]
Yalcin [16]Wheat, barley, lentil, cotton, and pepperRe-trained CNN AlexNet modelNot87.1488.1287.2487.28
Qin et al. [46]RiceRe-trained CNN ResNet-50 modelYes87.33---
Reeb et al. [20]Alliaria petiolataRe-trained CNN ResNet-18 modelNot86.4---
Taylor et al. [22]Corn, wheat, barley, soybean, and alfalfaRe-trained CNN VGG16 model and temporal model of MarkovNot-929091
Han et al. [47]RiceCNN model with support vector machine (SVM)Yes91.3--93
This studyBeanRe-trained CNN GoogleNet modelYes96.7195.197.4596.1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ibarra-Pérez, T.; Jaramillo-Martínez, R.; Correa-Aguado, H.C.; Ndjatchi, C.; Martínez-Blanco, M.d.R.; Guerrero-Osuna, H.A.; Mirelez-Delgado, F.D.; Casas-Flores, J.I.; Reveles-Martínez, R.; Hernández-González, U.A. A Performance Comparison of CNN Models for Bean Phenology Classification Using Transfer Learning Techniques. AgriEngineering 2024, 6, 841-857. https://doi.org/10.3390/agriengineering6010048

AMA Style

Ibarra-Pérez T, Jaramillo-Martínez R, Correa-Aguado HC, Ndjatchi C, Martínez-Blanco MdR, Guerrero-Osuna HA, Mirelez-Delgado FD, Casas-Flores JI, Reveles-Martínez R, Hernández-González UA. A Performance Comparison of CNN Models for Bean Phenology Classification Using Transfer Learning Techniques. AgriEngineering. 2024; 6(1):841-857. https://doi.org/10.3390/agriengineering6010048

Chicago/Turabian Style

Ibarra-Pérez, Teodoro, Ramón Jaramillo-Martínez, Hans C. Correa-Aguado, Christophe Ndjatchi, Ma. del Rosario Martínez-Blanco, Héctor A. Guerrero-Osuna, Flabio D. Mirelez-Delgado, José I. Casas-Flores, Rafael Reveles-Martínez, and Umanel A. Hernández-González. 2024. "A Performance Comparison of CNN Models for Bean Phenology Classification Using Transfer Learning Techniques" AgriEngineering 6, no. 1: 841-857. https://doi.org/10.3390/agriengineering6010048

APA Style

Ibarra-Pérez, T., Jaramillo-Martínez, R., Correa-Aguado, H. C., Ndjatchi, C., Martínez-Blanco, M. d. R., Guerrero-Osuna, H. A., Mirelez-Delgado, F. D., Casas-Flores, J. I., Reveles-Martínez, R., & Hernández-González, U. A. (2024). A Performance Comparison of CNN Models for Bean Phenology Classification Using Transfer Learning Techniques. AgriEngineering, 6(1), 841-857. https://doi.org/10.3390/agriengineering6010048

Article Metrics

Back to TopTop