Next Article in Journal
Leveraging Technology to Break Barriers in Public Health for Students with Intellectual Disabilities
Previous Article in Journal
From Transformers to Voting Ensembles for Interpretable Sentiment Classification: A Comprehensive Comparison
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model for Agricultural Production in Colombia Using a Neuro-Fuzzy Inference System

by
Andrea C. Gómez
*,
Lilian A. Bejarano
and
Helbert E. Espitia
Facultad de Ingeniería, Universidad Distrital Francisco José de Caldas, Bogotá 110231, Colombia
*
Author to whom correspondence should be addressed.
Computers 2025, 14(5), 168; https://doi.org/10.3390/computers14050168
Submission received: 3 April 2025 / Revised: 27 April 2025 / Accepted: 28 April 2025 / Published: 1 May 2025

Abstract

:
As mentioned by the Food and Agriculture Organization of the United Nations, agriculture has a primary role in food security. Given the advantageous conditions that Colombia has as a biodiverse country, creating and implementing sustainable and comprehensive agricultural systems is essential to generate agricultural decision-making tools. Therefore, this paper displays the design and deployment (training–validation) of a neuro-fuzzy model for the relevant agricultural production in Colombia. Four different configurations are proposed according to the data collected and the variables identified. The results show that a remarkable prediction of the models (configurations) is achieved by using training and validation data.

1. Introduction

According to [1], the diverse process of globalization integrating various markets and countries has been studied for its impact on climate change. The ecological modernization theory considers that environmental degradation is not necessarily the product of industrialization and economic growth [2]. However, it has been observed that economic growth, the use of fossil fuels, technological innovations, and globalization drive the emission of greenhouse gases [3], especially in low-income countries, which are the ones that suffer severely from the consequences [4]. Climate change is a global problem [5] that affects the agricultural sector, putting food security and ecosystem resilience at risk [6,7].
As a result of global warming and climate change, crops such as coffee [8], wheat [9], rice [10], cocoa [11], corn [12], cashews, and avocado [13] could disappear. Crop production is already suffering the consequences of climate change, and the degree of impact on a country depends on geographic location, adaptability, and Gross Domestic Product (GDP) [14]. For example, nearly half of the food in the UK is imported; thus, the food security and health of the inhabitants are susceptible to climate change in food-exporting countries [15]. In Colombia, the deterioration of ecosystems such as the Amazon contributes to climate change [16]. Forests lose carbon storage capacity; climatic phenomena become more prolonged or intense [17].
Colombia is located in a tropical zone, crossed by three mountain ranges, which provide advantageous conditions that position the country as one of the most biodiverse in the world, with regulating ecosystems such as wetlands [18], and with the third-largest number of plants [19]. Due to the variety of soils and temperatures [20], together with water availability, numerous crops such as coffee, corn, rice, sugar cane, and palm oil [21] grow easily, which are usually available all year round. Agriculture is one of the most representative sectors in Colombia. According to the 2019 National Agricultural Survey, the agricultural sector contributed 8.7% to the Gross Domestic Product, and the estimated total agricultural production for 2019 was 63 million tons [22]. Coffee is one of the star products, being the third-largest Colombian export product in 2022, with a total of USD 4.15 billion dollars, making Colombia the second-largest coffee exporter in the world [23].
Agriculture has a primary lead in food security, a concept that, regarding the Food and Agriculture Organization of the United Nations (FAO), is defined as follows: “at the individual, household, national and global levels, it is achieved when all persons, at all times, have economic and physical access to sufficient, safe and nutritious food to satisfy their dietary needs and preferences, in order to maintain a healthy and active life” [24]. The agricultural sector expands the productive and export base, generates employment, and provides material that serves as input for other sectors and, most importantly, ensures that food supply is adequate, safe, and affordable [25].
Colombian agriculture is characterized by extreme agroclimatic heterogeneity and often suffers from sparse, noisy data, posing significant challenges to accurately yield forecasting and decision-making. This study displays the advantages of the Adaptive Neuro-Fuzzy Inference System (ANFIS) in addressing these challenges. This approach can dynamically tune fuzzy membership functions to local microclimates—from the Caribbean coast to the Amazon rainforest and embedding linguistic rules that capture uncertainty (high rainfall, acidic soils, and intensive management). Its hybrid architecture robustly handles missing or biased records; for example, in an analysis of observations of agricultural employment, ANFIS generated clear decision rules for subsidy allocation and extension programs [26]. Regarding Stevia crop trials, ANFIS produces transparent, field-ready rules (e.g., if soil pH is medium and bulk density is high, then harvest is timely) that extension agents can readily interpret and adjust on-site [27]. Built on a lightweight five-layer Takagi–Sugeno structure with hybrid least squares (backpropagation) training, ANFIS trains rapidly on low-cost hardware and integrates new predictors without full retraining [28]. The attributes such as adaptability, interpretability, efficiency, and solid theoretical grounding make ANFIS an ideal tool for yield projection, investment guidance, and policy calibration across the various Colombian agricultural landscapes.

1.1. Related Works

This subsection covers different related works regarding data-driven prediction and decision-making. Among the main characteristics identified are traditional statistical methods, machine learning (ML) applications, hybrid statistical ML approaches, adaptive neuro-fuzzy inference systems, and agro-sustainability applications. Figure 1 shows the relationship of the identified characteristics of the related works, which are addressed in more detail below.
Technological development has made it possible to collect and store abundant data, from which information is gathered to suggest solutions and make decisions. Classification and prediction are two common actions that can be carried out with data [29,30,31]. In this order, traditional statistical methods have been used, such as the study by Parra et al. [19], which identifies cropping systems with greater flexibility in response to alterations in both climatic and non-climatic elements, or the work of Kim et al., which predicts the growth of garlic and onion bulbs [32], or Bai et al., who evaluate the impact of the foreign trade survey on the accuracy of foreign trade forecasting [33].
With the wide progress of artificial intelligence and the rise of machine learning, it is recommended to combine traditional statistical techniques with new tools that imitate human behavior and learn from the data they consume [34,35], from a metric that quantifies the error between the prediction obtained and the ideal response and feedback strategies that use the error to improve behavior in future events [36]. This approach also has the ability to analyze a large amount of data, and identify patterns, and it also has greater precision and captures the nonlinearity and sporadicity [37,38,39]. For example, González et al. use both statistical models (the linear regression model and Bayesian additive regression model) and machine learning approaches (neural networks, support vector machines, random forests, and the presence–absence approach) to predict possible coffee production scenarios [40].
Machine learning has demonstrated suitable predictive performance in different fields such as energy [41,42], environment [43,44], sustainability [45,46,47], medicine [48,49], economics [50,51], and agriculture [52]. These predictive models protrude for their capacity to produce inputs for well-informed policy and decision-making by analyzing, interpreting, and relating factors [50,53,54,55].
Using digital technology for sustainability is essential to find a balance between technological developments and sustainable practices, such as agroecology [56]. According to [57], rural sustainability, economic growth, and global food security depend on agricultural production. Numerous studies are under this approach, such as Xiong et al. [58], whose research offers various perspectives on food security and possible remedies for international organizations and governments, using machine learning methods such as random forest, K-nearest neighbors, artificial neural networks, and support vector machines. With a different approach, Florez et al. [59] developed a mobile application for plant recognition, using a family of convolutional neural networks to promote sustainable tourism.
The accurate and timely prediction of agricultural production behavior contributes to crop and harvest management, policy formulation, strategies, and decision-making in food security and sustainable development [60]. The neural network model can learn patterns of complex systems, an advantage for making predictions [61,62], as in the case of Kittichotsatsawat et al. [63], Abrougui et al. [64], Almady et al. [65], Thimmegowda et al. [66] and Satpath et al. [67], and Abraham [68], who study the production of coffee, citrus plants, potato, rice, and soybean, respectively. In addition to predicting crop yield or productivity, this machine learning method allows predicting soil quality with water deficit [69] or with specific nutrients [70], predicting outbreaks of fruit diseases [71], and even the composition of an agricultural product using nitrogen fertilization [72].
Neural networks can improve prediction accuracy using historical data and current variables. In this regard, the structure of neural networks has been modified with feedback to process data sequences [73], or to automate the learning of complex data relationships [74], or with convolutional layers to process data with special structures [75]. Mainly, the neural network design depends on the data to be worked with and the objective to obtain. The learning capacity of artificial neural networks can be leveraged by combining it with the reasoning capacity of fuzzy logic, thus providing improved forecasting capabilities compared to traditional models [76]. This combination is called Adaptive Neuro-Fuzzy Inference System (ANFIS).
The model resulting from the combination of fuzzy logic and neural networks has been used for predictive approach in the agricultural sector. For example, Dutta et al. faced a lack of sufficient and accurate collective data to forecast agricultural product inflation and capture the virtue of fuzzy systems and neural networks to rely on time-series analysis with restricted data [77]. Combined with optimization methods, Abd Elfattah uses an ANFIS system to predict coal prices [78]. Factors that influence agricultural productivity, such as soil quality [79], oil spills on land [80], and floods [81], are also predicted with neuro-fuzzy systems. Garg presents a time-series model (based on fuzzy logic) to predict rice crop yield, which facilitates government entities in proper crop resource planning and management [82]. On the other hand, Gunasundari in [83] employs a bee swarm intelligence-inspired algorithm to classify and predict crop yield. As can be seen in these investigations using neuro-fuzzy systems, the interpretability of the prediction is favored, i.e., the results can be conceptually associated with human language. However, it is necessary to explore the advantages that models built from neuro-fuzzy systems can provide in their predictive, and interpretative capacity.

Forecasting Approach

Forecasting consists of determining weather behavior in the future relevant for prediction models. Existing approaches that combine environmental forecasting with fuzzy logic can inspire future extensions of the proposed model. For example, in [84], regarding a pump storage operation, a proposal using energy price as a key input to fuzzy waste water level control is shown. The authors propose a pump control strategy for storage tanks considering energy cost fluctuations. This work demonstrates the importance of energy demand forecasting in balancing supply and demand as it considers a system that incorporates fluctuations in energy prices, liquid level, inflow rate, and storm forecasts as input variables. According to [85], since weather directly impacts human life and is significant in sectors such as agriculture, transportation, energy, and natural disaster management, it is a relevant scientific, social, and economic concern. The authors in [85] analyzed the use of fuzzy logic for weather condition analysis, focused on handling imprecise meteorological data through fuzzy sets. In [86], a combination of fuzzy logic and other multi-criteria decision-making techniques to identify and evaluate the primary climate transition hazards are employed. This method enables the handling of the inherent imprecision of these hazards and the improvement of their description by using language phrases. This analysis emphasizes how crucial it is to have a strong framework to predict and alleviate the effects of climate change. Finally, an approach to evaluating the financial risk to crops in areas with extremely variable weather is presented in [87]. In a low-income economy, authors employ data-driven strategies that utilize Singular Spectrum Analysis (SSA) and Principal Component Analysis (PCA). The risk measure was initially developed for Colombian coffee plantations productive from 2010 to 2019. The research is focused on comprehending how index insurance technologies will affect the strategic economic crops of the nation in the future.

1.2. Approach and Paper Organization

The link among knowledge, technological tools, and the advantageous conditions that Colombia has as a biodiverse country is essential for the creation and implementation of sustainable and comprehensive agricultural systems. In this sense, and in order to generate input for decision-making in the agricultural sector, a neuro-fuzzy model is proposed to predict the behavior of production as a pivotal element.
This paper presents the design and deployment (training–validation) of a neuro-fuzzy model for the production of relevant agricultural products in Colombia. The performance and features are observed regarding the applicability of the fuzzy models (interpretability level).
The document is organized as follows: Section 2 describes the employed techniques and Section 3 the methodology utilized; Section 4 displays the dataset utilized. Meanwhile, Section 5 describes the model design and the implementation process. Then, Section 6 contains the comparison results; Section 7 displays an alternative for selecting training and testing data, and, in Section 8, the interpretability of neuro-fuzzy systems obtained is addressed. Finally, in Section 9 and Section 10, the discussion and conclusions are given.

2. Neuro-Fuzzy Systems and Clustering

The Adaptive Network Fuzzy Inference System (ANFIS) merges the tuning capacity of neural networks and the qualitative approach provided by fuzzy logic, resulting in a technique for learning membership function parameters [88].
ANFIS is a multilayer feedforward network where every node serves a specific purpose both in the input signals and in the set of parameters that belong to that node. The operation can change from neuron to neuron depending on the function (input–output) required by the network. In Figure 2, the connections (arrows) in the network indicate the direction flow of signals [28]. The architecture of an ANFIS network consists of the following layers:
  • Layer 1: The function parameters are fitted in every node in this layer. The membership degree value provided by the membership functions’ input is each node’s output. At this stage, A i and B i correspond to fuzzy sets.
  • Layer 2: Each node located this layer is not adaptive-type. Outputs are the result of multiplying the signals entering the respective node.
  • Layer 3: In this layer, each node is fixed (not adaptive-type). The normalized firing strength of the i-th rule in each node is normalized.
  • Layer 4: All nodes in this layer are matched to an output-defined function. These nodes have a function defined as O 4 , i = w ¯ i f i , where f i is the output function of the respective rule i.
  • Layer 5: In this layer, the node is fixed and generates the system output by summing all signals coming from the preceding nodes [89].

Fuzzy C-Means

Fuzzy C-Means is an unsupervised clustering algorithm that treats a data item as belonging to multiple clusters based on membership. This algorithm allows the clustering of data in a multidimensional space. By using fuzzy partitioning, data items can be assigned to more than one cluster according to the membership degree matrix, which contains values ranging from 0 to 1. At each iteration, the cluster centers and membership values are updated, and the algorithm finds the cluster center that minimizes the dissimilarity function.
For each datum in each cluster is given a membership value according to how distant it is from the cluster center. The datum with the highest membership value within a cluster is located near that cluster center. All of the data items’ membership values within a cluster must add up to one.
One advantage of the algorithm is that each data item can have a membership value to more than two clusters; therefore, tolerance measures can be established based on the clustering accuracy required for each problem.
To perform the classification using the Fuzzy C-Mean algorithm, the object function J given in Equation (1) is minimized at each iteration.
J = i = 1 N j = 1 C ( μ i j ) m x i c j 2
where C is the quantity of clusters, N is the total data, μ i j is the membership degree of data item x i in cluster j, the parameter m is the fuzzification coefficient, and c j is the center vector for cluster j. The metric x i c j computes the proximity of the datum to the center of the cluster. The algorithm modifies the center of the vector for every cluster at each iteration. The membership degree for each datum is calculated as follows:
μ i j = 1 k = 1 C x i c j x i c k 2 m 1
The fuzzification coefficient m allows us to establish the clustering tolerance where 1 < m < ; therefore, m controls the level of overlapping between clusters [90]. Finally, the center of the respective cluster c j is calculated as follows:
c j = i = 1 N ( μ i j ) m x i i = 1 N ( μ i j ) m

3. Methodology

The research carried out has a quantitative approach since it uses data from the agricultural sector and employs metrics to determine the capacity of the developed predictive models and their interpretability. It also has an experimental part in which different configurations for neural networks and neuro-fuzzy systems are evaluated. Figure 3 presents a summary of the phases carried out.
  • Data collection: Data collected by the National Administrative Department of Statistics in the 2019 National Agricultural Survey were used. These data provide information on the agricultural production of various products in the five natural Colombian regions from 2012 to 2019.
    To carry out the training process, all input and output variables are normalized to have values between 0 and 1, and then for the simulation, they are scaled to their real values. Data imputation is also performed.
  • Model development: The model is implemented using neuro-fuzzy systems. The output corresponds to the production of the respective agricultural product in tons. The neuro-fuzzy models are developed using 80 % of the data for training. Different input–output configurations are proposed.
  • Validation and evaluation: In this stage, 20 % of the data are used for validation. Tables are built to show the mean squared error (MSE) results for training and validation data, where the minimum, maximum, average, and standard deviation (STD) values are presented. To determine the best configuration of the neuro-fuzzy system, an experimental design is performed considering different configurations.
  • Analysis of results: The best model can be determined by considering the best values obtained from the MSE in validation. Additionally, to obtain better interpretability, the Sugeno-type system obtained is converted to Mamdani, with linear functions and output constants. According to the output membership functions resulting from the systems, their capacity to fit the data and interpret it is analyzed.

4. Dataset

The dataset employed is obtained from DANE (https://www.dane.gov.co/index.php/estadisticas-por-tema/agropecuario/encuesta-nacional-agropecuaria-ena (accessed on 5 August 2024)), which provides a dataset with 215 usable entries acquired from 2012 to 2019 and records the production of traditional agricultural products (in tons) as displayed in Figure 4.
The model is implemented using neuro-fuzzy systems. The output corresponds to the production of the respective agricultural product in tons. The considered inputs of the system are as follows:
  • Type of product: banana, cocoa, coffee, sugar cane, orange, and plantain.
  • Natural region of Colombia: Andean, Caribbean, Pacific, Orinoco, and Amazon.
  • Year of production (optional): from 2012 to 2019.
  • Planted area: in hectares.
  • Productive area: in hectares.
  • Previous production value (optional): in tons.
For model training and validation, the data for inputs and outputs are normalized in the range [ 0 , 1 ] , which avoids disproportionate values and shapes of the fuzzy sets used in the model.

5. Models Description and Implementation

For the model development, different configurations of the previous variables can be used; therefore, different alternatives are proposed. Figure 5 displays the models considered according to the input–output configurations.
Figure 5a shows the configuration of model M 1 , where a total of 14 inputs are employed. The description of the respective inputs and outputs of the model is as follows:
  • X 1 , X 6 : entries associated with each product encoded in binary.
  • W 1 , W 5 : entries associated with each region encoded in binary.
  • U: input associated with the year in which the production measurement is made.
  • Z 1 : planted area (input).
  • Z 2 : productive area (input).
  • Y: output corresponding to production.
The schematic representation of model M 2 is observed in Figure 5b; similar to the M 1 model, 14 inputs are used; however, instead of using U (the year of production), T is used corresponding to the previous output value. The description of the input variables and the output of this model is as follows:
  • X 1 , X 6 : entries associated with each product encoded in binary.
  • W 1 , W 5 : entries associated with each region encoded in binary.
  • T: the previous value of the output.
  • Z 1 : planted area (input).
  • Z 2 : productive area (input).
  • Y: output corresponding to production.
Figure 5c displays the configuration of model M 3 , which has a smaller number of inputs, five in total. In this way, the aim is to have a more compact model. The description of the respective inputs and outputs of the model is as follows:
  • X: input associated with each product coded according to its production level (from lowest to highest).
  • W: input associated with each region coded according to its production level (from lowest to highest).
  • U: input associated with the year in which the production measurement is made.
  • Z 1 : planted area (input).
  • Z 2 : productive area (input).
  • Y: output corresponding to production.
The representation of model M 4 is presented in Figure 5d; in the same way, to obtain a compact model, the M 4 model employs five inputs. In contrast to the M 3 model that utilizes U (year of production), model M 4 employs T, which is the previous value of the output. The description of the input variables and the output of this model is as follows:
  • X: input associated with each product coded according to its production level (from lowest to highest).
  • W: entries associated with each region coded according to its production level (from lowest to highest).
  • T: previous value of the output.
  • Z 1 : planted area (input).
  • Z 2 : productive area (input).
  • Y: output corresponding to production.
The ordinary codification for product and geographic region is given considering the amount of production (tons); it first appears the product and regions with low production and in the end products and regions with the maximum production. The binary codification for products is displayed in Table 1; meanwhile, Table 2 shows the binary codification for geographic region.

5.1. Implementation Process

This section shows the results of implementing neuro-fuzzy systems for each model considered. For neuro-fuzzy models, an initial assignment based on Fuzzy C-Means (FCM) is used, which performs a random assignment of the clusters’ initial centers. Considering the random initialization of the model parameters, each configuration is trained 20 times.
For implementation, 80% of the data are used in training process and 20% for validation. In this way, the respective tables show the MSE results for training and validation data, where the minimum, maximum, average, and standard deviation values are presented.
To carry out the training process, all input–output variables are normalized to have values between 0 and 1, and then they are scaled to their real values for the simulation.
Linear and constant output functions are employed for neuro-fuzzy systems with the Sugeno model. For these systems, the configurations are as follows:
  • Output membership functions: linear and constant.
  • Number of clusters: 2, 3, 4, and 5.
  • Fuzzy partition exponent: 1.1, 2, 3, and 4.

5.1.1. Implementation for Model M 1

Regarding the training stage, the statistical values as minimum, maximum, mean, and STD obtained for each configuration is displayed in Table 3. Meanwhile, Table 4 encloses the values from using validation data. For the M 1 model, the best values obtained are as follows:
  • Linear:
    MSE training: 0.3625 × 10 4 .
    MSE validation: 2.6265 × 10 4 .
  • Constant:
    MSE training: 3.7288 × 10 4 .
    MSE validation: 5.3098 × 10 4 .

5.1.2. Implementation for Model M 2

For the M 2 configuration, the MSE training results can be seen in Table 5 where the statistical summary (min, max, mean, and STD) is detailed. Table 6 was obtained by employing validation data, the best values obtained are as follows:
  • Linear:
    MSE training: 0.1469 × 10 4 .
    MSE validation: 2.7913 × 10 4 .
  • Constant:
    MSE training: 3.1731 × 10 4 .
    MSE validation: 5.4551 × 10 4 .

5.1.3. Implementation for Model M 3

Regarding the training stage, Table 7 displays the max, min, mean, and STD metrics obtained for the considered configurations. Meanwhile, the results in Table 8 were gathered through validation data. The best values obtained for model M 3 are as follows:
  • Linear:
    MSE training: 2.6350 × 10 4 .
    MSE validation: 6.7322 × 10 4 .
  • Constant:
    MSE training: 17.0000 × 10 4 .
    MSE validation: 40.0000 × 10 4 .

5.1.4. Implementation for Model M 4

Considering the implementation for model M 4 , Table 9 contains the results (max, min, average, and STD) after the training process for each configuration of the fuzzy system. After the training procedure, Table 10 was obtained through validation data. The best obtained values are as follows:
  • Linear:
    MSE training: 3.9517 × 10 4 .
    MSE validation: 6.6033 × 10 4 .
  • Constant:
    MSE training: 16.0000 × 10 4 .
    MSE validation: 13.0000 × 10 4 .

6. Comparison Results

Regarding the four proposed models, the results comparison is shown in this section. The simulation outcomes for the best configuration found for each model utilizing training and validation data are shown in Figure 6. In general, a superior fit is observed with fuzzy systems using linear functions in the output.
Table 11 shows the MSE value obtained for each model. Regarding the results with validation data, the best model with linear and constant neuro-fuzzy systems is M 1 .

7. Alternative for Training and Testing Data Selection

Some limitations identified for model training and validation are the small amount of data and unavailable data segments, which is why random sampling was used to select the training and test data. However, for a more forecasting approach, data from previous years should be used to predict the next. For this reason, an experimental test is carried out to demonstrate this approach, where data from the 2012–2017 period are used for training and data from 2018–2019 for testing. Given the limitations, this test is carried out with the best configuration found in the previous section (model M 1 ). For subsequent work where a complete set of historical data can be obtained, the models can be adjusted using this approach.
Using this selection for training and validation data, the statistical values as minimum, maximum, mean, and STD obtained for each configuration are displayed in Table 12. Meanwhile, Table 13 contains the values obtained using validation data. In this way, for the M 1 model the best values obtained are as follows:
  • Linear:
    MSE training: 0.05714 × 10 4 .
    MSE validation: 21.0000 × 10 4 .
  • Constant:
    MSE training: 2.0679 × 10 4 .
    MSE validation: 15.0000 × 10 4 .
Figure 7 displays the simulation results taking the best system obtained in the validation process. This figure shows the spaces associated with the lack of available data for training, which lowers the performance of the system to make predictions.
Table 14 shows the MSE value obtained for model M 1 using random sampling data and segmented data (regarding years). As can be seen in this table, the strategy of segmenting data for training and validation according the period (2012–2017 for training and 2018–2019 for testing) presents a better MSE in the training process, but, in validation, the MSE worsens, which is why it is used random sampling for training and validation due to the low amount of data and incomplete data available in this application.

8. Interpretability of Neuro-Fuzzy Models

In order to obtain the best interpretability of neuro-fuzzy systems, the Sugeno-type system is converted into Mamdani-type using the Fuzzy Logic Designer tool of MATLAB (R2023a) [91]. When the Sugeno system is transformed into a Mamdani system, the obtained Mamdani system is composed in the output with triangular membership functions where the center is located on the value of the constant term of the output functions of the Sugeno system. The width of the triangular fuzzy set depends on the coefficients of the linear functions of the Sugeno system. By performing the conversion, the following Mamdani-type fuzzy systems are obtained:
  • FIS-L: system attained by converting the Sugeno system with linear functions at the output.
  • FIS-C: system determined from the conversion of the Sugeno system with constant functions at the output.
Considering the system achieved from the conversion of the Sugeno system with linear functions at the output, Figure 8a shows the structure of the Mamdani-type fuzzy system; Figure 8b presents the output fuzzy sets, and, finally, Figure 9 contains the set of rules of the Mamdani-type fuzzy system FIS-L.
On the other hand, having the system obtained from the conversion of the Sugeno system with constant functions at the output (FIS-C), Figure 10a shows the structure of the Mamdani-type fuzzy system. Additionally, Figure 10b contains the output fuzzy sets, and, finally, Figure 11 presents the set of rules of the Mamdani-type fuzzy system FIS-C. As seen, the Mamdani FIS-C fuzzy system allows for a better interpretation of the rules given the shape of the fuzzy sets used at the output. For example, Figure 10b shows that, although the fuzzy sets cover the entire output range, greater activation at lower output values are present.
Regarding the interpretability of the fuzzy system (for input variables), as an example, the case shown in Figure 9 and Figure 11 can be considered where it is observed that the inputs W 1 to W 4 have a greater influence on the inference and, therefore, on the output calculation. This depends on the approach of the rules. For example, it could be taken to the interpretation of the form: if the region is Andean and the crop is banana, the production will be high. The choice of model depends on the the preference of the user, and the accuracy of the prediction provides the necessary information to observe the model performance. On the other hand, the interpretative capacity of the models can make these technological tools closer to the users in the sense of handling a linguistic structure similar to the normal language. In contrast, the variable W 5 has little influence on the activation of the rules; this variable represents the Amazon region. On the other hand, the membership function for variable U in all the rules exhibits significant activation.
In addition, most of the membership functions observed in Figure 9 and Figure 11 cover the entire input range, indicating that each input has a continuous impact on rule activation. In the case of FIS-C, rules 2 and 4, having a wider output range, may have a greater influence on the overall inference.
Table 15 presents the MSE results for the Mamdani fuzzy systems. These results show that the best performance of the MSE is achieved with the FIS-L fuzzy system.
Figure 12a shows the simulation outcomes using training and validation data for the FIS-L system, while in Figure 12b, they are presented for the Mamdani FIS-C system. As can be seen, FIS-L presents an exceeding data fit; however, the FIS-C system allows better interpretability of the rules, where concepts such as high, medium-low, very low, and close to zero output can be associated (which can be used in an authentic context).
In summary of all results obtained in testing four versions of the neuro-fuzzy predictor ( M 1 M 4 ), it is found that the simplest input layout ( M 1 ) gives the most reliable results and that using linear output functions instead of fixed value outputs slashes forecast errors. Adding more clusters (mini groups of similar field conditions) can nudge accuracy even higher; however, each extra cluster also adds more “if this then that” rules and makes the system harder to interpret. For the model selection, the sweet spot is M 1 with a moderate number of clusters allowing sharp yield predictions from the fuzzy model (Sugeno) while still keeping a manageable rule set and can be transformed in the Mamdani version to obtain clear, plain language guidelines for agronomists and farmers.

9. Discussion

As seen, the fuzzy system with linear functions at the output allows for greater adaptability, which is why a better MSE value is obtained with this system. However, when greater interpretability is sought, the fuzzy system with constant functions at the output allows the association of output sets with linguistic labels according to their location.
Proposing several models and different configurations for each model made, it is possible to choose the best predictive performance. Ordering the data in ascending order proved not to provide an advantage for the accuracy of the prediction. The choice between fuzzy systems with linear and constant output functions depends on the trade-off between adaptability and interpretability. The FIS-L system achieves better adaptability and a lower MSE but at the cost of reduced interpretability due to less-defined fuzzy sets. In contrast, the FIS-C system enhances interpretability by associating output sets with linguistic labels and providing more structured membership functions. Therefore, selecting the appropriate model should consider factors such as interpretability, computational efficiency, and ease of implementation, ensuring an optimal balance between adaptability and clarity.
Neuro-fuzzy systems offer both accurate forecasts and human-readable rules for Colombian agriculture by balancing Sugeno-type models’ precision with Mamdani-type models’ interpretability. While the Sugeno approach minimizes error through adaptive linear functions, Mamdani translates those outputs into straightforward “if this then that” rules. Choosing the right model involves trading off raw accuracy for interpretability and considering hardware constraints, user expertise, and ease of updating rule sets to ensure both reliable predictions and actionable decision support on the farm.
It is also observed that the FIS-L system presents a smaller set of rules compared to the FIS-C system, which would allow its generalization; however, the fuzzy sets obtained in the FIS-L output make interpretability more complex. The membership functions of the FIS-C system, observed in Figure 10b, are more well defined than those displayed in Figure 8b, which may indicate a more structured and possibly more detailed distribution of the output data, allowing for better interpretation. For choosing the suitable model, aspects including interpretability, computational requirements, and ease of implementation might be taken into account. A balance must be sought between system adaptability (better MSE) and interpretability (rules and sets).
Regarding the reduced amount of data (215 records) and the use of 80% of the data for training models and validating them with 20% the risk of overfitting is possible; however, evaluating and choosing the models considering the interpretability feature allows one to avoid this risk since the labels of the fuzzy sets obtained must have an associated output concept. In this way, a model with better interpretability is chosen (usable) instead of a model with better MSE. This case is displayed in Section 8 where the FIS-L model has a better MSE value; however, the model FIS-C is useful from the interpretability feature. Comparison with other traditional prediction methods is an aspect to consider in order to determine the most suitable model usable in a real context; for this, it is necessary to set guidelines to obtain a fair comparison. In this way, aspects such as different performance metrics, complexity, interpretability, ease of use, and scalability must be regarded for comparison. For these models and techniques, an evaluation can be performed in a future work since this paper is oriented to display a proposal of codifications to implement the model for the production of relevant agricultural products in Colombia.
In order to obtain a more accurate forecasting model, climate information is necessary to be included for each agricultural product; however, these specialized data are not available since Colombia has wide variety of thermal zones making it difficult to obtain a uniform and usable dataset. The proposal displayed in this paper addresses the problem in a general way so that when specialized climate data become available, they can be used for the training models.
In addition, the use of a single model to predict the production of different crops may be more effective by training separate models for each product (e.g., cocoa and sugar cane). Nevertheless, this work wanted to test a coding approach that considered several products to identify what level of performance the model can achieve and thus be able to improve it for future developments.
Data are a relevant issue to train the models; in this work, different aspects were addressed. The first is the low amount of data (215 records), and the second is the missing data imputed (in three cases). In addition, data for region and product are not always available. It should be also noted that the collection of this type of data depends on the government institution in charge, which is subject to change depending on the policies of the current government (affecting data quality). Even with the data limitation, the models were trained and evaluated obtaining suitable results, especially in terms of interpretability.

10. Conclusions

This research can be regarded as an investigation of the connections among the attributes specified for agricultural production in Colombia and the strengths and weaknesses of each model, as well as the correlation between the input and output variables.
The suggested models could be clearly qualified thanks to the methodology and metrics taken into consideration. Different encodings were proposed for the input data, these being models M 1 , M 2 , M 3 , and M 4 . Considering the results, it was determined that model M 1 is the most convenient encoding for the fuzzy system. Comparing the results obtained across the four models, it is observed that arranging the inputs in ascending order has no contribution to improving prediction accuracy.
When analyzing the MSE values over 20 training processes (runs) with different configurations, a trend is observed in which the MSE decreases as the number of clusters increases. However, when more clusters are employed, the number of rules increases, which also increases the complexity of the fuzzy system and decreases its interpretability.
It was observed that the FIS-L system is more adaptable and provides a better MSE value; however, when converting the Sugeno system to a Mamdani, the FIS-C system achieves better interpretability. With the FIS-C system, linguistic labels can be associated with membership functions, considering their location in the output discourse universe. This feature allows for an exceeding interpretability rule.
Future works may consider developing a software tool to establish rules that allow human decision-making. Additionally, the method of building the fuzzy logic system can be investigated, considering the preliminary knowledge of a group of experts.

Author Contributions

Conceptualization, A.C.G., L.A.B., and H.E.E.; Methodology, A.C.G., L.A.B., and H.E.E.; Project Administration, A.C.G., L.A.B., and H.E.E.; Supervision, H.E.E.; Validation, A.C.G.; Writing—Original Draft, A.C.G., L.A.B., and H.E.E.; Writing—Review and Editing, A.C.G., L.A.B., and H.E.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

All data are included in the article.

Acknowledgments

The authors express gratitude to the Universidad Distrital Francisco José de Caldas.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Uddin, I.; Azam Khan, M. Global Evidence on the Impact of Globalization, Governance, and Financial Development on Economic Growth. J. Knowl. Econ. 2024, 15, 14546–14577. [Google Scholar] [CrossRef]
  2. Vlahinic Lenz, N.; Fajdetic, B. Does Economic Globalisation Harm Climate? New Evidence from European Union. Energies 2022, 15, 6699. [Google Scholar] [CrossRef]
  3. Matenda1, F.R.; Raihan, A.; Zhou, H.; Sibanda, M. The infuence of economic growth, fossil and renewable energy, technological innovation, and globalisation on carbon dioxide emissions in South Africa. Carb. Res. 2024, 3, 69. [Google Scholar] [CrossRef]
  4. Ahmed, Z.; Zhang, B.; Cary, M. Linking economic globalization, economic growth, financial development, and ecological footprint: Evidence from symmetric and asymmetric ARDL. Ecol. Indic. 2021, 121, 107060. [Google Scholar] [CrossRef]
  5. Awosusi, A.A.; Ozdeser, H.; Ojekemi, O.S.; Adeshola, I.; Ramzan, M. Environmental sustainability in Vietnam: Evaluating the criticality of economic globalisation, renewable energy, and natural resources. Environ. Sci. Pollut. Res. 2023, 30, 75581–75594. [Google Scholar] [CrossRef]
  6. Jameel, R.; Naqve, M.; Zia, M.A.; Mahmood, A.; Javaid, M.M.; Nadeem, M.A. Crop Management for Sustainable Wheat Production; Springer International Publishing: Cham, Switzerland, 2023; pp. 107–123. [Google Scholar] [CrossRef]
  7. Kamble, S.S.; Gunasekaran, A.; Gawankar, S.A. Achieving sustainable performance in a data-driven agriculture supply chain: A review for research and applications. Int. J. Prod. Econ. 2020, 219, 179–194. [Google Scholar] [CrossRef]
  8. Pham, Y.; Reardon-Smith, K.; Mushtaq, S. The impact of climate change and variability on coffee production: A systematic review. Clim. Chang. 2019, 156, 609–630. [Google Scholar] [CrossRef]
  9. Hussain, J.; Khaliq, T.; Ahmad, A. Wheat Responses to Climate Change and Its Adaptations: A Focus on Arid and Semi-arid Environment. Int. J. Environ. Res. 2018, 12, 117–126. [Google Scholar] [CrossRef]
  10. Hussain, J.; Hussain, S.; Tahir, N.; Rasool, I.; Ullah, A.; Ahmad, S. Climate Change and Rice Production: Impacts and Adaptations. Modern Techniques of Rice Crop Production; Springer: Singapore, 2022. [Google Scholar] [CrossRef]
  11. Lahive, F.; Hadley, P.; Daymond, A. The physiological responses of cacao to the environment and the implications for climate change resilience. A review. Agron. Sustain. Dev. 2019, 39, 5. [Google Scholar] [CrossRef]
  12. Simon, A.; Moraru, P.; Ceclan, A.; Russu, F.; Cheţan, F.; Bardaş, M.; Popa, A.; Rusu, T.; Pop, A.I.; Bogdan, I. The Impact of Climatic Factors on the Development Stages of Maize Crop in the Transylvanian Plain. Agronomy 2023, 13, 1612. [Google Scholar] [CrossRef]
  13. Grüter, R.; Trachsel, T.; Laube, P.; Jaisli, I. Expected global suitability of coffee, cashew and avocado due to climate change. PLoS ONE 2022, 17, e0261976. [Google Scholar] [CrossRef] [PubMed]
  14. Anderson, R.; Bayer, P.E.; Edwards, D. Climate change and the need for agricultural adaptation. Curr. Opin. Plant Biol. 2020, 56, 197–202. [Google Scholar] [CrossRef] [PubMed]
  15. Azam-Ali, S.N.; Gregory, P.J.; Jahanshiri, E. Diversifying the UK Agrifood System: A Role for Neglected and Underutilised Crops. Agronomy 2024, 14, 853. [Google Scholar] [CrossRef]
  16. Reina, J.V.; Garzón, J.C.; Schulz, K.; Rüttinger, L.; Mosello, B.; Ivleva, D.; Buderath, M.; Gorricho, J. Un Clima Peligroso: Deforestación, Cambio Climático y Violencia Contra los Defensores Ambientales en la Amazonía Colombiana; Technical Report; WWF Colombia, WWF Alemania, Tropenbos, FCDS— Fundación Conservación y Desarrollo Sostenible, FZS—Frankfurt Zoological Society, GAIA Amazonas, ACT (Amazon Conservation Team) Colombia: Berlin, Germany, 2021. [Google Scholar]
  17. Instituto de Hidrología. Informe del Estado del Ambiente y los Recursos Naturales Renovables, 2022; Technical Report; IDEAM: Bogotá, Colombia, 2024.
  18. Martín-López, J.M.; Verchot, L.V.; Martius, C.; da Silva, M. Modeling the Spatial Distribution of Soil Organic Carbon and Carbon Stocks in the Casanare Flooded Savannas of the Colombian Llanos. Wetlands 2023, 43, 65. [Google Scholar] [CrossRef]
  19. Parra-Londono, S.; Tigreros, J.A.; Montoya-Correa, C.A. Colombian Crop Resilience: Evaluating National Yield Stability for Fruit and Vegetable Systems. Agriculture 2024, 14, 1546. [Google Scholar] [CrossRef]
  20. Anacona Mopan, Y.; Solis Pino, A.; Rubiano-Ovalle, O.; Paz, H.; Ramirez Mejia, I. Spatial Analysis of the Suitability of Hass Avocado Cultivation in the Cauca Department, Colombia, Using Multi-Criteria Decision Analysis and Geographic Information Systems. ISPRS Int. J. Geo. Inf. 2023, 12, 136. [Google Scholar] [CrossRef]
  21. Cortés-Cataño, C.F.; Foronda-Tobón, Y.; Paez-Ricardo, J.A.; Parra-Herrera, J.E.; Cañon Ayala, M.J. The effect of environmental variations on the production of the principal agricultural products in Colombia. PLoS ONE 2024, 19, e0304035. [Google Scholar] [CrossRef]
  22. de Estadística, D.A.N. Encuesta Nacional Agropecuaria; Technical Report; DANE: Bogotá, Colombia, 2019.
  23. The Observatory of Economic Complexity. Coffee in Colombia; Technical Report; OEC: Taipei, Taiwan, 2022. [Google Scholar]
  24. Organización de las Naciones Unidas Para la Alimentación y Agricultura. Una Introducción a los Conceptos Básicos de la Seguridad Alimentaria; Technical Report; Programa CE-FAO: Rome, Italy, 2011. [Google Scholar]
  25. Iwegbu, O.; de Mattos, L.B. Financial development, trade globalisation and agricultural output performance among BRICS and WAMZ member countries. SN Bus. Econ. 2022, 2, 89. [Google Scholar] [CrossRef]
  26. Sánchez, J.; Rodríguez, J.; Espitia, H. Design of a neuro-fuzzy model for agricultural employment in Colombia using fuzzy clustering. AIMS Environ. Sci. 2024, 11, 759–775. [Google Scholar] [CrossRef]
  27. Gutiérrez-Magaã, S.M.; García-Díaz, N.; Soriano-Equigua, L.; Mata-López, W.A.; García-Virgen, J.; Brizuela-Ramírez, J.E. Neuro-Fuzzy System to Predict Timely Harvest in Stevia Crops. Agriculture 2025, 15, 840. [Google Scholar] [CrossRef]
  28. Jang, J.-S.R. ANFIS: Adaptive-network-based fuzzy inference system. IEEE Trans. Syst. Man Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
  29. Mondal, S.; Mondal, S.; Adhikari, S. Applications of Big Data in Various Fields: A Survey. Recent Trends Intell. Enabled Res. 2023, 1446, 221–233. [Google Scholar] [CrossRef]
  30. Raisinghani, M.; Idemudia, E.; Wang, F. From Big Data to Big Insights: A Synthesis of Real-World Applications of Big Data Analytics. Development Methodologies for Big Data Analytics Systems. Transactions on Computational Science and Computational Intelligence; Springer: Cham, Switzerland, 2024. [Google Scholar] [CrossRef]
  31. Badshah, A.; Daud, A.; Alharbey, R.; Banjar, A.; Bukhari, A.; Alshemaimri, B. Big data applications: Overview, challenges and future. Recent Trends Intell. Enabled Res. 2024, 57, 290. [Google Scholar] [CrossRef]
  32. Kim, D.; Cho, W.; Na, I.; Na, M.H. Prediction of Live Bulb Weight for Field Vegetables Using Functional Regression Models and Machine Learning Methods. Agriculture 2024, 14, 754. [Google Scholar] [CrossRef]
  33. Bai, Y.; Wang, S.; Zhang, X. Foreign Trade Survey Data: Do They Help in Forecasting Exports and Imports? J. Syst. Sci. Complex. 2022, 35, 1839–1862. [Google Scholar] [CrossRef]
  34. Athey, S.; Imbens, G.W. Machine Learning Methods That Economists Should Know About. Annu. Rev. Econ. 2019, 11, 685–725. [Google Scholar] [CrossRef]
  35. Çağlayan Akay, E.; Yılmaz Soydan, N.T.; Kocarık Gacar, B. Bibliometric analysis of the published literature on machine learning in economics and econometrics. Soc. Netw. Anal. Min. 2022, 12, 109. [Google Scholar] [CrossRef]
  36. Joshi, A.V. Machine Learning and Artificial Intelligence; Springer Nature Switzerland AG: Cham, Switzerland, 2023. [Google Scholar] [CrossRef]
  37. Zhu, H. Oil Demand Forecasting in Importing and Exporting Countries: AI-Based Analysis of Endogenous and Exogenous Factors. Sustainability 2023, 15, 13592. [Google Scholar] [CrossRef]
  38. Frison, L.; Gölzhäuser, S.; Bitterling, M.; Kramer, W. Evaluating different artificial neural network forecasting approaches for optimizing district heating network operation. Energy 2024, 307, 132745. [Google Scholar] [CrossRef]
  39. Das, P.K.; Das, P.K. Forecasting and Analyzing Predictors of Inflation Rate: Using Machine Learning Approach. J. Quant. Econ. 2024, 22, 439–517. [Google Scholar] [CrossRef]
  40. González-Orozco, C.E.; Porcel, M.; Byrareddy, V.M.; Rahn, E.; Cardona, W.A.; Salinas Velandia, D.A.; Araujo-Carrillo, G.A.; Kath, J. Preparing Colombian coffee production for climate change: Integrated spatial modelling to identify potential robusta coffee (Coffea canephora P.) growing areas. Clim. Chang. 2024, 177, 67. [Google Scholar] [CrossRef]
  41. Abdul Baseer, M.; Almunif, A.; Alsaduni, I.; Tazeen, N. Electrical Power Generation Forecasting from Renewable Energy Systems Using Artificial Intelligence Techniques. Energies 2023, 16, 6414. [Google Scholar] [CrossRef]
  42. Ayaz Atalan, Y.; Atalan, A. Testing the Wind Energy Data Based on Environmental Factors Predicted by Machine Learning with Analysis of Variance. Appl. Sci. 2025, 15, 241. [Google Scholar] [CrossRef]
  43. Chin, M.Y.; Qin, Y.; Hoy, Z.X.; Farooque, A.A.; Wong, K.Y.; Mong, G.R.; Tan, J.P.; Woon, K.S. Assessing carbon budgets and reduction pathways in different income levels with neural network forecasting. Energy 2024, 305, 132331. [Google Scholar] [CrossRef]
  44. Wang, D.; Cao, J.; Zhang, B.; Zhang, Y.; Xie, L. A Novel Flexible Geographically Weighted Neural Network for High-Precision PM2.5 Mapping across the Contiguous United States. ISPRS Int. J. Geo. Inf. 2024, 13, 217. [Google Scholar] [CrossRef]
  45. Sarwar, S.; Aziz, G.; Balsalobre-Lorente, D. Forecasting Accuracy of Traditional Regression, Machine Learning, and Deep Learning: A Study of Environmental Emissions in Saudi Arabia. Sustainability 2023, 15, 14957. [Google Scholar] [CrossRef]
  46. Kumar, A.; Singh, S.K.; Kumari, P. A Machine Learning Approach to Forecast the Food Prices for Food Security Issues. In Proceedings of the 11th International Conference on Intelligent Systems and Embedded Design (ISED), Dehradun, India, 15–17 December 2023; pp. 1–6. [Google Scholar] [CrossRef]
  47. Liu, J.; Li, M.; Li, R.; Shalamzari, M.J.; Ren, Y.; Silakhori, E. Comprehensive Assessment of Drought Susceptibility Using Predictive Modeling, Climate Change Projections, and Land Use Dynamics for Sustainable Management. Land 2025, 14, 337. [Google Scholar] [CrossRef]
  48. Alhussaini, A.J.; Steele, J.D.; Jawli, A.; Nabi, G. Radiomics Machine Learning Analysis of Clear Cell Renal Cell Carcinoma for Tumour Grade Prediction Based on Intra-Tumoural Sub-Region Heterogeneity. Cancers 2024, 16, 1454. [Google Scholar] [CrossRef]
  49. Barbieri, F.; Pfeifer, B.E.; Senoner, T.; Dobner, S.; Spitaler, P.; Semsroth, S.; Lambert, T.; Zweiker, D.; Neururer, S.B.; Scherr, D.; et al. A Neuronal Network-Based Score Predicting Survival in Patients Undergoing Aortic Valve Intervention: The ABC-AS Score. J. Clin. Med. 2024, 13, 3691. [Google Scholar] [CrossRef]
  50. Alzubaidi, L.; Bai, J.; Al-Sabaawi, A.; Santamaría, J.; Albahri, A.S.; Al-dabbagh, B.S.N.; Fadhel, M.A.; Manoufali, M.; Zhang, J.; Al-Timemy, A.H.; et al. A survey on deep learning tools dealing with data scarcity: Definitions, challenges, solutions, tips, and applications. J. Big Data 2023, 10, 46. [Google Scholar] [CrossRef]
  51. Felix, J.; Alexandre, M.; Lima, G.T. Applying Machine Learning Algorithms to Predict the Size of the Informal Economy. Comput. Econ. 2024, 65, 1169–1189. [Google Scholar] [CrossRef]
  52. El Sakka, M.; Ivanovici, M.; Chaari, L.; Mothe, J. A Review of CNN Applications in Smart Agriculture Using Multimodal Data. Sensors 2025, 25, 472. [Google Scholar] [CrossRef] [PubMed]
  53. Kayaci Çodur, M. Ensemble Machine Learning Approaches for Prediction of Türkiye’s Energy Demand. Energies 2024, 17, 74. [Google Scholar] [CrossRef]
  54. Panahi, F.; Ahmed, A.N.; Singh, V.P.; Ehtearm, M.; Elshafie, A.; Torabi Haghighi, A. Predicting freshwater production in seawater greenhouses using hybrid artificial neural network models. J. Clean. Prod. 2021, 329, 129721. [Google Scholar] [CrossRef]
  55. Das, P.; Jha, G.K.; Lama, A.; Parsad, R. Crop Yield Prediction Using Hybrid Machine Learning Approach: A Case Study of Lentil (Lens culinaris Medik.). Agriculture 2023, 13, 596. [Google Scholar] [CrossRef]
  56. Alshdaifat, S.M.; Aziz, N.H.A.; Alhasnawi, M.Y.; Alharasis, E.E.; Al Qadi, F.; Al Amosh, H. The Role of Digital Technologies in Corporate Sustainability: A Bibliometric Review and Future Research Agenda. J. Risk Finan. Manag. 2024, 17, 509. [Google Scholar] [CrossRef]
  57. Huang, W.; Wang, X. The Impact of Technological Innovations on Agricultural Productivity and Environmental Sustainability in China. Sustainability 2024, 16, 8480. [Google Scholar] [CrossRef]
  58. Xiong, R.; Peng, H.; Chen, X.; Shuai, C. Machine learning-enhanced evaluation of food security across 169 economies. Environ. Dev. Sustain. 2024, 26, 26971–27000. [Google Scholar] [CrossRef]
  59. Flórez, M.; Becerra, O.; Carrillo, E.; Villa, M.; Álvarez, Y.; Suárez, J.; Mendes, F. Deep Learning Application for Biodiversity Conservation and Educational Tourism in Natural Reserves. ISPRS Int. J. Geo. Inf. 2024, 13, 358. [Google Scholar] [CrossRef]
  60. Banerjee, R.; Bharti; Das, P.; Khan, S. Crop Yield Prediction Using Artificial Intelligence and Remote Sensing Methods. Artificial Intelligence and Smart Agriculture. Advances in Geographical and Environmental Sciences; Springer: Singapore, 2024. [Google Scholar] [CrossRef]
  61. Jabed, M.A.; Azmi Murad, M.A. Crop yield prediction in agriculture: A comprehensive review of machine learning and deep learning approaches, with insights for future research and sustainability. Heliyon 2024, 10, e40836. [Google Scholar] [CrossRef]
  62. Castillo-Girones, S.; Munera, S.; Martínez-Sober, M.; Blasco, J.; Cubero, S.; Gómez-Sanchis, J. Artificial Neural Networks in Agriculture, the core of artificial intelligence: What, When, and Why. Comput. Electron. Agric. 2025, 230, 109938. [Google Scholar] [CrossRef]
  63. Kittichotsatsawat, Y.; Tippayawong, N.; Tippayawong, K. Prediction of arabica coffee production using artificial neural network and multiple linear regression techniques. Soil Tillage Res. 2022, 12, 14488. [Google Scholar] [CrossRef] [PubMed]
  64. Abrougui, K.; Gabsi, K.; Mercatoris, B.; Khemis, C.; Amami, R.; Chehaibi, S. Prediction of organic potato yield using tillage systems and soil properties by artificial neural network (ANN) and multiple linear regressions (MLR). Soil Tillage Res. 2019, 190, 202–208. [Google Scholar] [CrossRef]
  65. Almady, S.S.; Abdel-Sattar, M.; Al-Sager, S.M.; Al-Hamed, S.A.; Aboukarima, A.M. Employing an Artificial Neural Network Model to Predict Citrus Yield Based on Climate Factors. Agronomy 2024, 14, 1548. [Google Scholar] [CrossRef]
  66. Thimmegowda, M.N.; Manjunatha, M.H.; Huggi, L.; Shivaramu, H.S.; Soumya, D.V.; Nagesha, L.; Padmashri, H.S. Weather-Based Statistical and Neural Network Tools for Forecasting Rice Yields in Major Growing Districts of Karnataka. Agronomy 2023, 13, 704. [Google Scholar] [CrossRef]
  67. Satpathi, A.; Setiya, P.; Das, B.; Nain, A.S.; Jha, P.K.; Singh, S.; Singh, S. Comparative Analysis of Statistical and Machine Learning Techniques for Rice Yield Forecasting for Chhattisgarh, India. Sustainability 2023, 15, 2786. [Google Scholar] [CrossRef]
  68. Abraham, E.R.; Mendes dos Reis, J.G.; Vendrametto, O.; Oliveira Costa Neto, P.L.D.; Carlo Toloi, R.; Souza, A.E.D.; Oliveira Morais, M.D. Time Series Prediction with Artificial Neural Networks: An Analysis Using Brazilian Soybean Production. Agriculture 2020, 10, 475. [Google Scholar] [CrossRef]
  69. Varga, I.; Radocaj, D.; Jurišic, M.; Markulj Kulundžic, A.; Antunovic, M. Prediction of sugar beet yield and quality parameters with varying nitrogen fertilization using ensemble decision trees and artificial neural networks. Comput. Electron. Agric. 2023, 212, 108076. [Google Scholar] [CrossRef]
  70. Dhanaraj, R.K.; Chandraprabha, M. Ant lion optimization in deep neural network for forecasting the rice crop yield based on soil nutrients. Prog. Artif. Intell. 2024, 14, 101–116. [Google Scholar] [CrossRef]
  71. Sannakki, S.; Rajpurohit, V.S.; Sumira, F.; Venkatesh, H. A neural network approach for disease forecasting in grapes using weather parameters. In Proceedings of the 2013 Fourth International Conference on Computing, Communications and Networking Technologies (ICCCNT), Tiruchengode, India, 4–6 July 2013; pp. 1–5. [Google Scholar] [CrossRef]
  72. Taheri-Garavand, A.; Beiranvandi, M.; Ahmadi, A.; Nikoloudakis, N. Predictive modeling of Satureja rechingeri essential oil yield and composition under water deficit and soil amendment conditions using artificial neural networks (ANNs). Comput. Electron. Agric. 2024, 222, 109072. [Google Scholar] [CrossRef]
  73. Sudhamathi, T.; Perumal, K. A novel hybrid DNN-RNN framework for precise crop yield prediction. Int. J. Syst. Assur. Eng. Manag. 2024, 1–13. [Google Scholar] [CrossRef]
  74. Chavan, Y.; Swamikan, B.; Gupta, M.; Bobade, S.; Malhan, A. Enhanced Crop Yield Forecasting Using Deep Reinforcement Learning and Multi-source Remote Sensing Data. Remote Sens. Earth Syst. Sci. 2024, 7, 426–442. [Google Scholar] [CrossRef]
  75. Aslan, M.F.; Sabanci, K.; Aslan, B. Artificial Intelligence Techniques in Crop Yield Estimation Based on Sentinel-2 Data: A Comprehensive Survey. Sustainability 2024, 16, 8277. [Google Scholar] [CrossRef]
  76. Atsalakis, G. Agriculture Commodity Prices Forecasting Using a Fuzzy Inference System. Agricultural Cooperative Management and Policy. Cooperative Management; Springer: Cham, Switzerland, 2014. [Google Scholar] [CrossRef]
  77. Dutta, A.; Nayak, A.; Aditya; Panda, R.R.; Nagwani, N.K. A Neuro Fuzzy System Based Inflation Prediction of Agricultural Commodities. In Proceedings of the 2020 11th International Conference on Computing, Communication and Networking Technologies (ICCCNT), Kharagpur, Inida, 1–3 July 2020; pp. 1–6. [Google Scholar] [CrossRef]
  78. Abd Elfattah, M.; Ewees, A.; Sayed, G.; Darwish, A.; Ella Hassanien, A. Carbon price time series forecasting utilizing an optimized ANFIS model. Evol. Intell. 2024, 17, 3745–3765. [Google Scholar] [CrossRef]
  79. Remya, S. An adaptive neuro-fuzzy inference system to monitor and manage the soil quality to improve sustainable farming in agriculture. Soft Comput. 2022, 26, 13119–13132. [Google Scholar] [CrossRef]
  80. Burmakova, A.; Kalibatiene, D. An ANFIS-based Model to Predict the Oil Spill Consequences on the Ground. In Proceedings of the IEEE Open Conference of Electrical, Electronic and Information Sciences, Vilnius, Lithuania, 22–22 April 2021. [Google Scholar] [CrossRef]
  81. Ghose, D.K.; Tanaya, K.; Sahoo, A.; Kumar, U. Performance Evaluation of hybrid ANFIS model for Flood Prediction. In Proceedings of the 8th International Conference on Advanced Computing and Communication Systems (ICACCS), Coimbatore, India, 25–26 March 2022. [Google Scholar] [CrossRef]
  82. Garg, B.; Sah, T. Prediction of Crop Yield Using Fuzzy-Neural System. In EAI International Conference on Big Data Innovation for Sustainable Cognitive Computing; EAI/Springer Innovations in Communication and Computing; Springer: Cham, Switzerland, 2020. [Google Scholar] [CrossRef]
  83. Ananthara, M.G.; Arunkumar, T.; Hemavathy, R. CRY—An improved crop yield prediction model using bee hive clustering approach for agricultural data sets. In Proceedings of the 2013 International Conference on Pattern Recognition, Informatics and Mobile Engineering, Salem, India, 21–22 February 2013; pp. 473–478. [Google Scholar] [CrossRef]
  84. Euzébio, T.A.M.; Ramirez, M.A.P.; Reinecke, S.F.; Hampel, U. Energy Price as an Input to Fuzzy Wastewater Level Control in Pump Storage Operation. IEEE Access 2023, 11, 93701–93712. [Google Scholar] [CrossRef]
  85. Małlolepsza, O.; Mikołlajewski, D.; Prokopowicz, P. Using Fuzzy Logic to Analyse Weather Conditions. Electronics 2025, 14, 85. [Google Scholar] [CrossRef]
  86. Pérez-Pérez, J.F.; Gómez, P.I.; Bonet, I.; Sánchez-Pinzón, M.S.; Caraffini, F.; Lochmuller, C. Assessing Climate Transition Risks in the Colombian Processed Food Sector: A Fuzzy Logic and Multi-Criteria Decision-Making Approach. Mathematics 2024, 12, 2713. [Google Scholar] [CrossRef]
  87. Abrego-Perez, A.L.; Pacheco-Carvajal, N.; Diaz-Jimenez, M.C. Forecasting Agricultural Financial Weather Risk Using PCA and SSA in an Index Insurance Model in Low-Income Economies. Appl. Sci. 2023, 13, 2425. [Google Scholar] [CrossRef]
  88. Zakri, A.A.; Mustafa, M.W.; Tribowo, I. ANFIS Design Based on Prediction Models for The Photovoltaic System. In Proceedings of the 2019 International Conference on Sustainable Information Engineering and Technology (SIET), Lombok, Indonesia, 28–30 September 2019. [Google Scholar] [CrossRef]
  89. Suparta, W.; Alhasa, K.M. Modeling of Tropospheric Delays Using ANFIS; Springer: Cham, Switzerland, 2016. [Google Scholar] [CrossRef]
  90. Khalid, E.T.; Jassim, S.A.; Saqaeeyan, S. Fuzzy C-mean clustering technique based visual features fusion for automatic video summarization method. Multimed. Tools Appl. 2024, 83, 87673–87696. [Google Scholar] [CrossRef]
  91. Fuzzy Logic Designer. Available online: https://la.mathworks.com/help/fuzzy/types-of-fuzzy-inference-systems.html (accessed on 19 July 2024).
Figure 1. Related work features.
Figure 1. Related work features.
Computers 14 00168 g001
Figure 2. Representation of the neuro-fuzzy system.
Figure 2. Representation of the neuro-fuzzy system.
Computers 14 00168 g002
Figure 3. Research summary.
Figure 3. Research summary.
Computers 14 00168 g003
Figure 4. Presentation of all the collected data.
Figure 4. Presentation of all the collected data.
Computers 14 00168 g004
Figure 5. Model configurations proposed.
Figure 5. Model configurations proposed.
Computers 14 00168 g005
Figure 6. Model results using training and validation data.
Figure 6. Model results using training and validation data.
Computers 14 00168 g006
Figure 7. Model results using training (2012–2017) and validation (2018–2019) data.
Figure 7. Model results using training (2012–2017) and validation (2018–2019) data.
Computers 14 00168 g007
Figure 8. Features of the Mamdani-type fuzzy system structure FIS-L.
Figure 8. Features of the Mamdani-type fuzzy system structure FIS-L.
Computers 14 00168 g008
Figure 9. Set of rules of the Mamdani-type fuzzy system FIS-L.
Figure 9. Set of rules of the Mamdani-type fuzzy system FIS-L.
Computers 14 00168 g009
Figure 10. Features of the Mamdani-type fuzzy system structure FIS-C.
Figure 10. Features of the Mamdani-type fuzzy system structure FIS-C.
Computers 14 00168 g010
Figure 11. Set of rules of the Mamdani-type fuzzy system FIS-C.
Figure 11. Set of rules of the Mamdani-type fuzzy system FIS-C.
Computers 14 00168 g011
Figure 12. Fuzzy system simulations employing training and validation data.
Figure 12. Fuzzy system simulations employing training and validation data.
Computers 14 00168 g012
Table 1. Binary codification for products.
Table 1. Binary codification for products.
Product X 1 X 2 X 3 X 4 X 5 X 6
Banano100000
Cocoa010000
Coffee001000
Sugar cane000100
Orange000010
Plantain000001
Table 2. Binary codification for geographic region.
Table 2. Binary codification for geographic region.
Region W 1 W 2 W 3 W 4 W 5
Andean10000
Caribbean01000
Pacific00100
Orinoco00010
Amazon00001
Table 3. MSE results for model M 1 using training data.
Table 3. MSE results for model M 1 using training data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00809700670.00065202870.00058586190.00061639370.02491757130.00527642930.00527067510.0052861664
Min0.00144640260.00035010090.00034900700.00034476100.01171084930.00516972090.00518062230.0052047847
Mean0.00671062250.00044713620.00050472890.00048563410.01893762980.00522712510.00523294610.0052357157
STD0.00166047610.00010581630.00009070920.00011150180.00367775100.00002777300.00002447480.0000259180
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00732187110.00024746750.00025647120.00026378260.02176464020.00079077620.00076288820.0008177421
Min0.00239023380.00014350860.00013562830.00015500740.01230384520.00055105380.00054914280.0005835823
Mean0.00566265720.00018423010.00017356530.00018300930.01646224600.00068062500.00065403210.0007090027
STD0.00158626180.00002628250.00003227250.00002417940.00421000600.00007009640.00006312870.0000615479
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00453801440.00010905440.00013424050.00011318430.02118409930.00065552160.00061083460.0010487058
Min0.00165597190.00004861690.00004989800.00004803080.01201515090.00039432220.00042078060.0004225335
Mean0.00386910200.00007851510.00008558170.00008451270.01309400590.00051170560.00051370230.0005594526
STD0.00076217380.00001652330.00001613950.00001559930.00245347770.00006241630.00005002550.0001286913
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00608558590.00006060320.00006115750.00006711400.02310512170.00104585410.00054001850.0005874693
Min0.00094246330.00003624970.00003666520.00003697070.00279848440.00042739850.00038334000.0003728754
Mean0.00362567800.00004959810.00004499430.00004633800.01237074840.00051401090.00045502870.0004679564
STD0.00121754360.00000743090.00000581800.00000766880.00411706210.00013407490.00004061670.0000628095
Table 4. MSE for model M 1 using validation dada.
Table 4. MSE for model M 1 using validation dada.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01772706650.00173068650.00165361810.00168251470.03807653290.00336760670.00333513920.0033845704
Min0.00496532960.00074098560.00077249990.00093645120.02623344130.00329050590.00328604760.0032862555
Mean0.01479514780.00134726090.00144478940.00139080570.03232049180.00330777200.00330547940.0033150417
STD0.00293406390.00023709940.00024617510.00022741880.00389521200.00001595080.00001345080.0000277304
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01616306090.00150183390.00162314310.00314283540.03862118410.00135008150.00144010970.0011474682
Min0.00495832930.00047872220.00032094760.00031927370.02622200380.00071674210.00070712090.0007056395
Mean0.01297647640.00086831760.00111537130.00096963540.03049807980.00087742510.00090086480.0008391093
STD0.00292326160.00028872910.00036972600.00060320620.00494990880.00019415840.00019998130.0001262611
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01152359270.00485868010.00262098450.01310336310.03731929620.00117249280.00113743210.0020260329
Min0.00301840900.00033396080.00036913880.00026265200.02595554610.00077800040.00068013320.0007437698
Mean0.00942945560.00173914070.00106675810.00211946220.02757218570.00093836720.00091518440.0010106556
STD0.00253447180.00143250400.00058532780.00279791290.00322726850.00011312310.00010311280.0002552771
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01461524990.02398277290.00234098060.00924910200.03724507220.00194717660.00105862340.0017299126
Min0.00274747700.00037470910.00031728130.00043262320.00720523740.00058632900.00053098270.0005327444
Mean0.00949497460.00207647640.00095566530.00159368950.02577687220.00094736660.00080349980.0008209868
STD0.00294798440.00516990350.00059136550.00208337940.00621855450.00034909050.00012124170.0002468106
Table 5. MSE results for model M 2 using training data.
Table 5. MSE results for model M 2 using training data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00439347540.00062337360.00062951720.00053843440.02938035430.00550458470.00547722330.0057506063
Min0.00298555570.00035550750.00034831900.00034941330.01151641880.00499612350.00501004470.0050115597
Mean0.00412321710.00040687350.00044250670.00040435670.02366133010.00511797600.00509685060.0051524248
STD0.00032761220.00006669210.00009073330.00005079820.00525784720.00012384070.00010378680.0001743111
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00416147890.00046740370.00041653980.00047812490.02556416110.00349713030.00352959270.0035163343
Min0.00277587550.00013587770.00012397470.00011738330.01635708510.00073982820.00070509220.0007424269
Mean0.00327611980.00022616460.00021608830.00022965710.02033407100.00246609620.00288018680.0025841186
STD0.00060509750.00011302030.00009844710.00013058770.00367248290.00109687790.00104787540.0011448758
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00281672010.00011711990.00010983300.00028453190.02397022700.00357293870.00368088330.0033936204
Min0.00100434210.00003520870.00004905940.00004985440.01442422140.00053738980.00055985280.0005331879
Mean0.00260570860.00007269080.00006568490.00008000690.01784066820.00123691860.00113073900.0011542260
STD0.00043074050.00002186250.00001581730.00005022820.00161703920.00083457790.00079687890.0007561475
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00283878530.00004111170.00004686950.00004365160.01767774790.00333428120.00326328470.0031787204
Min0.00055569580.00001469770.00001928860.00001534750.00818917460.00031730750.00032157090.0005053861
Mean0.00232733600.00003049440.00002874210.00002979590.01625583650.00088945510.00097215530.0007936620
STD0.00077630690.00000651860.00000756740.00000716280.00230879810.00073530020.00070135330.0006100003
Table 6. MSE results for model M 2 using validation data.
Table 6. MSE results for model M 2 using validation data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00250040150.00121717110.00124862710.00125447320.02004081720.00805205300.00796362840.0082941530
Min0.00171515530.00035845660.00034290690.00034137250.00448354180.00766211320.00767181620.0076801546
Mean0.00197996320.00055806120.00066266260.00058903280.01469967050.00773239820.00771818620.0077600574
STD0.00023966720.00020217010.00026788700.00025649240.00506790740.00009311650.00006562140.0001422642
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00253349580.00578419690.00863955080.02359163600.01738798820.00555059350.00555944460.0055526054
Min0.00176283290.00032113190.00038936550.00027913300.00759447270.00098491620.00091870420.0009509613
Mean0.00231362210.00139525950.00167400560.00236780430.01028140950.00386961040.00466986540.0040424918
STD0.00023838730.00125320800.00178254210.00505872830.00365208390.00190608350.00162006370.0019767727
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00259960790.00720614280.00932404270.15917692090.01779977230.00579657300.00580954320.0055260254
Min0.00127749510.00082664690.00048736390.00068632350.00233966450.00086523280.00095361770.0009072779
Mean0.00233624330.00193671010.00231857160.00980467230.00787980240.00213845810.00199680720.0020126783
STD0.00025432600.00149028490.00204168540.03519406410.00261742780.00159323820.00130301620.0011910029
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00663379750.00698802240.07579562050.00315327540.01686509850.00610983270.00664251540.0062429654
Min0.00126314180.00088398870.00068159370.00124932170.00227555680.00087858040.00066458630.0005455133
Mean0.00250817020.00241209510.00639111990.00201313470.00697204180.00162379770.00221322200.0015372202
STD0.00104028840.00168868260.01673361340.00048079280.00311215600.00150111460.00206948470.0012245736
Table 7. MSE results for model M 3 using training data.
Table 7. MSE results for model M 3 using training data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00699823250.00440729950.00291441930.00439358600.01193708780.01093443340.00785180590.0078569119
Min0.00531278150.00253526970.00234327210.00233046910.01132770920.00794649220.00765660940.0077534165
Mean0.00645767090.00364258950.00243851450.00251729390.01165417210.00925275650.00779443490.0078061320
STD0.00064556090.00057788700.00017914520.00046768730.00027827270.00139882120.00005365570.0000208386
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00648575360.00232600110.00173550630.00175125980.01231698490.00795845070.00488006100.0046525032
Min0.00648556930.00112459560.00085496820.00090702890.01231195060.00640359640.00432569840.0043299794
Mean0.00648561870.00171653990.00128710680.00153730400.01231434370.00732395010.00440976600.0044146060
STD0.00000005830.00023269320.00034445910.00023687920.00000237600.00029988930.00012406840.0000902811
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00648571490.00181031450.00115404610.00113947660.01231880790.00503742710.00396183730.0040610837
Min0.00646439140.00121505840.00054621010.00046705730.01231406310.00451868050.00231647270.0023652725
Mean0.00648445570.00156223060.00088993170.00085832130.01231474560.00480445100.00329539120.0029579023
STD0.00000472370.00016047050.00016242480.00018836880.00000105220.00013794900.00045827610.0005270630
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00648466530.00146719570.00079415150.00071461690.01231097280.00455262890.00276947230.0027418833
Min0.00571981220.00121116820.00029038290.00026350310.00823442460.00447036540.00170808550.0017304822
Mean0.00636991650.00131115110.00047965750.00048979280.01169941370.00451107430.00226794960.0022310550
STD0.00028018820.00006225670.00012404000.00011497680.00149338090.00002253220.00026203960.0002977945
Table 8. MSE results for model M 3 using validation data.
Table 8. MSE results for model M 3 using validation data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01723522780.01227659670.00649461910.01217385730.02243409450.02914274870.01206446210.0121034889
Min0.01528035320.00635407960.00561544450.00558035490.02124737220.01247338720.00983702460.0110405045
Mean0.01639545710.00883723790.00575462360.00604672220.02195639220.02008062910.01145459830.0116125450
STD0.00083845010.00183615480.00028538810.00146354420.00047875920.00771781750.00062505790.0002143604
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01332202010.00720652950.00497443390.00528394710.02672057160.01657100260.01042299220.0094742808
Min0.01331702250.00251374050.00152242110.00188131950.02671135350.01240206120.00888451560.0087871646
Mean0.01331927860.00555449900.00329270330.00435091560.02671696610.01303554950.00913371050.0091131271
STD0.00000159800.00091323020.00134819990.00098468650.00000340360.00109748230.00032281450.0001548588
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01334650210.05489082120.00553347310.00569751410.02672707030.01275826960.00883049110.0081842984
Min0.01332531040.00534072060.00137082500.00102000570.02671123820.01117288860.00527019500.0052908074
Mean0.01334163340.01113831260.00324251360.00301374430.02671391640.01238137230.00690638480.0062151911
STD0.00000399350.01192011630.00110198210.00135882330.00000374120.00046180510.00106596020.0009436170
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01333569160.16062648660.00582423260.00622928370.02670698150.01264510560.00646003900.0064658082
Min0.01175941770.00456365190.00086882790.00067321610.01596526130.01253855540.00439198960.0040299573
Mean0.01309937670.03717430210.00276456680.00295585100.02509562810.01259472960.00510140850.0051233568
STD0.00057657630.05030183520.00143792770.00165494290.00393457950.00002753800.00060256980.0007006443
Table 9. MSE results for model M 4 using training data.
Table 9. MSE results for model M 4 using training data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00485065960.00154692460.00154809400.00174015810.01784325910.01119861060.00978203310.0088869809
Min0.00485063040.00127544790.00126264860.00126024580.01784325000.00960691310.00865119950.0081913207
Mean0.00485063670.00140555940.00136555490.00152855140.01784325480.00997092770.00935963130.0084057658
STD0.00000000850.00011495380.00010672310.00018591970.00000000180.00031424960.00034904820.0003096864
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00484843630.00163294690.00169358480.00161389730.01785229680.00616285520.00568934640.0079089000
Min0.00484843220.00151032150.00166441750.00086475550.01785138260.00614921220.00561519030.0048539310
Mean0.00484843460.00158291700.00167530890.00101926590.01785180550.00615529880.00566523030.0077344605
STD0.00000000180.00004444770.00000864120.00014663480.00000024610.00000417010.00001415650.0006782212
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00484829910.00140186010.00144099980.00135825770.02872527440.00574053220.01087529870.0105797261
Min0.00334563020.00122243270.00122991910.00070795240.00502234100.00560095900.00543904670.0015664727
Mean0.00399606720.00128715170.00136485690.00121485700.01057793370.00563589860.00987775990.0059346292
STD0.00070450220.00006228070.00009142130.00021531730.00738815200.00002580460.00185727490.0026029491
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00477920540.00117564340.00120088340.00121848890.02872504890.00577518630.01060437220.0105243163
Min0.00333739370.00111373240.00104824600.00039516830.00502838070.00572615090.01047066300.0017787307
Mean0.00392158310.00114690640.00117788160.00085406070.00622377910.00574143410.01056541770.0047988146
STD0.00049660800.00001972390.00003877200.00027310920.00529626890.00001467020.00002652860.0023025050
Table 10. MSE results for model M 4 using validation data.
Table 10. MSE results for model M 4 using validation data.
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00138535270.00270970160.00278029990.00245132820.00762084830.01016757040.01011370090.0107280161
Min0.00138254540.00105217650.00093167010.00088753210.00762083730.00927214060.00929777640.0089434370
Mean0.00138518730.00199125510.00207390870.00145396950.00762083890.00996828140.00961486230.0101331653
STD0.00000062280.00069700580.00058019070.00050322520.00000000290.00024410210.00017893230.0007498761
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00139591730.00406783500.00246295580.04187544540.00762448100.00266261050.00359965670.0108567520
Min0.00139578710.00104397090.00223456750.00090701540.00762341500.00263396860.00324971700.0056295300
Mean0.00139583980.00212728490.00226221970.00510854850.00762379150.00264458060.00347181680.0104920583
STD0.00000006310.00115290930.00004842400.00912676810.00000043320.00001005150.00006138380.0011464662
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00139732320.07365932880.00739062630.00372146650.01821435490.00650447860.00425409440.0086863591
Min0.00073388150.00085591710.00086699780.00089569070.00130619900.00586930400.00242639600.0014432050
Mean0.00114260880.00570904440.00203405950.00176101290.00425775310.00626527350.00399595520.0058096363
STD0.00021324480.01603620960.00178698340.00070811610.00440126900.00017502380.00058204120.0021182686
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00136380190.01343763270.00391512030.00398821570.01822996480.00674778030.00440404270.0075041045
Min0.00094203900.00086325600.00085593500.00066033410.00130471840.00373141600.00412327600.0016793748
Mean0.00117644020.00187698060.00296989300.00150418570.00225706140.00548661170.00422315880.0046657596
STD0.00012701350.00278394590.00114503230.00091187540.00376321590.00064303640.00004794200.0020469260
Table 11. MSE results taking the best configuration determined from each model.
Table 11. MSE results taking the best configuration determined from each model.
LinearConstant
ModelTrainingValidationTrainingValidation
M 1 0.3625 × 10 4 2.6265 × 10 4 3.7288 × 10 4 5.3098 × 10 4
M 2 0.1469 × 10 4 2.7913 × 10 4 3.1731 × 10 4 5.4551 × 10 4
M 3 2.6350 × 10 4 6.7322 × 10 4 17.0000 × 10 4 40.0000 × 10 4
M 4 3.9517 × 10 4 6.6033 × 10 4 16.0000 × 10 4 13.0000 × 10 4
Table 12. MSE results for model M 1 using training data (2012–2017).
Table 12. MSE results for model M 1 using training data (2012–2017).
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00532926000.00040601090.00044388630.00041803770.01969789420.00504092840.00503866930.0050158100
Min0.00278035120.00018271360.00018195010.00018121790.00811604340.00317465580.00291272590.0034310978
Mean0.00464396700.00026449680.00023160520.00025397990.01560706910.00415958980.00411266370.0043089835
STD0.00095316460.00007844350.00007459140.00009300470.00322389890.00059189650.00062065300.0004791798
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00507509710.00012433410.00014051200.00012575830.01759157090.00058253970.00054928230.0005491278
Min0.00099802690.00007124860.00008102410.00007698460.00904547960.00041885700.00043524000.0004177740
Mean0.00321131830.00009899690.00010784960.00009732040.01195982960.00049160830.00047608290.0004715419
STD0.00114089350.00001567220.00001723120.00001612720.00298407360.00004917220.00003019380.0000337984
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00274855770.00005495210.00006935900.00005972830.01532486920.00043501210.00052366970.0005168433
Min0.00049093240.00002688310.00002748180.00002630480.00856587750.00028513970.00027517760.0003024105
Mean0.00233121640.00003883820.00003994440.00004034650.01041763570.00034111690.00036567650.0003709913
STD0.00059359420.00000797750.00001036650.00001021010.00135592530.00004795960.00007419690.0000568863
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.00254554640.00002318070.00004214530.00003662120.01009994300.00084877270.00039517130.0005972200
Min0.00049776810.00001291530.00000571400.00000633980.00513965640.00026457140.00020679190.0002349152
Mean0.00210958400.00001784950.00001834180.00001675610.00950586170.00035585660.00029849970.0003407384
STD0.00079720640.00000299690.00000740280.00000667780.00132872220.00012564670.00005048450.0000877377
Table 13. MSE for model M 1 using validation dada (2018–2019).
Table 13. MSE for model M 1 using validation dada (2018–2019).
Clusters2
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.02355638190.01121569160.01057377600.01174306020.04552013800.02295390380.02305662780.0232422471
Min0.01734228910.00250909570.00214030510.00263658290.02876840570.00474718630.00478369780.0046726179
Mean0.02184172860.00552349110.00591555670.00629530600.03863899900.01937988640.02019326010.0199095693
STD0.00235117210.00270382970.00243400780.00299428470.00517064660.00523055450.00406125980.0053811967
Clusters3
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.02269245090.01348786410.02536832130.06300202250.04277416500.00310739420.00318190430.0030090038
Min0.00671139560.00280040500.00367401690.00308205020.02918679680.00171047030.00180806740.0018142282
Mean0.01760360470.00802456920.00825419290.01485931110.03249975010.00232987850.00226102040.0022041344
STD0.00366039900.00315636870.00521811740.01594693330.00474155600.00043912310.00036050950.0002996196
Clusters4
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.01701295440.14057474920.06469456470.16867025440.04217682140.00494807630.00334800700.0033080639
Min0.00266430980.00279321220.00276565250.00385774440.02449816290.00194797740.00180246090.0015143868
Mean0.01421028100.02128335870.01269349670.02538352090.02980683480.00281205170.00266894200.0025681594
STD0.00443225750.03331016150.01521462400.04136677370.00312553020.00061559310.00043539780.0004861269
Clusters5
Output FunctionsLinearConstant
Exponent1.12.03.04.01.12.03.04.0
Max0.26183357990.09870984400.13138527270.07520322690.03683875400.00611348870.01082681250.0063217016
Min0.00225463670.00331176120.00236465100.00407329590.02235927120.00215699270.00228363790.0017891166
Mean0.02616133240.01687061480.01911416600.02232572170.02897796440.00310980900.00414980950.0033573824
STD0.05562223570.02075701550.02758373220.02085472040.00266566010.00098408580.00264679510.0011534380
Table 14. MSE results taking best configurations for M 1 .
Table 14. MSE results taking best configurations for M 1 .
LinearConstant
Data SelectionTrainingValidationTrainingValidation
Random sampling 0.3625 × 10 4 2.6265 × 10 4 3.7288 × 10 4 5.3098 × 10 4
Segmented data 0.0571 × 10 4 21.0000 × 10 4 2.0679 × 10 4 15.0000 × 10 4
Table 15. MSE results for Mamdani fuzzy systems.
Table 15. MSE results for Mamdani fuzzy systems.
SystemTrainingValidation
FIS-L 0.9580 × 10 4 2.6265 × 10 4
FIS-C 4.2950 × 10 4 5.3098 × 10 4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Gómez, A.C.; Bejarano, L.A.; Espitia, H.E. Model for Agricultural Production in Colombia Using a Neuro-Fuzzy Inference System. Computers 2025, 14, 168. https://doi.org/10.3390/computers14050168

AMA Style

Gómez AC, Bejarano LA, Espitia HE. Model for Agricultural Production in Colombia Using a Neuro-Fuzzy Inference System. Computers. 2025; 14(5):168. https://doi.org/10.3390/computers14050168

Chicago/Turabian Style

Gómez, Andrea C., Lilian A. Bejarano, and Helbert E. Espitia. 2025. "Model for Agricultural Production in Colombia Using a Neuro-Fuzzy Inference System" Computers 14, no. 5: 168. https://doi.org/10.3390/computers14050168

APA Style

Gómez, A. C., Bejarano, L. A., & Espitia, H. E. (2025). Model for Agricultural Production in Colombia Using a Neuro-Fuzzy Inference System. Computers, 14(5), 168. https://doi.org/10.3390/computers14050168

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop