Machine Learning-Based Probabilistic Lithofacies Prediction from Conventional Well Logs: A Case from the Umiat Oil Field of Alaska

A good understanding of different rock types and their distribution is critical to locate oil and gas accumulations in the subsurface. Traditionally, rock core samples are used to directly determine the exact rock facies and what geological environments might be present. Core samples are often expensive to recover and, therefore, not always available for each well. Wireline logs provide a cheaper alternative to core samples, but they do not distinguish between various rock facies alone. This problem can be overcome by integrating limited core data with largely available wireline log data with machine learning. Here, we presented an application of machine learning in rock facies predictions based on limited core data from the Umiat Oil Field of Alaska. First, we identified five sandstone reservoir facies within the Lower Grandstand Member using core samples and mineralogical data available for the Umiat 18 well. Next, we applied machine learning algorithms (ascendant hierarchical clustering, self-organizing maps, artificial neural network, and multi-resolution graph-based clustering) to available wireline log data to build our models trained with core-driven information. We found that self-organizing maps provided the best result among other techniques for facies predictions. We used the best self-organizing maps scheme for predicting similar reservoir facies in nearby uncored wells—Umiat 23H and SeaBee-1. We validated our facies prediction results for these wells with observed seismic data.


Introduction
The success of any conventional hydrocarbon exploration program primarily depends upon identifying and mapping porous and permeable sandstone reservoirs where commercial volumes of hydrocarbons are stored. In these reservoir rocks, the spatial heterogeneity of porosity and permeability has been known to be affected by the geological character of the sediments, including their depositional environment, stratigraphic position, and mineral composition [1][2][3]. Detailed knowledge of sedimentary characteristics of high-quality sandstone reservoirs is, therefore, crucial in the early stages of hydrocarbon exploration.
Core samples are widely used in the petroleum industry for comprehensive analysis of lithologic and petrophysical properties of sandstone reservoirs that provide valuable insight into past depositional environments. Sandstones with consistent lithologic characteristics, such as grain size, appearance, and composition, are usually referred to as lithofacies [4]. Although core analysis provides rock information with high precision and spatial resolution, continuous coring is both an expensive and time-consuming process. Largely available conventional geophysical logs, including gamma-ray, density, neutron, and resistivity, can be a cheaper and reliable alternative to predicting rock petrophysical parameters. However, these well logs do not provide direct observations of sand lithofacies and, therefore, cannot be linked directly to the depositional environments. Over the past ten years, research efforts are focused on the application of machine learning techniques to integrate conventional well data with limited core data, enabling reliable lithologic interpretations for sandstone reservoirs [5][6][7]. The key advantage of this machine learning-based approach is that it can be quickly applied to any number of nearby wells with limited or no core data once the data model is trained.
In this study, we demonstrated how advanced machine learning algorithms, including ascendant hierarchical clustering (AHC), self-organizing maps (SOM), artificial neural network (ANN), and multi-resolution graph-based clustering (MRGC), could be applied to conventional well log measurements from wells with no or poor core data to predict sandstone lithofacies. Our approach involved generating machine learning-based lithofacies models that were initially trained using limited core data and a suite of conventional well logs from a reference well (labeled data). These data models were then individually applied to well log data (unlabeled data) from nearby wells that had no core data to predict similar lithofacies. These predicted reservoir facies were then validated using seismic data. To demonstrate our approach, we analyzed the sandstones from the Lower Grandstand Member in the Umiat Oil Field-an emerging oil and gas frontier region of Alaska. Based on the limited core data available, these rocks are considered to have the best reservoir characteristics and are thought to be deposited by progradational wave-dominated deltaic river systems [8]. Five sandstone lithofacies were selected from the available core descriptions (761 ft to 868 ft) of the Lower Grandstand Member in an exploratory well, Umiat 18 [9]. These lithofacies were used to train the unsupervised machine learning models through machine learning techniques. We applied machine learning models to unclassified wells-Umiat 21H and SeaBee-1-for facies predictions, having only conventional wireline logs available and no core data. The resultant lithofacies prediction in two wells were validated through comparison with seismic data. Our results show vertical variability in deltaic sandstone facies observed within the Lower Grandstand Member, and they appear laterally continuous throughout the central portion of the field area. Our facies validation efforts further suggest that SOM is the best among the examined machine learning techniques and can be used as a supplementary tool to traditional core-driven interpretations of sandstone lithofacies.

Umiat Oil Field
The Umiat Oil Field of Alaska is one of the largest, undeveloped light-oil fields in the United States, with an estimated 1 billion barrels of original oil in place within the Cretaceous Nanushuk Formation [10] (Figure 1). The Umiat field was first discovered in 1946 and remains largely undeveloped over the past 70 years due to a lack of infrastructure at Umiat. Only 14 wells have been drilled into sandstone reservoirs of the Nanushuk Formation, with no significant oil or gas production. Previous studies suggest that the gas-oil-water contact exists in the Nanushuk Formation at about 1500 ft, with all sandstone reservoirs located within the hydrocarbon zone above the free-water level [11].
The Cretaceous Nanushuk Formation primarily consists of mudstone and coarsening-upward sequences of sandstone deposits that were deposited in fluvio-deltaic, shoreface, and marine-shelf environments [12,13] (Figure 2). Limited core data combined with outcrop observations show that the Nanushuk Formation has been subdivided into five lithologic units-the uppermost shallow marine Ninuluk sandstone unit, the deltaic Killik Tongue of the Chandler Formation, the underlying shallow marine to deltaic Upper and Lower Grandstand sandstone units that are separated by marine mudstones, and shales of the Tuktu [8]. Deltaic sandstones of the Lower Grandstand Formation are the primary exploration target in the Umiat Oil Field. Our study focused on the Lower Grandstand Formation, which is further comprised of two upward coarsening deltaic sandstone bodies that are separated by silty mudstones associated with lower offshore/shelf flooding surfaces (Table 1). The Cretaceous Nanushuk Formation primarily consists of mudstone and coarsening-upward sequences of sandstone deposits that were deposited in fluvio-deltaic, shoreface, and marine-shelf environments [12,13] (Figure 2). Limited core data combined with outcrop observations show that the Nanushuk Formation has been subdivided into five lithologic units-the uppermost shallow marine Ninuluk sandstone unit, the deltaic Killik Tongue of the Chandler Formation, the underlying shallow marine to deltaic Upper and Lower Grandstand sandstone units that are separated by marine mudstones, and shales of the Tuktu [8]. Deltaic sandstones of the Lower Grandstand Formation are the primary exploration target in the Umiat Oil Field. Our study focused on the Lower Grandstand Formation, which is further comprised of two upward coarsening deltaic sandstone bodies that are separated by silty mudstones associated with lower offshore/shelf flooding surfaces (Table 1).

Data
In this study, we analyzed publicly available Umiat Oil Field data, including digital well logs, core descriptions, and seismic data. Digital well logs and well history files for three wells (Umiat 18, Umiat 23H, and SeaBee-1) were obtained from the Alaska Oil and Gas Conservation Commission (AOGCC) website (http://doa.alaska.gov/ogc/data.html). Digital logs include gamma-ray, neutron, density, sonic, and resistivity logs. Core data, including facies types [9], mineral composition,

Data
In this study, we analyzed publicly available Umiat Oil Field data, including digital well logs, core descriptions, and seismic data. Digital well logs and well history files for three wells (Umiat 18, Umiat 23H, and SeaBee-1) were obtained from the Alaska Oil and Gas Conservation Commission (AOGCC) website (http://doa.alaska.gov/ogc/data.html). Digital logs include gamma-ray, neutron, density, sonic, and resistivity logs. Core data, including facies types [9], mineral composition, porosity, and permeability, from routine core analyses of the Lower Grandstand Formation, are also available publicly for Umiat 18 (depth range 710 ft to 1014 ft) on the website. Both prestack and poststack 2D and 3D seismic data within the Umiat Oil Field are available at the Alaska Geologic Materials Center (AGMC) (http://dggs.alaska.gov/gmc/seismic-well-data.php).

Lithofacies of the Lower Grandstand Member (LGST)
LePain et al.
[9] examined cores (710 ft-1014 ft) from Umiat 18 and described five lithofacies within LGST that together comprise of prodelta and river-dominated delta-front facies (Table 1). These five reservoir lithofacies of LGST include SI-horizontal planer laminated sandstone, Sxtrough cross-bedded sandstone, Sm-massive sandstone, Sr-ripple cross-laminated sandstone, and FIcarbonaceous mudstone. These five lithofacies suggest a transition from an upper shoreface deltaic environment (SI, Sx, Sm, and Sr) to a distal shelf/prodelta environment. Porosity and permeability values follow subtle changes in lithofacies and are highest in high-energy facies of upward-coarsening deltaic-shoreface sandstones. Lithofacies data from Umiat 18 core were then used as a reference to train well log data from Umiat wells using machine learning methods.

Well Log Data
Geophysical well logs represent single-point measurements of rock physical properties with depth recorded in a well. A suite of conventional well logs from three wells (Umiat 18, Umiat 23H, and SeaBee-1) of the Umiat Oil Field were considered for machine learning-based lithofacies modeling and predictions. For each well, we utilized four scalar attributes from well log measurements, including:  Table 2 shows the depth range, minimum, maximum, and mean log values from all wells used in this study.

Pre-Processing
Prediction of lithofacies using well log data relies on the assumption that each rock facies generates a unique set of log readings. In order to perform reliable facies prediction, all of the well logs need to be consistent and unbiased throughout the target lithological formation across all wells. Initial preprocessing of the original raw logs is, therefore, needed to remove any depth-related discrepancies, as well as any effect from borehole logging and the presence of hydrocarbons or other fluids.
We used the Paradigm TM formation evaluation product-Geolog TM -to perform initial log processing. All gamma-ray logs from Umiat wells were aligned in depth and plotted from 0 to 200 units. Calibrated GR values were then used for lithology predictions. The opposing response of neutron and density logs are commonly used to indicate the presence of hydrocarbons in the formation. Neutron porosity increases with increasing hydrocarbon content in the formation with a decrease in the formation's bulk density. If this hydrocarbon effect is not corrected, raw density and neutron porosity values can lead to wrong facies prediction in a well. We used neutron-density cross-plots to identify the hydrocarbon zones in LGST of Umiat 18. The identified hydrocarbon zones were analyzed in Geolog TM to correct neutron and density logs for hydrocarbon effect corrections.

Machine Learning Algorithms
The workflow for the facies prediction includes four main steps: (1) selecting and pre-processing of training data, (2) building a machine learning model(s) to sort input data into clusters for facies prediction, (3) reviewing model(s) to identify an optimal number of clusters that fit the training data, and (4) applying model(s) to other wells in the area. For training purposes, we used three log attributes-GR, NPHI, and RHOB-as the model logs. We observed facies from core data as the reference log. We used DT as a reference log to create a synthetic seismic model for SeaBee-1. Initially, we applied machine learning workflows on the training data from LGST of Umiat 18 using Paradigm's Facimage TM module. For each machine learning technique, we varied all combinations of the number of clusters and training parameters to generate different facies models for comparison and to identify the best matching scheme. We used R 2 , the coefficient of determination, to analyze the goodness of fit between the observed and predicted facies. We employed the best machine-learning scheme to predict lithofacies from well logs in two un-cored wells-Umiat 23H and SeaBee-1-of the Umiat field. We examined the available seismic data to validate the predicted LGST lithofacies in test wells.

A. Ascendant Hierarchical Clustering (AHC)
AHC is an unsupervised machine learning algorithm used to find relatively homogeneous clusters of input data based on measured attributes [14,15]. This method employs a bottom-up way ("ascendant") to iteratively aggregate a pair of nearest clusters in order to have one final cluster containing all input variables. The method is outlined as follows: Let's consider two data clusters, x and y, with an inter-cluster distance as d xy . We assumed that cluster x contains N i observations and cluster y N y observations. If d xy is the smallest distance that satisfies the conditions, (i) x = y ⇔ d(x, y) = 0 , and (ii) d(x, y) = d(y, x), clusters x and y are merged into a new cluster, p. Once a new cluster is created, the new minimum distance between p and any other cluster q is calculated using the Lance-Williams formula: where the values of α x , α y , β, and γ are calculated using Ward's minimum variance method [16]: where N x , N q , and N q are the number of elements in the clusters x, p, and q, respectively. For the minimum distance, d pqmin , the closest two clusters are merged further into a new cluster. These steps are repeated until a single cluster with all observations remains.
For AHC, we gradually increased the number of classes into which the data was grouped from 10 to 50. At the beginning of the merging process, we considered each class as a separate cluster. Based on Ward's aggregation criteria, we iteratively merged the clusters until all the remaining clusters matched with the observed lithofacies classes.

B. Self-Organizing Maps (SOM)
Kohonen, in 1975, first introduced an unsupervised learning method that uses neural network-based algorithms for visualizing high-dimensional data in low dimensions. The SOM algorithm implemented in this study considers an adaptive learning process in which neurons are allowed to learn to represent different input data vectors. The neuron that best represents a selected input vector is considered as the winning neuron, and the neighboring neurons are gradually allowed to learn to represent similar inputs. Each neuron is then placed at the nodes of a one or two-dimensional lattice to transform multi-dimensional data into a low-dimensional discrete map.
This SOM-based data learning process can be expressed by the following equation [17]: where W i is the synaptic weight vector of the winning neuron (i ∈ (0 . . . number of neurons)), η is the learning rate that controls the size of the weight vector, x is the randomly identified input vector, and N(i, x) is a neighborhood function that determines the rate of change of the neighborhood around the winning neuron. N(i, x) is further defined as: where d(i, w) represents the Euclidean distance between the winning neuron and any ith neuron in the neighborhood, and λ is the iteration limit. It suggests that the neighborhood function returns to 1 for the winning neuron, which receives the most training, and is reduced to zero for neurons that are further away from the winning neuron, thereby receiving less training. These steps are repeated until the learning rates and neighborhood get smaller. When the neural network is fully trained, neurons that are similar are placed together, whereas neurons that are not similar are placed apart. The net result is a two-dimensional map of clusters, showing high-density nodes with similar inputs. We trained SOM classifiers with a rectangular topography and with Euclidean distance as the distance function. The topology of the output layer was selected based on the training data and involved three schemes: 2 × 2, 5 × 5, and 7 × 7 neurons. In each case, the initial number of clusters were iteratively merged to five to match with the observed facies classes.

C. Artificial Neural Network (ANN)
A conventional back-propagation ANN algorithm is widely used to identify complex nonlinear relationships between input and output data, especially in pattern recognition applications [18][19][20]. ANN is a three-layer neural network structure with many interconnected nodes consisting of an input layer that collects the input data values, a hidden layer that performs the learning process via a system of weighted connections, and an output layer that shows the desired output, i.e., predicted patterns based on ANN's response to the input parameters.
The number of hidden layers and neurons in the hidden layers primarily influence the weighted connections and are, therefore, the most critical parameters in the ANN design. During the training phase of ANN, initial weights are usually randomly assigned between nodes in the first layer and are adjusted to minimize the differences between the actual and predicted outcomes. A weight decides how much the input value affects the layer output. An activation function is further applied to introduce nonlinearity by computing the weighted sum of input and biases at each neuron before passing the output to the next hidden layer. The resulting non-linear output is given by, where x is the input, w is the corresponding weight to x, b is the bias value, and α is the activation function. When the backpropagation algorithm is used, the calculated error between the actual and desired output is propagated backward through hidden layers down to the input layer. Once the error for a hidden layer is known, the weights between the nodes are updated. This process is continued until the error is small enough across the hidden layers. When the model is trained to the desired level, the weight values are saved and stored before applying them to a new set of input data.
In this study, we implemented the ANN classifiers with one hidden layer and using a simple backpropagation algorithm. The neurons in the hidden layer were varied from 2 to 10 to examine the modes' prediction accuracy. We trained ANN models until we did not see any improvement in the model performance, compared to the observed data.

D. Multi-Resolution Graph-Based Clustering (MRGC)
Unlike conventional methods, MRGC is an unsupervised clustering technique that automatically calculates an optimal number of clusters independent of prior knowledge of input data. This method uses k-nearest neighbors and kernel index statistical parameters to identify the best number of dot clusters. Dots are representations of individual data points in a user-defined space to reveal data patterns in low-dimensions.
To develop a dot cluster, the proximity between each data point (x) to its nearest neighbor and other points in the space is identified using the nearest index (NI) [21,22] as expressed below: where y is pth nearest neighbor of x among all data points, q is the total data points in the space, and k = 1, 2, . . . , K − 1 for x's K nearest neighbors. The sum of the rank of x relative to its pth nearest neighbor is expressed as: The smallest sum of the limited ranks (σ(x)) is expressed as: where i represents the number of measurement points from 1 up to K. The largest sum of the limited ranks (σ(x)) is similarly expressed as: Although NI provides a rapid estimate of nearest neighbor density at each data point in large data sets, it does not learn from the training data to do any generalization.
To overcome this issue, MRGC considers another function-kernel representation index (KRI)-which uses the weighted distance (M) and distance (D) between data points x and y to generate cluster indices. KRI at the point x is expressed by Rahimpour-Bonab and Aliakbardoust [21] as: where M(x,y) is the number of neighbors, i.e., p, as y being pth neighbor of x, and D(x,y) is the distance between x and y. If we consider m as any point from the nearest neighbors of x, and n is a point whose NI value is greater than m, then m merges all points around it, but it does not merge the point n. In this way, the kernel of a cluster is identified. When an appropriate kernel and its nearby neighbors are identified, cluster boundaries are set. Kernels are, therefore, localized in space and do not influence data points beyond the cluster boundaries. To obtain MRGC models, we considered the optimal model number to be between 3 and 5.

Results and Discussion
Our study aimed to examine the potential of different machine learning methods to predict lithofacies from well logs using limited core data from the Umiat Oil Field. Figure 3 shows the predicted lithofacies from the well log data of Umiat 18 using different machine learning methods, compared to the observed core-based lithofacies. Our results show that SOM provides facies predictions, closely mimicking the observed facies among all examined machine learning methods.

Model Comparison
In this study, training sample data that we used to evaluate the performance of prediction models included 62 samples. Table 3 shows a range of parameters considered for a number of training trials for each machine learning technique. All models except MRGC were able to predict the lithofacies with a higher value of R 2 (R 2 > 0.70). Regarding the model performance, SOM was able to predict the facies with the highest accuracy compared to both AHC and ANN.

Model Comparison
In this study, training sample data that we used to evaluate the performance of prediction models included 62 samples. Table 3 shows a range of parameters considered for a number of training trials for each machine learning technique. All models except MRGC were able to predict the lithofacies with a higher value of R 2 (R 2 > 0.70). Regarding the model performance, SOM was able to predict the facies with the highest accuracy compared to both AHC and ANN. Table 3. R 2 values for each of the combinations of machine learning algorithm parameters used in this study. SOM with 7 neurons was observed to be the best among the examined methods. Our model outputs further suggest the training with SOM tends to provide better results with increasing numbers of neurons. From a rectangular topology, the best SOM scheme with 7 × 7 neurons achieved a better clustering quality (R 2 = 0.90) (Figure 4, Table 3). It is, therefore, important to select the neurons in the output layer equal to the number of input classes from the training data. For AHC, the best result was obtained when the number of input classes was relatively higher and close to the number of samples in the training set (R 2 = 0.85). Figure 5 shows that the number of classes strongly influences the aggregation style and the resulting facies models. We examined ANN using a smaller training set (62 samples) and one hidden layer. The number of neurons in the hidden layer was varied from 2 to 10. When only two were considered in a hidden layer, the model produced the lowest goodness-of-fit to the observed data (R 2 = 0.65). Increasing numbers of neurons from 2 to 10 in a hidden layer did not significantly affect the model output (R 2 = 0.76). Table 3 shows the influence of the number of neurons per hidden layer on the model output compared to the observed facies. For a smaller dataset, we recommend using at least two neurons in a hidden layer to improve the accuracy of ANN. The best value of R 2 for MRGC was obtained using five optimal models with 17 clusters (R 2 = 0.68). The poor performance of the MRGC was due to the small number of training samples and the irregularity in the training data, especially the uneven distribution of sandstone facies in the cored section of Umiat 18.

Lithofacies Prediction and Validation in Test Wells
In order to predict the lithofacies in uncored wells-Umiat 23H and SeaBee-1-we applied the best SOM model to the well data from these wells. Figure 6 shows the predicted lithofacies for both wells. Our initial observations suggest that lithofacies identified from the core data of Umiat 18 also occur in Umiat 23H and SeaBee-1 but at different depths. In both wells, sandstone lithofacies show a strong vertical heterogeneity.
Energies 2020, 13, x FOR PEER REVIEW 10 of 16 of ANN. The best value of R 2 for MRGC was obtained using five optimal models with 17 clusters (R 2 = 0.68). The poor performance of the MRGC was due to the small number of training samples and the irregularity in the training data, especially the uneven distribution of sandstone facies in the cored section of Umiat 18.   Figure 5 shows the dendrograms generated using a range of classes with Ward's aggregation criteria. Please refer to Table 1 for facies colors.

Lithofacies Prediction and Validation in Test Wells
In order to predict the lithofacies in uncored wells-Umiat 23H and SeaBee-1-we applied the best SOM model to the well data from these wells. Figure 6 shows the predicted lithofacies for both wells. Our initial observations suggest that lithofacies identified from the core data of Umiat 18 also occur in Umiat 23H and SeaBee-1 but at different depths. In both wells, sandstone lithofacies show a strong vertical heterogeneity.  Figure 5 shows the dendrograms generated using a range of classes with Ward's aggregation criteria. Please refer to Table 1 for facies colors.

Lithofacies Prediction and Validation in Test Wells
In order to predict the lithofacies in uncored wells-Umiat 23H and SeaBee-1-we applied the best SOM model to the well data from these wells. Figure 6 shows the predicted lithofacies for both wells. Our initial observations suggest that lithofacies identified from the core data of Umiat 18 also occur in Umiat 23H and SeaBee-1 but at different depths. In both wells, sandstone lithofacies show a strong vertical heterogeneity. Figure 6. Umiat test wells (Umiat 23H and SeaBee-1), showing predicted lithofacies from the available well data using the SOM algorithm with 7 × 7 neurons.
To validate the presence of our predicted lithofacies, we examined the available seismic data across SeaBee-1. The predicted facies in Umiat 23H were not cross-validated due to the poor seismic data quality in the vicinity of the well. Figure 7 shows a synthetic seismogram constructed using sonic and density logs of SeaBee-1 in comparison to the predicted lithofacies for LGST in the well. Our predicted sandstone facies within LGST can be seen on the seismogram as negative (trough) amplitudes. To evaluate the hydrocarbon storage potential of these sandstone facies, we performed seismic amplitude variation with offset (AVO) analysis on the top sand in LGST using pre-stack seismic gathers. Figure 8 shows a strong AVO response along the horizon near the sand top. The To validate the presence of our predicted lithofacies, we examined the available seismic data across SeaBee-1. The predicted facies in Umiat 23H were not cross-validated due to the poor seismic data quality in the vicinity of the well. Figure 7 shows a synthetic seismogram constructed using sonic and density logs of SeaBee-1 in comparison to the predicted lithofacies for LGST in the well. Our predicted sandstone facies within LGST can be seen on the seismogram as negative (trough) amplitudes. To evaluate the hydrocarbon storage potential of these sandstone facies, we performed seismic amplitude variation with offset (AVO) analysis on the top sand in LGST using pre-stack seismic gathers. Figure 8 shows a strong AVO response along the horizon near the sand top. The computed background gradient for this particular set of gathers was −0.003, whereas the computed gradient for the displayed anomaly at the Lower Grandstand was −0.250, which was significantly higher than the background (Figure 9). The average stacked amplitude for the Lower Grandstand was roughly −300, whereas the stacked amplitude for the gather, exhibiting a strong negative gradient, was −625, which doubled the average value. The combination of a strong negative gradient in the pre-stack domain with doubling of amplitude in the stacked domain are strong indicators for a Class III AVO response at this particular Lower Grandstand location. These Class III AVO anomalies along the southern upthrown fault block of the Umiat anticline suggest a potential gas accumulation zone within the Lower Grandstand Member (Figures 2 and 10).
was roughly −300, whereas the stacked amplitude for the gather, exhibiting a strong negative gradient, was −625, which doubled the average value. The combination of a strong negative gradient in the pre-stack domain with doubling of amplitude in the stacked domain are strong indicators for a Class III AVO response at this particular Lower Grandstand location. These Class III AVO anomalies along the southern upthrown fault block of the Umiat anticline suggest a potential gas accumulation zone within the Lower Grandstand Member (Figures 2 and 10).   in the pre-stack domain with doubling of amplitude in the stacked domain are strong indicators for a Class III AVO response at this particular Lower Grandstand location. These Class III AVO anomalies along the southern upthrown fault block of the Umiat anticline suggest a potential gas accumulation zone within the Lower Grandstand Member (Figures 2 and 10).

Conclusions
In our study, we demonstrate the application of four machine learning techniques-ascending hierarchical clustering, self-organizing maps, artificial neural network, and multi-resolution graph-