Next Article in Journal
Convolutional Neural Network Applications in Current Sensor Fault Classification Mechanisms in Permanent Magnet Synchronous Motor Drive Systems
Previous Article in Journal
The Correlation Between Self-Assessment and Observation in Driving Style Classification: An On-Road Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

G-TS-HRNN: Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network

by
Omar Bahou
,
Mohammed Roudani
and
Karim El Moutaouakil
*,†
Laboratory of Mathematics and Data Science, Polydisciplinary Faculty of Taza, Sidi Mohamed Ben Abdellah of Fez, Taza P.O. Box 1223, Morocco
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Information 2025, 16(2), 141; https://doi.org/10.3390/info16020141
Submission received: 24 December 2024 / Revised: 24 January 2025 / Accepted: 26 January 2025 / Published: 14 February 2025

Abstract

:
The Hopfield Recurrent Neural Network (HRNN) is a single-point descent metaheuristic that uses a single potential solution to explore the search space of optimization problems, whose constraints and objective function are aggregated into a typical energy function. The initial point is usually randomly initialized, then moved by applying operators, characterizing the discrete dynamics of the HRNN, which modify its position or direction. Like all single-point metaheuristics, HRNN has certain drawbacks, such as being more likely to get stuck in local optima or miss global optima due to the use of a single point to explore the search space. Moreover, it is more sensitive to the initial point and operator, which can influence the quality and diversity of solutions. Moreover, it can have difficulty with dynamic or noisy environments, as it can lose track of the optimal region or be misled by random fluctuations. To overcome these shortcomings, this paper introduces a population-based fuzzy version of the HRNN, namely Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network (G-TS-HRNN). For each neuron, the G-TS-HRNN associates an input fuzzy variable of d values, described by an appropriate Gaussian membership function that covers the universe of discourse. To build an instance of G-TS-HRNN(s) of size s, we generate s n-uplets of fuzzy values that present the premise of the Takagi–Sugeno system. The consequents are the differential equations governing the dynamics of the HRNN obtained by replacing each premise fuzzy value with the mean of different Gaussians. The steady points of all the rule premises are aggregated using the fuzzy center of gravity equation, considering the level of activity of each rule. G-TS-HRNN is used to solve the random optimization method based on the support vector model. Compared with HRNN, G-TS-HRNN performs better on well-known data sets.

1. Introduction

Preoccupations with computational accuracy, speed, and engineering completeness have prompted scientists to investigate soft computing approaches to modeling, forecasting, and monitoring nonlinear dynamic systems. Both artificial neural networks (ANNs) and fuzzy logic approaches are widely employed as techniques for soft computation. The marriage of such techniques is being applied in numerous areas of science and engineering to tackle real-world challenges, in particular, optimization problems. The utilization of fuzzy logic has the potential to enhance the ability of a machine learning system to reason and infer. Categorical insights, however vague, may be modeled to permit the symbolic articulation of automatic learning through fuzzy logic. The use of neural networks incorporates learning capability, robustness, and massive parallelism into the system. The representation of expert knowledge and the machine-learning capacities of the neuro-fuzzy network turn it into a potent framework to tackle automatic learning challenges [1,2]. Currently, training techniques and the architecture of traditional neuro-fuzzy nets have been refined to deliver higher levels of performance in terms of accuracy and learning time.
On the one hand, the Takagi–Sugeno–Kang (TSK) induction scheme is the leading fuzzy inference system and a powerful tool for building complex nonlinear dynamical systems. The main advantage of TSK modeling is its “multimodal” nature: it is able to merge nonlinear submodels to describe the global dynamics of a complex system [3]. On the other hand, the Hopfield Recurrent Neural Network (HRNN) is distinctive as a particular sort of recurrent neural network that has grown rapidly and gained significant interest among scientists in a range of fields, especially optimization [4,5,6,7,8]. This network has shown itself to be extremely efficient over other approaches, notably in delivering high-performance answers to mathematical optimization problems [9,10,11,12,13,14,15].
This paper introduces a population-based fuzzy version of the Hopfield recurrent neural network, namely the Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network (G-TS-HRNN).

1.1. Hopfield Recurrent Neural Network: Main Knowledge

The HRNN is a semantically recurrent neural network consisting of a monolayer of n units that are completely interconnected, where each neuron has an activating function. This network is considered a driven system characterized using a dynamic equation specified as follows [9]:
d u d t = T v + I
In Equation (1), T R n × n represents the link strength of the neurons, I R n is the bias, u is the state vector of different units, and v is the vector activations of the different units. The correlation that exists between the vector of states u and their activations v can be expressed by the activation function v = g ( u ) (generally a hyperbolic tangent), limited at the lower end with 0 and higher end with 1 and has a soft threshold.
Definition 1.
A state vector u e is referred to as the equilibrium point of Equation (1) when, for a given entry vector u 0 , u e fulfills the following conditions: u ( t ) = u e t t e at some t e 0 .
The E L y a p energy function can be employed to investigate the stability of HRNN. This function serves to demonstrate whether the network attains a steady state, as it must diminish as time goes by ( d E L y a p d t 0 ) in order for the system to tend towards a minimum. Such a Lyapunov function is expressible as shown below [9]:
E L y a p ( v ) = 1 2 v t T v I t v
The steady-state of the HRNN network would exist if the energy function exists. In 1984, Hopfield proved that if the matrix T is symmetric, then the energy function E L y a p holds.
The solution of optimization problems employing the HRNN usually requires the creation of an energy function. This function represents a mathematical description of the problem to be resolved [16]. The local minima of this energy function refers to the local optimal decision of the optimization problem [9,17,18]. If f and g j , j = 1 , , m , are the objective function and the set of constraints, respectively, of the optimization problem, a possible energy function can be defined as follows:
E ( v ) = α f ( v ) + i ϕ i g i ( v ) + i ϕ i , j g i ( v ) g j ( v )
where α , ϕ i , and ϕ i , j are the penalty parameters used to control the amplitude of f and the violated constraints. The optimization of the E cost function requires appropriate parameters that guarantee the feasibility of the solutions. These parameters can be extracted in a suitable way on the basis of the first and second derivatives of the E function, using the hyperplane method, which prevents the system from reaching stability outside the feasible set delimited by the constraints of the optimization problem [9,12,13,15].

1.2. State of the Art

The Hopfield Recurrent Neural Network (HRNN) is widely used to solve Optimization Problems (OPs) based on the appropriate energy function. A local solution of the OP is obtained by solving the differential equation governing the dynamics of HRNN using numerical methods. This method falls within the descent approaches that consider only one individual at each iteration. Recall that the dynamics of HRNN is governed by the equation:
d u ( t ) d t = T v ( t ) + I , u ( 0 ) = u 0 R n
where for a given u 0 > 0 ,
v i ( t ) = e 2 u i ( t ) u 0 1 e 2 u i ( t ) u 0 1 + 1 i = 1 , , n
To solve Equation (4), using the Euler–Cauchy method, each sub-equation is discretized using a time step called δ t [19]. This procedure implies a substantial local truncation error of the order of O ( δ t 2 ) per pass, requiring the implementation of a very small step width at the expense of reduced speed and higher accuracy errors in floating-point calculations. It is also vulnerable to modifications of any scale [10]. In this way, the HRNN size, the magnitudes of weights, biases, and the gain parameter of the HRNN must be considered to assign the step size in the Euler method. In spite of all these objections, some authors have proposed a very small time step, which requires a lot of CPU time to obtain a local minimum and prevent the HRNN from escaping unpromising regions [20]. This is why some authors suggest increasing δ t to save computing time and explore more promising regions. For example, the authors of [21,22,23] believe that δ t [ 0.0001 , 0.01 ] is an accurate value in their calculation trials. However, with such a wide range of values, the trajectories can be irregular, skipping towards other attractors in the basin. This last property, however, can be used to improve solution quality by escaping from local minima, see Wang and Smith [24]. Given the increasing size of the HRNN, the Euler method and the Runge–Kutta method can be very slow in the search for a local optimum [11]. In [10], the authors introduce an algorithm with the optimal time step for HRNN to obtain an HRNN equilibrium point. This algorithm computes an optimal time-step in each iteration so that a rapid convergence to an equilibrium point is achieved. This technique has been widely used in real-world problems [12,14,15,25]. However, in the case of a non-quadratic objective function or constants, calculating the optimal time step becomes very difficult, requiring the use of fixed-point or Newton methods or quasi-Newton methods to estimate the optimal time step of the Euler–Cauchy algorithm [26,27,28]. However, calling several numerical functions at each iteration can make the search procedure very slow and does not guarantee convergence to a local solution.
Through these works, HRNN can be thought of as a single-point descent metaheuristic that uses a single potential solution to explore the search space. The point is usually randomly initialized and then moved by applying operators that modify its position or direction. Like all single-point metaheuristics, HRNN has certain drawbacks, such as being more likely to get stuck in local optima or miss global optima due to the use of a single point to explore the search space. In addition, it is more sensitive to the initial point and operator, which can influence the quality and diversity of solutions. Moreover, it can have difficulty with dynamic or noisy environments, as it can lose track of the optimal region or be misled by random fluctuations.

1.3. Contributions

To overcome the shortcomings discussed in Section 1.2, this paper introduces a fuzzy-neuro optimization method, namely Gaussian Takaji–Sugeno Hopfield Recurrent Neural Network (G-TS-HRNN). This is a population-based metaheuristic that has certain advantages over single-point metaheuristics, including classical HRNN.
To help classical HRNNs escape local optima while exploring promising regions of the search space, we have introduced a family of input fuzzy variables modeling variations in the quadratic factors of the system of Equation (4) governing the dynamics of HRNN. The initial HRNN stability point is approximated by fuzzy rules with linear differential equations as a consequence.
Sampling via the non-linear sector associated with the Takagi–Sugeno fuzzy model allows the initial HRNN to be decomposed into independent linear sub-problems, enabling parallel or distributed computing to speed up the search process and handle large-scale problems. Furthermore, to deal with the stochastic characteristics of real environments, we can strike an appropriate balance between exploration and exploitation when generating the fuzzy membership of Takaji–Sugeno fuzzy variables.
More precisely, to construct G-TS-HRNN, we associate a d-valued fuzzy input variable with each HRNN unit, described by an appropriate Gaussian membership function, that covers the universe of discourse determined by the slope of the HRNN neurons’ activation functions. To build an instance of G-TS-HRNN(s) of size s, we generate s n-uplets of fuzzy values that represent the premise of the Takagi–Sugeno system. The consequents are the differential equations governing the dynamics of the HRNN obtained by replacing each premise fuzzy value with the mean of different Gaussians. The steady points of all the rules consequents are aggregated using the fuzzy center of gravity equation, considering the level of activity of each rule. G-TS-HRNN is used to solve the random optimization method and solve the classification problem based on the support vector mathematical model. To ensure the feasibility of the decisions delivered by the considered models, we used the hyperplane technique introduced in [9] and used successfully in [12,15,25]. HRNN and G-TS-HRNN were tested on nine data sets found in the UCI (Machine Learning Repository) machine learning database [29]. The performance metrics considered in this work are accuracy, F1-measure, precision, and recall. Compared with the classical HRNN, G-TS-HRNN performs better on well-known data sets.
Towards the end, three choices need to be justified: (a) the nature of membership functions associated with the non-linear sector, (b) their number, and (c) their parameters. (a) Gaussian Mixture Models (GMMs) are considered universal approximators of densities, meaning they can approximate any smooth density function to an arbitrary degree of accuracy given a sufficient number of components [30]. This property makes GMMs highly versatile in modeling complex distributions in various fields, including statistics and machine learning [31]. (b) The choice of the number of Gaussians must strike a balance between exploration, diversification, and complexity of the G-TS-HRNN; in our case, this selection is made experimentally. (c) When designing membership functions for fuzzy variables, certain conditions must be satisfied to ensure system accuracy and coherence [32]. Membership functions must collectively cover the entire universe of discourse. Parameters must be chosen such that overlap between adjacent membership functions is essential to enable smooth transitions, while the principle of completeness ensures that sets that overlap at any point add up to a meaningful total, often 1 [33].
Outline: Section 2 introduces the sector nonlinearity of the Takagi–Sugeno fuzzy model. Section 3 presents the Gaussian Takagi–Sugeno model applied to the Hopfield Recurrent Neural Network (G-TS-HRNN). Section 4 discusses the application of G-TS-HRNN to the classification problem using Support Vector Machines (SVMs). Section 5 outlines the experimentation and implementation conditions. Finally, Section 6 provides conclusions and future directions.

2. Sector Nonlinearity to Takagi–Sugeno Fuzzy Model

To build a Takagi–Sugeno pattern, we first have to design a T-S fuzzy map of a nonlinear model. Therefore, building a fuzzy model is an essential and fundamental task in this process. In general, two different strategies can be adopted to develop fuzzy models [34,35,36]: (a) identification (fuzzy modeling) based on entry-exit data, (b) deducing the governing equations of a specified non-linear operating system [37]. The recognition process of fuzzy modeling is tailored to plants that can or can not sufficiently be captured by analytical and/or physical models. In addition, nonlinear dynamic models for mechanical engineering systems may be simply derived, for instance, using the Lagrange and the Newton–Euler procedure. In such cases, the second approach, which derives a fuzzy model from given nonlinear dynamical models, is more appropriate. This brief review covers the second approach. This is an approach that employs the concept of “sectoral non-linearity”, “local approximation”, or a mixture of the two to generate fuzzy models.
Sector non-linearity revolves around the next idea. Take the following non-linear model:
d x ( t ) d t = f ( x ) , x 0 R n
where f is a non-linear function. The aim is to find the global sector such that
d x ( t ) d t = f ( x ) A x ( t ) , x 0 R n
where A is a matrix free of x.
Figure 1 illustrates the sector nonlinearity approach in the case when A = [ a 1 , a 2 ] . This ensures the accurate building of the fuzzy model.
Suppose we can write Equation (5) as follows:
d x ( t ) d t = B ( x ) x ( t ) , x 0 R n
where B ( x ) = ( b i j ( x ) ) i , j = 1 , , n and b i j ( x ) [ b ̲ i j , b ¯ i j ] . Here, b i j , i , j = 1 , , n are nonlinear terms in Equation (5), which represent the fuzzy variables that we name z i j and are called premise variables. The starting point in any fuzzy design is to establish the fuzzy variables and the fuzzy sets or membership functions. While there is no generic way of doing this, and it may be carried out by a number of methods, primarily trial-and-error, in precise fuzzy modeling using sectorial nonlinearity, it is entirely routine. In this tutorial, we assume that the premise variables are only functions of the state variables for reasons of simplicity [38].
To acquire membership functions, we should calculate the minimum and maximum values of z i j , which are obtained as follows:
m i n ( z i j ) = b ̲ i j and m a x ( z i j ) = b ¯ i j Therefore, z i j can be represented by two membership functions M i j ( 1 ) and M i j ( 2 ) as follows: z i j = b ̲ i j M i j ( 1 ) + b ¯ i j M i j ( 2 ) , i , j such that M i j ( 1 ) + M i j ( 2 ) = 1 . We associate M i j ( 1 ) with the fuzzy value “Small” and M i j ( 2 ) with the fuzzy value “Big”.
To each sample s { S m a l l , B i g } n × n , we introduce the following Takaji–Sugeno fuzzy rule:
If z 1 , 1 is s 1 , 1 and… z n , n is s n , n , Then, d x ( t ) d t = B s x ( t ) , x 0 R n . where B i j s = b ̲ i j if s i j = S m a l l else B i j s = b ¯ i j if s i j = B i g .
d x ( t ) d t = B 1 x ( t ) , x ( 0 ) R n .
d x ( t ) d t = B 2 x ( t ) , x ( 0 ) R n .
For a random x 0 R n , to obtain an estimation x of the steady state of the system (5), we process it as follows:
(a) Generate k samples s 1 { S m a l l , B i g } n × n , , s k { S m a l l , B i g } n × n ;
(b) Calculate the steady state of the premises of the rules associated with these samples; let x ( 1 ) , , x ( k ) be these solutions obtained using the Euler–Cauchy method [9];
(c) Calculate the activation weight w s of a different rule for s = 1 , , k . As the elementary conditions of different rules are linked by the “and” operator, the w s = M 1 , 1 ( s ) ( z 1 , 1 ( s ) ) M n , n ( s ) ( z n , n ( s ) ) ; where z i j ( s ) = b i j ( x ( s ) )
(d) Use the fuzzy Center of gravity to estimate x: x = s = 1 k w s x ( s ) s = 1 k w s .

3. Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network (G-TS-HRNN)

Considering the equation of the activation function of the HRNN neurons (hyperbolic tangente tangent), the dynamical system (1) can also be expressed as:
d v i d t = 2 u 0 v i ( 1 v i ) ( j = 1 n T i j v j + I i ) , i = 1 , , n v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , , n
To transform Equation (6) into a linear equation, we introduce n fuzzy sets defined on the following variables:
z i = 2 u 0 v i ( 1 v i ) , i = 1 , , n
We have m i n ( z i ) = 0 and m a x ( z i ) = 1 u 0 ; we set U = m : s 0 : M and d = s i z e ( U ) , where s 0 = 1 p u 0 and p is a positive integer and 0 m M 1 u 0 . For each z i , we introduce d fuzzy values F 1 , , F d that we associate with d Gaussian G 1 ( m e a n = m ) , , G 1 ( m e a n = m + l s 0 ) , , G d ( m e a n = m + ( d 1 ) s 0 = M ) , respectively.
Example 1.
For U = 1 10 u 0 : 3 10 u 0 : 10 10 u 0 , we have d = 4 and we obtain the Gaussian membership functions shown in Figure 2 (obtained for a standard diviation of 1 7 u 0 ). In this case, we can say that F 1 = V e r y _ S m a l l , F 2 = S m a l l , F 3 = B i g , and F 4 = V e r y _ B i g .
We pull N samples π 1 = ( a 1 1 , , a n 1 ) , , π s = ( a 1 s , , a n s ) , , π N = ( a 1 N , , a n N ) from U n , where n is the number of neurons of a given HRNN. For each sample s, we denote by π s . a i s the fuzzy value associated with the Gaussian membership G ( mean = a i ) . We work with a Gaussian of the same standard deviation chosen to cover the universe of discourse.
Example 2.
We consider the same example as before and we pull the following five samples:
π 1 = ( a 1 1 , , a n 1 ) , π 2 = ( a 1 2 , , a n 2 ) , π 3 = ( a 1 3 , , a n 3 ) , π 4 = ( a 1 4 , , a n 4 ) , π 5 = ( a 1 5 , , a n 5 ) . If, in addition, the considered HRNN contains three neurons, we can express our sample as follows:
π 1 = ( 4 10 u 0 , 1 10 u 0 , 7 10 u 0 ) , π 2 = ( 4 10 u 0 , 4 10 u 0 , 4 10 u 0 ) , π 3 = ( 7 10 u 0 , 10 10 u 0 , 7 10 u 0 ) ) ,
π 4 = ( 1 10 u 0 , 4 10 u 0 , 7 10 u 0 ) ) , π 5 = ( 7 10 u 0 , 7 10 u 0 , 10 10 u 0 ) ) .
For each sample π s = ( a 1 s , , a n s ) , we introduce the following rule R ( π s ) :
if z 1 is π s . a 1 s and z i is π s . a i s and … and z n is π s . a n s ,
then d v i d t = a i s ( i = 1 n T i j v i + I i ) , i = 1 , , n v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , , n
Example 3.
To illustrate the introduced rules, we continue with the same example as before.
R ( π 1 ) : i f z 1 i s S m a l l a n d z 2 i s V e r y _ S m a l l a n d z 3 i s B i g t h e n d v 1 d t = 4 10 u 0 ( j = 1 3 T 1 , j v j + I 1 ) d v 2 d t = 1 10 u 0 ( j = 1 3 T 2 j v j + I 2 ) d v 3 d t = 7 10 u 0 ( j = 1 n T 3 , j v j + I 3 ) v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , 2 , 3
R ( π 2 ) : i f z 1 i s S m a l l a n d z 2 i s S m a l l a n d z 3 i s S m a l l t h e n d v 1 d t = 4 10 u 0 ( j = 1 3 T 1 , j v j + I 1 ) d v 2 d t = 4 10 u 0 ( j = 1 3 T 2 j v j + I 2 ) d v 3 d t = 4 10 u 0 ( j = 1 n T 3 , j v j + I 3 ) v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , 2 , 3
R ( π 3 ) : i f z 1 i s B i g a n d z 2 i s V e r y _ B i g a n d z 3 i s B i g t h e n d v 1 d t = 7 10 u 0 ( j = 1 3 T 1 , j v j + I 1 ) d v 2 d t = 10 10 u 0 ( j = 1 3 T 2 j v j + I 2 ) d v 3 d t = 7 10 u 0 ( j = 1 n T 3 , j v j + I 3 ) v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , 2 , 3
R ( π 4 ) : i f z 1 i s V e r y _ S m a l l a n d z 2 i s S m a l l a n d z 3 i s B i g t h e n d v 1 d t = 1 10 u 0 ( j = 1 3 T 1 , j v j + I 1 ) d v 2 d t = 4 10 u 0 ( j = 1 3 T 2 j v j + I 2 ) d v 3 d t = 7 10 u 0 ( j = 1 n T 3 , j v j + I 3 ) v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , 2 , 3
R ( π 5 ) : i f z 1 i s B i g a n d z 2 i s B i g a n d z 3 i s V e r y _ B i g t h e n d v 1 d t = 7 10 u 0 ( j = 1 3 T 1 , j v j + I 1 ) d v 2 d t = 7 10 u 0 ( j = 1 3 T 2 j v j + I 2 ) d v 3 d t = 10 10 u 0 ( j = 1 n T 3 , j v j + I 3 ) v i ( 0 ) = v i 0 [ 0 , 1 ] , i = 1 , 2 , 3
Thus, the number of rules is exactly the number of samples. For each sample s, let v s be the solution of the conclusion of the rule R ( π s ) .
For all s = 1 , , N and i = 1 , , n , we have z i s = 2 u 0 v i s ( 1 v i s ) . The degree of each premise “ z i is a i ” is given by:
G ( a i s , σ , z i s ) = 1 σ 2 π e 1 2 z i s a i s σ 2
As the elementary premises of R ( π s ) are linked via the “AND” operator, the weight of v s in the Sugeno mean solution is obtained by the “min” operator, denoted ∧. In this sense, an approximation of the solution of Equation (6) is given by:
v s = s = 1 N w s v s s = 1 N w s
where w s = i = 1 n G ( a i s , σ , z i s ) .
Example 4.
To illustrate the introduced rules, we continue with the same example as before.
w 1 = G 4 10 u 0 , σ , z 1 1 G 1 10 u 0 , σ , z 2 1 G 7 10 u 0 , σ , z 3 1 ,
w 2 = G 4 10 u 0 , σ , z 1 2 G 4 10 u 0 , σ , z 2 2 G 4 10 u 0 , σ , z 3 2 ,
w 3 = G 7 10 u 0 , σ , z 1 3 G 10 10 u 0 , σ , z 2 3 G 7 10 u 0 , σ , z 3 3 ,
w 4 = G 1 10 u 0 , σ , z 1 4 G 4 10 u 0 , σ , z 2 4 G 7 10 u 0 , σ , z 3 4 ,
w 5 = G 7 10 u 0 , σ , z 1 5 G 7 10 u 0 , σ , z i 5 G 10 10 u 0 , σ , z i 5 .
For a given initial state v ( 0 ) = ( v 1 ( 0 ) , v 2 ( 0 ) , v 3 ( 0 ) ) , we can use the Euler–Cauchy method to estimate v s , s = 1 , 2 , 3 , 4 , 5 , which will allow us to calculate z s , s = 1 , 2 , 3 , 4 , 5 . Then, for a given standard deviation σ, we can calculate w s , s = 1 , 2 , 3 , 4 , 5 . Finally, we obtain an estimation of the equilibrium point of Equation (6).
Figure 3 gives the main steps to obtain an approximation of the steady state of the HRNN using Gaussian Takagi–Sugeno HRNN sampling: (a) introduction of the n fuzzy z i sets, (b) generation of a set of N samples by considering a blind k-Gaussian membership function for each sample, (c) calculation of the steady state of different samples using the Euler–Cauchy method, (d) calculation of the activation level of the rule associated with each sample, and (e) estimation of the HRNN steady state using the fuzzy center of gravity.
Graphically, Figure 4 illustrates the idea behind the G-TS-HRNN method. Indeed, a direct calculation of the HRNN equilibrium point may lead to an Euler–Cauchy attraction to a bad local minimum due to a bad initialization or an inappropriate time step. Despite poor initialization, the generation of several G-TS-HRNNs allows us to cover multiple regions of the search space. The fuzzy combination, using the degree of activation of the different rules, provides an excellent approximation of the HRNN equilibrium state.

4. G-TS-HRNN for the Classification Problem Using Vector Support Machine

Let D R n be a set of N samples x 1 , , x N labeled, respectively, by y 1 , , y N , distributed across K classes C 1 , , C K . In our case, K = 2 and y i { 1 , + 1 } .
The soft support vector machine (SVM) looks for a separator of the equation w · x i + b = 0 , where w is the weight such that i = 1 , , N , y i ( x i · w + b ) 1 . To ensure a maximum margin, you need to maximize 2 w . As the patterns are not linearly separable, a kernel function K (which satisfies the Mercer conditions [39]) is introduced.
By introducing the Lagrange relaxation and writing the Kuhn–Tucker conditions, we obtain a quadratic optimization problem with a single linear constraint, which we solve to determine the support vectors [40,41,42]:
Max < 1 , α > 1 2 ( α y ) K ( α y ) Subject to : < α , y > = 0 0 α C , i = 1 , , N
where α = ( α 1 , , α N ) are the Lagrange parameters, and C is a constant. To solve the obtained dual problem via HRNN [15], we propose the following energy function:
E ( α ) = β 0 < 1 , α > β 0 2 ( α y ) K ( α y ) + β 1 < α , y > + β 2 2 ( < α , y > ) 2
To determine the vector of the neuron biases, we calculate the partial derivatives of E:
E α i ( α ) = β 0 β 0 i = 1 N α j y i y j K ( x i x j ) + β 1 y i + β 2 y i i = 1 N α j y j 2 E α j α i ( α ) = β 0 y i y j K ( x i , x j ) + β 2 y i y j
The bias and weights of the built HRNN are given by:
I i = E α i ( 0 ) = β 0 β 1 y i i = 1 , , N T i , j = 2 E α j α i ( 0 ) = β 0 y i y j K ( x i , x j ) β 2 y i y j
The dynamical system governing the SVM-HRNN is given by:
d α i d t = 2 u 0 α i ( 1 α i ) j = 1 n ( β 0 y i y j K ( x i , x j ) β 2 y i y j ) α j β 0 β 1 y i ) α i ( 0 ) = α i 0 [ 0 , 1 ] , i = 1 , , n
To transform (8) into a linear equation, we introduce n fuzzy sets defined on the following variables:
z i = 2 u 0 α i ( 1 α i ) , i = 1 , , n
We have min ( z i ) = 0 and max ( z i ) = 1 u 0 ; we set U = m : s 0 : M and d = size ( U ) , where s 0 = 1 p u 0 and p is a positive integer, and 0 m M 1 u 0 . For each z i , we introduce d fuzzy values F 1 , , F d that we associate with d Gaussians G 1 ( mean = m ) , , G 1 ( mean = m + l s 0 ) , , G d ( mean = m + ( d 1 ) s 0 = M ) , respectively.
We pull N samples π 1 = ( a 1 1 , , a n 1 ) , , π s = ( a 1 s , , a n s ) , , π N = ( a 1 N , , a n N ) from U n , where n is the number of neurons of a given SVM-HRNN. For each sample π s = ( a 1 s , , a n s ) , we introduce the following rule R ( π s ) :
  • if z 1 is π s . a 1 s , z i is π s . a i s , and z n is π s . a n s ,
  • then d α i d t = a i s j = 1 n ( β 0 y i y j K ( x i , x j ) β 2 y i y j ) α j β 0 β 1 y i ) α i ( 0 ) = α i 0 [ 0 , 1 ] , i = 1 , , n

5. Experimentation

In order to test the algorithm introduced in this paper, some computer programs have been designed to calculate the equilibrium points of several HRNNs randomly generated to solve the classification problem using support vector machines. Two programs, which have been coded in Matlab, compute one equilibrium point of the HRNN and of the G-TS-HRNN by the Euler method. The system used in this study is a Dell desktop computer named DESKTOP-N1UTCKF, equipped with an 11th Gen Intel(R) Core(TM) i5-1145G7 processor (4 cores, 8 threads, 2.60 GHz base frequency), 16 GB of RAM (15.7 GB available), and running a 64-bit operating system.

5.1. Testing and Benchmarking on Blind Entities

The properties of energy function decrease and convergence were verified through a first set of computational experiences based on 100 random HRNNs of different sizes. Similar to the study in [10], the weight and bias of the HRNN and G-TS-HRNN have been generated blindly as follows:
T = U [ 10 , 0 ] d × d , I = U [ 0 , 5 ] d , and u 0 [ 0.02 , 2 ] .
We have calculated the steady points of the G-TS-HRNN for different sample sizes (6 to 100). Figure 5 gives the energy of the G-TS-HRNN for different sample sizes vs. the energy of the HRNN at steady points. The tests were repeated several times, and the curve presents the means of the energies. The red line represents the energy of the HRNN at the steady point. We remark that the curve has several local minima. For some instances, the energy of the HRNN is better than the ones produced by the proposed system. However, the most important thing is that for some instances, the proposed system can produce steady states with better energy compared to classical HRNNs. In this sense, when the optimization problem has several local minima, the Gaussian Takaji–Sugeno sampling technique shows a very high capability to escape the bad valleys thanks to the uniform search at the realizable domain.
To measure the robustness of the G-TS-HRNN, we calculate the steady point of each sample 100 times and we present the boxplots of the obtained populations; see Figure 6. The maximum dispersion (=0.41) is the one associated with the sample of size 85 and 100; the minimum dispersion (=0.1) is the one associated with the sample of size 5. However, this latter is not interesting because a sample with a small size permits the exploration of only a few regions. In this regard, several other instances (instances of size 30 to 70) have a very small dispersion and are capable of exploring more promising regions in the meantime.

5.2. Application to the Classification Problem

In this section, we compare G-TS-HRNN-SVM to HRNN-SVM. The algorithm underwent testing across eight commonly used data sets, divided randomly into 80% training and 20% testing sets [29]. Attributes for each data set are summarized in Table 1.
In machine learning, classification models are evaluated through a series of metrics such as the following:
  • Precision: Precision is the number of positive samples compared to the ones that the classifier labels positive and is used to evaluate the accuracy of the classifier’s predictions. P r e c i s i o n = T P T P + F P
  • Recall: Recall is the number of samples that were correctly predicted as positive compared to actual positive samples. The importance of such a metric can be observed in sensitive applications whereby accurate predictions of rare instances are essential. R e c a l l = T P T P + F N
  • F-Measure: The F-Measure is F - M e a s u r e = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n
  • Geometric mean: The geometric mean is a metric that is employed to calculate the balance of performances in terms of the classification of minority and majority classes. It is effective when evaluating classifiers when addressing imbalanced data. G - m e a n s = s e n s i t i v i t y × s p e c i f i c i t y
We have used HRNN-SVM and G-TS-HRNN-SVM to classify samples from different domains (Physics and Chemistry, Biology, Health and Medicine, Social Science); see Table 1. In this regard, we have considered different sizes of G-TS-HRNN samples (10, 20, 30, 40, 50, 60, 70, 80, 90). The tests were repeated several times and Table A1 and Table A2 give the mean values of the performance measures (accuracy, F1-score, precision, and recall) associated with different data sets; see Appendix A. A rough comparison considering the nine data sets based on these two tables reveals the following:
(a) The G-TS-HRNN-SVM classifiers, with sample sizes between 30 and 70, allow a significant modification of the performance of HRNN-SVM for all the metrics (accuracy, F1-measure, precision, and recall).
(b) The improvements provided by the G-TS-HRNN-SVM classifiers, with sample sizes between 30 and 70, allow a significant modification of the performance of HRNN-SVM for all the metrics (accuracy, F1-measure, precision, and recall). The improvements provided by the G-TS-HRNN-SVM classifiers, with sample sizes of 10, 20, 80, and 90, allow small improvements, and in some cases, HRNN-SVM outperforms these versions of G-TS-HRNN-SVM.
To realize more deep comparisons between the proposed method and the HRNN-CVM, we use the following formulas to detect the method with the best accuracy, F1-score, precision, and recall, on each data set d and for each metric m e t r i c :
m e t r i c d = Max { m e t r i c ( m t e s t e d o n d ) / m { H R N N S V M , G T S H R N N S V M ( s s ) , s s = 10 , , 90 } } .
Then, we obtain the methods m associated with the optimal performances using the following equation:
m d = argmin ( m e t r i c d )
To calculate the percentage p d m e t r i c of the improvement in performances of HRNN-SVM, for each data set d and metric m e t r i c , we use the following equation:
p d m e t r i c % = m d m e t r i c d ( H R N N ) m e t r i c d ( H R N N )
Table 2 provides the obtained results using these three statistical equations; from left to right, the best performances, methods associated with these optimal metrics, and the percentage of the improvement in performances compared to the HRNN-SVM classification method. Compared to the performance table of HRNN-SVM, the proposed family of methods strongly improve the accuracy, F1-score, precision, and recall on all the data sets. We note that HRNN-SVM is absent from the table of best methods. In addition, the best performance was obtained by the G-TS-HRNN-SVM with samples of 50, 60, and 70. In this sense, G-TS-HRNN-SVM(50) appears more times in the best methods table, followed by G-TS-HRNN-SVM(60) then G-TS-HRNN-SVM(60). In this regard, the proposed family of methods permits an improvement of the HRNN-SVM performance that we provide as the maximum and minimum improvement percentages in terms of intervals:
p a c c u r a c y [ 1.31 , 2.99 ] , p F 1 - s c o r e [ 1.11 , 4.16 ] , p p r e c i s i o n [ 1.68 , 3.83 ] , and p r e c a l l [ 0.82 , 6.69 ] .
To compare different methods independently of the type and complexity of data, we calculate the mean of different metric values considering the nine data sets. Figure 7 gives the mean accuracy (Figure 7a), mean f1-score (Figure 7b), mean precision (Figure 7c), and mean recall (Figure 7d) of all methods on all data sets. We note that HRNN-SVM always has the smallest performance and the GTS-HRNN-SVM with very big or very small sample sizes has a performance very close to HRNN-SVM. The best performance is obtained with GTS-HRNN-SVM using mean sample sizes.
To study the stability of the best versions of the proposed method, G-TS-HRNN-SVM(50) and G-TS-HRNN-SVM(60), we repeat the classification tests several times and estimate the mean and standard deviation of each population; then we calculate the width of the confidence interval of each sample. Table 3 gives the widths of these intervals associated with each data set. In this context, we prefer to give the global confidence interval [ l b f i e l d m e t r i c , U B f i e l d m e t r i c ] for each field associated with the best version (G-TS-HRNN-SVM-50), which we can obtain using the formulas, where l b f i e l d m e t r i c (resp. U B f i e l d m e t r i c ) is the minimum of the negative deviation (resp. the maximum of the positive deviation) from the mean of the metric “metric” of all the data from the field “field”:
  • Physics and Chemistry: the only data set from this field is wine. From Table 3, we have CI(G-TS-HRNN-SVM-50) = ±0.12; CI(G-TS-HRNN-SVM-60) = ±0.09; CI(G-TS-HRNN-SVM-70) = ±0.12. Thus, the most stable version in this field is G-TS-HRNN-SVM-60.
  • Biology: the data sets from this field are Iris, Abalone, Ecoli, and Seed. From Table 3, we have CI(G-TS-HRNN-SVM-50) = ±0.20; CI(G-TS-HRNN-SVM-60) = ±0.17; CI(G-TS-HRNN-SVM-70) = ±0.14. Thus, the most stable version in this field is G-TS-HRNN-SVM-70.
  • Health and Medicine: the data sets from this field are Liver, Spect, and Pima. From Table 3, we have CI(G-TS-HRNN-SVM-50) = ±0.18; CI(G-TS-HRNN-SVM-60) = ±0.17; CI(G-TS-HRNN-SVM-70) = ±0.13. Thus, the most stable version in this field is G-TS-HRNN-SVM-70.
  • Social Science: the only data set from this field is Balance. From Table 3, we have CI(G-TS-HRNN-SVM-50) = ±0.19; CI(G-TS-HRNN-SVM-60) = ±0.16; CI(G-TS-HRNN-SVM-70) = ±0.12. Thus, the most stable version in this field is G-TS-HRNN-SVM-70.
We have computed the CPU time required by HRNN and TS-G-HRNN, considering several sizes (10 to 90), to teach SVM the classes of the UCI data sets given in Table 1; see Table 4. The convergence criterion considered in this context is the stability of neuron activation. To visualize the difference in computation time between the different methods, we calculated the average computation time required for all data sets; see Figure 8. We note that TS-G-HRNN-SVM versions with sample sizes below 30 require less computation time than HRNN-SVM. TS-G-HRNN-SVM versions with sample sizes between 40 and 60 require more computation time than HRNN-SVM but with a slight difference (0.01 s). This difference becomes more significant for instances with sample sizes above 70 (0.1 s). For all TS-G-HRNN-SVM instances with sample sizes below 90, the difference in computation time remains acceptable. Moreover, this difference almost explodes when the sample size exceeds 100. In fact, the time gap increased from 0.01 s to 0.1 s for the samples considered, showing the exponential evolution of cpu time as a function of sample size. This remains the major drawback of our method.
Finally, the proposed family of classification methods G-TS-HRNN-SVM performs better than HRNN-SVM in terms of accuracy, F1-Measure, precision, and recall. The most performant versions are G-TS-HRNN-SVM-50, G-TS-HRNN-SVM-60, and G-TS-HRNN-SVM-70. The most stable version, among all the considered fields (Biology, Social science, and Health and Medicine), is G-TS-HRNN-SVM-70 except in Physics and Chemistry, where G-TS-HRNN-SVM-60 is the most stable. The reason for the poor performance of low-sample versions of G-TS-HRNN-SVM is that the complexity of this model is not sufficient to reproduce the under-studied phenomena: this is the underlearning phenomenon. The cause of the poor performance of G-TS-HRNN-SVM versions with very high sample sizes is the high complexity of this model, which leads to rote learning of the phenomena studied: this is the overlearning phenomenon.
Furthermore, a better compromise between performance and computation time is to consider versions of G-TS-HRNN-SVM with sample sizes between 50 and 70. Moreover, given the exponential behavior of computation time as a function of sample size, 100 samples should not be exceeded.

6. Conclusions

Problem: Like all single-point metaheuristics, HRNN has certain drawbacks, such as being more likely to get stuck in local optima or miss global optima due to the use of a single point to explore the search space. Additionally, it is more sensitive to the initial point and operator, which can influence the quality and diversity of solutions. Furthermore, it can have difficulty with dynamic or noisy environments, as it can lose track of the optimal region or be misled by random fluctuations. Despite attempts, based on the choice of the Euler–Cauchy discretization step, these problems persist, and the initial activations still have a significant influence on the quality of the HRNN’s stability point.
Proposed solution: To overcome these shortcomings, this paper introduces a fuzzy HRNN population-based optimization method, namely Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network (G-TS-HRNN-SVM). G-TS-HRNN-SVM associates a fuzzy variable with each neuron, with suitable fuzzy values described by Gaussian membership functions that cover the universe of discourse of the introduced fuzzy variables by adopting appropriate means and standard deviations. To build a sample of G-TS-HRNN-SVM, we randomly generate a Gaussian mean for each neuron. This leads to a Takagi–Sugeno rule with a linear differential equation as the premise. The fuzzy center of gravity is used to estimate the steady state of the initial nonlinear HRNN.
Tests and results: G-TS-HRNN-SVM is used to solve the random optimization method and solve the classification problem based on the support vector mathematical model. HRNN and G-TS-HRNN-SVM were tested on nine data sets found in UCI (Machine Learning Repository). The performance metrics considered in this work are accuracy, F1-measure, precision, and recall. Compared with the classical HRNN, G-TS-HRNN-SVM performs better on well-known data sets. First, the best G-TS-HRNN-SVM is the one with the mean number of rules (i.e., 40, 50, and 60). In addition, G-TS-HRNN-SVM permits an improvement of the classical HRNN accuracy (by 2.99%), F1-measure (by 4.16%), precision (by 3.83%), and recall (by 6.69%). Based on the confidence interval, the most stable version of G-TS-HRNN-SVM in the physics and chemistry fields is G-TS-HRNN-SVM(60), with CI = ±0.09; the most stable version in “biology”, with CI = ±0.14; and “health and medicine”, with CI = ±0.13, and “social science”, with CI = ±0.12, fields G-TS-HRNN-SVM(70).
Limitations: An instance of G-TS-HRNN-SVM with a very small sample size leads to underlearning and premature convergence, whereas an instance of G-TS-HRNN-SVM with a large sample size leads to overlearning and requires a very large memory size. In this context, selecting the best sample size remains a very difficult issue. In this respect, to avoid an explosion in calculation time, we must not exceed 100 samples, which is the main drawback of our method.
Future scope: In the future, we will use the genetic algorithm metaheuristic to realize the crossover and mutation operator between the rules within each sample to improve the quality of the solution. In addition, we will introduce the fractional version of G-TS-HRNN-SVM to handle more information about the problem to be solved. In addition, we will use G-TS-HRNN to perform the transformer and CNN learning tasks and perform self-association memory.

Author Contributions

Conceptualization, O.B. and K.E.M.; methodology, K.E.M. and M.R.; software, M.R. and O.B.; writing—original draft preparation, M.R., K.E.M. and O.B.; writing—review and editing, M.R. and K.E.M.; supervision, K.E.M. and O.B.; project administration, K.E.M. and O.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data set used in the experiments is publicly available in [29].

Acknowledgments

This work was supported by the Ministry of National Education, Professional Training, Higher Education and Scientific Research (MENFPESRS), and the Digital Development Agency (DDA) and CNRST of Morocco (No. Alkhawarizmi/2020/23).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1. Accuracy, F1-score, precision, and recall, on all data sets, of the HRNN-SVM and the samples (10 to 50) of G-TS-HRNN-SVM.
Table A1. Accuracy, F1-score, precision, and recall, on all data sets, of the HRNN-SVM and the samples (10 to 50) of G-TS-HRNN-SVM.
SVM-HRNNT-S SVM-HRNN sample size = 10
AccuracyF1-scorePrecisionRecallAccuracyF1-scorePrecisionRecall
Iris95.9394.6589.8894.8796.9293.1087.9595.61
Abalone81.2242.6683.2833.0479.6740.9383.4331.57
Wine80.7977.8673.5374.6479.2777.3074.3874.86
Ecoli87.3397.0497.4397.5987.3595.2596.3598.40
Balance79.7070.6156.0264.9780.2771.2855.7863.46
Liver80.4277.7577.9070.1478.9577.9978.7470.11
Spect94.6993.1791.9771.6294.3292.9690.6772.08
Seed85.3583.3492.4175.4185.7183.1191.7275.06
Pima79.2061.7884.3249.5578.2562.7783.4150.45
T-S SVM-HRNN sample size = 20T-S SVM-HRNN sample size = 30
AccuracyF1-scorePrecisionRecallAccuracyF1-scorePrecisionRecall
Iris94.5095.4889.7095.2696.8394.5890.2794.51
Abalone80.6743.9282.4532.6780.2344.0583.3932.09
Wine80.2276.4873.8575.6581.9477.5274.6074.38
Ecoli87.6096.6396.3596.4487.1396.4497.4798.86
Balance81.1470.7354.7764.2481.3671.7455.7565.33
Liver79.5577.1979.1170.2082.2178.6077.6771.03
Spect93.6892.1090.6672.6895.7893.3891.5571.54
Seed86.7784.6392.1376.6286.4384.2392.1674.54
Pima79.1561.6983.3248.3079.7961.3184.8851.16
T-S SVM-HRNN sample size = 40T-S SVM-HRNN sample size = 50
AccuracyF1-scorePrecisionRecallAccuracyF1-scorePrecisionRecall
Iris97.2794.9691.9494.8496.2397.0192.1995.11
Abalone81.5044.1783.4635.0282.2643.4785.4035.41
Wine81.2778.8375.3276.3582.6279.5775.1475.43
Ecoli88.7497.2999.1498.1189.1697.4399.3397.82
Balance80.1070.4955.5567.2582.1272.1358.2565.46
Liver82.6578.4079.5071.6582.8978.3578.4671.14
Spect94.5793.2794.1773.7996.6495.2792.9972.62
Seed85.9883.5993.7176.7987.2883.9694.5777.12
Pima78.7663.7485.3751.7881.6462.2584.8650.33
Table A2. Accuracy, F1-score, precision, and recall on all data sets of the HRNN-SVM and the samples (60 to 90) of G-TS-HRNN-SVM.
Table A2. Accuracy, F1-score, precision, and recall on all data sets of the HRNN-SVM and the samples (60 to 90) of G-TS-HRNN-SVM.
T-S SVM-HRNN sample size = 60T-S SVM-HRNN sample size = 70
AccuracyF1-scorePrecisionRecallAccuracyF1-scorePrecisionRecall
Iris97.5195.2791.0995.5396.6995.6589.8894.72
Abalone82.3043.3284.6532.7880.7844.5184.5233.93
Wine82.5378.4475.0276.7481.0377.4775.1775.63
Ecoli88.3098.4098.2199.4286.8898.8796.9698.69
Balance81.0371.5756.6466.3881.1771.4256.2466.04
Liver80.3777.5779.7771.4780.9878.6277.8871.50
Spect95.9293.9292.9371.7394.7292.9293.7173.17
Seed85.2884.7592.4676.4387.0683.0693.6675.34
Pima78.9062.5584.5850.9080.7863.1085.7650.96
T-S SVM-HRNN sample size = 80T-S SVM-HRNN sample size = 90
AccuracyF1-scorePrecisionRecallAccuracyF1-scorePrecisionRecall
Iris95.2895.4391.4195.4196.5394.0490.0695.65
Abalone81.7442.5285.0133.7580.9643.2884.7333.52
Wine82.5177.1773.2275.1981.5979.4174.8775.33
Ecoli88.2097.1398.3297.7288.8996.4797.7698.22
Balance81.1171.2256.9764.8480.7870.6955.1665.32
Liver79.9377.0579.4171.3079.7777.7078.7169.47
Spect95.9793.5291.3671.2295.7392.3093.0972.87
Seed85.1982.7593.8275.0284.8683.9294.0074.52
Pima80.0062.3684.0150.7278.5661.3684.8350.29

References

  1. Mitra, S.; Hayashi, Y. Neuro-fuzzy rule generation: Survey in soft computing framework. IEEE Trans. Neural Netw. 2000, 11, 748–768. [Google Scholar] [CrossRef] [PubMed]
  2. Shihabudheen, K.V.; Pillai, G.N. Recent advances in neuro-fuzzy system: A survey. Knowl.-Based Syst. 2018, 152, 136–162. [Google Scholar] [CrossRef]
  3. Yen, J.; Wang, L.; Gillespie, C.W. Improving the interpretability of TSK fuzzy models by combining global learning and local learning. IEEE Trans. Fuzzy Syst. 1998, 6, 530–537. [Google Scholar] [CrossRef]
  4. Du, M.; Behera, A.K.; Vaikuntanathan, S. Active oscillatory associative memory. J. Chem. Phys. 2024, 160, 055103. [Google Scholar] [CrossRef]
  5. Abdulrahman, A.; Sayeh, M.; Fadhil, A. Enhancing the analog to digital converter using proteretic hopfield neural network. Neural Comput. Appl. 2024, 36, 5735–5745. [Google Scholar] [CrossRef]
  6. Roudani, M.; El Moutaouakil, K.; Palade, V.; Baïzri, H.; Chellak, S.; Cheggour, M. Fuzzy Clustering SMOTE and Fuzzy Classifiers for Hidden Disease Predictions; Lecture Notes in Networks and Systems; Springer: Cham, Switzerland, 2024; Volume 1093. [Google Scholar]
  7. El Moutaouakil, K.; Bouhanch, Z.; Ahourag, A.; Aberqi, A.; Karite, T. OPT-FRAC-CHN: Optimal Fractional Continuous Hopfield Network. Symmetry 2024, 16, 921. [Google Scholar] [CrossRef]
  8. Uykan, Z. On the Working Principle of the Hopfield Neural Networks and its Equivalence to the GADIA in Optimization. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 3294–3304. [Google Scholar] [CrossRef] [PubMed]
  9. Talaván, P.M.; Yáñez, J. The generalized quadratic knapsack problem. A neuronal network approach. Neural Netw. 2006, 19, 416–428. [Google Scholar] [CrossRef] [PubMed]
  10. Talaván, P.M.; Yáñez, J. A continuous Hopfield network equilibrium points algorithm. Comput. Oper. Res. 2005, 32, 2179–2196. [Google Scholar] [CrossRef]
  11. El Ouissari, A.; El Moutaouakil, K. Intelligent optimal control of nonlinear diabetic population dynamics system using a genetic algorithm. Syst. Res. Inf. Technol. 2024, 1, 134–148. [Google Scholar]
  12. Kurdyukov, V.; Kaziev, G.; Tattibekov, K. Clustering Methods for Network Data Analysis in Programming. Int. J. Commun. Netw. Inf. Secur. 2023, 15, 149–167. [Google Scholar]
  13. Singh, H.; Sharma, J.R.; Kumar, S. A simple yet efficient two-step fifth-order weighted-Newton method for nonlinear models. Numer. Algorithms 2023, 93, 203–225. [Google Scholar] [CrossRef]
  14. Bu, F.; Zhang, C.; Kim, E.H.; Yang, D.; Fu, Z.; Pedrycz, W. Fuzzy clustering-based neural network based on linear fitting residual-driven weighted fuzzy clustering and convolutional regularization strategy. Appl. Soft Comput. 2024, 154, 111403. [Google Scholar] [CrossRef]
  15. Sharma, R.; Goel, T.; Tanveer, M.; Al-Dhaifallah, M. Alzheimer’s Disease Diagnosis Using Ensemble of Random Weighted Features and Fuzzy Least Square Twin Support Vector Machine. IEEE Trans. Emerg. Top. Comput. Intell. 2025; early access. [Google Scholar] [CrossRef]
  16. Smith, K.A. Neural networks for combinatorial optimization: A review of more than a decade of research. Informs J. Comput. 1999, 11, 15–34. [Google Scholar] [CrossRef]
  17. Joya, G.; Atencia, M.A.; Sandoval, F. Hopfield neural networks for optimization: Study of the different dynamics. Neurocomputing 2002, 43, 219–237. [Google Scholar] [CrossRef]
  18. Wang, L. On the dynamics of discrete-time, continuous-state Hopfield neural networks. IEEE Trans. Circuits Syst. II Analog Digit. Signal Process. 1998, 45, 747–749. [Google Scholar]
  19. Demidowitsch, B.P.; Maron, I.A.; Schuwalowa, E.S. Metodos Numericos de Analisis; Paraninfo: Madrid, Spain, 1980. [Google Scholar]
  20. Bagherzadeh, N.; Kerola, T.; Leddy, B.; Brice, R. On parallel execution of the traveling salesman problem on a neural network model. In Proceedings of the IEEE International Conference on Neural Networks, San Diego, CA, USA, 21–24 June 1987; Volume III, pp. 317–324. [Google Scholar]
  21. Brandt, R.D.; Wang, Y.; Laub, A.J.; Mitra, S.K. Alternative networks for solving the traveling salesman problem and the list-matching problem. In Proceedings of the International Conference on Neural Network (ICNN’88), San Diego, CA, USA, 24–27 July 1988; Volume II, pp. 333–340. [Google Scholar]
  22. Wu, J.K. Neural Networks and Simulation Methods; Marcel Dekker: New York, NY, USA, 1994. [Google Scholar]
  23. Abe, S. Global convergence and suppression of spurious states of the Hopfield neural networks. IEEE Trans. Circuits Syst. 1993, 40, 246–257. [Google Scholar] [CrossRef]
  24. Wang, L.; Smith, K. On chaotic simulated annealing. IEEE Trans. Neural Netw. 1998, 9, 716–718. [Google Scholar] [CrossRef]
  25. Jin, Q.; Mokhtari, A. Non-asymptotic superlinear convergence of standard quasi-Newton methods. Math. Program. 2023, 200, 425–473. [Google Scholar] [CrossRef]
  26. Ogwo, G.N.; Alakoya, T.O.; Mewomo, O.T. Iterative algorithm with self-adaptive step size for approximating the common solution of variational inequality and fixed point problems. Optimization 2023, 72, 677–711. [Google Scholar] [CrossRef]
  27. Gharbia, I.B.; Ferzly, J.; Vohralík, M.; Yousef, S. Semismooth and smoothing Newton methods for nonlinear systems with complementarity constraints: Adaptivity and inexact resolution. J. Comput. Appl. Math. 2023, 420, 114765. [Google Scholar] [CrossRef]
  28. Candelario, G.; Cordero, A.; Torregrosa, J.R.; Vassileva, M.P. Generalized conformable fractional Newton-type method for solving nonlinear systems. Numer. Algorithms 2023, 93, 1171–1208. [Google Scholar] [CrossRef]
  29. Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2019; Available online: http://archive.ics.uci.edu/ml (accessed on 24 December 2024).
  30. Breiding, P.; Rose, K.; Timme, S. Certifying zeros of polynomial systems using interval arithmetic. ACM Trans. Math. Softw. 2023, 49, 1–14. [Google Scholar] [CrossRef]
  31. Bhuvaneswari, S.; Karthikeyan, S. A Comprehensive Research of Breast Cancer Detection Using Machine Learning, Clustering and Optimization Techniques. In Proceedings of the 2023 IEEE International Conference on Data Science and Network Security (ICDSNS), Tiptur, India, 28–29 July 2023. [Google Scholar]
  32. Zimmermann, H.-J. Fuzzy Set Theory—And Its Applications, 4th ed.; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  33. Rajafillah, C.; El Moutaouakil, K.; Patriciu, A.M.; Yahyaouy, A.; Riffi, J. INT-FUP: Intuitionistic Fuzzy Pooling. Mathematics 2024, 12, 1740. [Google Scholar] [CrossRef]
  34. Sugeno, M.; Kang, G. Structure identification of fuzzy model. Fuzzy Sets Syst. 1988, 28, 15–33. [Google Scholar] [CrossRef]
  35. Sugeno, M. Fuzzy Control; North-Holland, Publishing Co.: Amsterdam, The Netherlands, 1988. [Google Scholar]
  36. Mehran, K. Takagi-sugeno fuzzy modeling for process control. Ind. Autom. Robot. Artif. Intell. 2008, 262, 1–31. [Google Scholar]
  37. Kawamoto, S.; Tada, K.; Ishigame, A.; Taniguchi, T. An approach to stability analysis of second order fuzzy systems. In Proceedings of the IEEE International Conference on Fuzzy Systems, San Diego, CA, USA, 8–12 March 1992; pp. 1427–1434. [Google Scholar]
  38. Tanaka, K. A Theory of Advanced Fuzzy Control; Springer: Berlin, Germany, 1994. [Google Scholar]
  39. Mercer, J. Functions of positive and negative type, and their connection the theory of integral equations. Philos. Trans. R. Soc. Lond. 1909, 209, 415–446. [Google Scholar] [CrossRef]
  40. Schölkopf, B.; Smola, A.J.; Bach, F. Learning with Kernels: Support Vector Machines, Regularization, Optimization, and Beyond; MIT Press: Cambridge, MA, USA, 2002. [Google Scholar]
  41. Pandey, R.K.; Agrawal, O.P. Comparison of four numerical schemes for isoperimetric constraint fractional variational problems with A-operator. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Boston, MA, USA, 2–5 August 2015; American Society of Mechanical Engineers: New York City, NY, USA, 2015; Volume 57199. [Google Scholar]
  42. Kumar, K.; Pandey, R.K.; Sharma, S. Approximations of fractional integrals and Caputo derivatives with application in solving Abel’s integral equations. J. King Saud Univ.-Sci. 2019, 31, 692–700. [Google Scholar] [CrossRef]
Figure 1. Illustration of the sector nonlinearity approach.
Figure 1. Illustration of the sector nonlinearity approach.
Information 16 00141 g001
Figure 2. Membership functions for u 0 = 0.2 .
Figure 2. Membership functions for u 0 = 0.2 .
Information 16 00141 g002
Figure 3. Gaussian Takagi–Sugeno HRNN sampling, fuzzification, and equilibrium state approximation.
Figure 3. Gaussian Takagi–Sugeno HRNN sampling, fuzzification, and equilibrium state approximation.
Information 16 00141 g003
Figure 4. Illustration of the approximation of the optimal steady state via the Takagi–Sugeno HRNN sampling method.
Figure 4. Illustration of the approximation of the optimal steady state via the Takagi–Sugeno HRNN sampling method.
Information 16 00141 g004
Figure 5. The energy of the G-TS-HRNN for different sample sizes vs. the energy of the HRNN at a steady point.
Figure 5. The energy of the G-TS-HRNN for different sample sizes vs. the energy of the HRNN at a steady point.
Information 16 00141 g005
Figure 6. Takagi–Sugeno HRNN robustness.
Figure 6. Takagi–Sugeno HRNN robustness.
Information 16 00141 g006
Figure 7. Mean accuracy (a), mean F1-score (b), mean precision (c), and mean recall (d), on all data sets of the HRNN-SVM and the nine samples of G-TS-HRNN-SVM.
Figure 7. Mean accuracy (a), mean F1-score (b), mean precision (c), and mean recall (d), on all data sets of the HRNN-SVM and the nine samples of G-TS-HRNN-SVM.
Information 16 00141 g007
Figure 8. Mean cpu time required by HRNN and G-TS-HRNN.
Figure 8. Mean cpu time required by HRNN and G-TS-HRNN.
Information 16 00141 g008
Table 1. UCI data description.
Table 1. UCI data description.
Data SetFeaturesSamplesNomber of ClassesSubject Aria
Wine131783Physics and Chemistry
Iris41504Biology
Pima87682Health and Medicine
Ecoli73367Biology
Abalone841173Biology
Haberman33062Health and Medicine
Balance46255Social Science
Liver63452Health and Medicine
Seed72103Biology
Table 2. Best performances, methods associated with these best performances, and the percent of performance improvements of different methods ((G-TS)-HRNN-SVM).
Table 2. Best performances, methods associated with these best performances, and the percent of performance improvements of different methods ((G-TS)-HRNN-SVM).
Data SetBest PerformanceBest MethodPerformance Improvement Percent
AccuracyF1-ScorePrecisionRecallAccuracyF1-ScorePrecisionRecallAccuracyF1-ScorePrecisionRecall
Iris97.5197.0192.1995.65GTS-HRNN-60GTS-HRNN-50GTS-HRNN-50GTS-HRNN-801.622.432.510.82
Abalone82.344.5185.435.41GTS-HRNN-60GTS-HRNN-70GTS-HRNN-50GTS-HRNN-501.314.162.486.69
Wine82.6279.5775.3276.74GTS-HRNN-50GTS-HRNN-50GTS-HRNN-40GTS-HRNN-602.212.152.382.74
Ecoli89.1698.8799.3399.42GTS-HRNN-50GTS-HRNN-70GTS-HRNN-50GTS-HRNN-602.051.851.911.84
Balance82.1272.1358.2567.25GTS-HRNN-50GTS-HRNN-50GTS-HRNN-50GTS-HRNN-402.952.113.833.39
Liver82.8978.6279.7771.65GTS-HRNN-50GTS-HRNN-70GTS-HRNN-90GTS-HRNN-402.981.112.342.11
Spect96.6495.2794.1773.79GTS-HRNN-50GTS-HRNN-50GTS-HRNN-40GTS-HRNN-402.022.202.342.94
Seed87.2884.7594.5777.12GTS-HRNN-50GTS-HRNN-60GTS-HRNN-50GTS-HRNN-502.211.662.282.22
Pima81.6463.7485.7651.78GTS-HRNN-50GTS-HRNN-40GTS-HRNN-70GTS-HRNN-402.993.071.684.31
Table 3. Confidence interval associated with different measures of G-TS-HRNN-SVM for sample sizes of 50 to 70.
Table 3. Confidence interval associated with different measures of G-TS-HRNN-SVM for sample sizes of 50 to 70.
Data SetG-TS-HRNN-SVM Sample Size = 50G-TS-HRNN-SVM Sample Size = 60G-TS-HRNN-SVM Sample Size = 70
Accuracy
(CI)
F1-Score (CI)Precision (CI)Recall (CI)Accuracy (CI)F1-Score (CI)Precision (CI)Recall (CI)Accuracy (CI)F1-Score (CI)Precision (CI)Recall (CI)
Iris±0.14±0.20±0.03±0.03±0.08±0.17±0.05±0.07±0.09±0.02±0.13±0.09
Abalone±0.15±0.02±0.09±0.18±0.02±0.09±0.08±0.15±0.05±0.14±0.12±0.10
Wine±0.10±0.09±0.06±0.12±0.05±0.09±0.03±0.01±0.12±0.12±0.10±0.09
Ecoli±0.03±0.03±0.16±0.11±0.03±0.07±0.03±0.02±0.04±0.08±0.06±0.04
Balance±0.05±0.19±0.09±0.04±0.04±0.16±0.17±0.04±0.11±0.07±0.12±0.05
Liver±0.18±0.01±0.18±0.17±0.05±0.07±0.17±0.12±0.04±0.07±0.09±0.08
Spect±0.04±0.16±0.04±0.13±0.08±0.03±0.11±0.13±0.06±0.05±0.06±0.04
Seed±0.17±0.17±0.06±0.08±0.02±0.14±0.02±0.12±0.10±0.08±0.14±0.13
Pima±0.11±0.17±0.04±0.11±0.16±0.08±0.05±0.09±0.12±0.08±0.13±0.04
Table 4. CPU time required by HRNN and G-TS-HRNN to teach SVM the classes of UCI data sets.
Table 4. CPU time required by HRNN and G-TS-HRNN to teach SVM the classes of UCI data sets.
CPU time (s)
Data set/MethodHRNN-SVMTS-HRNN-SVM(10)TS-HRNN-SVM(20)TS-HRNN-SVM(30)TS-HRNN-SVM(40)
iris1.715251.712341.713021.713631.72697
Wine0.16760.164690.165370.165980.17932
Pima2.83062.827692.828372.828982.84232
Haberman0.02340.020490.021170.021780.03512
Balance0.00860.005690.006370.006980.02032
Libra0.00870.005790.006470.007080.02042
Liver0.12940.126490.127170.127780.14112
Seed0.00330.000390.001070.001680.01502
Spect0.00540.002490.003170.003780.01712
Segment0.06240.059490.060170.060780.07412
Vehicle0.15810.155190.155870.156480.16982
CPU time (s)
Data set/MethodTS-HRNN-SVM(50)TS-HRNN-SVM(60)TS-HRNN-SVM(70)TS-HRNN-SVM(80)TS-HRNN-SVM(90)
iris1.744331.748551.761421.727251.75843
Wine0.196680.200900.213770.179600.21078
Pima2.859682.863902.876772.842602.87378
Haberman0.052480.056700.069570.035400.06658
Balance0.037680.041900.054770.020600.05178
Libra0.037780.042000.054870.020700.05188
Liver0.158480.162700.175570.141400.17258
Seed0.032380.036600.049470.015300.04648
Spect0.034480.038700.051570.017400.04858
Segment0.091480.095700.108570.074400.10558
Vehicle0.187180.191400.204270.169100.20128
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bahou, O.; Roudani, M.; El Moutaouakil, K. G-TS-HRNN: Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network. Information 2025, 16, 141. https://doi.org/10.3390/info16020141

AMA Style

Bahou O, Roudani M, El Moutaouakil K. G-TS-HRNN: Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network. Information. 2025; 16(2):141. https://doi.org/10.3390/info16020141

Chicago/Turabian Style

Bahou, Omar, Mohammed Roudani, and Karim El Moutaouakil. 2025. "G-TS-HRNN: Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network" Information 16, no. 2: 141. https://doi.org/10.3390/info16020141

APA Style

Bahou, O., Roudani, M., & El Moutaouakil, K. (2025). G-TS-HRNN: Gaussian Takagi–Sugeno Hopfield Recurrent Neural Network. Information, 16(2), 141. https://doi.org/10.3390/info16020141

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop