1. Introduction
Neural networks occupy an important position in the field of artificial intelligence. Neurons have important research value as components of neural networks. Since the neuron model was proposed, it has been continuously developed and improved. The first artificial neuron model was proposed by McCulloch and Pitts in 1943. However, this model is too simple and does not take into account the nonlinear mechanism of dendrites [
1]. With the development of neurobiology, researchers have discovered the importance of dendritic structures in neural computation [
2]. Koch et al. proposed a dendritic neuron model called the
cell model [
3,
4]. However, this model cannot identify the type of synapse and cannot determine which dendrite is needed. This is because the dendritic structure of this model is fixed [
5]. Later, Legenstein and Maass also proposed a neuron model with dendritic structures [
6]. This model possesses nonlinear computational power but cannot solve problems that are not linearly separable.
The dendritic neuron model, called the logic dendritic neuron model (LDNM), has emerged as a potential machine learning model in recent years. A series of dendritic neuron models have been proposed to address practical problems. Ji et al. proposed a logic dendritic neuron model that has a dendritic structure and can be pruned after training [
7]. This model can effectively solve problems that are not linearly separable. Later, Ji et al. focused on the classification ability of LDNM and simulated a pruned LDNM with logic circuits [
8]. Tang et al. applied LDNM to disease diagnosis [
9]. Zhou et al. used LDNM to predict financial time series [
10]. Song et al. used LDNM to predict wind speed [
11]. Luo et al. combined machine learning methods with LDNM [
12], and the performance of LDNM was further improved by using a decision tree to initialize LDNM parameters. These studies demonstrated that LDNM is a type of neuronal model with excellent performance. Currently, researchers are still working on the training algorithm of LDNM and exploring the application areas of LDNM.
The common training algorithm of LDNM is the BP algorithm, which is widely used to train neuronal models [
13,
14]. The BP algorithm uses error backpropagation to adjust the parameter values. However, the BP algorithm is sensitive to the initial values of the parameters and easily falls into local minima [
15,
16]. Moreover, it is not easy to set the learning rate for BP [
17]. These disadvantages limit the performance of LDNM. In some experiments, LDNM trained by the BP algorithm can achieve good results in solving small-scale problems. However, the performance of LDNM trained by the BP algorithm decreases sharply as the problem size increases [
18].
In recent years, heuristic algorithms have been increasingly used and have achieved positive results in several fields [
19,
20,
21]. Specifically, the combination of heuristic algorithms and neural networks is becoming increasingly popular. For example, Cheng et al. proposed an improved artificial electric field algorithm and used it for neural network optimization [
22]. Soni et al. combined hybrid heuristic algorithms with neural networks for face recognition [
23]. In their work, optimal feature extraction was performed using a hybrid heuristic algorithm. Mathe et al. used heuristic algorithms to adjust the parameters of convolutional neural networks to remove artifacts from electroencephalography signals [
24]. In these studies, heuristic algorithms were used to optimize neural networks or as auxiliary methods for neural networks with satisfactory results. These studies are recognized as belonging to the field of neuroevolution [
25]. In the research on LDNM, heuristic algorithms have shown some advantages in training LDNM, especially in avoiding falling into local optima [
8,
26]. However, the datasets used in these experiments have been relatively small. Therefore, training LDNM with heuristic algorithms to solve high-dimensional classification problems is still a challenging task.
In this study, we attempt to find a more suitable heuristic algorithm for training LDNM, and we try to solve high-dimensional classification problems. For this purpose, we use the GBO algorithm [
27], which is a gradient-based heuristic algorithm. Compared to other heuristic algorithms, this algorithm uses a gradient approach and is able to accelerate the training of LDNM. Moreover, the algorithm has a unique operator that can avoid local convergence and further improve the classification performance of LDNM. To demonstrate the advantage of this algorithm in training LDNM, we compare it with four heuristic algorithms. These algorithms are the particle swarm optimization algorithm (PSO) [
28,
29], the differential evolution algorithm (DE) [
30], the genetic algorithm (GA) [
31], and the equilibrium optimizer algorithm (EO) [
32]. The experimental results demonstrate that the GBO algorithm can serve as a powerful algorithm for training LDNM. Finally, we compare LDNM trained by the GBO algorithm with five conventional classifiers to verify the effectiveness of LDNM. Moreover, the effectiveness of the neuronal structure pruning mechanisms is verified. The results showed that compared with the classic machine learning methods, LDNM is a very competitive classifier. The contribution of this paper is threefold. First, a biology-inspired LDNM with neuronal structure pruning mechanisms is proposed. Second, we employee a novel gradient-based heuristic algorithm GBO as the training method of LDNM. Third, the performance of the proposed LDNM/GBO is evaluated on seven datasets. We compare GBO with BP and four heuristic algorithms to verify its superior performance. In addition, the classification performance of LDNM is verified in comparison with five classic classifiers.
The remainder of this paper is organized as follows:
Section 2 introduces the structure of LDNM and its neuronal structure pruning mechanisms.
Section 3 describes the training algorithm, namely, GBO, and the training process.
Section 4 presents the experimental study. Finally, the conclusions are presented in
Section 5.
2. Logic Dendritic Neuron Model
LDNM is a neuron model with a dendritic structure consisting of four layers: a synaptic layer, dendritic layer, membrane layer, and soma body. The structure of LDNM is shown in
Figure 1.
The function of the synaptic layer is to receive input signals from other neurons and to process these input signals by using the following equation:
where
is the output of the
i-th (
) synapse on the
j-th (
) dendritic branch;
k is a fixed parameter, which is set to 5;
is the input to the synapse, which is in [0, 1]; and
and
are the connection parameters that we need to train.
The dendritic layer receives the output signal from the synapse and performs a multiplication operation, as expressed in Equation (
2), on each branch. The dendritic layer plays an important role in the transmission and processing of neural information.
Information about all dendritic branches is collected in the membrane layer. The sum of this information is calculated by Equation (
3), and the output is transmitted to the soma body.
The output signal of the membrane layer is calculated in the soma body using a modified sigmoid function, which is expressed as follows:
where
O is the final output;
c is a fixed parameter, which is set to 5; and
is the threshold of the soma body, which is set to 0.5. If this signal is greater than the threshold of the soma body, the neuron is fired.
2.1. Connection States
The connection parameters
and
of a synapse can determine the connection state of the synapse. There are four types of connection states, namely, direct connection, inverse connection, constant 1 connection, and constant 0 connection. After training, each synapse will be in one of the four connection states. These connection states are shown in
Figure 2.
When , the connection state of the corresponding synapse is direct connection. When , the connection state of the synapse is inverse connection. When or , the connection state of the synapse is constant 1 connection. When or , the connection state of the synapse is constant 0 connection. In addition, a threshold is defined at each synapse, which is calculated as .
In the direct connection state, the synaptic output is approximately 1 when the input is greater than the threshold; otherwise, the output is approximately 0. In the inverse connection state, the synaptic output is approximately 0 when the input is greater than the threshold; otherwise, the output is approximately 1. In the constant 1 connection state, the synaptic output is always approximately 1, and in the constant 0 connection state, the synaptic output is always approximately 0.
2.2. Neuronal Structure Pruning
According to the connection states of the synaptic layer, LDNM can be pruned to remove unnecessary synapses and dendrites. The structure of LDNM is simplified. There are two pruning mechanisms operating on LDNM: synaptic pruning and dendritic pruning.
Synaptic pruning: When a synapse is in the constant 1 connection state, the synaptic output is always approximately 1, so the output value of the synapse has no effect on the result of a multiplication operation. Thus, any synapse in the constant 1 connection state can be omitted.
Dendritic pruning: When a synapse is in the constant 0 connection state, the synaptic output is always approximately 0. When 0 is involved in a multiplication operation, the result of the operation is always 0. In this case, the output of the dendritic branch is always 0, regardless of the values of the other synapses on the branch. When an addition operation is performed in the membrane layer, 0 can be omitted without affecting the result. Thus, any dendritic branch that contains a synapse in the constant 0 connection state can be omitted.
Figure 3 depicts examples of the synaptic pruning and dendritic pruning operations.
5. Conclusions
In this paper, we used the GBO algorithm to train a logic dendritic neuron model to solve classification problems. GBO is a heuristic algorithm that is effective in avoiding falling into local optima and enables LDNM to solve high-dimensional problems. In our experiments, we used seven datasets to verify the effectiveness of the GBO algorithm. First, the GBO algorithm was compared with the BP algorithm, which is a common training algorithm of LDNM. The results showed that compared with LDNM trained by BP, LDNM trained by the GBO algorithm showed greatly improved performance. Then, we compared the GBO algorithm with four heuristic algorithms, and the results verified the superiority of the GBO algorithm. In addition, we compared LDNM/GBO with five classic machine learning methods, and the experimental results proved that LDNM/GBO is a very competitive classifier. Finally, we pruned the trained LDNM to investigate whether the proposed neuronal structure pruning mechanisms are effective. The fine classification results and the simplified structures of the pruned LDNMs proved the reliability of the neuronal structure pruning mechanisms.
In the future, we plan to apply LDNM to solve other problems to verify its effectiveness. In addition, the classification result of LDNM requires further explanation.