Next Article in Journal
The Progress in Bioprinting and Its Potential Impact on Health-Related Quality of Life
Previous Article in Journal
Real-Time Tracking of Laryngeal Motion via the Surface Depth-Sensing Technique for Radiotherapy in Laryngeal Cancer Patients
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Learning Causal Biological Networks with Parallel Ant Colony Optimization Algorithm

Beijing Municipal Key Laboratory of Multimedia and Intelligent Software Technology, Beijing Institute of Artificial Intelligence, Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(8), 909; https://doi.org/10.3390/bioengineering10080909
Submission received: 12 June 2023 / Revised: 22 July 2023 / Accepted: 27 July 2023 / Published: 31 July 2023
(This article belongs to the Section Biosignal Processing)

Abstract

:
A wealth of causal relationships exists in biological systems, both causal brain networks and causal protein signaling networks are very classical causal biological networks (CBNs). Learning CBNs from biological signal data reliably is a critical problem today. However, most of the existing methods are not excellent enough in terms of accuracy and time performance, and tend to fall into local optima because they do not take full advantage of global information. In this paper, we propose a parallel ant colony optimization algorithm to learn causal biological networks from biological signal data, called PACO. Specifically, PACO first maps the construction of CBNs to ants, then searches for CBNs in parallel by simulating multiple groups of ants foraging, and finally obtains the optimal CBN through pheromone fusion and CBNs fusion between different ant colonies. Extensive experimental results on simulation data sets as well as two real-world data sets, the fMRI signal data set and the Single-cell data set, show that PACO can accurately and efficiently learn CBNs from biological signal data.

1. Introduction

The rapid development of science and technology yields a huge amount of biological signal data and also drives the development of related fields in biological systems [1,2]. For example, the invention of functional magnetic resonance imaging (fMRI) facilitates to learn causal brain networks from brain activity [3,4,5], the creation of intracellular multicolor flow cytometry allows more quantitative simultaneous observations of multiple signaling molecules in many thousands of individual cells and making it easier to infer causal protein signaling networks among protein biomolecules [6,7] and the invention of Single-cell RNA sequencing (ScRNA-seq) yields large amounts of gene expression data, bringing new research perspectives to learn the causal regulatory relationships between different genes [8,9]. A causal biological network (CBN) is a set of nodes and directed edges that can succinctly represent the causal relationships between different types of biological nodes mentioned above [10]. Learning CBNs accurately and efficiently from biological signaling data has been an important issue in recent years [11], and is important for a deeper understanding of the underlying principles in biological mechanisms. In recent years, many CBN learning methods have been proposed, which can be broadly classified into two categories, one based on traditional machine learning and the other on deep learning methods with complex model structures.
Traditional machine learning methods include Linear non-Gaussian Acyclic Model (LiNGAM) [12] based methods, Bayesian Network (BN) based methods [13] and Granger Causality (GC) [14] based methods. etc. Recently, Wei et al. [15] proposed a method for learning CBNs based on BN with pruning strategies. Zhang et al. [16] proposed a CBNs learning method based on truncated matrix power iteration. Gao et al. [17] proposed a Gaussian model for optimal learning of CBNs and showed significant performance improvement. The advantages of these methods are simple models, relative flexibility, and short running times, but the learned CBNs are often not accurate enough and easily fall into local optima.
With the rapid development of deep learning, many deep learning methods are also successfully used to learn CBNs. Yu et al. [18] proposed a graph neural network-based CBNs learning method and successfully applied it to learn causal protein signaling networks. Fan et al. [8] used 3D convolutional neural networks in successfully learning more accurate causal gene regulatory networks. Lu et al. [19] proposed a deep reinforcement learning-based framework and Liu et al. used generative adversarial network [20] and recurrent generative adversarial network [21] for learning causal brain networks. Compared with traditional machine learning methods, the accuracy of learning CBNs from biological signal data using deep learning methods will be improved, but it will cost a lot of time because the model structure is mostly complex.
To solve the above-mentioned problems of existing methods, in this paper, we propose a novel CBNs learning algorithm called parallel ant colony optimization (PACO), which utilizes a parallel ant colony optimization algorithm to learn CBNs from biological signal data. The PACO algorithm consists of three main phases: initialization, parallel ant colony optimization, and pheromone fusion and CBNs fusion phase. During the initialization phase, PACO initializes the parallel ant colony and sets some initial parameters for the ant colonies. In the parallel ant colony optimization phase, the K2 metric is used to measure the quality of the learned CBNs and guides the ant colony search. PACO employs multiple ant colonies to learn the best CBN with the highest K2 metric in parallel. In the pheromone fusion and CBNs fusion phase, all ant colonies are guided to perform a more accurate search by sharing pheromones from the colony with the highest K2 metric to other colonies. Finally, PACO obtains the best CBN from all ant colonies according to the extraction rule. Extensive experimental results on simulation data sets as well as on two real-world data sets, the fMRI signal data set and the Single-cell data set, show that PACO outperforms other state-of-the-art or classical methods in learning CBNs from biological signal data. The main contributions of this paper can be summarized as follows:
  • To the best of our knowledge, this is the first study to employ a parallel ant colony optimization algorithm to learn CBNs from biological signal data. The incorporation of parallelization allows for more accurate and efficient learning of CBNs, which will provide a significant reference for the causal discovery and bioinformatics fields.
  • PACO incorporates the parallel ant colony optimization and information fusion strategy. This approach not only enhances the algorithm’s efficiency and reduces time complexity, but also facilitates the extraction of shared information from multiple data sets, thereby improving the accuracy of learn CBNs and more fully utilizes global information, effectively reducing the probability of falling into a local optima.
  • Numerous experiments conducted on simulation data sets, fMRI signal data sets and Single-cell data set have demonstrated that the proposed method is capable of learning CBNs from different biological signal data, thereby improving inference performance, which has significant implications for a deeper understanding of the underlying causal relationships in biological systems.

2. Related Work

2.1. Causal Biological Networks

The CBNs learned from different types of biological signal data can be specifically subdivided into many types, such as causal brain networks, causal protein signaling networks, causal gene regulatory networks and other CBNs, and we will describe the related work of causal brain networks and causal protein signaling networks in detail in the following. Table 1 shows the introduction of different CBN learning methods.

2.1.1. Causal Brain Networks

Causal brain networks consist of multiple brain nodes and causal interactions between different nodes, and accurate learning of causal brain networks is valuable for understanding the functioning of brain cognition and gaining insight into the pathogenesis of brain diseases [32,33]. In recent years, many studies have emerged to learn causal brain networks from fMRI signal data, Friston et al. [22] first proposed a spectral dynamic causal modeling for learning causal brain networks from fMRI signal data. Zhang et al. [30] first proposed a amortization transformer model for learning causal brain networks from fMRI signal data. Li et al. [28] and Razi et al. [34] extended the model to learn the causal brain networks on large-scale brain regions from fMRI signal data. Ji et al. first proposed to learn causal brain networks using an artificial immune algorithm (AIA) [25] and a recurrent generative adversarial network (RGAN) model [21], with greatly performance. Li et al. [35] explored the dynamic abnormalities of brain function in Parkinson’s disease and the pathophysiological significance behind them by constructing causal brain networks from fMRI signal data.

2.1.2. Causal Protein Signaling Networks

Causal protein signaling networks consist of multiple protein biomolecule nodes and causal relationships between different nodes. Learning causal protein signaling networks accurately from Single-cell data is important for understanding the causal relationships of biomolecules in cells and for gaining insight into the pathogenesis of cell-based diseases. Recently, Zhu et al. [26] designed a causal discovery model based on reinforcement learning that employs a reinforcement learning framework for learning causal protein signaling networks. Zheng et al. [23] first transformed the causal discovery problem from a combinatorial optimization problem to a continuous optimization problem by proposing a continuous optimization structure approach (NoTears) and successfully used it for learning causal protein signaling networks. Baek et al. [27] proposed to learn causal protein signaling networks using a three-track neural network. Li et al. [31] first proposed to learn causal protein signaling networks based on the Deconfounded Functional Structure Estimation. Squires et al. [29] proposed using latent factor causal models to learn causal protein signaling networks. Whitaker et al. [36] discussed the effect of p38 MAPK protein biomolecules on the relationship between cell cycle and apoptotic signaling pathways by constructing and analyzing the BCL2 family of causal protein signaling networks.

2.2. Ant Colony Optimization Algorithm

The ant colony optimization algorithm was originally proposed by Dorigo et al. [37] based on the intelligent behavior of ant colonies during the foraging process. After more than a decade of development, the ant colony algorithm has become one of the most effective algorithms for solving combinatorial optimization problems in swarm intelligence and has been widely used in various fields. Liu et al. [24] first used the ant colony algorithm to learn the effective connectivity of causal brain networks from fMRI signal data and to quantitatively characterize the strength of the connectivity. Liang et al. [38] proposed an improved context-based ant colony optimization algorithm and applied it to travel route planning. However, the tendency to fall into local optima is still one of the main factors limiting the performance of the algorithm.

3. The Parallel Ant Colony Optimization Algorithm

In this section, we will introduce a new algorithm to learn CBNs more accurately and efficiently from biological signal data.

3.1. Main Idea

To accurately and efficiently learn CBNs from biological signal data, we propose a novel algorithm called the parallel ant colony optimization (PACO). The PACO algorithm comprises three phases: initialization, parallel ant colony optimization, and pheromone fusion and CBNs fusion. Specifically, the PACO algorithm is a score-and-search approach for learning CBNs from biological signal data, utilizing the K2 metric to evaluate the quality of CBNs and guide the parallel ant colony to search for the global optimal CBNs. Additionally, we introduce a new information fusion mechanism that merges and updates the pheromones of all colonies after the completion of the same iteration of all ant colonies, serving as the initial pheromones of the colonies in the next iteration. When all iterations are completed, the optimal CBNs learned by all colonies are merged into an adjacency matrix. The final CBN is obtained by setting the extraction rules. Figure 1 illustrates the flowchart of the PACO algorithm.

3.2. Initialization

In the parallel ant colony optimization algorithm, a CBN can be represented as G = < V , E > , where V is a set of biological nodes and E is a set of arcs with each arc representing a causal interaction between two biological nodes. A CBN uses a graph structure and a set of parameters to encode uniquely the joint probability distribution of the domain variables X = { X 1 , X 2 , X 3 , · · · X n } :
P ( X 1 , X 2 , X 3 , · · · X n ) = i = 1 n P ( X i | ( X i ) ) .
First we initialize N ant colonies with N u m ants in each colony, then we initialize an empty CBN G i ( 0 ) ( 1 i N ) for each colony. Since ants produce pheromones and the concentration of pheromones changes continuously during the movement, we need to initialize a pheromone matrix for each colony. Finally, we need to initialize some parameters for the algorithm, such as the number of iterations N C , the number of iterations l s t e p for local search, the initial information concentration τ 0 , etc. Additionally, we employ the K2 metric in PACO to evaluate the quality of CBNs. The K2 metric is a famous evaluation measure for learning CBNs from biological signal data, and the initial expression for the K2 metric is:
P ( G , D ) = P ( G ) · i = 1 n j = 1 q i ( r i 1 ) ! ( N i j + r i 1 ) ! k = 1 r i N i j k !
where D is a given training set, G is a possible CBN, r i is the number of possible values of the variable X i , q i is the number of possible configurations for the variables in ( X i ) , and N i j k is the number of cases in D where X i has its kth value and ( X i ) is instantiated to its jth value.

3.3. Parallel Ant Colony Optimization

In this phase of the search process, we open N threads corresponding to N ant colonies searching N biological signal data sets in parallel. In PACO, each ant k ( k = 1 , 2 , · · · N u m ) in N ant colonies starts from the empty CBN G i ( 0 ) ( 1 i N ) and increases one arc at a time until it is impossible to make the K2 metric of the CBNs higher by adding one arc. At time t, the probabilistic transition rule that an ant selects a directed arc a i j between two biological nodes X i and X j from the current set of candidate arcs is defined as:
a i , j = a r g m a x i , j D A k ( t ) { [ τ i j ( t ) ] · [ η i j ( t ) ] β } , i f q q 0 , a I , J , o t h e r w i s e ,
where τ i j ( t ) is the pheromone concentration, η i j ( t ) represents the heuristic information of a i , j , and β is the weighted coefficient which controls η i j ( t ) to influence the selection of arcs. D A k ( t ) ( i , j D A k ( t ) ) is the set of all candidate arcs whose heuristic information is larger than zero; q 0 (0 q 0 < 1) is an initial parameter that determines the relative importance of exploitation versus exploration (exploitation means selecting arcs by pheromone intensity and heuristic information, and exploration means global random selecting arcs); q is a random number uniformly sampled in [0,1]; and I and J are a pair of biological nodes randomly selected according to the probability in the following way:
p i , j k ( t ) = [ τ i j ( t ) ] α · [ η i j ( t ) ] β r , l D A k ( t ) [ τ r l ( t ) ] α ] · [ η r l ( t ) ] β , i f i , j D A k ( t ) , 0 , o t h e r w i s e ,
where α denotes the relative importance of τ r l ( t ) left by ants. The heuristic function η i j is defined as follows:
η i j ( t ) = ω · f ( X i , ( X i X j ) f ( X i , ( X i ) )
where ω is a weighted factor concerned with the arc connecting intensity whose value is defined as:
ω = 1 + I n f ( X i , X j )
where I n f ( X i , X j ) represents the mutual information between the two biological nodes X i and X j . Because the mutual information I n f ( X i , X j ) can objectively reflect whether the two biological nodes in a CBN are dependent and how much the dependency is, thus when the dependency intensity is stronger and the score-increase is larger, the heuristic information becomes greater, and vice versa. The mutual information between two biological nodes X i and X j is defined as:
I n f ( X i , X j ) = x i , x j P ( x i , x j ) l o g P ( x i , x j ) P ( x i ) P ( x j )
after each iteration of the ant colony is performed, the PACO algorithm will carry out the pheromone updating process, which includes local and global updating steps. For the local optimization process, when an ant selects an arc a i j , the pheromone level of the corresponding arc is changed in the following way:
τ i j ( t + 1 ) = ( 1 ρ ) τ i j ( t ) + ρ τ 0
where 0 < ρ 1 is a parameter that controls the pheromone evaporation.

3.4. Pheromone Fusion and CBNs Fusion

The above search process is performed by N colonies in parallel. Each colony finds the best solution from all feasible CBNs learned so far by means of the K2 metric, and performs the global updating for each arc of the current best CBN. The global updating rules at the t t h iteration are shown in Equations (9) and (10):
τ i j ( t ) = ( 1 ρ ) τ i j ( t 1 ) + ρ Δ τ i j ( t )
Δ τ i j ( t ) = 1 | f ( G + : D ) | , i f a i j G + τ i j ( t 1 ) , o t h e r w i s e
When all N colonies completed the above process, the algorithm enters the pheromone fusion and CBNs fusion phase, which is divided into two parts: when the number of iterations is not satisfied, the algorithm selects the pheromone matrix M m a x corresponding to the CBN with the highest K2 metric from all ant colonies, and updates the pheromone matrices M 1 to M N of all ant colonies to M m a x to guide the continued search of the ant colony during the next iteration. When the number of iterations reaches N C , the algorithm merges all the optimal CBNs learned by N ant colonies into the adjacency matrix G. Then we set the new extraction rule such that G i j = 1 , ( 1 i , j N ) when the value G i j 50 % · N , otherwise G i j = 0. Finally, we extract the optimal CBN G learned from the N biological signal data sets.
Based on the above description, the termination process of PACO is as follows: when the current iteration number t reaches the preset iteration number N C , PACO merges the N CBNs learned by the parallel ant colony, and then extracts and outputs the optimal CBN G according to the designed rules, and the algorithm terminates.

3.5. Algorithm Description and Analysis

The PACO in this paper consists of three main phases: initialization, parallel ant colony optimization, and pheromone fusion and CBNs fusion phase, which are summarized in Algorithm 1. For the initialization phase, first, the PACO algorithm initialize N ant colonies and opens N threads for N colonies simultaneously. Then PACO generates an initial set of empty CBNs G i ( 0 ) ( 1 i N ) and set some parameters for each colony. Finally, N biological signal data sets are input to the algorithm. For the parallel ant colony optimization phase, all ants in N colonies perform the search CBNs in parallel, starting with one empty CBN per ant and adding one arc at a time until the CBN cannot be constructed to have a higher K2 metric. The CBN learned is then locally optimized using O p t i m a z a t i o n ( ) , a function that uses the standard addition, deletion, and inversion operators for arcs. At this point, each ant colony obtains the optimal CBN for the current number of iterations t and updates the global pheromone. For the pheromone fusion and CBNs fusion phase, the algorithm selects the pheromone matrix M m a x corresponding to the CBN with the highest K2 metric from the N CBNs learned in the parallel ant colony optimization phase, and then updates M m a x to the pheromone matrix M 1 to M N to guide each ant colony to continue the search in the next iteration. When the N C iterations are completed, the algorithm merges N optimal CBNs learned by N ant colonies and extracts the optimal CBN G according to the extraction rules we designed.
Algorithm 1: PACO
Bioengineering 10 00909 i001
Based on the description of Algorithm 1, the complexity of PACO can be simply analyzed as follows: let the algorithm input N data sets, N C is the number of iterations, and the number of ants per colony is N u m . Then the time complexity of PACO can be expressed as O ( N ) + O ( N C ) · O ( 3 · N u m + N ) O ( n 2 ) . The time complexity of the non-parallel Ant Optimization Colony (ACO) algorithm can be expressed as O ( 1 ) + O ( N ) · O ( N C ) · O ( 3 · N u m ) O ( n 3 ) . It is obvious that PACO increases the time cost O ( N ) mainly in the pheromone fusion and CBNs fusion process compared to ACO, but for learning CBNs with multiple biological signal data sets, the time cost of PACO is much less than that of ACO, and the advantage becomes more and more obvious the larger the number N of biological signal data sets is.

4. Experimental Result of Learning CBNs

4.1. Data Description

4.1.1. Simulation Data Sets

The generation of the simulation data sets is briefly described here, and more details can be found in [26]. Given a number of nodes v, we generate a d × d upper triangular matrix as the graph binary adjacency matrix, in which the upper entries are sampled independently from Bern(0.5). We assign edge weights independently from Unif ([−1.5, −0.5] ∪ [0.5, 1.5]) to obtain a weight matrix W R d × d , and then sample x = W T x + λ R d from both Gaussian and non-Gaussian noise models. We choose unit noise variance in both models and use m = 200 samples as the sub-data set. The variables are then randomly ordered. Finally, we generate 10 simulation data sets sim1 to sim10 and each simulation data set consists of different numbers of nodes (v = 5, 10, 30, 50, 100) and different numbers of sub-data sets (N = 20, 50) in series. The input of each colony in PACO corresponds to each sub-data set of a simulation data set. Other algorithms take a complete simulation data set as input.

4.1.2. fMRI Signal Data Sets

A causal brain network for representing the effective connectivity of different brain regions can be represented by a CBN. The Smith data set [39] is a set of fMRI signal data sets given by Smith et al. The data set is availableat http://www.fmrib.ox.ac.uk/datasets/netsim/index.html (accessed on 11 June 2023), which can be used to verify the accuracy of different methods to identify functional and effective brain connectivity and contains 28 data sets. Each data set contains different number of brain regions (Nodes), scan time (Session), repetition time of the pulse, noise, and several other influencing factors. We select the second of 28 data sets, which has 10 nodes and contains 50 subjects with 11 edges in the ground-truth to validate the performance of our method.

4.1.3. Single-Cell Data Sets

Learning causal protein signaling networks from human immune Single-cell data can also be considered as the process of learning CBNs. The real multi-parameter fluorescence-activated cell sortera data set [6] to learn causal protein signaling networks based on expression levels of proteins and phospholipids, and the Single-cell data sets are availableat https://www.science.org/doi/10.1126/science.1105809#supplementary (accessed on 11 June 2023). This is a widely used bioinformatics data set for research on graphical models. The data set offers continuous measurements of expression levels of multiple phosphorylated proteins and phospholipid components in human immune system cells. There are 14 sub-data sets with respect to 14 different biochemical experiments, and the number of data points in each sub-data set ranges from 723 to 917. There are 11 signaling nodes in each sub-data set, and each signaling node represents a phosphorylated protein molecule in the research of the human primary T cell signaling pathway. Over the past two decades, classical biochemistry and genetic analysis have constructed a protein network that can be taken as a ground-truth network. The ground-truth network contains 17 high-confidence causal edges.

4.2. Evaluation Metrics

We compared the learned CBN to ground-truth CBN on the for most common graph metrics: (1) Precision; (2) Recall; (3) F1-measure( F 1 ); (4) Structural Hamming Distance ( S H D ); (5) T i m e . Specifically, Precision, Recall, and F 1 can be defined asfollows:
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l .
In this paper, we use edges to denote the directed connectivity relationship between two nodes in a CBN. T P denotes the number of edges that exist in both the learned CBN and the ground-truth CBN; F P denotes the number of edges that exist only in the learned CBN compared to the ground-truth CBN; and F N denotes the number of edges that exist in the ground-truth CBN but are not learned. Thus, Precision and Recall range from 0 to 1, and F 1 is their harmonic. S H D is the total number of edge additions, deletions, and reversals needed to convert the learned CBN into the ground-truth network. The S H D can be calculated as
S H D = R e d u + M i s s + R e v e ,
where R e d u represents the number of redundant edges that need to be removed, M i s s represents the number of missing edges that need to be added, and R e v e represents the number of edges in the opposite direction that need to be reversed. To indicate the time spent by the algorithm in seconds, we use the T i m e metric.

4.3. Contrast Algorithm Introduction and Experimental Setup

To intuitively illustrate the competitiveness of our PACO, we compare with 5 other state-of-the-art or classic algorithms. These algorithms include: continuous optimization for structure learning (NoTears) [23], deep reinforcement learning (DRL) [19], greedy equivalence search (GES) [40], DAG Structure Learning with Graph Neural Networks (DAG-GNN) [18], Artificial Immune Algorithm (AIA) [25]. In addition, to demonstrate the superiority of the parallel strategy of the PACO algorithm, we also compare it with the non-parallel ant optimization colony (ACO) [24] as our ablation experiment. Among the above algorithms, NoTears and DAG-GNN are two that have been successfully applied to learn causal protein signaling networks and achieved good performance on Single-cell data sets. DRL, GES, AIA and ACO are four algorithms to learn causal brain networks that achieved good performance on Smith’s fMRI data set. We use the gCastle toolbox proposed by [41] for the implementation of all publicly available comparison algorithms, and the code is availableat https://github.com/huawei-noah/trustworthyAI (accessed on 11 June 2023). The code for PACO is availableat https://github.com/ZJH66/PACO (accessed on 11 June 2023).
To compare with other algorithms in a fair and appropriate way, we set the parameters of all comparison algorithms to the default values in the citation. The parameters of PACO include the weights for the pheromone trail ( α ) and for the heuristic information ( β ), the controls of the pheromone evaporation ( ρ ), the relative importance of the exploitation versus exploration ( q 0 ), the number of iterations ( N C ), and the number of ants ( N u m ). After a large number of experimental tests, we find that the set of parameters α = 1.8 , β = 2 , ρ = 0.35 , q 0 = 0.75 performed well on most of simulation data sets, and the parameters N C and N u m are mainly associated with the stability and convergence speed of the algorithm. Thus we test on the simulation data to determine a better parameter configuration of N C and N u m for PACO. We find that as Num and N C increase, the learning performance of PACO gets better and better and takes more and more time, and the algorithm achieves the highest accuracy and stabilizes when N u m = 20 and N C = 10, from which we determine the final parameters. During all experiments, we employ the control variate technique that the value of a single parameter is changed, while keeping the values of other parameters fixed. The parameter settings of all algorithms are shown in Table 2.
We choose 5 evaluation metrics Precision, Recall, F 1 , S H D and T i m e to evaluate the performance of different algorithm and compare PACO with 6 other algorithms using 10 simulation data sets, a set of real fMRI signal data set and a set of Single-cell data sets. To reduce the effect of algorithm randomness on the experimental results, we run all algorithms 10 times on all data sets and take the average. After the validation of the performance and effectiveness of the PACO algorithm, we further compare it on real fMRI signal data sets and real Single-cell data set. The experimental platform is a PC with Intel Core i5-8300, 16 GB RAM, 2.30 GHz CPU, and Windows 10.

4.4. The Results of Learning CBNs from Simulation Data Sets

We comprehensively test and compare the above 7 algorithms on the generated 10 simulation data sets, each consisting of a different number of nodes v and a different number of sub-data sets N. An algorithm performs well when it gets higher values of Precision, Recall and F 1 and lower values of S H D and T i m e . Note that when we test PACO, we use the data set of all the sub-data sets concatenated, the number of colonies N in the PACO algorithm is equal to the number of sub-data sets N in a simulation data set, and the input of each colony is a sub-data set in a simulation data set. When we test other algorithms, we need to concatenate all the sub-data sets of a simulation data set as the input.
From Table 3, we can find that PACO outperforms the other 6 algorithms in all metrics on the 10 simulation data sets. Specifically, following the two chains Sim1-Sim3-Sim5-Sim7-Sim9 and Sim2-Sim4-Sim6-Sim8-Sim10, each chain has 5 data sets with gradually increasing number of nodes, 5, 10, 30, 50, and 100, respectively, and the difference between the two chains is that each data set in the first chain contains 20 sub-data sets, while each data set in the second chain consists of 50 sub-data sets. Overall, the F 1 of all algorithms, including PACO, decreases as the number of nodes v and the number of sub-data sets N increase. While the T i m e increases with the number of nodes as well as the number of sub-data sets. When the number of nodes increases to 100, PACO still achieves F 1 values of 0.71 and 0.70 on both Sim9 and Sim10 with T i m e of 5.87 s and 7.66 s, respectively, which is far ahead of the other algorithms and keeps the highest performance. Then we divide the number of nodes into 5 groups to see the impact of the number of sub-data sets on the performance of the algorithm. It is obvious that as the number of sub-data set increases from 20 to 50, the F 1 and S H D of most of the compared algorithms get worse and the time cost of all algorithms increases, but the F 1 and S H D of PACO get better instead, and we think that the pheromone fusion and CBNs fusion rules we set play a role. In addition, we found that the larger the number of nodes and the larger the number of sub-data set, the more obvious the time advantage of the PACO algorithm. For example, for the Sim10 data set, the number of nodes are 50 and sub-data sets are 100, the time cost of ACO is 13.67 s, while the time cost of PACO is 7.66 s, which is a significant improvement in time performance, which demonstrates the superiority of our parallel strategy on large-sample multi-data set data. For a more visual representation of the stability of the algorithm, we plot the average results on the above 10 simulation data sets in a box plot, as shown in Figure 2.
The above experimental results can fully demonstrate that our proposed algorithm has a more stable performance in all evaluation metrics and a significant improvement in time performance and learning accuracy compared to other algorithms. Next, we will further discuss the performance of the algorithm on real fMRI signal data sets and Single-cell data sets.

4.5. The Results of Learning Causal Brain Networks from fMRI Signal Data Sets

In this section, we test the performance of algorithms to learn brain effective connectivity networks from fMRI signal data. The results are shown in Table 4, and the causal brain networks learned by each algorithm are shown in Figure 3. Note that for PACO, we input 50 subjects simultaneously into 50 ant colonies for parallel search, and for the other algorithms, we input 50 sub-data sets in series into the algorithm consecutively.
Combining Table 4 and Figure 3, we can find that DAG-GNN identifies 5 true effective connectivity edges, and the time cost 22.6 s, and the performance of each metric is much lower than other algorithms, which we speculate that this is due to the complex deep learning model used by the algorithm and the assumption that the acyclic graph constraint will lead to a significant performance degradation when generating cyclic graphs. NoTears and DRL identify 7 and 8 true effective connectivity edges, respectively, but the 2 algorithms themselves take more time due to the deep learning and reinforce learning model used. GES identifies 8 true effective connectivity edges, but the Precision is low due to the large number of redundant edges generated; AIA identifies the 7 true effective connectivity edges, and performance is moderate in all metrics, with no outstanding advantages. ACO identifies 9 true effective connectivity edges, Recall, Precision and F 1 are 0.82. PACO identifies 9 true effective connectivity edges and outperforms ACO on all evaluation metrics. Moreover, ACO consumes 0.98 s while PACO consumes 0.49 s, which is almost a times improvement in time performance, which proves that our pheromone fusion and CBNs fusion mechanism and parallel search strategy have achieved significant results. In summary, PACO can learn causal brain networks more accurately and efficiently.

4.6. The Results of Learning Causal Protein Signaling Networks from Simulation Data Sets

We further validate the performance of the PACO algorithm on the Single-cell data set. For PACO, we input 14 sub-data sets simultaneously into 14 ant colonies for parallel search, and for other algorithms, we input 14 sub-data sets in series into the algorithm consecutively. The results are shown in Table 5, and the causal protein signaling networks learned by each algorithm are shown in Figure 4.
Combining Table 5 and Figure 4, we can find that PACO outperforms all other algorithms in all evaluation metrics, where S H D is 15 with both NoTears and DRL, but PACO is far ahead of NoTears and DRL in terms of time. NoTears, DRL, and DAG-GNN are at the same level of accuracy in learning causal protein signaling networks, and all learn nearly half of the connected edges of protein signals correctly, but all three algorithms have a significant disadvantage in time performance due to the relatively complex neural networks and deep learning models they all use. The Recall of GES is 0.41, but the Precision is 0.2, which leads to an F 1 of only 0.27 and a S H D of up to 30, obviously learning a large number of redundant signal connection edges. AIA and ACO algorithms also have the same characteristics as GES algorithm in terms of evaluation metrics. The above three algorithms have short running time but less accuracy. PACO has an obvious advantage in accuracy with Precision, Recall and F 1 of 0.53, which proves that our pheromone fusion and CBNs fusion strategy can improve the accuracy of the algorithm in learning CBNs, and also has an obvious advantage in time, which proves that our parallel strategy is effective. In summary, PACO can learn causal protein signaling networks accurately and efficiently.

5. Conclusions

In this paper, we introduce a novel parallel ant colony optimization algorithm (PACO) for learning CBNs from biological signal data. Our experiments on the generated simulation data set, the real fMRI signal data set and the real Single-cell data set show that PACO has significant improvements in accuracy performance and time performance compared to the state-of-the-art algorithm and the advantage of this method will be more obvious when dealing with large scale multiple data sets. Compared with the non-parallel ant colony optimization (ACO) algorithm, PACO shows a significant improvement in accuracy and time performance, which proves the effectiveness of the parallel strategy and pheromone fusion and CBNs fusion mechanism. In future work, we plan to further investigate the acyclicity constraint during parallel ant colony optimization to reduce the search space to improve the performance.

Author Contributions

Conceptualization, J.Z., J.J. and J.L.; methodology, J.Z. and J.L.; software, J.Z.; validation, J.Z.; formal analysis, J.L.; investigation, J.J. and J.L.; resources, J.J. and J.L.; data curation, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, J.L.; visualization, J.Z.; supervision, J.J. and J.L.; project administration, J.L.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly funded by National Natural Science Foundation of China Research Program (62106009, 62276010), in part by R&D Program of Beijing Municipal Education Commission (KM202210005030, KZ202210005009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Babur, Ö.; Luna, A.; Korkut, A.; Durupinar, F.; Siper, M.C.; Dogrusoz, U.; Jacome, A.S.V.; Peckner, R.; Christianson, K.E.; Jaffe, J.D.; et al. Causal interactions from proteomic profiles: Molecular data meet pathway knowledge. Patterns 2021, 2, 100257. [Google Scholar] [CrossRef]
  2. Paul, I.; Bolzan, D.; Youssef, A.; Gagnon, K.A.; Hook, H.; Karemore, G.; Oliphant, M.U.; Lin, W.; Liu, Q.; Phanse, S.; et al. Parallelized multidimensional analytic framework applied to mammary epithelial cells uncovers regulatory principles in EMT. Nat. Commun. 2023, 14, 688. [Google Scholar] [CrossRef]
  3. Ji, J.; Zou, A.; Liu, J.; Yang, C.; Zhang, X.; Song, Y. A survey on brain effective connectivity network learning. IEEE Trans. Neural Netw. Learn. Syst. 2021, 34, 1879–1899. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, J.; Ji, J.; Jia, X.; Zhang, A. Learning Brain Effective Connectivity Network Structure Using Ant Colony Optimization Combining with Voxel Activation Information. IEEE J. Biomed. Health Inform. 2020, 24, 2028–2040. [Google Scholar] [CrossRef]
  5. ElNakieb, Y.; Ali, M.T.; Elnakib, A.; Shalaby, A.; Mahmoud, A.; Soliman, A.; Barnes, G.N.; El-Baz, A. Understanding the Role of Connectivity Dynamics of Resting-State Functional MRI in the Diagnosis of Autism Spectrum Disorder: A Comprehensive Study. Bioengineering 2023, 10, 56. [Google Scholar] [CrossRef]
  6. Sachs, K.; Perez, O.; Pe’er, D.; Lauffenburger, D.A.; Nolan, G.P. Causal protein-signaling networks derived from multiparameter single-cell data. Science 2005, 308, 523–529. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Shalek, A.K.; Satija, R.; Adiconis, X.; Gertner, R.S.; Gaublomme, J.T.; Raychowdhury, R.; Schwartz, S.; Yosef, N.; Malboeuf, C.; Lu, D.; et al. Single-cell transcriptomics reveals bimodality in expression and splicing in immune cells. Nature 2013, 498, 236–240. [Google Scholar] [CrossRef] [Green Version]
  8. Fan, Y.; Ma, X. Gene regulatory network inference using 3d convolutional neural network. In Proceedings of the the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 2–9 February 2021; Volume 35, pp. 99–106. [Google Scholar]
  9. Shu, H.; Zhou, J.; Lian, Q.; Li, H.; Zhao, D.; Zeng, J.; Ma, J. Modeling gene regulatory networks using neural network architectures. Nat. Comput. Sci. 2021, 1, 491–501. [Google Scholar] [CrossRef]
  10. Lages, J.; Shepelyansky, D.L.; Zinovyev, A. Inferring hidden causal relations between pathway members using reduced Google matrix of directed biological networks. PLoS ONE 2018, 13, e0190812. [Google Scholar] [CrossRef] [Green Version]
  11. Badsha, M.B.; Fu, A.Q. Learning causal biological networks with the principle of Mendelian randomization. Front. Genet. 2019, 10, 454043. [Google Scholar] [CrossRef]
  12. Hoyer, P.; Janzing, D.; Mooij, J.M.; Peters, J.; Schölkopf, B. Nonlinear causal discovery with additive noise models. Adv. Neural Inf. Process. Syst. 2008, 21, 689–696. [Google Scholar]
  13. Ji, J.; Wei, H.; Liu, C. An artificial bee colony algorithm for learning Bayesian networks. Soft Comput. 2013, 17, 983–994. [Google Scholar] [CrossRef]
  14. Shimizu, S.; Hoyer, P.O.; Hyvärinen, A.; Kerminen, A.; Jordan, M. A linear non-Gaussian acyclic model for causal discovery. J. Mach. Learn. Res. 2006, 7, 2003–2030. [Google Scholar]
  15. Wei, X.; Zhang, Y.; Wang, C. Bayesian Network Structure Learning Method Based on Causal Direction Graph for Protein Signaling Networks. Entropy 2022, 24, 1351. [Google Scholar] [CrossRef] [PubMed]
  16. Zhang, Z.; Ng, I.; Gong, D.; Liu, Y.; Abbasnejad, E.; Gong, M.; Zhang, K.; Shi, J.Q. Truncated Matrix Power Iteration for Differentiable DAG Learning. Adv. Neural Inf. Process. Syst. 2022, 35, 18390–18402. [Google Scholar]
  17. Gao, M.; Tai, W.M.; Aragam, B. Optimal estimation of Gaussian DAG models. In Proceedings of the International Conference on Artificial Intelligence and Statistics, PMLR, Virtual, 28–30 March 2022; pp. 8738–8757. [Google Scholar]
  18. Yu, Y.; Chen, J.; Gao, T.; Yu, M. DAG-GNN: DAG structure learning with graph neural networks. In Proceedings of the International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 7154–7163. [Google Scholar]
  19. Lu, Y.; Liu, J.; Ji, J.; Lv, H.; Huai, M. Brain Effective Connectivity Learning with Deep Reinforcement Learning. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 6–8 December 2022; pp. 1664–1667. [Google Scholar]
  20. Liu, J.; Ji, J.; Xun, G.; Yao, L.; Huai, M.; Zhang, A. EC-GAN: Inferring brain effective connectivity via generative adversarial networks. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 4852–4859. [Google Scholar]
  21. Ji, J.; Liu, J.; Han, L.; Wang, F. Estimating Effective Connectivity by Recurrent Generative Adversarial Networks. IEEE Trans. Med. Imaging 2021, 40, 3326–3336. [Google Scholar] [CrossRef]
  22. Friston, K.J.; Kahan, J.; Biswal, B.; Razi, A. A DCM for resting state fMRI. Neuroimage 2014, 94, 396–407. [Google Scholar] [CrossRef] [Green Version]
  23. Zheng, X.; Aragam, B.; Ravikumar, P.K.; Xing, E.P. Dags with no tears: Continuous optimization for structure learning. Adv. Neural Inf. Process. Syst. 2018, 31, 9492–9503. [Google Scholar]
  24. Liu, J.; Ji, J.; Zhang, A.; Liang, P. An ant colony optimization algorithm for learning brain effective connectivity network from fMRI data. In Proceedings of the 2016 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Shenzhen, China, 15–18 December 2016; pp. 360–367. [Google Scholar]
  25. Ji, J.; Liu, J.; Liang, P.; Zhang, A. Learning effective connectivity network structure from fMRI data based on artificial immune algorithm. PLoS ONE 2016, 11, e0152600. [Google Scholar] [CrossRef]
  26. Zhu, S.; Ng, I.; Chen, Z. Causal discovery with reinforcement learning. arXiv 2019, arXiv:1906.04477. [Google Scholar]
  27. Baek, M.; DiMaio, F.; Anishchenko, I.; Dauparas, J.; Ovchinnikov, S.; Lee, G.R.; Wang, J.; Cong, Q.; Kinch, L.N.; Schaeffer, R.D.; et al. Accurate prediction of protein structures and interactions using a three-track neural network. Science 2021, 373, 871–876. [Google Scholar] [CrossRef]
  28. Li, G.; Liu, Y.; Zheng, Y.; Li, D.; Liang, X.; Chen, Y.; Cui, Y.; Yap, P.T.; Qiu, S.; Zhang, H.; et al. Large-scale dynamic causal modeling of major depressive disorder based on resting-state functional magnetic resonance imaging. Hum. Brain Mapp. 2020, 41, 865–881. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Squires, C.; Yun, A.; Nichani, E.; Agrawal, R.; Uhler, C. Causal structure discovery between clusters of nodes induced by latent factors. In Proceedings of the Conference on Causal Learning and Reasoning, PMLR, Eureka, CA, USA, 11–13 April 2022; pp. 669–687. [Google Scholar]
  30. Zhang, Z.; Zhang, Z.; Ji, J.; Liu, J. Amortization Transformer for Brain Effective Connectivity Estimation from fMRI Data. Brain Sci. 2023, 13, 995. [Google Scholar] [CrossRef] [PubMed]
  31. Li, C.; Shen, X.; Pan, W. Nonlinear causal discovery with confounders. J. Am. Stat. Assoc. 2023, 1–10. [Google Scholar] [CrossRef]
  32. Abazid, M.; Houmani, N.; Dorizzi, B.; Boudy, J.; Mariani, J.; Kinugawa, K. Weighted brain network analysis on different stages of clinical cognitive decline. Bioengineering 2022, 9, 62. [Google Scholar] [CrossRef]
  33. Chiarion, G.; Sparacino, L.; Antonacci, Y.; Faes, L.; Mesin, L. Connectivity Analysis in EEG Data: A Tutorial Review of the State of the Art and Emerging Trends. Bioengineering 2023, 10, 372. [Google Scholar] [CrossRef]
  34. Razi, A.; Seghier, M.L.; Zhou, Y.; McColgan, P.; Zeidman, P.; Park, H.J.; Sporns, O.; Rees, G.; Friston, K.J. Large-scale DCMs for resting-state fMRI. Netw. Neurosci. 2017, 1, 222–241. [Google Scholar] [CrossRef]
  35. Li, K.; Tian, Y.; Chen, H.; Ma, X.; Li, S.; Li, C.; Wu, S.; Liu, F.; Du, Y.; Su, W. Temporal Dynamic Alterations of Regional Homogeneity in Parkinson’s Disease: A Resting-State fMRI Study. Biomolecules 2023, 13, 888. [Google Scholar] [CrossRef]
  36. Whitaker, R.H.; Cook, J.G. Stress relief techniques: p38 MAPK determines the balance of cell cycle and apoptosis pathways. Biomolecules 2021, 11, 1444. [Google Scholar] [CrossRef]
  37. Dorigo, M.; Birattari, M.; Stutzle, T. Ant colony optimization. IEEE Comput. Intell. Mag. 2006, 1, 28–39. [Google Scholar] [CrossRef]
  38. Liang, S.; Jiao, T.; Du, W.; Qu, S. An improved ant colony optimization algorithm based on context for tourism route planning. PLoS ONE 2021, 16, e0257317. [Google Scholar] [CrossRef] [PubMed]
  39. Smith, S.M.; Miller, K.L.; Salimi-Khorshidi, G.; Webster, M.; Beckmann, C.F.; Nichols, T.E.; Ramsey, J.D.; Woolrich, M.W. Network modelling methods for FMRI. Neuroimage 2011, 54, 875–891. [Google Scholar] [CrossRef] [PubMed]
  40. Chickering, D.M. Optimal structure identification with greedy search. J. Mach. Learn. Res. 2002, 3, 507–554. [Google Scholar]
  41. Zhang, K.; Zhu, S.; Kalander, M.; Ng, I.; Ye, J.; Chen, Z.; Pan, L. gcastle: A python toolbox for causal discovery. arXiv 2021, arXiv:2111.15155. [Google Scholar]
Figure 1. The flowchart of the PACO algorithm.
Figure 1. The flowchart of the PACO algorithm.
Bioengineering 10 00909 g001
Figure 2. The average results of 7 algorithms on 10 simulation data sets corresponding to 5 metrics. The box represents the middle 50% of the data, with a horizontal line inside the box representing the mean. The horizontal axis represents 7 algorithms, and the vertical axis is the value of the evaluation metrics.
Figure 2. The average results of 7 algorithms on 10 simulation data sets corresponding to 5 metrics. The box represents the middle 50% of the data, with a horizontal line inside the box representing the mean. The horizontal axis represents 7 algorithms, and the vertical axis is the value of the evaluation metrics.
Bioengineering 10 00909 g002
Figure 3. The ground-truth and the causal brain networks learned from fMRI signal data set. The horizontal and vertical coordinates indicate the corresponding brain regions of interest and the blue grid indicates effective connectivity between the two corresponding brain regions.
Figure 3. The ground-truth and the causal brain networks learned from fMRI signal data set. The horizontal and vertical coordinates indicate the corresponding brain regions of interest and the blue grid indicates effective connectivity between the two corresponding brain regions.
Bioengineering 10 00909 g003
Figure 4. The ground-truth and the causal protein signaling networks learned from Single-cell data. The connection curves in the graph represent the signaling networks between the 11 phosphorylated protein biomolecules.
Figure 4. The ground-truth and the causal protein signaling networks learned from Single-cell data. The connection curves in the graph represent the signaling networks between the 11 phosphorylated protein biomolecules.
Bioengineering 10 00909 g004
Table 1. The introduction of different CBN learning methods.
Table 1. The introduction of different CBN learning methods.
CategoryMethodsYearsCategoryMethodsYears
Causal
Brain
Network
spectral Dynamic
Causal Modeling (spDCM)
2014 [22]Causal
Protein
Signaling
Network
Continuous
Optimization (NoTears)
2018 [23]
Ant Colony
Optimization (ACO)
2016 [24]Graph Neural
Network (DAG-GNN)
2019 [18]
Artificial Immune
Algorithm (AIA)
2016 [25]Reinforcement
Learning (RL)
2019 [26]
Generative Adversarial
Network (GAN)
2020 [20]Three Track
Neural Network (TTNN)
2021 [27]
Large-scale Dynamic
Causal Mode (PEB)
2020 [28]Latent Factor
Causal Models (LFCMs)
2022 [29]
Recurrent Generative
Adversarial Network (RGAN)
2021 [21]Truncated Matrix
Power Iteration (TMPI)
2022 [16]
Deep Reinforcement
Learning (DRL)
2022 [19]BN with Pruning
Strategies (CO-CDG)
2022 [15]
Amortization
Transformer (AT-EC)
2023 [30]Deconfounded Functional
Structure Estimation (DeFuSE)
2023 [31]
Table 2. Parameter settings of 7 algorithms.
Table 2. Parameter settings of 7 algorithms.
AlgorithmsParameters
NoTears [23] λ 1 = 0.1 ,   m a x i t e r = 100 ,   t h r e s h o l d = 0.3
DRL [19] e p o c h = 50 ,   α = 0.99
GES [40] k = 0.01 ,   N = 10
DAG-GNN [18] e p o c h = 300 ,   η = 10, γ = 1
AIA [25] P s = 0.5 ,   P c = 0.6 ,   P m = 0.4 ,
T = 150 ,   N = 80 ,   M = 70
ACO [24] α = 1.8 ,   β = 1.5 ,   ρ = 0.6 ,  
q 0 = 1.2 ,   N C = 10 ,   N u m = 30
PACO α = 1.2 ,   β = 2 ,   ρ = 0.35 ,   q 0 = 0.75
N C = 10 ,   N u m = 20
Table 3. Comparisons of 7 algorithms on 10 simulation data sets.
Table 3. Comparisons of 7 algorithms on 10 simulation data sets.
Data (v,N)MetricsAlgorithms
NoTearsDRLGESDAG-GNNAIAACOPACO
Sim1(5,20)Precision0.620.610.550.380.600.570.79
Recall0.660.510.350.270.600.620.75
F 1 0.640.560.450.320.600.590.77
S H D 3456432
T i m e (s)2.7614.125.1936.121.980.790.51
Sim2 (5,50)Precision0.590.620.510.350.600.560.75
Recall0.610.520.350.350.500.620.77
F 1 0.600.570.430.350.550.590.76
S H D 3467532
T i m e (s)4.5119.2510.3245.342.731.420.78
Sim3 (10,20)Precision0.590.620.530.400.600.560.77
Recall0.530.510.430.300.540.560.75
F 1 0.560.560.480.350.570.560.76
S H D 9912131074
T i m e (s)4.1217.228.9139.732.361.340.68
Sim4 (10,50)Precision0.570.550.490.380.600.520.75
Recall0.550.510.450.400.520.540.75
F 1 0.560.530.470.390.560.530.75
S H D 91013159103
T i m e (s)7.8925.6413.6748.264.412.860.91
Sim5 (30,20)Precision0.550.620.540.370.580.530.73
Recall0.500.520.340.270.590.530.71
F 1 0.530.570.440.320.590.530.72
S H D 1715192314179
T i m e (s)7.5424.3314.9854.664.613.641.56
Sim6 (30,50)Precision0.560.630.550.390.610.540.75
Recall0.560.610.350.310.540.560.73
F 1 0.560.620.450.350.580.550.74
S H D 1513182115167
T i m e (s)9.1430.557.1650.375.694.791.95
Sim7 (50,20)Precision0.550.610.480.360.620.600.73
Recall0.530.510.440.270.520.620.71
F 1 0.540.560.460.310.570.610.72
S H D 36344348333019
T i m e (s)13.3645.6218.9675.335.975.652.73
Sim8 (50,50)Precision0.540.590.510.380.630.570.72
Recall0.520.530.320.270.540.610.70
F 1 0.530.560.430.320.580.590.71
S H D 37344149333221
T i m e (s)16.6457.8923.6786.517.157.113.15
Sim9 (100,20)Precision0.520.560.550.380.610.550.71
Recall0.520.540.350.420.520.530.71
F 1 0.520.550.450.400.570.540.71
S H D 68607986576148
T i m e (s)32.7698.7541.63139.8710.3610.885.87
Sim10 (100,50)Precision0.490.510.450.350.580.540.71
Recall0.510.510.470.330.600.540.69
F 1 0.500.510.460.340.590.540.70
S H D 71717790556046
T i m e (s)42.36135.552.36159.6216.9813.677.66
The bold values indicate that the algorithm achieved the best results.
Table 4. Comparisons of 7 algorithms on the fMRI signal data set.
Table 4. Comparisons of 7 algorithms on the fMRI signal data set.
AlgorithmsPrecisionRecall F 1 SHDTime (s)
NoTears0.470.640.5499.20
DRL0.570.730.64616.10
GES0.380.730.50132.91
DAG-GNN0.330.450.381322.60
AIA0.640.640.6451.35
ACO0.820.820.8230.98
PACO0.830.910.8720.49
The bold values indicate that the algorithm achieved the best results.
Table 5. Comparisons of 7 algorithms on the Single-cell data set.
Table 5. Comparisons of 7 algorithms on the Single-cell data set.
AlgorithmsPrecisionRecall F 1 SHDTime (s)
NoTears0.440.470.451535.30
DRL0.470.470.471563.20
GES0.200.410.273011.60
DAG-GNN0.440.410.4217122.70
AIA0.190.350.24313.80
ACO0.250.470.33263.10
PACO0.530.530.53151.90
The bold values indicate that the algorithm achieved the best results.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhai, J.; Ji, J.; Liu, J. Learning Causal Biological Networks with Parallel Ant Colony Optimization Algorithm. Bioengineering 2023, 10, 909. https://doi.org/10.3390/bioengineering10080909

AMA Style

Zhai J, Ji J, Liu J. Learning Causal Biological Networks with Parallel Ant Colony Optimization Algorithm. Bioengineering. 2023; 10(8):909. https://doi.org/10.3390/bioengineering10080909

Chicago/Turabian Style

Zhai, Jihao, Junzhong Ji, and Jinduo Liu. 2023. "Learning Causal Biological Networks with Parallel Ant Colony Optimization Algorithm" Bioengineering 10, no. 8: 909. https://doi.org/10.3390/bioengineering10080909

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop