Next Article in Journal
Analytical and Numerical Treatments of Conservative Diffusions and the Burgers Equation
Next Article in Special Issue
A Measure of Information Available for Inference
Previous Article in Journal
A Spatio-Temporal Entropy-based Framework for the Detection of Trajectories Similarity
Previous Article in Special Issue
A Moment-Based Maximum Entropy Model for Fitting Higher-Order Interactions in Neural Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Novel Brain Complexity Measures Based on Information Theory

Graphics and Imaging Laboratory, University of Girona, 17003 Girona, Spain
*
Authors to whom correspondence should be addressed.
Entropy 2018, 20(7), 491; https://doi.org/10.3390/e20070491
Submission received: 26 April 2018 / Revised: 6 June 2018 / Accepted: 19 June 2018 / Published: 25 June 2018
(This article belongs to the Special Issue Information Theory in Neuroscience)

Abstract

:
Brain networks are widely used models to understand the topology and organization of the brain. These networks can be represented by a graph, where nodes correspond to brain regions and edges to structural or functional connections. Several measures have been proposed to describe the topological features of these networks, but unfortunately, it is still unclear which measures give the best representation of the brain. In this paper, we propose a new set of measures based on information theory. Our approach interprets the brain network as a stochastic process where impulses are modeled as a random walk on the graph nodes. This new interpretation provides a solid theoretical framework from which several global and local measures are derived. Global measures provide quantitative values for the whole brain network characterization and include entropy, mutual information, and erasure mutual information. The latter is a new measure based on mutual information and erasure entropy. On the other hand, local measures are based on different decompositions of the global measures and provide different properties of the nodes. Local measures include entropic surprise, mutual surprise, mutual predictability, and erasure surprise. The proposed approach is evaluated using synthetic model networks and structural and functional human networks at different scales. Results demonstrate that the global measures can characterize new properties of the topology of a brain network and, in addition, for a given number of nodes, an optimal number of edges is found for small-world networks. Local measures show different properties of the nodes such as the uncertainty associated to the node, or the uniqueness of the path that the node belongs. Finally, the consistency of the results across healthy subjects demonstrates the robustness of the proposed measures.

Graphical Abstract

1. Introduction

The human brain is a complex system composed of a set of regions, which are segregated in order to perform specific tasks and are also efficiently integrated in order to share information [1]. The mapping of the structure and the functionality of brain networks is therefore a main challenge in understanding the functioning, as it cannot be studied as a group of independent elements. An important first step to understand how the information is shared, is the generation of a comprehensive map. Felleman and Essen [2] represented the connections of different regions of the human brain by defining a connectivity matrix. Later, the idea of a connectome [1,3] was introduced, which mapped the neural connections in the brain using networks and graph theory [4,5,6].
In a brain network or graph, nodes correspond to brain regions and edges to structural or functional connections [7,8,9]. To model the brain, different graphs can be used: un-directed binary graphs which are the most popular; weighted graphs that assign weights to the edges according to the degree of connectivity between the nodes; and directed graphs that take the influence of one region in another [7]. Once the graph is built, it needs to be analyzed to describe the hidden information in this dense network.
From the connectome, it has been shown that each brain region has a unique pattern of connections (known as connectional fingerprint) [10] that varies across subjects [6] but preserves a similar structure. Different techniques have been applied to describe the topological features of brain networks [11,12,13]. For instance, the independence of large areas, denoted as integration, has been studied by the path length measure, the characteristic path length [14], or the global efficiency [15]. The independence of small subsets, defined as segregation, can be analyzed by the clustering coefficient [14], the transitivity [16] or the modularity [17]. The importance of individual nodes can be defined with centrality measures such as the degree [18], or the density. A good summary of these measures can be found in [12].
Global measures have also been proposed to describe the overall network structure of the brain. Studies such as Kennedy et al. [19] suggested that a functional and structural central circuit with different areas acting as a cluster governed the information distribution and integration in the brain. Clusters are densely interconnected areas and are defined as a rich-club [20,21,22]. Sporns et al. [23] evidenced small-world properties of human brain networks. Small-world networks are systems with a high level of clusterization, like lattice networks, and with small path lengths, like random graphs.
Information theory has been previously used to study the integration and segregation of brain networks [24,25,26]. For instance, the neuronal complexity measure ( C N ) showed a balance between segregation and integration [24,27,28]. Additionally, other measures were proposed such as the matching complexity measure ( C M ) that shows the change in C N after receiving signals from the environment [29], the functional clustering, which finds groups of regions that are more connected among themselves than with the rest [30], the degeneracy measure ( D N ) that describes how structurally different elements are able to perform the same function, and the redundancy measure (R), that describes how identical elements perform the same function [31,32].
Brain network measures are also able to associate different diseases with disruptions [6,33,34,35]. As an example, Sato et al. [36] used the assessment of the graphs entropy to distinguish subjects with and without hyperactivity [36]. Unfortunatelly, the measures that best describe a brain network are still unknown. Therefore, new network measures showing new properties are required to better understand brain networks and their functioning [37].
In this paper, we use a brain network model, where regions correspond to states of a Markov process, to model impulses as random walks on the brain network [38]. Please note that this model differs from the previous ones [24,27,28], where correlations between subsets are used to study the centrality and segregation. This Markov process-based interpretation provides a solid theoretical framework from which global and local measures can be derived. Global measures provide quantitative values to characterize the whole brain network while local measures, which are based on different decompositions of the global measures, are used to quantify the informativeness associated to each node. To evaluate the proposed measures different synthetic model networks, and structural and functional human networks at different scales are considered.

2. Method

2.1. Information Theory Basis

Let the alphabet X be a finite set and X a random variable taking values x in X . The Shannon entropy H ( X ) of a random variable X is defined by
H ( X ) = x X p ( x ) log p ( x ) ,
where p ( x ) = P r [ X = x ] is the probability of the value x. Entropy measures the average uncertainty of a random variable X. All logarithms are base 2 and entropy is expressed in bits. In this paper, the convention 0 log 0 = 0 is used.
Likewise, let Y be a random variable taking values y in Y . The conditional entropy is defined by
H ( Y | X ) = x X p ( x ) y Y p ( y | x ) log p ( y | x ) ,
where p ( y | x ) = P r [ Y = y | X = x ] is the conditional probability. The conditional entropy H ( Y | X ) measures the average uncertainty associated with Y if we know the outcome of X. In general, H ( Y | X ) H ( X | Y ) , and H ( X ) H ( X | Y ) 0 .
The mutual information (MI) between X and Y is defined by
I ( X ; Y ) = H ( X ) H ( X | Y ) = x X y Y p ( x , y ) log p ( x , y ) p ( x ) p ( y ) = x X p ( x ) y Y p ( y | x ) log p ( y | x ) p ( y ) ,
where p ( x , y ) = P r [ X = x , Y = y ] is the joint probability. MI measures the shared information between X and Y. It can be seen that I ( X ; Y ) = I ( Y ; X ) 0 [39].
The relative entropy or Kullback-Leibler distance, D K L ( p , q ) , between two probability distributions p and q, that are defined over the same alphabet X , is defined by
D K L ( p , q ) = x X p ( x ) log p ( x ) q ( x ) .
The relative entropy satisfies that D K L ( p , q ) 0 , with equality if and only if p = q . Kullback-Leibler distance is a basic information theory measure to quantify the dissimilarity between two probability distributions, and other measures, like entropy or mutual information, can be reformulated in terms of this.
A stochastic process or a discrete-time information source X is an indexed sequence of random variables characterized by the joint probability distribution p ( x 1 , x 2 , , x L ) = P r { ( X 1 , X 2 , , X L ) = ( x 1 , x 2 , , x L ) } with ( x 1 , x 2 , , x L ) X L for L 1 [39,40]. The entropy rate or entropy density h of a stochastic process X is defined by
h = lim L 1 L H ( X 1 , X 2 , , X L )
when the limit exists. The entropy rate represents the average information content per symbol in a stochastic process. It is the “uncertainty associated with a given symbol if all the preceding symbols are known” and can be viewed as “the intrinsic unpredictability” or “the irreducible randomness” associated with the chain [41].
A stochastic Markov process [39], or Markov chain, is a discrete stochastic process defined over a set of states X which is described by a transition probability matrix P. In each step, the process makes a transition from its current state i to a new state j with transition probability P i j = p ( x j t + 1 | x i t ) = P r [ X t + 1 = x j | X t = x i ] .
For a stationary Markov process (that is, a Markov process whose statistical properties are invariant to a shift in time), the probability of each state i converge to a stationary distribution μ = { μ 1 , , μ n } after several steps. The stationary or equilibrium probabilities μ i fulfill the relation μ i = j = 1 n μ j P j i and also the reciprocity relation μ i P i j = μ j P j i .
In particular, a Markov process can be considered as a chain of random variables complying with
H ( X L | X 1 , X 2 , , X L 1 ) = H ( X L | X L 1 ) .
An important result is the following theorem: for a stationary Markov chain with stationary distribution μ i , the entropy rate or information content is given by
h = H ( X t + 1 | X t ) = i = 1 n μ i j = 1 n P i j log P i j ,
where μ i is the stationary distribution and P i j is the transition probability from state i to state j.
The excess entropy [42,43,44,45] of an infinite chain is defined by
E = lim L ( H ( X 1 , X 2 , , X L ) L h )
= L = 1 ( H ( L ) h ) ,
where h is the entropy rate of the chain, L is the length of this chain, and H ( L ) = H ( X L | X L 1 , , X 1 ) . The excess entropy can be interpreted as the mutual information between two semi-infinite halves of the chain. Another way of viewing this is that excess entropy is a measure of the apparent memory or structure in the system, that is, the excess entropy measures how much more random the system would become if we suddenly forgot all information about the left half of the string [46]. For a stationary Markov process, excess entropy coincides with mutual information, and, hence, in this case, mutual information can be seen as a measure of the system structure.
The erasure entropy [47] measures the information content of each symbol knowing its context, i.e., the previous and posterior samples. For any stationary process, the erasure entropy is given by
H = lim L H ( X 0 | X L 1 , X 1 L ) ,
where X L 1 symbolizes the previous samples (past) and X 1 L the posterior samples (future).

2.2. Markov Process-Based Brain Model

A brain graph can be defined as a pair of sets G = ( N , E ) , where N is a brain parcellation of n nodes labelled { N 1 , , N n } , and E is a set of m edges between two nodes of N. This graph can be represented by a connectivity matrix C with n × n elements, where C i j gives the connectivity weight between node x i and node x j . Please note that for undirected graphs C i j = C j i .
In this work, brain functions are modeled as a random walk of a particle on the connectivity graph, where the particle randomly goes from node to node defining a path or a sequence of nodes. From node x i , the next node x j is chosen among all nodes connected to node x i , with a probability proportional to the weight C i j . By introducing this model, we are assuming that the next step in the random walk of a neural impulse is determined only by the region and its connections, but not by previous steps of the random walk.
This model leads to a conditional probability P i j = p ( x j t + 1 | x i t ) given by C i j / i C i j . The stationary distribution for this Markov chain assigns probability to node x i proportional to the total weight of the edges emanating from node x i [39]. Thus, the stationary distribution of a node x i is given by
μ i = p ( x i ) = C i C T ,
where C i = j C i j is the total weight of the edges emanating from node i and C T = i j C i j is the sum of the weights of all the edges. Observe that this stationary distribution has an interesting property of locality: it depends only on the total weight and the weight of edges connected to the node and hence, it does not change if the weights in some other part of the graph are changed while keeping constant the total weight.
The definition of this model allows to propose new global and local measures to characterize brain networks. Global measures describe by a single value the whole connectivity of the brain, while local measures assign a value to each brain region, by considering the contribution of the region to the corresponding global measure. In this work, we propose new measures in three different levels: stationary measures, causal measures, and contextual measures. Stationary measures are based on the stationary distribution (i.e., current state). Causal measures are based on how the previous states influence the current state in the random walk. Finally, contextual measures describe how the context (i.e., the previous and future states) is related to the current state. Table 1 summarizes these measures, which are described in the next subsections.

2.3. Global Informativeness Measures

Global measures provide quantitative values to typify the brain connectome as a whole. Depending on which level is considered (stationary, causal, or contextual), three different measures are given: entropy, mutual information, and erasure mutual information.

2.3.1. Entropy

From the stationary distribution μ Equation (11), the Shannon entropy H ( μ ) Equation (1) measures the average uncertainty of the stationary distribution:
H ( μ ) = i = 1 n μ i log μ i .
Since the probability of each region depends on the weight of their edges, this measure will take high values when all nodes in a network have similar connectivity (weights) and will take low values when there is large variability in terms of number of connections or weights. For instance, in the graph shown in Figure 1a, all nodes have the same number of connections (in this case, each connection has the same weight). Thus, the entropy takes the maximum value given by log 2 N = log 2 4 = 2 , where N is the number of nodes. For the graphs of Figure 1b,c, the value of the entropy decreases since the connectivity of the nodes is not equal for all nodes.

2.3.2. Mutual Information

As we have previously mentioned, mutual information measures the shared information between two random variables. From our Markov process-based brain model, we propose as a global connectivity measure the mutual information between two consecutive states of the process:
I ( X t ; X t + 1 ) = H ( X t + 1 ) H ( X t + 1 | X t ) = x i t X x j t + 1 X p ( x i t , x j t + 1 ) log p ( x i t , x j t + 1 ) p ( x i t ) p ( x j t + 1 ) = i = 1 n μ i j = 1 n P i j log P i j μ j .
From Equations (3) and (17), MI can also be seen as the difference between the uncertainty of the states without any knowledge ( H ( X t + 1 ) ) and the uncertainty of the states when the past is known ( H ( X t + 1 | X t ) ). In other words, MI measures the information gained when the previous node is known. The higher the MI, the less random the connections. MI can be seen as a measure of brain structure, since it coincides with excess entropy [48].
In the graph of Figure 1a, the fact of knowing the state at a given time t (present) leads to the states for the time t + 1 (future). For instance, if a given time t the random walk is on the state 1, for the next time t + 1 , the random walk would be either in state 2 or in state 4. Thus, the conditional entropy H ( X t + 1 | X t ) is log 2 2 = 1 . The mutual information, given by the entropy (which corresponds to log 2 4 = 2 ) minus the conditional entropy, is also 1. In the other graphs of Figure 1, the conditional entropy is higher, since there are multiple paths, so the uncertainty of the future step is higher. This fact leads to lower values of MI.

2.3.3. Erasure Mutual Information

The idea of the mutual information measure can be extended by considering not only past states, but also future states. Erasure entropy [47] measures the uncertainty of a system when past and future is known. For a Markov process, this measure can be simplified as
H ( X ) = H ( X t | X t 1 , X t + 1 ) ,
where X symbolizes the whole process. Please note that, in this case, X t 1 symbolizes the past, X t the present, and X t + 1 the future. From this measure and Equation (3), we can extend mutual information as a measure of the decrease of information when the context (i.e., past and future) is known. Thus, we propose a new global measure, called erasure mutual information, defined as
I ( X ) = I ( X t ; X t 1 , X t + 1 ) = H ( X t ) H ( X t | X t 1 , X t + 1 ) .
While the mutual information of Equation (13) measures the loss of information taking into account only the previous node in the random walk (past), erasure mutual information measures the loss of information taking into account the previous node (past) and the next node (future). High values of this measure will show a network with predictable paths to go from node to node, and low values will define a network with several possible paths to go from node to node.
In the graph of Figure 1a, the fact of knowing both past and future states does not reduce the uncertainty of the present (compared with only knowing the past). For instance, if in t 1 the random walk is the state 1 and in time t + 1 in 3, the state in t can be either 2 or 4. This uncertainty is the same for all possible pairs of past and future states. Thus, the conditional entropy H ( X t | X t 1 , X t + 1 ) is log 2 2 = 1 and erasure mutual information is also 1. In the other graphs of Figure 1, the knowledge of future and past states, reduces the uncertainty compared with only knowing the past. For instance, in the graph of Figure 1b, if the past state is 4, the present state can be either 1 or 2, but, if the future state is 3, therefore, the present state is, without uncertainty, state 2. Thus, erasure mutual information takes higher values than mutual information.

2.4. Local Informativeness Measures

In this section, we describe how global measures can be decomposed in order to characterize the degree of informativeness of each state i. When applied to the human connectome, since each state corresponds to an anatomical or functional region, these measures can be seen as the contribution of each node to the whole graph structure, thus, they can describe specific topology of brain areas.

2.4.1. Entropic Surprise

The entropy of X can also be interpreted as the expected value of log p ( x ) , where X is drawn according to probability mass function p ( x ) . Then, in our Markov process-based brain model, the entropic surprise value associated to a brain region x i is defined as
E ( x i ) = log μ i ,
where μ i is the stationary probability of the region x i .
This measure uses the stationary probability of a node without taking into account the previous or the next node in the random walk. Nodes with a low value will be nodes with a large number of connections or weights in its connections. Therefore, high values will define nodes with a low number of connections to other brain regions. This measure is inversely proportional to the logarithm of the well known strength measure, which is defined as the sum of the edge weights emanating from the node.
Some examples are shown in Figure 2. In the left graph, all nodes have the same entropic surprise value, which is given by log 2 2 8 = 2 . Please note that each node has 2 edges and there are 4 bidirectional edges (remember that bidirectional edges are counted twice). In the graphs of Figure 2b,c, it can be seen that nodes with high connectivity take lower entropic values.

2.4.2. Mutual Surprise

The interpretation of mutual information explained in Section 2.3 can be extended to define the information associated with a single node x i X , that is, the information gained on X t + 1 by knowing the original node x i of the impulse. The definition of mutual surprise, denoted by I 1 , can be directly derived from the formula of mutual information Equation (3), taking the contribution of a node x i to I ( X t ; X t + 1 ) , as follows:
I ( X ; Y ) = H ( X ) H ( X | Y ) = x X p ( x ) y Y p ( y | x ) log p ( y | x ) p ( y ) , = x X p ( x ) I 1 ( x ; Y ) .
Mutual surprise was used by DeWeese and Meister [49] to emphasize the fact that the observation of x has moved the estimate of another variable Y towards values that seemed very unlikely prior to the observation. I 1 always takes positive values and it can be shown that I 1 is the only positive decomposition of MI [49].
We reformulate Equation (17) in the framework of the Markov process as follows:
I ( X t ; X t + 1 ) = x i t X p ( x i t ) x j t + 1 X p ( x j t + 1 | x i t ) log p ( x j t + 1 | x i t ) p ( x j t + 1 ) = x i t X p ( x i t ) I 1 ( x i t ; X t + 1 ) ,
where
I 1 ( x i t ; X t + 1 ) = x j t + 1 X p ( x j t + 1 | x i t ) log p ( x j t + 1 | x i t ) p ( x j t + 1 ) = j = 1 n P i j log P i j μ j
expresses the surprise about X t + 1 from observing x i t , i.e., how “surprising” are the nodes connected with the original node. Observe that surprise I 1 ( x i t ; X t + 1 ) is high when p ( X t + 1 | x i t ) is very different from p ( X t + 1 ) (i.e., the stationary distribution), thus, the region x i is connected with regions which are less connected considering all the connections.
I 1 can be seen as the Kullback-Leibler distance see Equation (4) between p ( X t + 1 ) (i.e., the stationary distribution) and p ( X t + 1 | x i t ) (i.e., the distribution of future states if, in the present state, the random walk is on node x i ). Thus, those nodes that are connected with more likely nodes (in terms of the stationary distribution) will lead to low values of I 1 , while those with very specific connections or connected with few unlikely nodes will have high I 1 values. This can be seen, for instance, in node 3 of graph shown in Figure 2b.

2.4.3. Mutual Predictability

DeWeese and Meister [49] defined the specific information I 2 , which we call mutual predictability, using another decomposition of mutual information obtained from Equation (3):
I ( X ; Y ) = H ( Y ) H ( Y | X ) = x X p ( x ) H ( Y ) x X p ( x ) H ( Y | x ) = x X p ( x ) I 2 ( x ; Y ) ,
where
I 2 ( x ; Y ) = H ( Y ) H ( Y | x ) = y Y p ( y ) log p ( y ) + y Y p ( y | x ) log p ( y | x )
expresses the change in uncertainty of Y when x is observed. In our case, we reformulate I 2 in the framework of the Markov process as follows
I 2 ( x i t ; X t + 1 ) = H ( X t + 1 ) H ( X t + 1 | x i t ) = x j t + 1 X p ( x j t + 1 ) log p ( x j t + 1 ) + x j t + 1 X p ( x j t + 1 | x i t ) log p ( x j t + 1 | x i t ) = H ( μ ) + j = 1 n P i j log P i j .
Observe that this measure expresses the difference between global entropy of the graph (i.e., the entropy of the stationary distribution) and entropy of future states of the random walk from node x i . So, this comparison is done globally and, contrarily to the I 1 measure, it is not affected by the stationary probability of the nodes that is connected to. Another property that fulfills I 2 is additivity, i.e., the information obtained about X from two observations, y Y and z Z , is equal to that obtained from y plus that obtained from z when y is known. Additivity is a desirable property that responds to the intuitive notion that information accumulates additively over a sequence of observations. Because of the additivity property, DeWeese and Meister [49] prefer I 2 against I 1 .
Please note that I 2 ( x i t ; X t + 1 ) can take negative values. In this case, this means that a certain region x i is connected with more uncertainty than the mean connectivity of the whole brain. Regions with high values of I 2 (like node 3 in the graph of Figure 2b) greatly reduce the uncertainty in X t + 1 and, thus, they are very significant in the relationship between two consecutive steps in the random walk, X t and X t + 1 . Regions with low values of I 2 (like node 2 in the graph of Figure 2b) are assumed to be broadly connected with other brain regions. From this interpretation, we can say that I 2 expresses the capacity of prediction for a given brain region.

2.4.4. Erasure Surprise

In this section, we propose a novel measure based on the decomposition of the erasure mutual information Equation (15) measure. Remember that erasure mutual information represents the reduction of uncertainty when the context (i.e., both past and future) is known.
Then, we can decompose the erasure mutual information measure as:
I ( X ) = H ( X t ) H ( X t | X t 1 , X t + 1 ) = x i t X p ( x i t ) I 1 ( x i t ; X ) ,
where
I 1 ( x i t ; X ) = x j t 1 X x k t + 1 X p ( x j t 1 , x k t + 1 | x i t ) log p ( x j t 1 , x k t + 1 | x i t ) p ( x j t 1 , x k t + 1 )
= j = 1 n μ j P j i k = 1 n P i k μ i log P j i P i k μ i Q j k
and Q j k = i = 1 n P j i P i k . I 1 is the erasure surprise associated to the region x i and it always takes positive values. Observe that I 1 can be seen as the Kullback-Leibler distance see Equation (4) between p ( X t 1 , X t + 1 ) (i.e., joint probability of being at t 1 on node x j and at t + 1 on node k) and p ( X t 1 , X t + 1 | x i t ) (i.e., the same as the latter but conditioned to the fact that at t the random walk is on node x i . Thus, those nodes that connect brain regions which are already connected will lead to low values of I 1 , and are likely to belong to the same cluster. This can be seen, for instance, in node 2 of the graph shown in Figure 2b. Instead, nodes that connect nodes which would not be connected otherwise (unique paths), will have high values (node 3 of graph shown in Figure 2b).

3. Material

3.1. Synthetic Network Models

The human connectome has been defined as a network with an average short path length which gives a high efficiency in transferring information, a high clustering which provides robustness to random errors, a degree distribution similar to networks with hubs, and a modular community structure [18]. According to these properties, random, lattice, and small-world networks are models that can represent the human connectome. If efficiency was the only property used in the network design, the network would be random [18], with low clustering, short path length [50], and all connections equally probable. However, it is clear that the cortex is not just a uniform system of random connected neurons since random graphs cannot encode and process information [19]. If wiring cost was the priority, the network would be similar to a lattice graph with long paths and high clustering. If we aim for a balance between high clustering and average short path length, then small-world networks are the more accurate representation for both structural and functional networks. For this reason, to illustrate features of the proposed measures, we created three datasets containing random, lattice, ring lattice, and small-world networks.
The first dataset contained random, lattice, ring lattice, and small-world networks with 128 nodes and different number of edges ranging from 128 to 8192 with a step of 128 edges. The second dataset contained the same network models with 256 nodes and edges ranging from 256 to 8192 with a step of 128 edges. Please note that these two datasets provide equivalent networks but with different densities, since the number of nodes was fixed and the number of edges varied. Additionally, a third dataset was created with nodes ranging from 32 to 512 with a step of 32 and a fixed density of 0.4 (varied number of edges). For all graphs, a random weight ranging from 0 to 1 was assigned to all the edges.
The network models were created using the Brain Connectivity Toolbox (BCT) [12]. This toolbox contains a large selection of reference network models and measures that have been previously used in several studies [51,52,53]. To create the undirected random networks, we used the function m a k e r a n d C I J _ u n d which generates graphs with no connections on the diagonal (see Figure 3a). The directed non-ring lattice networks were created with the function m a k e l a t t i c e C I J . This lattice is made by placing connections as close as possible to the main diagonal, without wrapping around, and with no connections on the diagonal (see Figure 3b). The ring directed lattice networks were created with the function m a k e r i n g l a t t i c e C I J . In this case, the lattice is also made by placing connections close to the diagonal, but wrapping around (see Figure 3c). Finally, directed small-world networks were created with the function m a k e e v e n C I J . These networks have a specific number of fully connected nodes linked together but with a balanced random connections (see Figure 3d). To transform directed graphs to undirected graphs, all values above diagonal were copied below the diagonal, therefore, all synthetic networks used in this work are weighted and undirected.

3.2. Human Datasets

Human datasets were used to test the proposed measures with real data. To show the applicability of our method, we considered both functional and structural brain networks.

3.2.1. Anatomic Dataset

To study the human structural network, we used normalized connection matrices created from MRI tractography [54]. The connectivity matrices were from 10 different subjects at 5 different scales, corresponding to 83, 129, 234, 463 and 1015 cortical and subcortical ROIs. Subjects were all males aged 22 ± 1.3 years old. Edge weights were given by the connectivity density which corresponds to the number of fibers divided by the average of the region surface and by the average length of the fibers. All values were positive, and values on the diagonal were eliminated. The average matrices of the 10 patients for each scale were also created. Figure 4 shows the averaged matrices for the 5 different scales. Edges were resorted to place more edges closer to the diagonal for visualization purposes only.

3.2.2. Functional Dataset

Independent component analysis (ICA) is a widely used method to generate functional brain networks of the brain during rest and task. For our analysis, we used the HCP500-PTN functional dataset which belongs to the Human Connectome Project (HCP) beta-release of group-ICA maps [55,56,57]. This dataset contains functional network matrices of 461 subjects at 5 different scales (25, 50, 100, 200 and 300). For our experiment, we used the approach where the principal eigen-timeseries are estimated and a full normalized temporal correlation has been used. The original matrices contain positive and negative values and no values on the diagonal, but for our experiments, the matrices were thresholded ( Z > 5 ) and the negative values were eliminated. The averaged networks were also used. Figure 5 shows the averaged functional matrices at different scales.

3.3. Standard Network Measures

The BCT toolbox [12] provides different complex network measures to describe either structural or functional brain connectivity. To evaluate the proposed approach we compared our measures with standard measures included in the BCT. The clustering coefficient is a measure of segregation and expresses the fraction of triangles around a node. The node eccentricity is a measure of distance defined as the maximal shortest path length between a node and any other node. Finally, the node strength is a measure of similarity defined as the sum of weights of links connected to the node.

4. Results and Discussion

In this section, we apply the proposed measures to the synthetic network models and to the human structural and functional connectomes. The results with the global measures and local measures are shown and a comparison with standard measures is presented.

4.1. Global Measures

Firstly, to show the behavior of the global measures, we characterize the synthetic network models (random, lattice, ring lattice and small-world) from the first and second dataset defined in Section 3.1, with 128 and 256 nodes. We apply the global measures (entropy, mutual information, and erasure mutual information) which give a single value per graph.
The first column of Figure 6 presents the entropy measure results. Observe that, when the number of edges increases, the entropy measure tends to a constant value for all types of graphs. This is due to the fact that the higher the number of edges, the more similar the node probability. Thus, the entropy tends to log 2 N , where N is the number of nodes (i.e., for 128 nodes the entropy tends to 7 and for 256 nodes to 8). The slightly decreasing tendency of high values in lattice networks is due to the boundary conditions of extreme nodes which have a lower number of connections which leads to an entropy drop.
The second column of Figure 6 shows the behavior of the mutual information measure. In this case, when the number of edges increases, the mutual information of the graph decreases for all types of networks. This is due to the fact that the higher the number of connections, the lower the correlation between consecutive states. For a very low number of edges, we can see that first, the mutual information increases and then decreases. This is due to the fact that for low densities, there are nodes not connected with any node, leading to a decrease of the overall mutual information. Since different tracking methods may provide different number of fibers for a given parcellation [58], the optimal point found with the mutual information measure may allow to find the minimum number of fibers needed for a given brain parcellation to study ring lattice and lattice properties. For a low number of edges, we can also observe that lattice and ring lattice graphs have a slightly higher mutual information than random and small-world graphs. This is due to the higher degree of structure of these kind of graphs, which is what the mutual information measure quantifies.
The third column of Figure 6 presents the values of the erasure mutual information measure. In this case, when the number of edges increases, for all networks, the measure tends to decrease. Note that when there are only a few edges, the uncertainty when past and future states are known is very low ( H ( X t | X t 1 , X t + 1 ) , the second term of Equation (15)), leading to high I ( X ) values. When the number of edges increases, the uncertainty tends to increase, thus, the I ( X ) tends to decrease. For this measure, different behaviors can be observed depending on the graph type. For instance, the lattice and ring lattice graphs have a lower erasure mutual information compared to random and small-world graphs. This is due to the fact that the erasure mutual information takes into account the previous node and next node, and for lattice networks, nodes tend to be connected with the closest ones, thus, globally there is more uncertainty. An interesting behavior can be observed for the random and small-world graphs where the measure reaches an optimal point with a larger number of connections compared to lattice and ring lattice networks. In this case, for a low number of edges, there are nodes which are not connected or only connected with intra-module nodes. Thus, all the paths are within the same module. When the number of edges slightly increases, there are more paths that connect different modules but the probability of these paths is very low. Therefore, the erasure mutual information slightly increases. After the optimal point, the erasure mutual information decreases due to the larger number of connections between different modules that increase the uncertainty.
Secondly, we generated different graphs, in this case, modifying the number of nodes but preserving the density (number of edges divided by the number of edges of the complete graph) to 0.4 , which is the third dataset described in Section 3.1. Figure 7 shows the behavior of the global measures when the number of nodes increases. As it can be seen, the entropy value increases with the logarithm of the size for a constant edge density. This is consistent with the results of the first experiment where entropy tends to log 2 N , being N the number of nodes. On the contrary, mutual information is not very sensitive for random networks since its connections are randomly placed, so fixing the graph density, the structure of the networks remains similar. A comparable behavior can be observed for the small-world networks. In this case, graphs with a low number of nodes, have a higher mutual information due to more intra-module connections, and, as a consequence, if we increase the number of nodes, the number of edges also increases. On the other side, we can observe that while ring lattice network have a high value, lattice network have a very small value. This is because two nodes of the lattice network are not connected and, the rest of the nodes, have a higher degree compared to ring lattice. Consequently, there are less unique paths. Finally, erasure mutual information is not very sensitive to the graph size but to the graph topology. Random and small world have higher values compared to ring lattice and lattice. This is due to the existence of a large amount of connecting paths for neighbor nodes in ring and lattice networks, so paths are not unique. Since for lattice graphs two nodes are not connected, the rest of the nodes have a slightly higher degree, and, as a result the overall predictability is lower. If we increase the number of nodes we have to increase also the number of edges, thus, as a result, the degree of the nodes increases. Because of this, if we focus on the values for a low number of nodes, we can observe that the erasure mutual information for lattice and ring lattice slightly decreases, and, for random and small-world, increases. The erasure mutual information measure takes into account the next node but also the previous one. Therefore, increasing the degree in the ring lattice and the lattice networks, the overall uncertainty increases. On the contrary, for random and small-world networks with a low node degree, paths are more unique for a low number of nodes. Increasing the number of nodes while keeping the density the same, the erasure mutual information tends to stabilize.
To evaluate the global measures with anatomical data, we applied the global measures to the anatomic and functional datasets at different scales described in Section 3.2. Figure 8 shows the result of entropy, mutual information, and erasure mutual information for the 10 structural networks with 83, 129, 234 and 1015 partitions. Observe that all measures have a similar behavior for all the patients which demonstrates that the measures are consistent among all patients. Figure 9 shows the result of the global measures applied to 468 functional networks with 25, 50, 100, 200 and 300 nodes. In this case, the entropy measure has the same behavior as the structural network. Moreover, since the density is similar between different partitions, the mutual information and the erasure measures have a more uniform value. The same effect has been shown in the behavior of the mutual information and the erasure for model graphs with a constant density.

4.2. Local Measures

In this section, we compare local measures (entropic surprise, mutual surprise, mutual predictability and erasure surprise) with standard measures provided in the BCT. In addition, we show the result of the measures applied to the structural and functional human connectivity networks.
First of all, we provide a comparison of the proposed local measures with strength, eccentricity and clustering measures. Strength measures the sum of the weights for each node, eccentricity the maximal shortest path length between a node and any other node, and clustering the fraction of triangles in the node. To carry out this experiment, we have considered the averaged connectivity matrix created from the 10 structural networks with 1015 nodes of the anatomic dataset described in Section 3.2.1. The purpose of this experiment is to demonstrate the feasibility and application of the method in a real clinical scenario. Results are shown in Figure 10. From left to right, columns represent entropic surprise (E), mutual surprise ( I 1 ), mutual predictability ( I 2 ), and erasure surprise ( I 1 ), and, from top to down, rows represent the value of our measure with respect to strength, eccentricity, and clustering, respectively. For each scatter plot, x-axis represents the standard measure value and the y-axis the value of our measure. In each plot, the logarithmic curve ( f ( x ) = a + b log 2 ( x ) ) that best fits to the data and the determination coefficient, R 2 , of the data model are also shown. We can see that the surprise measure is directly related to the strength since both depend on the weight of the node and the surprise measure is mathematically defined as minus the logarithm of the strength see Equation (16). As it can be seen, the other measures are moderately correlated (mainly I 1 and I 1 ) to the strength. This is not directly related to their mathematical definition, but by the fact that those nodes with more connections (high strength) tend to have more uncertainty on their connections and, thus, lower measure values. Comparing with the eccentricity measure, we can observe that nodes with a high maximal shortest path length (high eccentricity) tend to not be highly connected (low E value). On the other side, nodes with a low eccentricity are highly connected. The other measures do not demonstrate significant correlation with eccentricity. With respect to the clustering measure, I 2 is the only measure that slightly correlates with it. This can be explained by the fact that those nodes with a high clustering coefficient will tend to have less uncertainty on their connections.
Finally, we show the value of each measure for each node of the human structural and functional averaged networks, with 83 and 25 partitions, respectively. Figure 11a shows all the nodes for the structural network in yellow and the connections between nodes in black.
The value of the entropic surprise E for each node of the human structural averaged network, with 83 partitions is shown on the left histogram of Figure 12. This measure is equivalent to the classic strength measure, where nodes with high values are nodes not highly connected or with low weights, which lead to a low stationary probability. The maximum and minimum E values corresponds to the right hemisphere transverse temporal and the brain stem, respectively. These nodes have been represented on the brain network in green and orange (see Figure 12 top image of the central column). The value of the mutual surprise I 1 for each node is shown on the right histogram of Figure 12. High values correspond to nodes connected to poorly connected nodes (nodes with a low number of connections), while low values correspond to nodes connected to highly connected nodes. This fact is illustrated on the bottom image of the central column of Figure 12 where the right hemisphere transverse temporal, represented as a green node, has the maximum value and the left hemisphere thalamus proper, represented as an orange node, has the minimum one. Comparing entropic surprise and mutual surprise for the structural connectome, we observe an organization, where nodes highly connected are also nodes connected to similar nodes in terms of probability, and nodes not highly connected are connected to nodes with a very different probability compared to them.
The value of the mutual predictability I 2 for each node of the human structural averaged network, with 83 partitions is shown on left histogram of Figure 13. Remember that, for nodes with a high mutual predictability, the distribution of connections with other nodes have a low entropy. For instance, observe the first image of the central column in Figure 13, the green node, which corresponds to the right hemisphere temporal pole, has the highest I 2 value. On the other hand, nodes with low values have more uncertainty in predicting the next node. In this case, the lowest I 2 value corresponds to the right hemisphere putamen, represented as an orange node. The value of the erasure surprise I 1 is shown on the right histogram of Figure 13. Nodes with high values are nodes that connect different areas otherwise not connected or less connected, like a bridge or a hub. For example, the right hemisphere transverse temporal, shown in Figure 11b together with its neighbor connections, is the region with a higher value in the bottom image of the central column in Figure 13. On the other side, nodes with low values, are nodes that belong to a cluster since there are multiple paths connecting its neighbors. In this case, the lower value of the histogram corresponds to the left hemisphere thalamus proper, which is shown in Figure 11c together with its neighbor connections.
Figure 14 shows the results for the entropic surprise, the mutual surprise, the mutual predictability, and the erasure surprise applied to the human functional network with 25 partitions. An illustrative image of each partition is shown in Figure 15. Analyzing independently the measures, we found a behavior similar to the structural networks. However, evaluating all the measures and comparing them, we can observe interesting properties. For instance, regions 14 and 19 have both a high erasure surprise value, while mutual surprise is high for region 14 and low for region 19. Thus, these two regions belong to a unique path (due to a high erasure surprise value) but region 14 connects regions highly connected while region 19 connects regions poorly connected (due to the mutual information value). On the other side, region 1, which has a high sum of weights, is also connected to regions similar to itself, so regions with also a big amount of connections. Region 19 has also a low mutual predictability, which means that there is a high capacity to predict the regions which is connected to, on the contrary, region 1 has a lower mutual predictability, so even if it is highly connected to similar nodes it is difficult to predict which are the nodes. Finally, region 1 has a low erasure surprise, which indicates that is likely to belong to a cluster, and region 14 has a high erasure surprise, so it acts more as a bridge of areas which are not strongly connected with other areas.
These results show a proof of principle of the proposed brain model and the suggested set of measures, that provide robust results using structural or functional data. Prior to a further investigation with more clinical data, the proposed approach provides new insights into the brain complexity which may be of interest in studying the functioning of the brain and the connections between regions.

5. Conclusions

In this paper, we have used a Markov process-based brain model in which we apply existent and novel information theory-based measures to characterize new properties of complex brain networks. The main contribution of the paper is the proposal of new local and global measures to describe new properties of brain networks in terms of topology and organization, with the main novelty being the definition of erasure mutual information and the erasure surprise. The proposed measures have been tested on synthetic model networks, increasing the number of nodes and the number of edges, and with structural and functional human networks at different scales.
From our experiments and focusing on global measures, we showed that, given a network, the entropy, describes the overall uncertainty of the nodes connectivity. In addition, mutual information, which is a measure of structure, is able to differentiate the topology of network models. Finally, the erasure mutual information, which is a new measure defined by extending the mutual information, describes how unique the paths for a given network are. With this measure, we show an optimal point for small-world networks.
Focusing on local measures, we observed that the entropic surprise, which describes how connected is a node taking into account all the connections in the network, is inversely proportional to the logarithm of the standard strength measure. The mutual surprise, which defines the connectivity of the neighbor nodes for a given node, allows to identify nodes whose nodes have a high connectivity taking into account all connections. The mutual predictability, which given a node, determines the uncertainty associated to a node in predicting the next node, shows that regions with a high clustering tend to be more predictable. Finally, the erasure surprise, which takes into account previous and next nodes, defines how unique the path is which the node belongs to. Results show that regions with a high strength belong to a module where all nodes are strongly connected. The consistency of the results for structural and functional human networks demonstrates the robustness of the proposed measures.
In future work, we will analyze in detail the properties of specific anatomical areas of the human brain and we will study how it can help to detect different diseases. Furthermore, we will investigate clinically informative visualizations using the presented measures.

Author Contributions

E.B., A.B., M.F. and I.B. conceived and designed the experiments; E.B. performed the experiments; E.B. and A.B. analyzed the data; E.B., A.B. and I.B. wrote the paper.

Funding

This work was supported by the Spanish Government (Grant No. TIN2016-75866-C3-3-R) and by the Catalan Government (Grant No. 2017-SGR-1101). Data were provided, in part, by the Human Connectome Project, WU-Minn Consortium (Principal Investigators: David Van Essen and Kamil Ugurbil; 1U54MH091657) funded by the 16 NIH Institutes and Centers that support the NIH Blueprint for Neuroscience Research; and by the McDonnell Center for Systems Neuroscience at Washington University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sporns, O.; Tononi, G.; Kötter, R. The human connectome: A structural description of the human brain. PLoS Comput. Biol. 2005, 1, e42. [Google Scholar] [CrossRef] [PubMed]
  2. Felleman, D.J.; Essen, D.C.V. Distributed hierarchical processing in the primate cerebral cortex. Cereb. Cortex 1991, 1, 1–47. [Google Scholar] [CrossRef] [PubMed]
  3. Hagmann, P. From Diffusion MRI to Brain Connectomics. Ph.D. Thesis, École polytechnique fédérale de Lausanne (EPFL), Lausanne, Switzerland, 2005. [Google Scholar]
  4. Hagmann, P.; Kurant, M.; Gigandet, X.; Thiran, P.; Wedeen, V.J.; Meuli, R.; Thiran, J.P. Mapping human whole-brain structural networks with diffusion MRI. PLoS ONE 2007, 2, e597. [Google Scholar] [CrossRef] [PubMed]
  5. Hagmann, P.; Cammoun, L.; Gigandet, X.; Gerhard, S.; Grant, P.E.; Wedeen, V.; Meuli, R.; Thiran, J.P.; Honey, C.J.; Sporns, O. MR connectomics: Principles and challenges. J. Neurosci. Methods 2010, 194, 34–45. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Sporns, O. The human connectome: Origins and challenges. NeuroImage 2013, 80, 53–61. [Google Scholar] [CrossRef] [PubMed]
  7. Bullmore, E.T.; Bassett, D.S. Brain graphs: Graphical models of the human brain connectome. Ann. Rev. Clin. Psychol. 2011, 7, 113–140. [Google Scholar] [CrossRef] [PubMed]
  8. Sporns, O. The human connectome: A complex network. Ann. New York Acad. Sci. 2011, 1224, 109–125. [Google Scholar] [CrossRef] [PubMed]
  9. Wu, G.R.; Liao, W.; Stramaglia, S.; Ding, J.R.; Chen, H.; Marinazzo, D. A blind deconvolution approach to recover effective connectivity brain networks from resting state fMRI data. Med. Image Anal. 2013, 17, 365–374. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Passingham, R.E.; Stephan, K.E.; Kotter, R. The anatomical basis of functional localization in the cortex. Nat. Rev. Neurosci. 2002, 3, 606–616. [Google Scholar] [CrossRef] [PubMed]
  11. Kaiser, M. A tutorial in connectome analysis: Topological and spatial features of brain networks. NeuroImage 2011, 57, 892–907. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Rubinov, M.; Sporns, O. Complex network measures of brain connectivity: Uses and interpretations. NeuroImage 2010, 52, 1059–1069. [Google Scholar] [CrossRef] [PubMed]
  13. Stam, C.; Reijneveld, J. Graph theoretical analysis of complex networks in the brain. Nonlinear Biomed. Phys. 2007, 1, 3. [Google Scholar] [CrossRef] [PubMed]
  14. Watts, D.J.; Strogatz, S.H. Collective dynamics of ’small-world’ networks. Nature 1998, 393, 440–442. [Google Scholar] [CrossRef] [PubMed]
  15. Latora, V.; Marchiori, M. Efficient behavior of small-world networks. Phys. Rev. Lett. 2001, 87, 198701. [Google Scholar] [CrossRef] [PubMed]
  16. Newman, M. The structure and function of complex networks. SIAM 2003, 45, 167–256. [Google Scholar] [CrossRef]
  17. Newman, M. Fast algorithm for detecting community structure in networks. Phys. Rev. E 2003, 69, 066133. [Google Scholar] [CrossRef] [PubMed]
  18. Bullmore, E.; Sporns, O. Complex brain networks: Graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 2009, 10, 186–198. [Google Scholar] [CrossRef] [PubMed]
  19. Kennedy, H.; Knoblauch, K.; Toroczkai, Z. Why data coherence and quality is critical for understanding interareal cortical networks. NeuroImage 2013, 80, 37–45. [Google Scholar] [CrossRef] [PubMed]
  20. Colizza, V.; Flammini, A.; Serrano, M.A.; Vespignani, A. Detecting rich-club ordering in complex networks. Nat. Phys. 2006, 2, 110–115. [Google Scholar] [CrossRef] [Green Version]
  21. Harriger, L.; van den Heuvel, M.P.; Sporns, O. Rich club organization of macaque cerebral cortex and its role in network communication. PLoS ONE 2012, 7, e46497. [Google Scholar] [CrossRef] [PubMed]
  22. Van Den Heuvel, M.; Kahn, R.; Goñi, J.; Sporns, O. High-cost, high-capacity backbone for global brain communication. Proc. Natl. Acad. Sci. USA 2012, 109, 11372–11377. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Sporns, O.; Chialvo, D.R.; Kaiser, M.; Hilgetag, C.C. Organization, development and function of complex brain networks. Trends Cognit. Sci. 2004, 8, 418–425. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. Tononi, G.; Sporns, O.; Edelman, G.M. A measure for brain complexity: Relating functional segregation and integration in the nervous system. Proc. Natl. Acad. Sci. USA 1994, 91, 5033–5037. [Google Scholar] [CrossRef] [PubMed]
  25. Marrelec, G.; Bellec, P.; Krainik, A.; Duffau, H.; Pélégrini-Issac, M.; Lehéricy, S.; Benali, H.; Doyon, J. Regions, systems, and the brain: Hierarchical measures of functional integration in fMRI. Med. Image Anal. 2008, 12, 484–496. [Google Scholar] [CrossRef] [PubMed]
  26. Kitazono, J.; Kanai, R.; Oizumi, M. Efficient algorithms for searching the minimum information partition in integrated information theory. Entropy 2018, 20, 173. [Google Scholar] [CrossRef]
  27. Tononi, G.; Edelman, G.M.; Sporns, O. Complexity and coherency: Integrating information in the brain. Trends Cognit. Sci. 1998, 2, 474–484. [Google Scholar] [CrossRef]
  28. Sporns, O.; Tononi, G.; Edelman, G.M. Connectivity and complexity: The relationship between neuroanatomy and brain dynamics. Neural Netw. 2000, 13, 909–922. [Google Scholar] [CrossRef]
  29. Tononi, G.; Sporns, O.; Edelman, G.M. A complexity measure for selective matching of signals by the brain. Proc. Natl. Acad. Sci. USA 1996, 93, 3422–3427. [Google Scholar] [CrossRef] [PubMed]
  30. Tononi, G.; McIntosh, A.R.; Russell, D.P.; Edelman, G.M. Functional clustering: Identifying strongly interactive brain regions in neuroimaging data. NeuroImage 1998, 7, 133–149. [Google Scholar] [CrossRef] [PubMed]
  31. Edelman, G.M.; Gally, J.A. Degeneracy and complexity in biological systems. Proc. Natl. Acad. Sci. USA 2001, 98, 13763–13768. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Tononi, G.; Sporns, O.; Edelman, G.M. Measures of degeneracy and redundancy in biological networks. Proc. Natl. Acad. Sci. USA 1999, 96, 3257–3262. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Crossley, N.; Mechelli, A.; Scott, J.; Carletti, F.; Fox, P.; McGuire, P.; Bullmore, E. The hubs of the human connectome are generally implicated in the anatomy of brain disorders. Brain 2014, 137, 2382–2395. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Meskaldji, D.E.; Fischi-Gomez, E.; Griffa, A.; Hagmann, P.; Morgenthaler, S.; Thiran, J.P. Comparing connectomes across subjects and populations at different scales. NeuroImage 2013, 80, 416–425. [Google Scholar] [CrossRef] [PubMed]
  35. van den Heuvel, M.P.; Pol, H.E.H. Exploring the brain network: A review on resting-state fMRI functional connectivity. Eur. Neuropsychopharmacol. 2010, 20, 519–534. [Google Scholar] [CrossRef] [PubMed]
  36. Sato, J.R.; Takahashi, D.Y.; Hoexter, M.Q.; Massirer, K.B.; Fujita, A. Measuring network’s entropy in ADHD: A new approach to investigate neuropsychiatric disorders. NeuroImage 2013, 77, 44–51. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  37. Papo, D.; Buldú, J.M.; Boccaletti, S.; Bullmore, E.T. Complex network theory and the brain. Phil. Trans. R. Soc. B 2014, 369, 20130520. [Google Scholar] [CrossRef] [PubMed]
  38. Bonmati, E.; Bardera, A.; Boada, I. Brain parcellation based on information theory. Comput. Methods Programs Biomed. 2017, 151, 203–212. [Google Scholar] [CrossRef] [PubMed]
  39. Cover, T.M.; Thomas, J.A. Elements of Information Theory; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  40. Yeung, R.W. A First Course in Information Theory; Springer Science & Business Media: New York, NY, USA, 2002. [Google Scholar]
  41. Feldman, D.P.; Crutchfield, J.P. Discovering Noncritical Organization: Statistical Mechanical, Information Theoreticand Computational Views of Patterns in One-Dimensional Spin Systems; Working Paper 98-04-026; Santa Fe Institute: Santa Fe, NM, USA, 1998. [Google Scholar]
  42. Crutchfield, J.P.; Packard, N. Symbolic dynamics of noisy chaos. Physica D 1983, 7, 201–223. [Google Scholar] [CrossRef]
  43. Grassberger, P. Toward a quantitative theory of self-generated complexity. Int. J. Theor. Phys. 1986, 25, 907–938. [Google Scholar] [CrossRef]
  44. Shaw, R. The Dripping Faucet as a Model Chaotic System; Aerial Press: Santa Cruz, CA, USA, 1984. [Google Scholar]
  45. Szépfalusy, P.; Györgyi, G. Entropy decay as a measure of stochasticity in chaotic systems. Phys. Rev. A 1986, 33, 2852. [Google Scholar] [CrossRef]
  46. Feldman, D.P. A Brief Introduction to: Information Theory, Excess Entropy and Computational Mechanics; Lecture notes; Department of Physics, University of California: Berkeley, CA, USA, 1997. [Google Scholar]
  47. Verdú, S.; Weissman, T. The information lost in erasures. IEEE Trans. Inf. Theory 2008, 54, 5030–5058. [Google Scholar] [CrossRef]
  48. Feldman, D.; Crutchfield, J. Structural information in two-dimensional patterns: Entropy convergence and excess entropy. Phys. Rev. E 2003, 67, 051104. [Google Scholar] [CrossRef] [PubMed]
  49. DeWeese, M.R.; Meister, M. How to measure the information gained from one symbol. Network Comput. Neural Syst. 1999, 10, 325–340. [Google Scholar] [CrossRef]
  50. Fornito, A.; Zalesky, A.; Breakspear, M. Graph analysis of the human connectome: Promise, progress, and pitfalls. NeuroImage 2013, 80, 426–444. [Google Scholar] [CrossRef] [PubMed]
  51. Dennis, E.L.; Jahanshad, N.; Toga, A.W.; McMahon, K.; de Zubicaray, G.I.; Martin, N.G.; Wright, M.J.; Thompson, P.M. Test-retest reliability of graph theory measures of structural brain connectivity. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Nice, France, 1–5 October 2012; Ayache, N., Delingette, H., Golland, P., Mori, K., Eds.; Springer: Berlin, Germany, 2012; Volume 7512, pp. 305–312. [Google Scholar]
  52. Messé, A.; Rudrauf, D.; Giron, A.; Marrelec, G. Predicting functional connectivity from structural connectivity via computational models using MRI: An extensive comparison study. NeuroImage 2015, 111, 65–75. [Google Scholar] [CrossRef] [PubMed]
  53. Santos Ribeiro, A.; Miguel Lacerda, L.; Ferreira, H.A. Multimodal imaging brain connectivity analysis toolbox (MIBCA). PeerJ PrePrints 2014, 2, e699v1. [Google Scholar]
  54. Cammoun, L.; Gigandet, X.; Sporns, O.; Thiran, J.; Do, K.; Maeder, P.; Meuli, R.; Hagmann, P.; Bovet, P.; Do, K. Mapping the human connectome at multiple scales with diffusion spectrum MRI. J. Neurosci. Methods 2012, 203, 386–397. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  55. Essen, D.V.; Ugurbil, K.; Auerbach, E.; Barch, D.; Behrens, T.; Bucholz, R.; Chang, A.; Chen, L.; Corbetta, M.; Curtiss, S.; et al. The Human Connectome Project: A data acquisition perspective. NeuroImage 2012, 62, 2222–2231. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Glasser, M.F.; Sotiropoulos, S.N.; Wilson, J.A.; Coalson, T.S.; Fischl, B.; Andersson, J.L.; Xu, J.; Jbabdi, S.; Webster, M.; Polimeni, J.R.; et al. The minimal preprocessing pipelines for the Human Connectome Project. NeuroImage 2013, 80, 105–124. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  57. Hodge, M.R.; Horton, W.; Brown, T.; Herrick, R.; Olsen, T.; Hileman, M.E.; McKay, M.; Archie, K.A.; Cler, E.; Harms, M.P.; et al. ConnectomeDB—Sharing human brain connectivity data. NeuroImage 2016, 124, 1102–1107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  58. Christidi, F.; Karavasilis, E.; Samiotis, K.; Bisdas, S.; Papanikolaou, N. Fiber tracking: A qualitative and quantitative comparison between four different software tools on the reconstruction of major white matter tracts. Eur. J. Radiol. Open 2016, 3, 153–161. [Google Scholar] [CrossRef] [PubMed]
  59. Dai, D.; He, H. VisualConnectome: Toolbox for brain network visualization and analysis. In Proceedings of the Organization on human Brain Mapping, 2011, Québec City, QC, Canada, 26–30 June 2011. [Google Scholar]
Figure 1. Example values of the entropy (H), mutual information ( M I ) and erasure ( I ) measures for simple networks (ac), where N corresponds to the number of nodes and E to the number of edges. Networks are weighted and undirected, therefore each edge is counted twice.
Figure 1. Example values of the entropy (H), mutual information ( M I ) and erasure ( I ) measures for simple networks (ac), where N corresponds to the number of nodes and E to the number of edges. Networks are weighted and undirected, therefore each edge is counted twice.
Entropy 20 00491 g001
Figure 2. Example values of entropic surprise (E), mutual surprise ( I 1 ), mutual predictability ( I 2 ) and erasure surprise ( I 1 ) measures for simple networks (ac). Networks are weighted and undirected, therefore each edge is counted twice.
Figure 2. Example values of entropic surprise (E), mutual surprise ( I 1 ), mutual predictability ( I 2 ) and erasure surprise ( I 1 ) measures for simple networks (ac). Networks are weighted and undirected, therefore each edge is counted twice.
Entropy 20 00491 g002
Figure 3. Example of network models (synthetic dataset) used in this work. Each model has the corresponding connectivity matrix illustrated at the bottom. (a) Non directed random network (16 nodes, 120 edges); (b) Non directed lattice network (16 nodes, 118 edges); (c) Non directed ring lattice network (16 nodes, 122 edges); (d) Non directed small-world network (16 nodes, 116 edges and cluster size 2).
Figure 3. Example of network models (synthetic dataset) used in this work. Each model has the corresponding connectivity matrix illustrated at the bottom. (a) Non directed random network (16 nodes, 120 edges); (b) Non directed lattice network (16 nodes, 118 edges); (c) Non directed ring lattice network (16 nodes, 122 edges); (d) Non directed small-world network (16 nodes, 116 edges and cluster size 2).
Entropy 20 00491 g003
Figure 4. Illustration of the averaged structural connectivity matrices of the anatomic dataset with the corresponding number of nodes (N), edges (E) and density. 0 values are represented in white. Edges were resorted to place more edges closer to the diagonal for visualization purposes only.
Figure 4. Illustration of the averaged structural connectivity matrices of the anatomic dataset with the corresponding number of nodes (N), edges (E) and density. 0 values are represented in white. Edges were resorted to place more edges closer to the diagonal for visualization purposes only.
Entropy 20 00491 g004
Figure 5. Illustration of the averaged functional connectivity matrices of the functional dataset with the corresponding number of nodes (N), edges (E) and density.
Figure 5. Illustration of the averaged functional connectivity matrices of the functional dataset with the corresponding number of nodes (N), edges (E) and density.
Entropy 20 00491 g005
Figure 6. Behavior of the entropy, mutual information and erasure mutual information measures for each network model when the number of edges is increased (from 0 to 8192) and the number of nodes (N) is kept constant (128 nodes on top row and 256 nodes on bottom row).
Figure 6. Behavior of the entropy, mutual information and erasure mutual information measures for each network model when the number of edges is increased (from 0 to 8192) and the number of nodes (N) is kept constant (128 nodes on top row and 256 nodes on bottom row).
Entropy 20 00491 g006
Figure 7. Behavior of the (a) entropy; (b) mutual information; (c) and erasure mutual information measures when the number of nodes is increased (from 0 to 500) while the density is kept constant to 0.4 .
Figure 7. Behavior of the (a) entropy; (b) mutual information; (c) and erasure mutual information measures when the number of nodes is increased (from 0 to 500) while the density is kept constant to 0.4 .
Entropy 20 00491 g007
Figure 8. Box-plots showing median, 25th and 75th percentiles for global measures ((a) entropy; (b) mutual information; (c) and erasure mutual information) when applied to the 10 structural connectomes with 83, 129, 254, 463 and 1015 partitions.
Figure 8. Box-plots showing median, 25th and 75th percentiles for global measures ((a) entropy; (b) mutual information; (c) and erasure mutual information) when applied to the 10 structural connectomes with 83, 129, 254, 463 and 1015 partitions.
Entropy 20 00491 g008
Figure 9. Box-plots showing median, 25th and 75th percentiles for global measures ((a) entropy; (b) mutual information; (c) and erasure mutual information) when applied to 463 functional connectomes with 25, 50, 100, 200 and 300 partitions.
Figure 9. Box-plots showing median, 25th and 75th percentiles for global measures ((a) entropy; (b) mutual information; (c) and erasure mutual information) when applied to 463 functional connectomes with 25, 50, 100, 200 and 300 partitions.
Entropy 20 00491 g009
Figure 10. Relationship between the proposed local measures (entropic surprise (E), mutual surprise ( I 1 ), mutual predictability ( I 2 ) and erasure surprise ( I 1 ) and standard measures (strength, eccentricity and clustering) using the structural averaged connectivity matrix network with 1015 nodes.
Figure 10. Relationship between the proposed local measures (entropic surprise (E), mutual surprise ( I 1 ), mutual predictability ( I 2 ) and erasure surprise ( I 1 ) and standard measures (strength, eccentricity and clustering) using the structural averaged connectivity matrix network with 1015 nodes.
Entropy 20 00491 g010
Figure 11. (a) Illustration of all the connections in the structural dataset; (b) Right hemisphere transverse temporal region (green) connections including its neighbors connections; (c) Left hemisphere thalamus proper (orange) connections including its neighbors connections. This figure has been generated using the VisualConnectome software [59].
Figure 11. (a) Illustration of all the connections in the structural dataset; (b) Right hemisphere transverse temporal region (green) connections including its neighbors connections; (c) Left hemisphere thalamus proper (orange) connections including its neighbors connections. This figure has been generated using the VisualConnectome software [59].
Entropy 20 00491 g011
Figure 12. On the left, entropic surprise values obtained with the averaged structural network with 83 partitions. The maximum and minimum values have been represented on the brain network (first image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the brain stem. On the right, mutual surprise values obtained with the same network. The maximum and minimum values have been represented on the brain network (second image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the thalamus proper.
Figure 12. On the left, entropic surprise values obtained with the averaged structural network with 83 partitions. The maximum and minimum values have been represented on the brain network (first image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the brain stem. On the right, mutual surprise values obtained with the same network. The maximum and minimum values have been represented on the brain network (second image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the thalamus proper.
Entropy 20 00491 g012
Figure 13. On the left, mutual predictability values obtained with the averaged structural network with 83 partitions. The maximum and minimum values have been represented on the brain network (first image of the central column). The green node corresponds to the right hemisphere temporal pole area and the orange to the putamen. On the right, erasure surprise values obtained with the same network. The maximum and minimum values have been represented on the brain network (second image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the thalamus proper.
Figure 13. On the left, mutual predictability values obtained with the averaged structural network with 83 partitions. The maximum and minimum values have been represented on the brain network (first image of the central column). The green node corresponds to the right hemisphere temporal pole area and the orange to the putamen. On the right, erasure surprise values obtained with the same network. The maximum and minimum values have been represented on the brain network (second image of the central column). The green node corresponds to the right hemisphere transverse temporal area and the orange to the thalamus proper.
Entropy 20 00491 g013
Figure 14. Local measures values ((a) entropic surprise; (b) mutual surprise (c) mutual predictability and (d) erasure surprise) obtained with the averaged functional dataset with 25 partitions. An illustrative image of each partition is shown in Figure 15.
Figure 14. Local measures values ((a) entropic surprise; (b) mutual surprise (c) mutual predictability and (d) erasure surprise) obtained with the averaged functional dataset with 25 partitions. An illustrative image of each partition is shown in Figure 15.
Entropy 20 00491 g014
Figure 15. Illustrative images of the 25 regions from the averaged functional dataset [55,56,57].
Figure 15. Illustrative images of the 25 regions from the averaged functional dataset [55,56,57].
Entropy 20 00491 g015aEntropy 20 00491 g015b
Table 1. Summary of the proposed set of measures.
Table 1. Summary of the proposed set of measures.
GlobalLocal
StationaryEntropyEntropic surprise
CausalMutual InformationMutual surprise
Mutual predictability
ContextualErasure Mutual InformationErasure surprise

Share and Cite

MDPI and ACS Style

Bonmati, E.; Bardera, A.; Feixas, M.; Boada, I. Novel Brain Complexity Measures Based on Information Theory. Entropy 2018, 20, 491. https://doi.org/10.3390/e20070491

AMA Style

Bonmati E, Bardera A, Feixas M, Boada I. Novel Brain Complexity Measures Based on Information Theory. Entropy. 2018; 20(7):491. https://doi.org/10.3390/e20070491

Chicago/Turabian Style

Bonmati, Ester, Anton Bardera, Miquel Feixas, and Imma Boada. 2018. "Novel Brain Complexity Measures Based on Information Theory" Entropy 20, no. 7: 491. https://doi.org/10.3390/e20070491

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop