Next Article in Journal
Glass Transition, Crystallization of Glass-Forming Melts, and Entropy
Next Article in Special Issue
The Identity of Information: How Deterministic Dependencies Constrain Information Synergy and Redundancy
Previous Article in Journal
The Complex Neutrosophic Soft Expert Relation and Its Multiple Attribute Decision-Making Method
Previous Article in Special Issue
Information Entropy Production of Maximum Entropy Markov Chains from Spike Trains
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Mutual Information and Information Gating in Synfire Chains

1
Department of Mathematics, University of Arizona, Tucson, AZ 85721, USA
2
Center for Bioinformatics, National Laboratory of Protein Engineering and Plant Genetic Engineering, School of Life Sciences, Peking University, Beijing 100871, China
3
Yuanpei School, Peking University, Beijing 100871, China
4
Information Sciences, CCS-3, Los Alamos National Laboratory, Los Alamos, NM 87545, USA
5
Department of Mathematics, University of California, Davis, CA 95616, USA
6
Center for Quantitative Biology, Peking University, Beijing 100871, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2018, 20(2), 102; https://doi.org/10.3390/e20020102
Submission received: 22 December 2017 / Revised: 29 January 2018 / Accepted: 30 January 2018 / Published: 1 February 2018
(This article belongs to the Special Issue Information Theory in Neuroscience)

Abstract

:
Coherent neuronal activity is believed to underlie the transfer and processing of information in the brain. Coherent activity in the form of synchronous firing and oscillations has been measured in many brain regions and has been correlated with enhanced feature processing and other sensory and cognitive functions. In the theoretical context, synfire chains and the transfer of transient activity packets in feedforward networks have been appealed to in order to describe coherent spiking and information transfer. Recently, it has been demonstrated that the classical synfire chain architecture, with the addition of suitably timed gating currents, can support the graded transfer of mean firing rates in feedforward networks (called synfire-gated synfire chains—SGSCs). Here we study information propagation in SGSCs by examining mutual information as a function of layer number in a feedforward network. We explore the effects of gating and noise on information transfer in synfire chains and demonstrate that asymptotically, two main regions exist in parameter space where information may be propagated and its propagation is controlled by pulse-gating: a large region where binary codes may be propagated, and a smaller region near a cusp in parameter space that supports graded propagation across many layers.

1. Introduction

Faithful information transmission between neuronal populations is essential to computation in the brain. Correlated spiking activity has been measured experimentally between many brain areas [1,2,3,4,5,6,7,8]. Experimental and theoretical studies have shown that synchronized volleys of spikes can propagate within cortical networks and are thus capable of transmitting information between neuronal populations on millisecond timescales [9,10,11,12,13,14,15,16,17]. Many such mechanisms have been proposed for feedforward networks [14,15,17,18,19,20,21,22,23]. Commonly, mechanisms use transient synchronization to provide windows in time during which spikes may be transferred more easily from layer to layer [14,15,17,18,19,20,21,22].
For instance, the successful propagation of synchronous activity has been identified in “synfire chains” [18,19,21,24,25], wherein volleys of transiently synchronous spikes can be propagated through a predominantly excitatory feedforward architecture. Studies have shown that synchronous spike volleys can reliably drive responses through the visual cortex [13] with the temporal precision required for neural coding [26,27]. However, only sufficiently strong stimuli can elicit transient spike volleys that can successfully propagate through the network, and the waveform of spiking tends to an attractor with a single, fixed amplitude [21,28]. Thus, it is not possible to transfer graded information in the amplitudes of synchronously propagating spike volleys in this type of synfire chain.
Nonetheless, recent work has shown that synfire chains may be used as a pulse-gating mechanism coupled with a parallel “graded” chain (“synfire-gated synfire chain”—SGSC) to transfer arbitrary firing-rate amplitudes (graded information) through many layers in a feedforward neural circuit [14,15,16,17]. The addition of a companion gating circuit additionally provides a new mechanism for controlling information propagation in neural circuits [14,15,16,17]. Within a feedforward network, graded-rate transfer manifests as approximately time-translationally invariant spiking probabilities that propagate through many layers (guided by gating pulses) [17]. Using a Fokker–Planck (FP) approach, it has been demonstrated that this time-translational invariance arises near a cusp catastrophe in the parameter space of the gating current, synaptic strength and synaptic noise [17].
While many researchers have studied the dynamics of activity transmission in feedforward networks [21,23,25,29], few have examined these networks as information channels (however, see [30]). Shannon information [31,32] provides a natural framework with which to quantify the capacity of neural information transmission, by providing a measure of the correlation between input and output variables.
In particular, mutual information (MI), and measures based on MI, can be used to evaluate the expected reduction in entropy, for example, of the input, from the measurement of the output. Much work has focused on estimating probability distributions from spiking data (see, e.g., [33,34,35,36,37,38]). At the circuit level, studies have shown that MI is maximized in balanced networks (in cortical cultures, rats and macaques) that admit neuronal avalanches [39,40]. Furthermore, by examining the effects of gamma oscillations on MI in infralimbic and prelimbic cortex of mice, it has been shown that gamma rhythms enhance information transfer by reducing noise and signal amplification [7].
Theoretically, maximizing MI has been used to find nonlinear “infomax” networks [41] that can find statistically independent components capable of separating features in the visual scene [42]. MI has been used to assess the effectiveness and precision of population codes [43]. By combining decoding and MI, one can extract single-trial information from population activity and, at the same time, a quantitative estimate of how each neuron in the population contributes to the internal representation of the stimulus [44]. Furthermore, information-theoretic measures such as MI have been used on large-scale measurements of brain activity to estimate connectivity between different brain regions [45].
Here we consider information propagation properties in feedforward networks, and in particular, examine MI of mean firing-rate transfers across many layers of an SGSC neural circuit. We make use of an FP model to describe the dynamic evolution of membrane-potential probability densities in a pulse-gated, feedforward, integrate-and-fire (I&F) neuronal network [17]. We investigate the efficacy of information transfer in the parameter space near where graded mean firing-rate transfer is possible. We find that MI can be optimized by adjusting the strength of the gating (gating current), the feedforward synaptic strength, the level of synaptic noise, and the input distribution. Furthermore, our results reveal that via the coordination of pulse-gating and synaptic noise, a graded channel may be transformed into a binary channel. Our results demonstrate a wide range of possible information propagation choices in feedforward networks and the dynamic coding capacity of SGSCs.

2. Materials and Methods

We use a neuronal network model consisting of a set of j = 1 , , M populations, each with i = 1 , , N excitatory, current-based, I&F point neurons whose membrane potential, V i , j , and (feedforward) synaptic current, I i , j ff , are described by
d d t V i , j = g L ( V i , j V R ) + I g , j + I i , j ff
τ d d t I i , j ff = I i , j ff + S p N k p j k l δ t t j 1 , k l , j > 1 A δ ( t ) , j = 1
where V R is the reset voltage, τ is the synaptic timescale, S is the synaptic coupling strength, p j k is a Bernoulli distributed random variable and p = p j k is the mean synaptic coupling probability. The lth spike time of the kth neuron in layer j 1 is determined by V ( t j 1 , k l ) = V T h , that is, when the neuron reaches the threshold. The gating current, I g , j , is a white-noise process with a square pulse envelope, θ ( t ( j 1 ) T ) θ ( t j T ) , where θ is a Heaviside theta function and T is the pulse length [14] of pulse height I g and variance σ 0 2 . We note that with the j = 1 equation, an exponentially decaying current is injected into population 1, providing graded synchronized activity that subsequently propagates downstream through populations j = 2 , , M .
Assuming the spike trains in Equation (1b) to be Poisson-distributed, the collective behavior of this feedforward network may be described by the FP equations:
t ρ j V , t = V J j ( V , t )
τ d d t I j ff = I j ff + S m j 1 , j > 1 A δ ( t ) , j = 1
(While the output spike-train of a single neuron in general does not obey Poisson statistics, the spike train obtained from the summed output of all neurons in a single population does obey these statistics asymptotically for a large network size N. In the case of pulse-gating, the summed output spikes of a single population tend to a time-dependent Poisson process.) These equations describe the evolution of the probability density function (PDF), ρ j ( V , t ) , in terms of the probability density flux, J j ( V , t ) , the mean feedforward synaptic current, I j ff I i , j ff , and the population firing rate, m j . For each layer j, the probability density function gives the probability of finding a neuron with membrane potential V ( , V T h ] at time t.
The probability density flux is given by
J j V , t = g L V V R + I g + I j ff σ j 2 V ρ j V , t
where I g indicates the mean gating current. The effective diffusivity is
σ j 2 = σ 0 2 + 1 2 S 2 p N m j 1 t
In the simulations reported below, we have taken N , thus ignoring the second term in Equation (3) (i.e., ignoring diffusion due to finite size effects). The population firing rate is the flux of the PDF at threshold:
m j t = J j V T h , t
The boundary conditions for the FP equations are
J j V R + , t = J j V T h , t + J j V R , t
ρ j V R + , t = ρ j V T h , t + ρ j V R , t
and
ρ j V = , t = 0
To efficiently obtain solutions to the FP equations [17], we have used an approximate Gaussian initial distribution, ρ j ( V , t + j T ) = ( 1 / P ) exp ( ( V μ ( t ) ) 2 / 2 σ 2 ) , with width σ and mean μ ( t ) , where P = V Th ρ 0 ( V , 0 ) is a normalization factor that accounts for the truncation of the Gaussian at threshold, V Th . At the onset of gating, the distribution is advected toward the voltage threshold, V Th , and the population starts to fire. The advection neglects a small amount of firing due to a diffusive flux across the firing threshold; thus the fold bifurcation occurs at a slightly larger value of synaptic coupling, S, for this approximation, relative to numerical simulations [17]. Because the pulse is fast, neurons only have time to fire once (approximately). Thus, we neglect the re-emergent population at V R , which does not have enough time to advect to V Th and therefore does not contribute to firing during the transient pulse.
Using this approximation, Equation (2a) gives rise to μ ˙ = g L ( μ V R ) + I g + I u ff , where σ 2 = σ 0 2 / g L , with upstream current I u ff = A e t / τ . Setting V Th = 1 , this integrates to
μ ( t ) = μ 0 e g L t + I g g L ( 1 e g L t ) + A 1 τ g L ( e g L t e t / τ )
and from Equation (4), we have
m ( t ) = ( g L μ ( t ) + I g + A e t / τ ) 1 P e ( 1 μ ( t ) ) 2 / 2 σ 2 +
which, from Equation (2b), results in a downstream synaptic current at t = T :
I d ff = S e T / τ 0 T e t / τ m ( t ) d t τ
After the gating pulse terminates, the current decays exponentially. This decaying current feeds forward and is integrated by the next layer. For an exact transfer, I d ff ( S , I g , A , T ) = A .
To compute MI, we generated a distribution of upstream current amplitudes { I u } . These were typically within or near the range of fixed points of the map I d ff ( S , I g , A , T ) = A . Using Equations (8)–(10), we generated a distribution of downstream current amplitudes { I d } with { I u } as initial values. The joint probability distribution p ( I u , I d ) was estimated by forming a two-dimensional histogram with Δ I = 0.3 / s . MI, I ( I u ; I d ) = H ( I u ) + H ( I d ) H ( I u , I d ) , was computed (in bits) with H ( x ) = x p ( x ) log 2 ( p ( x ) ) and H ( x , y ) = x , y p ( x , y ) log 2 ( p ( x , y ) ) , where the marginal probabilities p ( I u ) and p ( I d ) were computed from the histogram.

3. Results

We consider the transfer of spiking activity between two successive layers of a pulse-gated feedforward network. The spiking activity of feedforward propagation quickly converges to a stereotypical firing-rate waveform with arbitrary amplitudes (within a certain range) and with associated dynamics of the membrane potential PDF, ρ ( V , t ) , for each layer. This waveform essentially represents a volley of spikes that propagate downstream within the network. Furthermore, the temporal evolution of the first two moments (i.e., mean and variance) of ρ ( V , t ) suffice to capture the dynamics during pulse-gating [17]. Therefore, as [17] showed, the population dynamics across a large region of parameter space can be mapped by using a Gaussian approximation to the membrane potential PDF (see Materials and Methods), revealing the bifurcation structure underlying a cusp catastrophe.
We first examined the cusp catastrophe in SGSC systems and its role in shaping the transfer of mean firing rates. Because of the existence of stable, attracting, translationally invariant firing-rate waveforms, we could capture and understand the dynamics in the SGSC system as an iterated map describing the firing-rate amplitude as it changed between successive network layers. Figure 1 shows two cases of the fixed points of this iterated map. Figure 1a plots the firing-rate amplitude, A, at the fixed points at the end of the pulse-gating period as a function of the strength of the gating current, I g , and synaptic coupling strength, S, for fixed σ = 1.05 . For small I g , there existed a range in S where the system was bistable; however, as we increased the strength of the gating, the region of bistability (two stable attracting solutions with an unstable solution in between) disappeared at a cusp where the manifold of the fixed point was nearly vertical. It was shown in [17] that near this cusp, the slow dynamics along the unstable manifold of the unstable fixed point allow for the nearly graded transfer of firing rates between successive layers, giving rise to an approximate line attractor in the amplitude of the output firing rates. Figure 1b shows that this cusp catastrophe also existed at fixed gating current I g = 24 as we varied both the feedforward synaptic coupling strength S and the variance of the initial PDF, ρ ( V , t = 0 ) , just before pulse-gating. As the variance, σ 2 , was increased, the bistable region in S also disappeared.
Next, we examine the evolution of MI through the network. Figure 2 shows the transmission of the mean firing rate, A / g L , in a pulse-gated feedforward network, as well as the associated MI for three types of channels. Figure 2a plots the firing rate for an idealized exact transfer across many layers ( j = 1 , , 100 ), where we utilized a thresholded linear f–I curve to model each layer [14]. In this toy model, the transfer was exact and the mean firing rate propagated indefinitely without change. Figure 2b shows the propagation of firing rates for an SGSC situated near the cusp of Figure 1a. The long-term propagation of the initially uniform input firing rates revealed the dynamical structure near the cusp, namely, two stable, attracting firing rates (top and bottom) with an unstable saddle in between. Because of the slowness of the dynamics along the unstable manifold of the saddle, the transfer of mean firing rates through the network was approximately graded for many layers ( j 1–30). Furthermore, this transfer of the mean firing rates was order preserving (i.e., the relative ordering of amplitudes was maintained) across many more layers (j up to 100). Figure 2c demonstrates the propagation of firing rates for a binary transfer, where the unstable saddle was strongly repelling; for most initial conditions, the rates converged to one of two rates within 10–15 layers. To investigate the effects of the SGSC dynamics on information propagation, we computed the MI for each of the three cases. Figure 2d demonstrates the effect of the SGSC dynamics on information propagation. In the exact transfer case, the MI between the input and each layer remained constant and represented an upper bound on the information transfer. In the binary transfer case, the MI quickly decayed to 1 bit, but, as for the exact case, was stable over long timescales. Near the cusp, where the firing-rate transfer was approximately graded, the MI decayed slowly, so that even after j = 100 transfers, the channel retained almost 4 bits of information. In Figure 2e, the joint probabilities from which the MI was computed for representative layers are plotted. For the exact case, the distribution is always along the diagonal. The fast transition from diagonal to binary is evident for binary transfers, and a slow transition from diagonal to binary is seen for the graded case.
An FP analysis of the SGSC system reveals four important system parameters: strength of the gating current, I g ; strength of the feedforward synaptic coupling, S; mean membrane potential distribution at the beginning of pulse-gating, ρ j ( V , t = j T ) ; and variance of the synaptic current, σ 2 (see Materials and Methods). Figure 3 shows where graded and binary codes are supported. Figure 3a,b plots the MI as a function of S and the network layer (which is equivalent to propagation time) for fixed I g , σ and ρ j ( V , t = j T ) (which equals 0 for all panels). As we can see, for both cases, there exists an optimal S for which MI (and approximate graded activity) can be maintained for many layers. As we move away from this optimal S, we quickly go into a binary coding regime ( MI = 1 ). Figure 3c demonstrates that similar qualitative behavior can be obtained by varying σ .
Out of the four system parameters, it is easiest to manipulate either the gating current or the variance of the synaptic current. Both can be viewed as controls independent of the SGSC system. Therefore, in Figure 4, we examine the evolution of MI as a function of I g and σ through the network. As we expect from our results thus far, large regions of parameter space support binary coding; however, there is a thin line that materializes towards the bottom of the binary region that corresponds to the location of the cusp, where MI can be maintained at high levels ( MI > 3 ) through many layers. We note that asymptotically, as j , MI approaches 1 (a binary channel) because of the existence of two attractors, even near the cusp.

4. Discussion

Although coherent activity has now been measured in many regions of the mammalian brain, the precise mechanism and the extent to which the brain can make use of synchronous spiking activity to transfer information have remained unclear. Many mechanisms have been proposed for information transfer via transiently synchronous spiking that relies on oscillations and gating [18,20,22,46,47,48,49,50,51,52]. These mechanisms make use of the fact that coherent input can provide temporal windows during which spiking activity may be more easily transferred between sending and receiving populations. However, from the theoretical perspective, how MI and other measures of communication capacity can be related to the underlying neuronal network architecture and the emergent network dynamics have remained unexplored.
Here we study the capacity for information transfer of feedforward networks by examining the evolution of mutual information through many layers of an SGSC system. Previous work has showed that by introducing suitably timed pulses, graded information could be transferred and controlled in a feedforward excitatory neuronal network [14,15,16,17]. In this context, pulse-gating has allowed us to understand information propagation as an iterated map. FP analysis of this map has enabled the identification of a cusp catastrophe in the relevant parameter space. Our results here demonstrate that the dynamics of the SGSC system naturally give rise to two different types of channel (as measured by MI). Large regions of parameter space support binary coding by using transiently synchronous propagation of high and low firing rates. (In the classical synfire chain case, the low firing-rate state is a silent state. In the SGSC case, as a result of the gating current, it is small, but non-zero.) In the binary regime, the distance between low and high firing rates is much larger than the variance of the distributions; therefore propagation is stable. Furthermore, by systematically varying the relevant SGSC system parameters, we were able to optimize graded-rate transfer near the fold of a cusp catastrophe, which enabled us to maintain a relatively high MI through many layers.
Evidence exists for graded information coding in visual and other cortices [53], and there is some evidence for binary coding in the auditory cortex [54]; sparse coding mechanisms for its use have been put forward [55,56]. Luczak et al. [51] have argued that spike packets and stereotypical and repeating sequences (similar to sequences observed in implementations of information processing algorithms in graded-transfer SGSCs [15]) underlie neural coding. More recently, Piet et al. [57] have used attractor networks to model frontal orienting fields in rat cortices to argue that bistable attractor dynamics can account for the memory observed in a perceptual decision-making task. Indeed, in many decision-making tasks, cortical activity appears to be holding graded information in working memory, before a decision forces the activity into a binary code [58]. Therefore, it is of interest that both graded and binary pulse-gated channels are supported by the SGSC mechanism and that it is fairly simple and rapid to convert a graded code to a binary code.
Examining the structure of the cusp catastrophe in the parameter space, it appears that in general, weaker I g and higher S are correlated with bistability, while stronger I g can support graded information transfer at lower S (see Figure 1a); at the same time, for fixed I g , a lower σ tends to bistable states (see Figure 1b).
In previous theoretical studies of graded propagation and line attractors, it has been shown that some fine tuning of system parameters is required [59,60,61]. However, graded propagation and line attractors have been observed in many areas of the brain (see, e.g., [62,63,64,65]). Our results here demonstrate that an important consideration for understanding graded propagation is the depth of the circuit being used to propagate the information. That is, graded information propagation circuits with a depth of 20 to 30 layers are not particularly fine-tuned. In this depth range, about 1 / 10 th of the parameter space (e.g., S or σ ; see Figure 3, top and bottom plots) is capable of propagating graded information (with relatively high MI). As the depth of the circuit grows, j , MI approaches 1 bit, and hence graded propagation in deep circuits is not possible (with an SGSC with parameters near the cusp).
A clear advantage of a graded information channel is that a vector of high-resolution graded information (resolution 2 n , where n is the number of bits of MI—up to 32 levels of resolution in the graded channel shown in Figure 2) could be rapidly processed in a network with linear synaptic connectivity. Thus, synaptic processing such as Gabor transforms, seen in the visual cortex, would most naturally operate on graded information, rapidly reducing the dimension of and orthogonalizing input data. However, the stable processing of information through deep and complicated neural logic circuits would take better advantage of binary channels. Here, essentially exact pulse-gated binary transformations and decisions [15] can make use of the attractor structure of the channel to reduce noise and maintain discrete states. In order to make use of high-bandwidth graded processing, at some point in a neural circuit, graded information would need to be transformed to binary information. Mechanisms to do this are beyond the scope of this paper, but they could make use of dimension-reducing transforms on the input (e.g., Gabor transform) and subsequent digitization of the data subspace.
Finally, we note that the symmetric nature of the binary SGSC channel may be an advantage for logic circuit gating, as either of the parallel synfire chains can operate as information or the gate [15]. Finally, it should be remembered that with a stable binary code, binary digit coding can be constructed, effectively increasing the information resolution propagating in a binary circuit.

5. Conclusions

We have performed an investigation of the communication capabilities of SGSCs using the metric of MI. The main conclusion of this investigation is that SGSCs sustain two types of channel: the first is binary transfer (MI = 1, in bits), which is supported for a wide range of parameters. For circuit depths of up to 30–40 layers, in a narrower range of parameters, graded transfer is also supported. A secondary conclusion is that, because of the dependence of MI on the depth of a circuit in some parameter regimes, circuit depth should be taken into account when considering the communication capacity of a neural circuit.

Acknowledgments

This work was supported by the Natural Science Foundation of China, Grant Nos. 31771147 and 91232715 (B.W. and L.T.); by the Open Research Fund of the State Key Laboratory of Cognitive Neuroscience and Learning, Grant No. CNLZD1404 (B.W. and L.T.); by the Beijing Municipal Science and Technology Commission under contract Z151100000915070 (B.W. and L.T.); by the SLS-Qidong Innovation Fund (B.W. and L.T.); and by the LANL ASC Beyond Moore’s Law project (A.T.S.).

Author Contributions

All authors conceived and designed the numerical simulations. A.T.S. and L.T. wrote the paper, with major input from Z.X. and B.W. All authors have read and approved the final manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gray, C.; König, P.; Engel, A.; Singer, W. Oscillatory responses in cat visual cortex exhibit inter-columnar synchronization which reflects global stimulus properties. Nature 1989, 338, 334–337. [Google Scholar] [CrossRef] [PubMed]
  2. Livingstone, M. Oscillatory firing and interneuronal correlations in squirrel monkey striate cortex. J. Neurophysiol. 1996, 66, 2467–2485. [Google Scholar] [CrossRef] [PubMed]
  3. Bragin, A.; Jandó, G.; Nádasdy, Z.; Hetke, J.; Wise, K.; Buzsáki, G. Gamma (40–100 Hz) oscillation in the hippocampus of the behaving rat. J. Neurosci. 1995, 15, 47–60. [Google Scholar] [PubMed]
  4. Brosch, M.; Budinger, E.; Scheich, H. Stimulus-related gamma oscillations in primate auditory cortex. J. Neurophysiol. 2002, 87, 2715–2725. [Google Scholar] [CrossRef] [PubMed]
  5. Bauer, M.; Oostenveld, R.; Peeters, M.; Fries, P. Tactile spatial attention enhances gamma-band activity in somatosensory cortex and reduces low-frequency activity in parieto-occipital areas. J. Neurosci. 2006, 26, 490–501. [Google Scholar] [CrossRef] [PubMed]
  6. Pesaran, B.; Pezaris, J.; Sahani, M.; Mitra, P.; Andersen, R. Temporal structure in neuronal activity during working memory in macaque parietal cortex. Nat. Neurosci. 2002, 5, 805–811. [Google Scholar] [CrossRef] [PubMed]
  7. Sohal, V.S.; Zhang, F.; Yizhar, O.; Deisseroth, K. Parvalbumin neurons and gamma rhythms enhance cortical circuit performance. Nature 2009, 459, 698–702. [Google Scholar] [CrossRef] [PubMed]
  8. Popescu, A.; Popa, D.; Paré, D. Coherent gamma oscillations couple the amygdala and striatum during learning. Nat. Neurosci. 2009, 12, 801–807. [Google Scholar] [CrossRef] [PubMed]
  9. Bair, W.; Koch, C. Temporal precision of spike trains in extrastriate cortex of the behaving macaque monkey. Neural Comput. 1996, 8, 1185–1202. [Google Scholar] [CrossRef] [PubMed]
  10. Butts, D.A.; Weng, C.; Jin, J.; Yeh, C.I.; Lesica, N.A.; Alonso, J.M.; Stanley, G.B. Temporal precision in the neural code and the timescales of natural vision. Nature 2007, 449, 92–95. [Google Scholar] [CrossRef] [PubMed]
  11. Varga, C.; Golshani, P.; Soltesz, I. Frequency-invariant temporal ordering of interneuronal discharges during hippocampal oscillations in awake mice. Proc. Natl. Acad. Sci. USA 2012, 109, E2726–E2734. [Google Scholar] [CrossRef] [PubMed]
  12. Reyes, A.D. Synchrony-dependent propagation of firing rate in iteratively constructed networks in vitro. Nat. Neurosci. 2003, 6, 593–599. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, H.P.; Spencer, D.; Fellous, J.M.; Sejnowski, T.J. Synchrony of thalamocortical inputs maximizes cortical reliability. Science 2010, 328, 106–109. [Google Scholar] [CrossRef] [PubMed]
  14. Sornborger, A.T.; Wang, Z.; Tao, L. A mechanism for graded, dynamically routable current propagation in pulse-gated synfire chains and implications for information coding. J. Comput. Neurosci. 2015, 39, 181–195. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, Z.; Sornborger, A.; Tao, L. Graded, dynamically routable information processing with synfire-gated synfire chains. PLoS Comput. Biol. 2016, 12, e1004979. [Google Scholar] [CrossRef] [PubMed]
  16. Shao, Y.; Sornborger, A.; Tao, L. A pulse-gated, predictive neural circuit. In Proceedings of the 50th Asilomar Conference on Signals, Systems and Computers, Pacific Grove, CA, USA, 6–9 November 2016. [Google Scholar]
  17. Xiao, Z.; Zhang, J.; Sornborger, A.; Tao, L. Cusps enable line attractors for neural computation. Phys. Rev. E 2017, 96, 052308. [Google Scholar] [CrossRef] [PubMed]
  18. Abeles, M. Role of the cortical neuron: Integrator or coincidence detector? Isr. J. Med. Sci. 1982, 18, 83–92. [Google Scholar] [PubMed]
  19. König, P.; Engel, A.; Singer, W. Integrator or coincidence detector? The role of the cortical neuron revisited. Trends Neurosci. 1996, 19, 130–137. [Google Scholar] [CrossRef]
  20. Fries, P. A mechanism for cognitive dynamics: Neuronal communication through neuronal coherence. Trends Cogn. Sci. 2005, 9, 474–480. [Google Scholar] [CrossRef] [PubMed]
  21. Diesmann, M.; Gewaltig, M.O.; Aertsen, A. Stable propagation of synchronous spiking in cortical neural networks. Nature 1999, 402, 529–533. [Google Scholar] [CrossRef] [PubMed]
  22. Rubin, J.; Terman, D. High frequency stimulation of the subthalamic nucleus eliminates pathological thalamic rhythmicity in a computational model. J. Comput. Neurosci. 2004, 16, 211–235. [Google Scholar] [CrossRef] [PubMed]
  23. Van Rossum, M.C.; Turrigiano, G.G.; Nelson, S.B. Fast propagation of firing rates through layered networks of noisy neurons. J. Neurosci. 2002, 22, 1956–1966. [Google Scholar] [PubMed]
  24. Kistler, W.; Gerstner, W. Stable propagation of activity pulses in populations of spiking neurons. Neural Comput. 2002, 14, 987–997. [Google Scholar] [CrossRef] [PubMed]
  25. Litvak, V.; Sompolinsky, H.; Segev, I.; Abeles, M. On the transmission of rate code in long feedforward networks with excitatory-inhibitory balance. J. Neurosci. 2003, 23, 3006–3015. [Google Scholar] [PubMed]
  26. Kumar, A.; Rotter, S.; Aertsen, A. Conditions for propagating synchronous spiking and asynchronous firing rates in a cortical network model. J. Neurosci. 2008, 28, 5268–5280. [Google Scholar] [CrossRef] [PubMed]
  27. Nemenman, I.; Lewen, G.D.; Bialek, W.; de Ruyter van Steveninck, R.R. Neural coding of natural stimuli: Information at sub-millisecond resolution. PLoS Comput. Biol. 2008, 4, e1000025. [Google Scholar] [CrossRef] [PubMed]
  28. Moldakarimov, S.; Bazhenov, M.; Sejnowski, T.J. Feedback stabilizes propagation of synchronous spiking in cortical neural networks. Proc. Natl. Acad. Sci. USA 2015, 112, 2545–2550. [Google Scholar] [CrossRef] [PubMed]
  29. Kumar, A.; Rotter, S.; Aertsen, A. Spiking activity propagation in neuronal networks: Reconciling different perspectives on neural coding. Nat. Rev. Neurosci. 2010, 11, 615–627. [Google Scholar] [CrossRef] [PubMed]
  30. Cannon, J. Analytical calculation of mutual information between weakly coupled poisson-spiking neurons in models of dynamically gated communication. Neural Comput. 2017, 29, 118–145. [Google Scholar] [CrossRef] [PubMed]
  31. Shannon, C. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef]
  32. Cover, T.; Thomas, J. Elements of Information Theory, 2nd ed.; Wiley: Hoboken, NJ, USA, 2006. [Google Scholar]
  33. Rieke, F.; Warland, D.; de Ruyter van Steveninck, R.; Bialek, W. Spikes: Exploring the Neural Code; The MIT Press: Cambridge, MA, USA, 1997. [Google Scholar]
  34. Paninski, L. Estimation of entropy and mutual information. Neural Comput. 2003, 15, 1191–1253. [Google Scholar] [CrossRef]
  35. Victor, J.D. Binless strategies for estimation of information from neural data. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 2002, 66, 051903. [Google Scholar] [CrossRef] [PubMed]
  36. Nemenman, I.; Bialek, W.; de Ruyter van Steveninck, R. Entropy and information in neural spike trains: progress on the sampling problem. Phys. Rev. E Stat. Nonlinear Soft Matter Phys. 2004, 69, 056111. [Google Scholar] [CrossRef] [PubMed]
  37. Shapira, A.H.; Nelken, I. Binless Estimation of Mutual Information in Metric Spaces. In Spike Timing: Mechanisms and Function; DiLorenzo, P., Victor, J., Eds.; CRC Press: Boca Raton, FL, USA, 2013; Chapter 5; pp. 121–135. [Google Scholar]
  38. Lopes-dos Santos, V.; Panzeri, S.; Kayser, C.; Diamond, M.E.; Quian Quiroga, R. Extracting information in spike time patterns with wavelets and information theory. J. Neurophysiol. 2015, 113, 1015–1033. [Google Scholar] [CrossRef] [PubMed]
  39. Shew, W.L.; Plenz, D. The functional benefits of criticality in the cortex. Neuroscientist 2013, 19, 88–100. [Google Scholar] [CrossRef] [PubMed]
  40. Shew, W.L.; Yang, H.; Yu, S.; Roy, R.; Plenz, D. Information capacity and transmission are maximized in balanced cortical networks with neuronal avalanches. J. Neurosci. 2011, 31, 55–63. [Google Scholar] [CrossRef] [PubMed]
  41. Bell, A.J.; Sejnowski, T.J. An information-maximization approach to blind separation and blind deconvolution. Neural Comput. 1995, 7, 1129–1159. [Google Scholar] [CrossRef] [PubMed]
  42. Bell, A.J.; Sejnowski, T.J. The “independent components” of natural scenes are edge filters. Vis. Res. 1997, 37, 3327–3338. [Google Scholar] [CrossRef]
  43. Yarrow, S.; Challis, E.; Series, P. Fisher and Shannon information in finite neural populations. Neural Comput. 2012, 24, 1740–1780. [Google Scholar] [CrossRef] [PubMed]
  44. Quiroga, R.Q.; Panzeri, S. Extracting information from neuronal populations: Information theory and decoding approaches. Nat. Rev. Neurosci. 2009, 10, 173–185. [Google Scholar] [CrossRef] [PubMed]
  45. Bullmore, E.; Sporns, O. Complex brain networks: Graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 2009, 10, 186–198. [Google Scholar] [CrossRef] [PubMed]
  46. Fries, P.; Reynolds, J.; Rorie, A.; Desimone, R. Modulation of oscillatory neuronal synchronization by selective visual attention. Science 2001, 291, 1560–1563. [Google Scholar] [CrossRef] [PubMed]
  47. Salinas, E.; Sejnowski, T. Correlated neuronal activity and the flow of neural information. Nat. Rev. Neurosci. 2001, 2, 539–550. [Google Scholar] [CrossRef] [PubMed]
  48. Csicsvari, J.; Jamieson, B.; Wise, K.; Buzsáki, G. Mechanisms of gamma oscillations in the hippocampus of the behaving rat. Neuron 2003, 37, 311–322. [Google Scholar] [CrossRef]
  49. Womelsdorf, T.; Schoffelen, J.; Oostenveld, R.; Singer, W.; Desimone, R.; Engel, A.; Fries, P. Modulation of neuronal interactions through neuronal synchronization. Science 2007, 316, 1609–1612. [Google Scholar] [CrossRef] [PubMed]
  50. Colgin, L.; Denninger, T.; Fyhn, M.; Hafting, T.; Bonnevie, T.; Jensen, O.; Moser, M.; Moser, E. Frequency of gamma oscillations routes flow of information in the hippocampus. Nature 2009, 462, 75–78. [Google Scholar] [CrossRef] [PubMed]
  51. Luczak, A.; McNaughton, B.L.; Harris, K.D. Packet-based communication in the cortex. Nat. Rev. Neurosci. 2015, 16, 745–755. [Google Scholar] [CrossRef] [PubMed]
  52. Yuste, R.; MacLean, J.; Smith, J.; Lansner, A. The cortex as a central pattern generator. Nat. Rev. Neurosci. 2005, 6, 477–483. [Google Scholar] [CrossRef] [PubMed]
  53. Hubel, D.; Wiesel, T. Receptive fields and functional architecture in two non striate visual areas (18 and 19) of the cat. J. Neurophysiol. 1965, 28, 229–289. [Google Scholar] [CrossRef] [PubMed]
  54. DeWeese, M.; Wehr, M.; Zador, A. Binary coding in auditory cortex. J. Neurosci. 2003, 23, 7940–7949. [Google Scholar] [PubMed]
  55. Olshausen, B.; Field, D. Sparse coding of sensory inputs. Curr. Opin. Neurobiol. 2004, 14, 481–487. [Google Scholar] [CrossRef] [PubMed]
  56. Hromádka, T.; DeWeese, M.; Zador, A. Sparse representation of sounds in the unanesthetized auditory cortex. PLoS Biol. 2008, 6, e16. [Google Scholar] [CrossRef] [PubMed]
  57. Piet, A.T.; Erlich, J.C.; Kopec, C.D.; Brody, C.D. Rat Prefrontal Cortex Inactivations during Decision Making Are Explained by Bistable Attractor Dynamics. Neural Comput. 2017, 29, 2861–2886. [Google Scholar] [CrossRef] [PubMed]
  58. Brody, C.D.; Hernandez, A.; Zainos, A.; Romo, R. Timing and neural encoding of somatosensory parametric working memory in macaque prefrontal cortex. Cereb. Cortex 2003, 13, 1196–1207. [Google Scholar] [CrossRef] [PubMed]
  59. Seung, H.S. How the brain keeps the eyes still. Proc. Natl. Acad. Sci. USA 1996, 93, 13339–13344. [Google Scholar] [CrossRef] [PubMed]
  60. Koulakov, A.A.; Raghavachari, S.; Kepecs, A.; Lisman, J.E. Model for a robust neural integrator. Nat. Neurosci. 2002, 5, 775–782. [Google Scholar] [CrossRef] [PubMed]
  61. Goldman, M. Memory without feedback in a neural network. Neuron 2008, 61, 621–634. [Google Scholar] [CrossRef] [PubMed]
  62. Brody, C.D.; Romo, R.; Kepecs, A. Basic mechanisms for graded persistent activity: Discrete attractors, continuous attractors, and dynamic representations. Curr. Opin. Neurobiol. 2003, 13, 204–211. [Google Scholar] [CrossRef]
  63. Major, G.; Tank, D. Persistent neural activity: Prevalence and mechanisms. Curr. Opin. Neurobiol. 2004, 14, 675–684. [Google Scholar] [CrossRef] [PubMed]
  64. Aksay, E.; Olasagasti, I.; Mensh, B.D.; Baker, R.; Goldman, M.S.; Tank, D.W. Functional dissection of circuitry in a neural integrator. Nat. Neurosci. 2007, 10, 494–504. [Google Scholar] [CrossRef] [PubMed]
  65. Mante, V.; Sussillo, D.; Shenoy, K.V.; Newsome, W.T. Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 2013, 503, 78–84. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Cusp enabling graded and binary propagation in a synfire-gated synfire chain (SGSC). (a) View for fixed noise, σ = 1.05 , as a function of firing-rate amplitude, A / g L ; synaptic strength, S; and gating current, I g . (b) View for fixed gating current, I g = 24 , as a function of A / g L , S, and σ . Plotted in both panels are zeros of the function I ( A , I g , σ ) A (the difference between input and output firing-rate amplitudes in a given layer). This function becomes zero at a fixed point. At a fixed point, a firing-rate amplitude propagates exactly from one layer to the next. When this function is small but non-zero, the firing-rate amplitude only changes slowly as it propagates. We note that the cusp is high-dimensional and can be viewed in various projections. Graded propagation (approximately exact amplitude propagation) is enabled as a result of ghost (slow) dynamics near the cusp, such that the firing-rate amplitude A varies only slowly from layer to layer. At the cusp, an approximate line attractor exists where an input firing rate in a given layer changes only slowly as it propagates downstream. Away from the cusp, firing rates rapidly approach attractors. For many parameter regions, there are two attractors giving rise to the propagation of a binary code.
Figure 1. Cusp enabling graded and binary propagation in a synfire-gated synfire chain (SGSC). (a) View for fixed noise, σ = 1.05 , as a function of firing-rate amplitude, A / g L ; synaptic strength, S; and gating current, I g . (b) View for fixed gating current, I g = 24 , as a function of A / g L , S, and σ . Plotted in both panels are zeros of the function I ( A , I g , σ ) A (the difference between input and output firing-rate amplitudes in a given layer). This function becomes zero at a fixed point. At a fixed point, a firing-rate amplitude propagates exactly from one layer to the next. When this function is small but non-zero, the firing-rate amplitude only changes slowly as it propagates. We note that the cusp is high-dimensional and can be viewed in various projections. Graded propagation (approximately exact amplitude propagation) is enabled as a result of ghost (slow) dynamics near the cusp, such that the firing-rate amplitude A varies only slowly from layer to layer. At the cusp, an approximate line attractor exists where an input firing rate in a given layer changes only slowly as it propagates downstream. Away from the cusp, firing rates rapidly approach attractors. For many parameter regions, there are two attractors giving rise to the propagation of a binary code.
Entropy 20 00102 g001
Figure 2. Mutual information (MI) for transfers across a 100 layer synfire-gated synfire chain (SGSC). (a) Exact transfer: Firing-rate amplitudes for theoretically perfect transfer. The range of A / g L is the same for (ac). Amplitudes transfer exactly from one layer to the next ( j = 1 , , 100 ). This practically unattainable communication mode is shown for comparison and attains the maximum possible information transfer (MI; see (d)) as a function of the layer. (b) Graded transfer: Firing-rate amplitudes from approximate Fokker–Planck (FP) solutions in a multi-layer SGSC. Parameters: I g = 24.03 ; S = 7.001 ; σ = 1.0496 . We note that as the number of layers through which the initial amplitudes propagate increases, the amplitudes drift slowly away from an unstable attractor in the center of the amplitude distribution towards stable attractors at the sides of the distribution. (c) Binary transfer: Firing-rate amplitudes from approximate FP solutions in a multi-layer SGSC. Parameters: I g = 5.0 ; S = 7 ; σ = 0.6 . For these parameters, the unstable attractor rapidly repels the amplitudes toward stable attractors on either side, resulting in binary transfer. Binary transfer is extremely stable in the SGSC. (d) MI as a function of layer, j, for each of the above types of channel (ac). (e) The joint probabilities, p ( I 1 , I 2 ) , p ( I 1 , I 20 ) , and p ( I 1 , I 100 ) , from which the MI was computed for exact, graded and binary transfer. Here, the exact case is as would be expected. The graded case shows a gradual deformation away from the diagonal that approaches binary transfer asymptotically. The binary case rapidly approaches the propagation of only two states. We note that in much of the parameter space, the approach to binary is much faster than that shown here. We used these parameters so that the transition was slower and therefore more evident.
Figure 2. Mutual information (MI) for transfers across a 100 layer synfire-gated synfire chain (SGSC). (a) Exact transfer: Firing-rate amplitudes for theoretically perfect transfer. The range of A / g L is the same for (ac). Amplitudes transfer exactly from one layer to the next ( j = 1 , , 100 ). This practically unattainable communication mode is shown for comparison and attains the maximum possible information transfer (MI; see (d)) as a function of the layer. (b) Graded transfer: Firing-rate amplitudes from approximate Fokker–Planck (FP) solutions in a multi-layer SGSC. Parameters: I g = 24.03 ; S = 7.001 ; σ = 1.0496 . We note that as the number of layers through which the initial amplitudes propagate increases, the amplitudes drift slowly away from an unstable attractor in the center of the amplitude distribution towards stable attractors at the sides of the distribution. (c) Binary transfer: Firing-rate amplitudes from approximate FP solutions in a multi-layer SGSC. Parameters: I g = 5.0 ; S = 7 ; σ = 0.6 . For these parameters, the unstable attractor rapidly repels the amplitudes toward stable attractors on either side, resulting in binary transfer. Binary transfer is extremely stable in the SGSC. (d) MI as a function of layer, j, for each of the above types of channel (ac). (e) The joint probabilities, p ( I 1 , I 2 ) , p ( I 1 , I 20 ) , and p ( I 1 , I 100 ) , from which the MI was computed for exact, graded and binary transfer. Here, the exact case is as would be expected. The graded case shows a gradual deformation away from the diagonal that approaches binary transfer asymptotically. The binary case rapidly approaches the propagation of only two states. We note that in much of the parameter space, the approach to binary is much faster than that shown here. We used these parameters so that the transition was slower and therefore more evident.
Entropy 20 00102 g002
Figure 3. Mutual information (MI) for parameters supporting graded and binary codes. (a) MI propagation across 100 layers for I g = 24 , σ = 1.05 and S ranging from 0 to 10. We note that graded transfer allows high values of MI to propagate across 100 layers at S 7 . (b) MI propagation across 100 layers for I g = 24 , σ = 0.6 and S from 0 to 10. We note that above S 5 , there is a large range of S for which binary propagation (MI of 1 bit; see colorbar) is supported. (c) MI propagation for I g = 24 , S = 7 and σ from 0.1 to 2. Here, both graded and binary information propagate depending on the value of σ .
Figure 3. Mutual information (MI) for parameters supporting graded and binary codes. (a) MI propagation across 100 layers for I g = 24 , σ = 1.05 and S ranging from 0 to 10. We note that graded transfer allows high values of MI to propagate across 100 layers at S 7 . (b) MI propagation across 100 layers for I g = 24 , σ = 0.6 and S from 0 to 10. We note that above S 5 , there is a large range of S for which binary propagation (MI of 1 bit; see colorbar) is supported. (c) MI propagation for I g = 24 , S = 7 and σ from 0.1 to 2. Here, both graded and binary information propagate depending on the value of σ .
Entropy 20 00102 g003
Figure 4. Regions of parameter space supporting graded and binary information propagation. Visualization of mutual information (MI) for σ between 0.5 and 1.05 and I g between 14 and 27 at successive layers j; σ and I g axes as denoted in the lower left panel are the same for all panels. Colorbar for lower right panel is the same for all panels. We note that across a few layers, MI remains high for a large region of parameter space, but by j = 21 , only binary and graded codes persist. By j = 91 , it is seen that a large region of parameter space supports binary propagation. Along a thin diagonal line at the bottom of the binary region, corresponding to the location of the cusp, graded information may be transferred.
Figure 4. Regions of parameter space supporting graded and binary information propagation. Visualization of mutual information (MI) for σ between 0.5 and 1.05 and I g between 14 and 27 at successive layers j; σ and I g axes as denoted in the lower left panel are the same for all panels. Colorbar for lower right panel is the same for all panels. We note that across a few layers, MI remains high for a large region of parameter space, but by j = 21 , only binary and graded codes persist. By j = 91 , it is seen that a large region of parameter space supports binary propagation. Along a thin diagonal line at the bottom of the binary region, corresponding to the location of the cusp, graded information may be transferred.
Entropy 20 00102 g004

Share and Cite

MDPI and ACS Style

Xiao, Z.; Wang, B.; Sornborger, A.T.; Tao, L. Mutual Information and Information Gating in Synfire Chains. Entropy 2018, 20, 102. https://doi.org/10.3390/e20020102

AMA Style

Xiao Z, Wang B, Sornborger AT, Tao L. Mutual Information and Information Gating in Synfire Chains. Entropy. 2018; 20(2):102. https://doi.org/10.3390/e20020102

Chicago/Turabian Style

Xiao, Zhuocheng, Binxu Wang, Andrew T. Sornborger, and Louis Tao. 2018. "Mutual Information and Information Gating in Synfire Chains" Entropy 20, no. 2: 102. https://doi.org/10.3390/e20020102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop