Next Article in Journal
Microcanonical Description of (Micro) Black Holes
Next Article in Special Issue
Information Theory and Dynamical System Predictability
Previous Article in Journal
Primitive Membrane Formation, Characteristics and Roles in the Emergent Properties of a Protocell
Previous Article in Special Issue
Information Theoretic Hierarchical Clustering
Entropy 2011, 13(2), 485-501; doi:10.3390/e13020485
Article

Estimating Neuronal Information: Logarithmic Binning of Neuronal Inter-Spike Intervals

Received: 9 December 2010 / Revised: 14 January 2011 / Accepted: 26 January 2011 / Published: 10 February 2011
(This article belongs to the Special Issue Advances in Information Theory)
Download PDF [143 KB, uploaded 24 February 2015]

Abstract

Neurons communicate via the relative timing of all-or-none biophysical signals called spikes. For statistical analysis, the time between spikes can be accumulated into inter-spike interval histograms. Information theoretic measures have been estimated from these histograms to assess how information varies across organisms, neural systems, and disease conditions. Because neurons are computational units that, to the extent they process time, work not by discrete clock ticks but by the exponential decays of numerous intrinsic variables, we propose that neuronal information measures scale more naturally with the logarithm of time. For the types of inter-spike interval distributions that best describe neuronal activity, the logarithm of time enables fewer bins to capture the salient features of the distributions. Thus, discretizing the logarithm of inter-spike intervals, as compared to the inter-spike intervals themselves, yields histograms that enable more accurate entropy and information estimates for fewer bins and less data. Additionally, as distribution parameters vary, the entropy and information calculated from the logarithm of the inter-spike intervals are substantially better behaved, e.g., entropy is independent of mean rate, and information is equally affected by rate gains and divisions. Thus, when compiling neuronal data for subsequent information analysis, the logarithm of the inter-spike intervals is preferred, over the untransformed inter-spike intervals, because it yields better information estimates and is likely more similar to the construction used by nature herself.
Keywords: neuron; neural information; inter-spike interval; logarithmic probability density; spike pattern entropy neuron; neural information; inter-spike interval; logarithmic probability density; spike pattern entropy
This is an open access article distributed under the Creative Commons Attribution License which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Share & Cite This Article

Further Mendeley | CiteULike
Export to BibTeX |
EndNote
MDPI and ACS Style

Dorval, A.D. Estimating Neuronal Information: Logarithmic Binning of Neuronal Inter-Spike Intervals. Entropy 2011, 13, 485-501.

View more citation formats

Article Metrics

For more information on the journal, click here

Comments

Cited By

[Return to top]
Entropy EISSN 1099-4300 Published by MDPI AG, Basel, Switzerland RSS E-Mail Table of Contents Alert