Back to Search
Start Over
Approximations of Shannon Mutual Information for Discrete Variables with Applications to Neural Population Coding
- Source :
- Entropy, Volume 21, Issue 3, Entropy, Vol 21, Iss 3, p 243 (2019)
- Publication Year :
- 2019
- Publisher :
- Multidisciplinary Digital Publishing Institute, 2019.
-
Abstract
- Although Shannon mutual information has been widely used, its effective calculation is often difficult for many practical problems, including those in neural population coding. Asymptotic formulas based on Fisher information sometimes provide accurate approximations to the mutual information but this approach is restricted to continuous variables because the calculation of Fisher information requires derivatives with respect to the encoded variables. In this paper, we consider information-theoretic bounds and approximations of the mutual information based on Kullback--Leibler divergence and R\'{e}nyi divergence. We propose several information metrics to approximate Shannon mutual information in the context of neural population coding. While our asymptotic formulas all work for discrete variables, one of them has consistent performance and high accuracy regardless of whether the encoded variables are discrete or continuous. We performed numerical simulations and confirmed that our approximation formulas were highly accurate for approximating the mutual information between the stimuli and the responses of a large neural population. These approximation formulas may potentially bring convenience to the applications of information theory to many practical and theoretical problems.<br />Comment: 31 pages, 6 figures
- Subjects :
- FOS: Computer and information sciences
Rényi divergence
Kullback–Leibler divergence
Kullback-Leibler divergence
Computer science
Approximations of π
Computer Science - Information Theory
General Physics and Astronomy
Mathematics - Statistics Theory
lcsh:Astrophysics
Context (language use)
Statistics Theory (math.ST)
Neural population
Information theory
01 natural sciences
Article
03 medical and health sciences
symbols.namesake
0302 clinical medicine
lcsh:QB460-466
0103 physical sciences
FOS: Mathematics
Applied mathematics
neural population coding
lcsh:Science
Divergence (statistics)
010306 general physics
Fisher information
mutual information
approximation
Information Theory (cs.IT)
Chernoff divergence
Mutual information
16. Peace & justice
lcsh:QC1-999
discrete variables
FOS: Biological sciences
Quantitative Biology - Neurons and Cognition
symbols
Neurons and Cognition (q-bio.NC)
lcsh:Q
lcsh:Physics
030217 neurology & neurosurgery
probability_and_statistics
Coding (social sciences)
Subjects
Details
- Language :
- English
- ISSN :
- 10994300
- Database :
- OpenAIRE
- Journal :
- Entropy
- Accession number :
- edsair.doi.dedup.....7555faff66daf2bb3def9b6af7174487
- Full Text :
- https://doi.org/10.3390/e21030243