-
Je něco špatně v tomto záznamu ?
How Far can Neural Correlations Reduce Uncertainty? Comparison of Information Transmission Rates for Markov and Bernoulli Processes
A. Pregowska, E. Kaplan, J. Szczepanski,
Jazyk angličtina Země Singapur
Typ dokumentu srovnávací studie, časopisecké články
- MeSH
- akční potenciály * MeSH
- informační teorie * MeSH
- Markovovy řetězce * MeSH
- modely neurologické * MeSH
- nejistota * MeSH
- Publikační typ
- časopisecké články MeSH
- srovnávací studie MeSH
The nature of neural codes is central to neuroscience. Do neurons encode information through relatively slow changes in the firing rates of individual spikes (rate code) or by the precise timing of every spike (temporal code)? Here we compare the loss of information due to correlations for these two possible neural codes. The essence of Shannon's definition of information is to combine information with uncertainty: the higher the uncertainty of a given event, the more information is conveyed by that event. Correlations can reduce uncertainty or the amount of information, but by how much? In this paper we address this question by a direct comparison of the information per symbol conveyed by the words coming from a binary Markov source (temporal code) with the information per symbol coming from the corresponding Bernoulli source (uncorrelated, rate code). In a previous paper we found that a crucial role in the relation between information transmission rates (ITRs) and firing rates is played by a parameter s, which is the sum of transition probabilities from the no-spike state to the spike state and vice versa. We found that in this case too a crucial role is played by the same parameter s. We calculated the maximal and minimal bounds of the quotient of ITRs for these sources. Next, making use of the entropy grouping axiom, we determined the loss of information in a Markov source compared with the information in the corresponding Bernoulli source for a given word length. Our results show that in the case of correlated signals the loss of information is relatively small, and thus temporal codes, which are more energetically efficient, can replace rate codes effectively. These results were confirmed by experiments.
Citace poskytuje Crossref.org
- 000
- 00000naa a2200000 a 4500
- 001
- bmc20006547
- 003
- CZ-PrNML
- 005
- 20200526134608.0
- 007
- ta
- 008
- 200511s2019 si f 000 0|eng||
- 009
- AR
- 024 7_
- $a 10.1142/S0129065719500035 $2 doi
- 035 __
- $a (PubMed)30841769
- 040 __
- $a ABA008 $b cze $d ABA008 $e AACR2
- 041 0_
- $a eng
- 044 __
- $a si
- 100 1_
- $a Pregowska, Agnieszka $u Institute of Fundamental Technological Research, Polish Academy of Sciences, ul. Pawinskiego 5B, 02-106 Warsaw, Poland.
- 245 10
- $a How Far can Neural Correlations Reduce Uncertainty? Comparison of Information Transmission Rates for Markov and Bernoulli Processes / $c A. Pregowska, E. Kaplan, J. Szczepanski,
- 520 9_
- $a The nature of neural codes is central to neuroscience. Do neurons encode information through relatively slow changes in the firing rates of individual spikes (rate code) or by the precise timing of every spike (temporal code)? Here we compare the loss of information due to correlations for these two possible neural codes. The essence of Shannon's definition of information is to combine information with uncertainty: the higher the uncertainty of a given event, the more information is conveyed by that event. Correlations can reduce uncertainty or the amount of information, but by how much? In this paper we address this question by a direct comparison of the information per symbol conveyed by the words coming from a binary Markov source (temporal code) with the information per symbol coming from the corresponding Bernoulli source (uncorrelated, rate code). In a previous paper we found that a crucial role in the relation between information transmission rates (ITRs) and firing rates is played by a parameter s, which is the sum of transition probabilities from the no-spike state to the spike state and vice versa. We found that in this case too a crucial role is played by the same parameter s. We calculated the maximal and minimal bounds of the quotient of ITRs for these sources. Next, making use of the entropy grouping axiom, we determined the loss of information in a Markov source compared with the information in the corresponding Bernoulli source for a given word length. Our results show that in the case of correlated signals the loss of information is relatively small, and thus temporal codes, which are more energetically efficient, can replace rate codes effectively. These results were confirmed by experiments.
- 650 12
- $a akční potenciály $7 D000200
- 650 12
- $a informační teorie $7 D007257
- 650 12
- $a Markovovy řetězce $7 D008390
- 650 12
- $a modely neurologické $7 D008959
- 650 12
- $a nejistota $7 D035501
- 655 _2
- $a srovnávací studie $7 D003160
- 655 _2
- $a časopisecké články $7 D016428
- 700 1_
- $a Kaplan, Ehud $u Icahn School of Medicine at Mount Sinai, One Gustave Levy Place, New York, NY 10029, USA. Department of Philosophy and History of Science, Faculty of Science, Charles University, Albertov 6, 128 43 Praha 2, Czech Republic. The National Institute of Mental Health, Topolová 748, 250 67 Klecany, Czech Republic.
- 700 1_
- $a Szczepanski, Janusz $u Institute of Fundamental Technological Research, Polish Academy of Sciences, ul. Pawinskiego 5B, 02-106 Warsaw, Poland.
- 773 0_
- $w MED00002342 $t International journal of neural systems $x 1793-6462 $g Roč. 29, č. 8 (2019), s. 1950003
- 856 41
- $u https://pubmed.ncbi.nlm.nih.gov/30841769 $y Pubmed
- 910 __
- $a ABA008 $b sig $c sign $y a $z 0
- 990 __
- $a 20200511 $b ABA008
- 991 __
- $a 20200526134604 $b ABA008
- 999 __
- $a ok $b bmc $g 1525405 $s 1096603
- BAS __
- $a 3
- BAS __
- $a PreBMC
- BMC __
- $a 2019 $b 29 $c 8 $d 1950003 $e 20190114 $i 1793-6462 $m International journal of neural systems $n Int J Neural Syst $x MED00002342
- LZP __
- $a Pubmed-20200511