A Tutorial for Information Theory in Neuroscience
dc.contributor.author | Timme, Nicholas M. | |
dc.contributor.author | Lapish, Christopher | |
dc.contributor.department | Psychology, School of Science | en_US |
dc.date.accessioned | 2019-05-02T15:03:19Z | |
dc.date.available | 2019-05-02T15:03:19Z | |
dc.date.issued | 2018-09-11 | |
dc.description.abstract | Understanding how neural systems integrate, encode, and compute information is central to understanding brain function. Frequently, data from neuroscience experiments are multivariate, the interactions between the variables are nonlinear, and the landscape of hypothesized or possible interactions between variables is extremely broad. Information theory is well suited to address these types of data, as it possesses multivariate analysis tools, it can be applied to many different types of data, it can capture nonlinear interactions, and it does not require assumptions about the structure of the underlying data (i.e., it is model independent). In this article, we walk through the mathematics of information theory along with common logistical problems associated with data type, data binning, data quantity requirements, bias, and significance testing. Next, we analyze models inspired by canonical neuroscience experiments to improve understanding and demonstrate the strengths of information theory analyses. To facilitate the use of information theory analyses, and an understanding of how these analyses are implemented, we also provide a free MATLAB software package that can be applied to a wide range of data from neuroscience experiments, as well as from other fields of study. | en_US |
dc.identifier.citation | Timme, N. M., & Lapish, C. (2018). A Tutorial for Information Theory in Neuroscience. eNeuro, 5(3), ENEURO.0052-18.2018. doi:10.1523/ENEURO.0052-18.2018 | en_US |
dc.identifier.uri | https://hdl.handle.net/1805/19075 | |
dc.language.iso | en_US | en_US |
dc.publisher | Society for Neuroscience | en_US |
dc.relation.isversionof | 10.1523/ENEURO.0052-18.2018 | en_US |
dc.relation.journal | eNeuro | en_US |
dc.rights | Attribution-NonCommercial-NoDerivs 3.0 United States | * |
dc.rights.uri | http://creativecommons.org/licenses/by-nc-nd/3.0/us/ | * |
dc.source | PMC | en_US |
dc.subject | Information flow | en_US |
dc.subject | Information theory | en_US |
dc.subject | Mutual information | en_US |
dc.subject | Neural computation | en_US |
dc.subject | Neural encoding | en_US |
dc.subject | Transfer entropy | en_US |
dc.title | A Tutorial for Information Theory in Neuroscience | en_US |
dc.type | Article | en_US |