- Poster presentation
- Open Access
Partial information decomposition as a unified approach to the characterization and design of neural goal functions
BMC Neuroscience volume 16, Article number: P199 (2015)
In many neural systems anatomical motifs are found repeatedly in different places. Despite this repetition these motifs often seem to serve a perplexing variety of functions. A prime example is the canonical microcircuit, which is repeated across multiple cortical areas, but supports a variety of functions from sensory processing and memory to executive functions and motor control. The multiplicity of functions served by a single anatomical motif suggests a common, but more abstract, information processing goal underlying all the different functions. Identifying this goal from neural recordings is a key challenge in understanding the general principles of neural information processing. The apparent diversity of functions makes it clear that this common goal cannot be described using function-specific language (e.g. "edge filters"), but calls for an abstract framework. Here, information theory is the obvious candidate. Notable past approaches using information theoretic descriptions of neural goal functions suggested to maximize the mutual information between input and output , maximize the coherent mutual information that all the inputs share about the output , or, very generally, to minimize the free energy . To facilitate these efforts, and to better dissect the implications of existing neural goal functions, we suggest to build on a recent progress in information theory, termed partial information decomposition (PID). PID allows to measure which of a set of inputs contributes either uniquely, redundantly or synergistically to the output of a (neural) processing unit [4–7], and which fraction of the output's entropy remains unexplained by the input set. We show how these measures can be used to identify an information theoretic footprint of a neural goal function. Most importantly, these measures can quantify how much of the information is modified rather than merely relayed when passing through the neural processor . This shifts the focus from information transmission to more complex processing and allows a much better understanding of the (theoretical?) capabilities of a neuron or neural circuit. Using this approach we show how to better understand existing neural goal functions using PID measures and provide an information theoretic framework for the design of novel goal functions for artificial neural networks.
Linsker R: Self-organization in a perceptual network. Computer. 1988, 21 (3): 105-117.
Kay JW, Phillips WA: Coherent Infomax as a computational goal for neural systems. Bull Math Biol. 2011, 73 (2): 344-372.
Friston K, Kilner J, Harrison L: A free energy principle for the brain. J Physiol Paris. 2006, 100 (1-3): 70-87.
Williams PL, Beer RD: Nonnegative Decomposition of Multivariate Information. ArXiv10042515 Math-Ph Physicsphysics Q-Bio. 2010
Bertschinger N, Rauh J, Olbrich E, Jost J, Ay N: Quantifying Unique Information. Entropy. 2014, 16 (4): 2161-2183.
Griffith V, Koch C: Quantifying Synergistic Mutual Information. Guided Self-Organization: Inception. Edited by: Prokopenko M. 2014, Springer Berlin Heidelberg, 159-190. [Emergence, Complexity and Computation, vol. 9]
Wibral M, Lizier JT, Priesemann V: Bits from Brains for Biologically-Inspired Computing. Frontiers in Robotics and AI. 2015
Lizier JT, Flecker B, Williams PL: Towards a synergy-based approach to measuring information modification. Artificial Life (ALIFE), 2013 IEEE Symposium on. IEEE. 2013, S43-S51.