bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. Neurobiology as information physics Sterling Street Independent Researcher Abstract This article reviews the preliminary results of an analysis of the brain that may supplement our current understanding of its structure and function, even if by falsification. This analysis seems to reveal that the information in a system of the brain can be quantified to an appreciable degree of objective precision, that many qualitative properties of the information content of a system of the brain can be inferred by observing matter and energy, and that many general features of the brain’s anatomy and architecture illustrate simple information-energy relationships. It is possible that the brain provides a unique window into the relationship between matter, energy, and physical information. Introduction That information is physical has been suggested by evidence for over a century (S Lloyd 2006). In recent years, Landauer’s principle (R Landauer 1961, CH Bennett et al. 2003), which relates informational entropy to thermodynamic entropy, has been confirmed (A Bérut et al. 2012; Y Jun et al. 2014; AO Orlov et al. 2012; B Piechocinska 2000), and information-energy equivalence (A Alfonso-Faus 2013) has been shown to qualify the existence of thermodynamic demons (S Toyabe et al. 2010; K Maruyama et al. 2009). The finding that information is conserved as event horizon area has led to the tentative resolution of major paradoxes in black hole thermodynamics (L Susskind and J Lindesay 2005; K Bradler and C Adami 2014, SW Hawking 2005), and the holographic principle is currently well-supported by evidence (R Bousso 2002, T Asselmeyer-Maluga 2015). Many classic mysteries of quantum mechanics, including entanglement and double-slit interference, may demonstrate properties of physical information (Č Brukner and A Zeilinger 2003; A Zeilinger 2004), and information-interpretive approaches in quantum gravity research (JW Lee et al. 2013; E Verlinde 2011) could even lead to the unification of quantum mechanics and general relativity. Surprisingly, however, “a formalization of the relationship between information and energy is currently lacking in neuroscience” (G Collell and J Fauquet 2015). The purpose of this article is to explore a few different sides of this relationship and, along the way, to suggest that many competing models and theories in neuroscience can be unified through the physics of information. bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. I. Information bounds “How can the events in space and time which take place within the spatial boundary of a living organism be accounted for by physics and chemistry?” – (E Schrödinger 1944) Information in any physical system is finite, and its flow is inevitably tied to thermodynamics. As a result, the information content of any system in the brain at any level of structure can be quantified by applying a modified form of the universal (JD Bekenstein 1981, 1984, 2001, 2005; J Oppenheim 2015) information-entropy bound: Isys = ζ √ ħ where I is the quantity of units of information in a physical system, A is area in m2, E is energy in joules,ħ is the sys reduced Planck constant, c is the speed of light in meters per second, and ζ is a factor of information saturation such that 0 ≤ ζ ≤ 1 In short, a system approaches its information bound as it approaches a state of maximum information saturation. In other words, the bound quantifies the absolute information present in a system, while the portion of this information available for inference (available for performing logical operations or receiving in an influence network node or observatorial apparatus) is necessarily observer-dependent and generally far lower than quantity given by the bound. It is critical to make this distinction; a number of researchers have used the bound to estimate the information capacity of the brain as a hypothetical upper limit rather than as an absolute quantity of information, and a central argument among their critics is that the bound overestimates the quantity of information that is functionally relevant in cognition. Further, in the case that quantum effects play a meaningful role in cognition, this bound includes both classical information in bits and quantum information in qubits. The bound quantifies units of information of any base or classification and varies only with the area and energy content of a system. The factor ζ partitions absolute information from observer-dependent information by separating entropic information from relative information. In other words, while absolute saturation (ζ = 1) specifies the absolute quantity of information present in a thermodynamic universe, relative saturation (0 ≤ ζ ≤ 1) specifies the quantity of absolute information that is available for influence into a referential system. Since entropy is hidden information (C Adami 2011; L Susskind 2008; S Lloyd 2006) with respect to a referential system, a system transferring none of its information to a referential system (resulting in maximal KL entropy) carries an entropic saturation factor of 1 and an informational saturation factor of 0. On the other hand, a system transferring the entirety of its information to a referential system (resulting in minimal KL entropy) carries an entropic saturation factor of 0 and an informational saturation factor of 1. The flow of information from an index to a referential system would parallel the crossing of Markov boundaries (KJ Friston 2013) and Fisher information spaces (BR Frieden 2015). bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. II. Stratification “If information forms the most basic layer of reality, what does it refer to?” – (A Aguirre et al. 2015) Consider the information processed by the four distinct bases of DNA. The quantity of information contained in a single base molecule includes the information in its constituent atoms and subatomic particles, all the way down to the quantum information evaluated in their spin states, angular momenta, vibrational quanta, and so forth. How is it that we can ignore this information when analyzing base pair sequences? Although the absolute information in a single DNA base far exceeds the relative information it transfers to its referents in a process such as protein synthesis, DNA lies within an effectively base-four layer of influence – the alternative evaluation of a qubit that determines the spin of a proton in an adenine molecule is very unlikely to alter the effective structure of adenine, much less a resultant base pair sequence or the structure of a protein, cell, or organism. So, information is stratified between layers of influence: only the value of a nucleotide’s macrostate is informative relative to a referential amino acid. On the other hand, while binary operations in an electronic device are nominally performed in a base-two layer of influence, a single bit of information registered by the presence or absence of electrons in a capacitor also contains the information within these electrons, and even in the atoms and subatomic particles of the capacitor itself. The absolute quantity of information processed by the entire device over a period of time would similarly include all of the information in any of its structural metals and plastics, in the photons that exit the screen, in the information lost as heat to the surroundings, and in the photons, air molecules, and external systems that interact with the device. The absolute information processed by the classical computing device clearly exceeds the quantity of information used in performing electronic computations. This stratification is arguably what allows the brain to contain a far greater quantity of relative information than an electronic device with equivalent physical dimensions. Some other properties of information should be noted in the context of stratification. Each degree of freedom in a physical system is essentially an informational influence space, so that the realization of one state or elimination of one degree of freedom (i.e., the induction of many microstates into a single macrostate, the occurrence of a probabilistic event over its complement, the transition from a prior to a posterior probability, the collapse of a wavefunction, etc.) registers a single unit of physical information. It is for this reason that the finite number of quantum states imposed on any physical system ensures its containing and processing of finite information (S Lloyd 2006; C Rovelli 2015), that storing many degrees of freedom in one computational macrostate confers greater statistical stability (S Lloyd 2000), that the occurrence of an unlikely physical event registers more Shannon information, and that a relative quantity of information is measurable as entropy (both as H = log N and S = k log W). In any case, the 2 e information in any layer of influence can be quantified even if the content of its internal evaluation is unattainable by an external system. bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. III. Information-energy equivalence “It is important to realize that in physics today, we have no knowledge what energy is.” – (RP Feynman 1964) Because a state of maximal internal energy corresponds to a state of maximal information saturation (TL Duncan et al. 2004), any transitions between energy levels occur as transitions between informational extrema. These transitions enclose a finite quantity of information, so the flow of information through a referential system is followed by an energy change: = where k is Boltzmann’s constant, E is energy in joules, and T is ambient temperature in kelvin Reintroducing the universal information-entropy bound as Iuniv = ζu √ ħ where A is the area of a thermodynamic universe in meters squared, E is energy in joules and ζu = 1 now allows us to partition the universal information saturation factor into a relative information component (ζi = 1 – ζs ) and an entropic information component (ζs = 1 – ζi ), Isys = ζi = (1 – ζs ) √ √ so that, in the case that information enters aħ ma croscopic phyħsica l system with an effectively constant mass and surface area, we can write ΔI = Δζ = ΔE sys i sys and, in the case that information exits a system as heat, ΔS = Δζ = ΔE sys s surr Various forms of similar relationships have been considered in neuroscience (P Sterling and S Laughlin 2015). This article simply proposes their restatement as fundamental physical relationships in the brain requiring further investigation and critique. On a side note, these relationships also seem to be consistent with the possibility that the quantum of gravity is (roughly) the energy of one bit of entropy (A Alfonso-Faus 2011), the possibility that the Planck constant is the minimal unit of information (C Rovelli 2015), and the possibility that the asymmetry of time is an entropy gradient (L Mlodinow and TA Brun 2014). bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. IV. An optimization principle "To heat also are due the vast movements which take place on the earth. It causes the agitations of the atmosphere, the ascension of clouds, the fall of rain and of meteors, the currents of water which channel the globe… It is necessary to establish principles applicable not only to steam engines but to all imaginable heat engines, whatever the working substance and whatever the method by which it is operated." — (S Carnot 1824) The observation that biological systems use a steady influx of energy to maintain order and structural integrity seems to show that, since entropy can be considered a referent-dependent measure of inaccessible information, the continuous influence of energy through the information bound of a biological system necessarily decreases entropy and increases internal energy. Since the maximization of internal energy and the maximization of relative information occur as coupled, mutually-reinforcing processes – and both processes offset the transition into equilibrium – living systems can be seen as thermodynamic demons that locally maximize information and internal energy. It has indeed been suggested that there should be no distinction made between formalized information and physical information in biology (C Adami 2004), that defining biological systems in terms of their management of information may condense a great deal of experimental evidence into a small number of theoretical models (S Brenner 2012, PM Binder et al. 2011), and that the development of a unified theory of biology may require the application of information thermodynamics (DR Brooks et al. 1989, BH Weber et al. 1989, L Demetrius 2000, LM Martyushev et al. 2006, ED Schneider et al. 2006, JL England 2013, DF Styer 2008). It appears that, as a general trend, living systems conform to a principle of information-energy maximization, or optimization. Interestingly, because this maximization would lead to the minimization of informational free energy, this principle is describing one side of the same relationship represented by the variational free energy principle (KJ Friston 2010, B Sengupta et al. 2013). In other words, this is simply an alternative way of considering the free energy principle as an informational principle of least action. bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. V. Neurobiology “… classical thermodynamics… is the only physical theory of a universal content concerning which I am convinced that, within the framework of applicability of its basic concepts, will never be overthrown.” – (A Einstein 1949, from JD Bekenstein 2001) Neurons Neuron activity is, to some extent, reducible to chemical and electrical activity, and such activity is necessarily accompanied by an exchange of energy that follows the transfer of information between systems. The occurrence of depolarization as a transition to E and the maintenance of max a resting state as E may reflect coupled information-energy dynamics, and the maintenance of min a negative resting membrane potential may show that depolarization requires the input of external information-energy. In addition, while the graded potential and analog influence is clearly important, a quintessential feature of neurons is their propagation of all-or-nothing action potentials. This feature seems to mirror the discrete, granular all-or-nothing mathematical structure of physical information. Although information is processed in the activity of the ions and neurotransmitters that generate an EPSP or IPSP, even at subthreshold levels, any transition into one state from two or more alternatives or degrees of freedom could be interpreted as carrying a single unit of information regardless of underlying dynamics. Further, that spatial and temporal summation of numerous dendritic spikes results in the propagation of a single axonal action potential – one macrostate – even suggests a compression algorithm. This operation may represent an analog of feature binding at the level of the neuron, particularly suitable for the layer III and IV pyramidal and stellate cells whose projections connect higher cortical areas. Additionally, the observation that the information in a sequence of action potentials is carried only by spike count and temporal spacing (E Kandel et. al 2000) may show that a frequency distribution code is partly an error-minimizing redundancy feature of neuronal information processing mechanisms. That signal strength increases linearly with frequency and is solely determined by frequency also seems to suggest that spikes represent discrete quantities of physical information. The possibility that bursting activity and repeated spiking responses may have evolved partly to allow error correction by repetition – perhaps the simplest and most rudimentary error correction algorithm, present everywhere from the organization of DNA (V Vedral 2010) to the structure of speech (J MacCormick 2012) – is supported by the observation that signal-to-noise ratios tend to increase with signal repetition (W Bialek and F Rieke 1992). Networks The general features of the various network architectures in the brain also appear to show information-energy relationships. Networks naturally stabilize by seeking energy minima, and the various attractor states in one of these networks define the geometry of its energy landscape (JJ Hopfield 1982). As a result, transitions into spontaneous states follow transitions into information-energy maxima, and such transitions have been associated with the generation of an informational entity such as a memory or a decision (G Deco et al. 2009). In this way, the local energy basins of attractor networks may be closely analogous to the resting potentials of neurons and receptors; a transition between any two distinctive energy states follows the passage of a bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. finite quantity of information. Even in the case of a massively integrated network managing a great deal of complex information, the observation that any energy landscape contains a finite number of basins at any moment may show that each energy transition in such a network carries a finite quantity of information. This architecture may even parallel the action potential as an mechanism for promoting stability (in the sense of minimizing prediction errors) simply by using many degrees of neuronal freedom to effect the transition into one macrostate from a large number of microstates. In other words, where the degrees of freedom governing the dynamics of a neuron rest on the degrees of freedom of its structural molecules, neurotransmitters, and ions, the degrees of freedom of a network rest (emergently) on the degrees of freedom in underlying neurons and their possible states and configurations. Of course, a network in the brain is still prone to chaos and nonlinear behavior, and the function of a single neuron could result in an altered network macrostate. For the very reason that this is true, a vastly connected network whose behavior is governed by the activity of large assemblies of integrated neurons is minimally likely to be altered by the macrostate of an individual subsystem (e.g., a single potassium ion) at a lower level of stratification. That a network’s number of degrees of freedom is limited by connection density is also suggested by the observation that vocabulary size and memory capacity appear to correspond to a network’s number of connections (ET Rolls 2012). If entities such as words and memories are considered to be units of information, there is clearly a correspondence between a network’s number of connections and its number of effective degrees of freedom as possible units of information. Network competition may also restate the importance of increasing outcome stability by maximizing conditional network responses, since inhibiting more weakly activated surrounding networks would reduce the overall number of degrees of freedom of a network at a higher level of architecture, and for the brain as a whole. Inhibition It is possible that inhibition only nominally reduces information processing in the brain. Since maintaining proper membrane potentials, energy gradients, neurotransmitter and signaling protein concentrations requires the expenditure of internal energy, it is advantageous for the brain to process only a minimal quantity of environmental information necessary for survival. For this reason, it is possible that inhibition in the brain still results in optimization, both by the brain itself and the organism as a whole. This possibility even seems to align with some very general observations at the level of behavior and cognition. For instance, the ability of a verbal command to rapidly inhibit the flow of information through non-salient networks (F Grabenhorst and ET Rolls 2008) may show that the affective nature of language parallels inhibitory neuromodulation: minimizing the expenditure of internal energy at any level of structure seems to maximize the internal information content of each influence node. In the case of social networks, integration enhances the efficiency of inhibition: the peripheral mediation of another’s inner state is clearly advantageous for a function such as calling attention to a stimulus in order to coordinate a group response, and the rapid evaluation of another’s inner state is inversely beneficial for the same reasons. The ability of disintegrated information flow to prevent the emergence of an optimized influence network may be further revealed by the observation that the maintenance of a stable internal state requires constant inhibitory modulation (K Lehmann et al. 2012), as well as the observation that large-scale network hyperexcitation can lead to the bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. rapid loss of consciousness. The “attentional spotlight” of awareness noted by Global Workspace Theory (BJ Baars 2005) may even describe a form of inhibition mediated largely by frontal and parietal network competition. In the case of volitional attention, the willful selection of a salient perceptual stream (i.e., through a chosen communication channel) requires the activation of large assemblies of inhibitory interneurons. Learning Lowering the firing threshold in order to increase the likelihood of individual cellular response to a stimulus (DO Hebb 1952) may optimize the activation of an engram by minimizing the quantity of energy needed to elicit maximal influence from its connections. Interestingly, the influence of a minimal energetic stimulus in an optimally integrated recurrent network has even been reported to enable a full energetic activation and an effectively perfect recall of a memory (ET Rolls 2012). For this reason, it is possible that a learned engram is not the storage of information as a dormant memory or response per se but rather a mechanism allowing the temporal convergence of otherwise separate units of information upon the engram’s future activation. It is also possible that neurons tend to form composite systems only if the internal energy and information management capacity of the composite system is likely to exceed that of the individual neurons in their prior state of separation. This possibility seems to be supported by the general molecular and cellular energetics of learning and plasticity. The minimization of free energy achieved through learning (KJ Friston 2010) would then maximize internal information, and the continual formation and disassembly of engrams during learning and forgetting would optimize the growth and pruning of a network in response to external states. Noise Any physical system whose state values are coupled to the exchange of energy and the transfer of information is inherently noisy, even if effectively non-stochastic and roughly predictable. Noise does appear to be an innate and unavoidable feature of information processing in the brain (ET Rolls 2012), scaling in proportion to influence network size (K Josic et al. 2009) and varying inversely with the coupling of network components (DJ Mar et al. 1999; B Kia et al. 2015). Even in the absence of any potential forms of stochastic resonance (F Moss et al. 2004), the noise-driven exploration of various state spaces could enhance the ability of a system in the brain to reach an optimal information-energy extremum. Interestingly, noise has been found to enhance learning and serve beneficial functions (MD McDonnell and LM Ward 2011); the finding that double-stranded DNA breaking (E Suberbielle et al. 2013) and reversible methylation (P Tognini et al. 2015) both occur as normal, everyday processes in the living brain could perhaps be explained by considering that any stochastically optimized structure would be more likely to survive over time as a strengthened connection or increased integration in a modifiable network. So, noise could enhance brain function in much the same way that noise at the level of the genome (i.e., recombination) enhances adaptation and evolution of species (JV Stone 2015; S Lloyd 2006). In other words, it possible that noise in the brain can be interpreted as informational entropy, which necessarily decreases when any systems integrate, couple their influence with mutual information, and jointly reduce KL and thermodynamic entropy. The signal-to-noise ratio could then be seen as a measurement of relative information. bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. Brain The brain’s dependence on energy (P Sterling and S Laughlin 2015) is readily apparent in its disproportionate consumption of energy substrates (e.g., oxygen, glucose, and ATP), in its vulnerability to hypoxic-ischemic damage, and in the rapid loss of consciousness conferred by the onset of an energy restriction. It is also interesting to consider that the information content of a large-scale system in the brain can change while mass and surface area remain effectively constant, and that the brain consumes a greater quantity of energy expected from its mass alone (ME Raichle and DA Gusnard 2002). These findings may even relate to the tradeoff of these parameters predicted by the information bound: a change in the information content of a system with constant mass and surface area must produce a change in internal energy. All fMRI and PET interpretation rests on the assumption that changes in the information content of a structure can be inferred by observing internal energy changes, and it is well known that the information processing capacities of systems in the brain are limited by energy supply. The living brain appears to dynamically maintain a state of optimization, while brain death could be said to occur as a rapid transition into a local information-energy minimum. bioRxiv preprint first posted online Jun. 24, 2016; doi: http://dx.doi.org/10.1101/060467. The copyright holder for this preprint (which was not peer-reviewed) is the author/funder, who has granted bioRxiv a license to display the preprint in perpetuity. It is made available under a CC-BY 4.0 International license. VI. Consciousness and free will “… science appears completely to lose from sight the large and general questions; but all the more splendid is the success when, groping in the thicket of special questions, we suddenly find a small opening that allows a hitherto undreamt of outlook on the whole.” – (L Boltzmann 1886) Unity At any given moment, awareness is experienced as a unified whole. It is possible that physical information is the substrate of consciousness (A Annila 2016) and it appears that any unit of information – the bit, trit, nat, qubit, and so forth – is necessarily transferred into a referential system as a discrete and temporally unitary quantity. In other words, it is possible that the passage of time itself occurs secondarily to the transfer of information, and that the information present in any system at any time is always cohesive and temporally unified. From this perspective, the dilation or contraction of time would vary as a rate of information influence and entropy generation – the flow of absolute information into a relative information space would increase the internal energy of a referential demon by kT J for every state observed, while the inverse would occur with the flow of relative information into an absolute information space (i.e., the generation of entropy). The possible requirement for any information to be unified at its time of influence into a referential system of the brain may even originate in fundamental conservation laws; any quantity of information is fully conserved during its course of influence, and the influence of unified quantities of information necessarily occurs alongside the influence of conserved quantities of energy. Although it is possible that all physical systems exchange information-energy in temporally unitary quantities, it is likely that many of the familiar features of unity in phenomenological human consciousness require the structure and function of neural networks in the brain. It appears that this unity, as an emergent informational structure (F Dretske 1981) requires an integration of thalamocortical and localized neocortical influence (S Dehaene et al. 2011; C Koch and S Greenfield 2007; S Greenfield and TFT Collins 2005). It is also possible that an emergent informational structure in the claustrum generates some of this experiential unity (FC Crick and C Koch 2005, MZ Koubeissi et al. 2014). However, it has been reported that complete unilateral resection of the claustrum performed in patients with neoplastic lesions of the region appears not produce any significant change in subjective awareness or observable functionality (Duffau et al. 2007). It appears unlikely that the flow of information into any isolated or compartmentalized network of the brain is responsible for producing all aspects of the unified nature of conscious experience. Complexity The subjective complexity of conscious experience may reveal that extensive influence network integration is needed for maximizing mutual information and internal energy (JS Torday et al. 2016). An exemplary structure enabling such integration, likely one of many that account for the sensory complexity of consciousness, is the cluster of projections that form the internal capsule. A structure such as this may show that, at any given moment in the influence space of a living brain, a wide range of integrated referential networks are sharing mutual sources of sensory information. This pattern of structure may reveal that the perceptual depth and complexity of
Description: