1. Introduction
Any classical physical system (by which we simply mean any deterministicfunction) can be taken as a measuring apparatus or input/output device.For example, a thermometer takes inputs from the atmosphere and outputsnumbers on a digital display. The thermometer categorizes inputs bytemperature and is blind to, say, differences
in air pressure.
Classical measurements are formalized as follows:
Definition 1.
Given a classical physical system with state space , ameasuring device is a function .The output is the reading and the pre-image is the measurement.
From this point of view a thermometer and a barometer are two functions, and , mapping the state space of configurations (positions and momenta) of atmospheric particlesto real numbers. When the thermometer outputs , it specifies thatthe atmospheric configuration was in the pre-image which,assuming the thermometer perfectly measures temperature, is exactlycharacterized as atmospheric configurations with temperature .Similarly, the pre-images generated by the barometer group atmosphericconfigurations by pressure.
The classical definition of measurement takes a thermometer as a monolithicobject described by a single function from atmospheric configurations toreal numbers. The internal structure of the thermometer – that is composedof countless atoms and molecules arranged in an extremely specific manner– is swept under the carpet (or, rather, into the function).
This paper investigates the structure of measurements performed bydistributed systems. We do so by adapting Definition 1to a large class of systems that contains networks of Boolean functions[10], Conway’s game of life [7, berlekamp:82]and Hopfield networks [9, 2] as special cases.
Our motivation comes from prior work investigating information processingin discrete neural networks [4, 5]. The brain can bethought of as an enormously complicated measuring device mapping sensory states and prior brain states to subsequent brain states. Analyzing the functional dependenciesimplicit in cortical computations reduces to analyzing how the measurementsperformed by the brain are composed out of submeasurements by subdevicessuch as individual neurons and neuronal assemblies. The cortex is ofparticular interest since it seemingly effortlessly integrates diversecontextual data into a unified gestalt that determines behavior. Themeasurements performed by different neurons appear to interact in sucha way that they generate more information jointly than separately. Toimprove our understanding of how the cortex integrates information we needto a formal language for analyzing how context affects measurements indistributed systems.
As a first step in this direction, we develop methods for analyzing thegeometry of measurements performed by functions with overlapping domains.We propose, roughly speaking, to study context-dependence in terms of thegeometry of intersecting pre-images. However, since we wish to work withboth probabilistic and deterministic systems, things are a bit morecomplicated.
We sketch the contents of the paper. Section §2 (http://planetmath.org/2stochasticmaps)lays the groundwork by introducing the category
of stochastic maps .Our goal is to study finite set
valued functions and conditional probabilitydistributions on finite sets. However, rather than work with sets, functionsand conditional distributions, we prefer to study stochastic maps (Markovmatrices) between function spaces on sets. We therefore introduce thefaithful functor
taking functions on sets to Markov matrices:
where is functions from to . Conditional probabilitydistributions can also be represented using stochastic maps.
Working with linear operators instead of set-valued functions is convenientfor two reasons. First, it unifies the deterministic and probabilistic casesin a single language. Second, the dual of a stochastic map provides a symmetric
treatment of functions and their corresponding inverseimage
functions. Recall the inverse
of function is, which takes values in thepowerset of , rather than itself. Dualizing a stochastic mapflips the domain and range of the original map, without introducing any newobjects:
(1) |
see Corollary 2 (http://planetmath.org/2stochasticmaps#Thmthm2)
Section §3 (http://planetmath.org/3distributeddynamicalsystems)introduces distributed dynamical systems. These extendprobabilistic cellular automata by replacing cells (spacecoordinates) with occasions (spacetime coordinates: cell at time ). Inspired by [8, 1], wetreat distributed systems as collections
of stochastic mapsbetween function spaces so that processes (stochastic maps)take center stage, rather than their outputs. framework bares a formal resemblance to the categoricalapproach to quantum mechanics developed in [1].Although the setting is abstract, it has the advantage that itis scalable: using a coarse-graining procedureintroduced in [3] we can analyze distributedsystems at any spatiotemporal granularity.
Distributed dynamical systems provide a rich class of toyuniverses. However, since these toy universes do not containconscious observers we confront Bell’s problem [6]:“What exactly qualifies some physical [system] to play therole of ‘measurer’?” In our setting, where we do not have toworry about collapsing wave-functions or the distinctionbetween macroscopic and microscopic processes, the solution issimple: every physical system plays the role ofmeasurer. More precisely, we track measurers via thecategory of subsystems of . Eachsubsystem is equipped with a mechanism which is constructed by gluing together the mechanisms of theoccasions in and averaging over extrinsic noise.
Measuring devices are typically analyzed by varying theirinputs and observing the effect on their outputs. By contrastthis paper fixes the output and varies the device overall its subdevices to obtain a family of submeasurementsparametrized by all subsystems in . The internalstructure of the measurement performed by is thenstudied by comparing submeasurements.
We keep track of submeasurements by observing that they aresections of a suitably defined presheaf. Sheaf theory providesa powerful machinery for analyzing relationships betweenobjects and subobjects [11], which we adapt toour setting by introducing the structure presheaf ,a contravariant functor
from to the categoryof measuring devices on . Importantly, is not a sheaf: although the gluing axiomholds, uniqueness fails, see Theorem 4 (http://planetmath.org/3distributeddynamicalsystems#Thmthm4).This is because the restriction
operator in is(essentially) marginalization, and of course there areinfinitely many joint distributions
that yieldmarginals and .
Section §4 (http://planetmath.org/4measurement)adapts Definition 1 to distributed systems andintroduces the simplest quantity associated with a measurement:effective information, which quantifies its precision, seeProposition 5 (http://planetmath.org/4measurement#Thmthm5).Crucially, effective information is context-dependent –it is computed relative to a baseline which may be completelyuninformative (the so-called null system) or provided by asubsystem.
Finally entanglement, introduced in §5 (http://planetmath.org/5entanglement),quantifies the obstruction (in bits) to decomposing ameasurement into independent submeasurements. It turns out,see discussion after Theorem 10 (http://planetmath.org/5entanglement#Thmthm10),that entanglement quantifies the extent to which a measurementis context-dependent – the extent to which contextualinformation provided by one submeasurement is useful inunderstanding another. Theorem 9 (http://planetmath.org/5entanglement#Thmthm9)shows that a measurement is more precise than the sum of itssubmeasurements only if entanglement is non-zero.Precision is thus inextricably bound to context-dependenceand indecomposability. The failure of unique descent is thusa feature, not a bug, since it provides “elbow room” tobuild measuring devices that are not products
ofsubdevices.
Space constraints prevent us from providing concrete examples;the interested reader can find these in [4, 5, 3]. Our running examples are the deterministicset-valued functions
which we use to illustrate the concepts as they are developed.
References
- 1 Samson Abramsky & Bob Coecke(2009): Categorical QuantumMechanics. In K Engesser, D M Gabbay &D Lehmann, editors: Handbook ofQuantum Logic and Quantum Structures: Quantum Logic,Elsevier.
- 2 DJ Amit (1989):Modelling brain function: the world of attractor neuralnetworks. Cambridge University Press.
- 3 David Balduzzi (2011):Detecting emergent processes in cellular automata withexcess information. preprint .
- 4 David Balduzzi & Giulio Tononi(2008): Integrated Information inDiscrete Dynamical Systems: Motivation and TheoreticalFramework. PLoS Comput Biol4(6), p. e1000091,doi:10.1371/journal.pcbi.1000091.
- 5 David Balduzzi & Giulio Tononi(2009): Qualia: the geometry ofintegrated information. PLoS Comput Biol5(8), p. e1000462,doi:10.1371/journal.pcbi.1000462.
- 6 J S Bell (1990):Against ‘Measurement’. Physics World August, pp.33–40.
- 7 Martin Gardner (1970):Mathematical Games - The Fantastic Combinations
of John Conway’s New Solitaire Game, Life. Scientific American 223,pp. 120–123.
- 8 G ’t Hooft (1999):Quantum gravity as a dissipative deterministicsystem. Classical and Quantum Gravity16(10).
- 9 JJ Hopfield (1982):Neural networks and physical systems with emergentcomputational properties. Proc. Nat. Acad. Sci. 79,pp. 2554–2558.
- 10 Stuart Kauffman, Carsten Peterson,Björn Samuelsson & Carl Troein(2003): Random Boolean network modelsand the yeast transcriptional network. Proc Natl Acad Sci U S A100(25), pp. 14796–9,doi:10.1073/pnas.2036429100.
- 11 S MacLane & Ieke Moerdijk(1992): Sheaves in Geometry andLogic: A First Introduction to Topos Theory. Springer.