The Brain's Connective Core and Its Role in Animal Cognition
The Brain's Connective Core and Its Role in Animal Cognition
The Brain's Connective Core and Its Role in Animal Cognition
Supplementary data
References
http://rstb.royalsocietypublishing.org/content/367/1603/2704.full.html#ref-list-1
Article cited in: http://rstb.royalsocietypublishing.org/content/367/1603/2704.full.html#related-urls
Subject collections
Articles on similar topics can be found in the following collections cognition (332 articles) neuroscience (451 articles)
Receive free email alerts when new articles cite this article - sign up in the box at the top right-hand corner of the article or click here
Research
1. INTRODUCTION The brain of an animal is a massively parallel, distributed control system that attempts to maintain certain metabolic variables within acceptable bounds while fullling a procreative mission. Evolution has solved a very difcult control problem, namely how to effectively exploit all those parallel, distributed resources when responding to a large variety of unpredictable circumstances. Moreover, because the brain of a bee contains fewer than one million neurons while that of a human contains around 100 billion, evolution has found a solution (or solutions) to this problem on a wide range of scales [1]. What are the underlying principles of the solution or solutions that evolution has found? The same question can be recast in more cognitive terms. Let us say that perfect cognitive integration is achieved when the animal brings the totality of what it knows to bear on the ongoing situationits grasp of the sensorimotor contingencies of multiple domains and its understanding of the associated affordances, plus the full contents of both its long-term (episodiclike) memory and its short-term (working) memory. Humans excel at cognitive integration. Yet failures of cognitive integration are commonplace even in humans, such as when I remove the U-bend of my sink then pour its dirty contents back into the plughole, causing an unwelcome deluge. On the other hand, apparent triumphs of cognitive integration have been demonstrated in non-human animals, such
as innovative tool manufacture [2,3]1 and sequential tool use [3,5,6].2 How is cognitive integration achieved in the biological brain? Drawing on the literature of computer science and articial intelligence (AI) as well as that of neuroscience, this paper surveys a range of mechanisms that might underlie some of the most impressive feats of cognitive integration performed by nonhuman animals. The focus is on achievements, such as those just cited, that appear to require combinatorial search of a space of possible sequences of action, or (to put it another way) that show signs of insight in the sense of Epstein [8, p. 627], who denes insightful behaviour as The spontaneous interconnection of two repertoires of behavior which [have] been established separately. (For a contemporary discussion of the concept, see Shettleworth [9].) The paper advocates a mix of serial and parallel processing to explain these accomplishments. However, in the context of a massively parallel neural substrate, the serial processing in question bears little resemblance to that of a conventional computer. Rather, each serial step results from the competitive formation of a (possibly novel) coalition of brain processes drawn from a large repertoire of potential combinations. To facilitate this blend of serial and parallel, it is hypothesized that a certain pattern of connectivity is required, wherein information and inuence funnel into and fan out from a connective core of cognitively important brain regions, resulting in an integrated response to the animals ongoing situation.
*[email protected] Electronic supplementary material is available at http://dx.doi.org/ 10.1098/rstb.2012.0128 or via http://rstb.royalsocietypublishing.org. One contribution of 14 to a Discussion Meeting Issue Animal minds: from computation to evolution.
2. EXPLAINING COGNITIVE PROWESS IN TERMS OF COMPUTATION We will begin with explanations that inherit their terms from classical AI and cognitive science, then move to
This journal is q 2012 The Royal Society
2704
The brains connective core M. Shanahan explanations in the language of neurodynamics. The cognitive feats of interest all feature an element of planningthat is to say, the discovery of a novel sequence of actions to achieve a goal.3 In the sequential tool use paradigm of Bird & Emery [3], a stone has to be dropped into a narrow tube to release a trap-door. But no suitably sized stone is available. Instead, the bird has to rst drop a large stone into a wide tube to release a small stone, which can then be used in the narrow tube. Similarly, in the wire-bending paradigm of Weir et al. [2], the bird has to retrieve a bucket from a tube. But the bucket is out of reach, and a tool is required to retrieve it. Moreover, no suitable ready-made tool is available, and success at the task requires a straight piece of wire rst to be bent into a hook. Success in such tasks only counts as planning if it is achieved on the rst trial, when the subject is unfamiliar with the problem. As soon as the subject has performed the task successfully once, planning is no longer strictly necessary. Moreover, even if the task is performed successfully on rst trial, this cannot be counted as planning if it is the result of extensive trial-and-error and/or luck. These explanations for success have to be ruled out. Of course, an innate behaviour involving sequences of actions, such as nest-building [11], does not necessarily require planning either, however complex it might be. The hallmark of a genuine planning task is the need to search a tree of alternative sequences of actions. The topic of planning has received extensive treatment in the eld of AI [12, ch. 10]; so it is instructive to consider the range of methods that can be deployed for solving planning problems on a computer while asking whether the brains of clever animals might operate in a similar way. Planning algorithms that are predominantly serial rather than parallel can be classied according to the order in which they explore a state space. A backward chaining technique is one that works backwards from a desired goal state, via an action that would achieve that goal state, to a subsidiary goal state. This process is then repeated until a state is reached that matches the current state of the world. The sequence of actions that is built up along the way is a plan. Conversely, a forward chaining technique begins with the current state of the world and works forwards, via an action that can be executed right away, to a successor state. This process is then repeated until a state is reached in which the desired goal is attained. Could it be the case that non-human animals who solve planning problems do so in a way that resembles an AI planner using either backward chaining or forward chaining? If they do, then it is possible in principle to discriminate experimentally between the backward-chaining hypothesis and the forward-chaining hypothesis, using an extension of the experimental setup described by Bird & Emery [3]. Instead of stones and wide and narrow tubes, the extended apparatus uses multiple tubes, each with a cap with a uniquely shaped hole cut in it (see electronic supplementary material, gure S1). Only an object that matches the hole can be used in that particular tube. Using this apparatus, it is possible to set up a variety of planning problems, and, by varying the combinations of caps and objects,
Phil. Trans. R. Soc. B (2012)
2705
to make every new trial effectively a rst trial. In particular, it is possible to devise a pair of planning problems that differ in difculty for a forward chaining method but not for backward chaining, as well as a complementary pair of planning problems that differ in difculty for a backward chaining method but not for a forward chaining method (see electronic supplementary material, gure S2). The range of options explored in AI is not exhausted by the forward-chaining/backward-chaining opposition, however. Planning techniques have also been devised that combine forward and backward chaining. For example, the GraphPlan algorithm interleaves a forward-chaining graph construction phase with a backward-chaining graph search phase [13]. Perhaps, when animals plan, their brains use a hybrid method of this sort. Indeed, if circumstances resembling the setup just described arise in the wild, in which the animal is cued with both an initial state (the sight of the tools) and a goal state (the sight of the food) before it acts, it seems unlikely that either a blind forward search (disregarding the goal state) or a blind backward search (disregarding the initial state) would be favoured by evolution. If an animals brain uses a hybrid forward and backward method, we would not expect to nd a marked difference in performance between the two paired conditions using the apparatus just described. All of the AI methods cited so far work at the level of the so-called primitive actions, that is to say, actions that are immediately executable (under the right conditions), such as picking up a twig or approaching a stone. In hierarchical planning (or hierarchical task network planning), by contrast, a suitable sequence of actions is found by rening a high-level action (such as get food) into a sequence of lower-level sub-actions (such as get a tool then retrieve the bucket), then repeating this process until a plan comprising just primitive actions is reached (such as get the wire, bend the wire, insert the wire into the tube, etc.). Hierarchical organization of behaviour is also implicit in many AI systems, such as those based on the cognitive architectures SOAR or ACT-R [14]. Could there, perhaps, be an element of hierarchical decomposition in the way the brain of an animal arrives at a novel sequence of actions to attain a goal? If this were true, we would again expect no performance differential between the two paired conditions in the proposed sequential tool-use experiment.
3. EXPLAINING COGNITIVE PROWESS IN TERMS OF NEURODYNAMICS The guiding metaphor in the kind of explanation envisaged above is that of classical, serial computation in the von Neumann style. To properly ll out such explanations, an account of how the proposed computational mechanisms are realized in the brain must be supplied. A variety of contemporary proposals seek to provide such an account [15,16]. However, a feature common to all general-purpose planning methods is the need to carry out combinatorial search, and this is where their computational burden chiey lies. The usual way to carry out combinatorial search on a conventional
2706
M. Shanahan The brains connective core examples is the Hopeld net. A Hopeld net is a network of articial neurons with recurrent (feedback) connections, that confer attractor dynamics on it [23,24]. The network is trained to remember a number of patterns (its attractors). Given an input pattern, the trained network converges to a state corresponding to the nearest attractorthat is to say, to the remembered pattern that most closely resembles the input. It can function, therefore, as a form of associative memory. A Hopeld net, like any neural network, is inherently parallel, and each successive state is the product of strictly local computation. In other words, when a neuron is updated, its new state is a function only of the state of its neighbours, the neurons to which it is directly connected. Attractor networks have many practical applications [25], and have been proposed as fundamental building blocks of the brain [24,26, ch. 4; 27,28]. However, our interest here is not in building blocks, not in mere components of the brain. Rather, we are interested in how the brains global dynamics can settle on a pattern of activity that results in effective motor output, given a novel challenge. To see how a global solution can arise from local processing with brain-scale parallelism, we need to look at networks with more complex topologies than a conventional attractor network. Moreover, the brain exhibits much richer dynamics than a conventional attractor network, including complex temporal patterns of spiking, metastability, rhythmic activity and synchronization phenomena. In 4, the challenge of achieving cognitive integration, as exemplied by Bettys wire-bending, is recast as a problem of coalition formation in a dynamical system of this sort.
computer is with the aid of a dynamic data structure such as a stack. A stack is a structure that grows and shrinks over time in so-called last-in-rst-out order, like a pile of plates only the top of which is accessible. In a search application, the stack is used to record branches of the tree of possibilities that remain to be explored. Although a plausible account can be constructed of how the brain might realize a form of serial processing resembling that of a conventional computer, it is less than obvious how such an account could convincingly be extended to describe the neural analogue of a stack. Moreover, to implement a search algorithm on a serial virtual machine realized in a neural substrate would be to squander the brains enormous potential for parallelism. Undoubtedly, there are cognitive processes that are best carried out in a strictly serial order so that the effects of one operation are available for the next, such as mental arithmetic [17] or the inner rehearsal of a sequence of actions [18]. Indeed, for planning problems with minimal search requirements, a few serial steps of inner rehearsal to anticipate the outcome of a two- or three-action plan might be adequate. Perhaps the sequential tool use paradigms of Bird & Emery [3] and Wimpenny et al. [5] fall into this category, because the challenges they involve can be met without departing from a narrow set of possibilities for action that are given by the learned affordances of the apparatus. In computer science terms, the branching factor of the tree of (obvious) possibilities is small, whether it is searched by backward chaining or forward chaining, and the reward can be obtained without moving beyond the obvious. By contrast, what is impressive about Bettys performance in the wire-bending experiment of Weir et al. [2] is that it apparently combines experience from two micro-domains of expertise that were not previously associated, namely the domain of bending pliable materials and the domain of using tools to retrieve food.4 That her brain was able to blend the relevant brain processes together is impressive enough. But even more striking is that her brain assembled the necessary raw ingredients for the blend in the rst place, given no indication of their mutual relevance to the problem at hand (or at beak, rather). The branching factor of the tree of non-obvious possibilities for action is huge, which puts us rmly back in combinatorial territory. Surely this tree of possibilities is not searched by a slow, serial virtual machine realized on a massively parallel substrate. Could it be instead that the underlying massively parallel system can effect the search without sifting through an infeasibly large number of combinatorial possibilities one at a time? Perhaps out of the maelstrom of competitive and co-operative local interactions between populations of neurons, a global network state corresponding to a suitable plan can emerge, coalescing out of the available dynamical elements rather than being the result of a sequence of discrete computational steps. This is a somewhat fanciful sounding description, but it is reminiscent of a wide class of computations that can be characterized in terms of local interactions among parallel elements that collectively converge to a minimal state (an attractor). One of the best known
Phil. Trans. R. Soc. B (2012)
4. THE COALITION FORMATION PROBLEM In computer science, we know how to exploit parallelism in various narrow domains. For example, the simulation of physical phenomena such as the diffusion of smoke in a burning building or the ow of air over a wing is most efciently carried out using parallel computation. Likewise, in computer vision, basic operations like edge detection or optical ow are straightforwardly expressed in terms of parallel algorithms. In these cases, the breakdown of the overall problem into numerous sub-problems that can be carried out independently falls right out of the spatial structure of the problem domain. These so-called embarrassingly parallel problems are solved in the brain too. The visual cortex, for example, exploits neural parallelism in order to detect edges and motion rapidly. But this sort of parallelism is not much use for achieving cognitive integration. When an animal confronts a novel situation, the problem of drawing on the totality of its past experience to come up with a sequence of motor outputs that is effective despite never having been tried before is not one that can be straightforwardly broken down into trivial sub-problems. To see the difculty in its proper context, we need to think of the brain not so much as a set of parallel, distributed computational processes that take input, produce output and transmit data to each other, but
2707
coupling
memory processes body and environment Figure 1. Competitive coalition formation. A coalition of coupled brain processes forms (shown as hatched circles) to the exclusion of rival coalitions, partly under the inuence of internal dynamics (including the preceding coalition) and partly under the inuence of external stimuli. Coalition membership is drawn from a large pool that includes sensory, motor and memory processes. Brain processes are anatomically localized, functionally specialized populations of neurons, but a coalition of brain processes can be anatomically distributed and functionally diverse. The dominant coalition governs the behaviour of the animal until it breaks up and is succeeded by a new coalition.
rather as a collection of parallel, distributed dynamical processes that unfold in continuous (rather than discrete) time and exert continuous inuence on each other (gure 1). This is the character of the wetware that is biologys provision to cognition. Within this wetware substrate, the past experience of an animal is rendered into a set of brain processes that, both individually and in combination, are specialized for particular situations. The ongoing construction, maintenance and expansion of this set of specialists and the increase in specialist expertise that results, are the responsibility of learning in the classical sense [9,29, ch. 4]. But specialist expertise is limited. A hallmark of cognitive sophistication is the capability to move beyond the tried-and-tested, to transcend domainspecic expertise, and thereby to tap the full combinatorial potential of an established sensorimotor repertoire.5 The basic question of cognitive integration, then, is how the elements of an ever-changing, everexpanding repertoire of specialist brain processes can be combined in previously untried coalitions. The question, in other words, is how open-ended coalition formation is possible in a dynamical system like the brain. Implicit in the idea of coalition formation here is the idea of a winning coalition. One coalition has to arise that dominates the dynamics of the brain, shuts out all rivals and dictates the animals behaviour. A winning coalition will be in the ascendant only briey. When events move on, it will be supplanted by a successor. But its victory, albeit temporary, must be complete. An animal whose brain allowed incompatible possibilities for action equal inuence on its nal motor output would not exhibit coherent behaviour. An animal cannot simultaneously move to the left and to the right, for example. It has just one body and this body is spatially conned and subject to various kinematic constraints.
Phil. Trans. R. Soc. B (2012)
So the question of how open-ended coalition formation is possible is really the question of how competitive open-ended coalition formation is possible. The basic question of open-ended coalition formation naturally leads to a number of subsidiary questions. In particular, how is it possible for a new coalition of sensory, motor and memory processes to form that is effective, that results in behaviour that is more benecial to the animal than if it had stuck to tried-and-tested combinations? There is no point in entertaining possibilities that are plain daft, for example. It does no good to an animal confronted with a food item that is just out of reach to initiate a courtship dance. But manipulating materials that resemble tools is promising. In short, candidate members of a successful coalition must be relevant to the situation at hand (or at beak). However, not every combination of relevant processes is effective. In variations of the trap-tube test [31], pulling the plunger one way is prima facie as relevant as pulling it the other way. But one way results in the loss of the food reward, while the other way results in its attainment. The ideal system will freely explore novel coalitions of relevant processes, but will bias the competition between them according to their expected outcomes. So the mechanism of coalition formation is not going to entertain every coalition with equal probability. The probability distribution in question will be shaped by experience. But our present interest is the connectivity and neurodynamics that make open-ended coalition formation possible in the rst place. How is it that a coalition of anatomically distributed brain processes can form whose constitution is not hardwired in advance? Cognitive integration is achieved when the animal is able to draw upon the full battery of its neuronal resources, mixing and matching them as required, to nd an effective, and sometimes innovative, response to unfamiliar situations. A winning coalition, a coalition of brain processes that issues in overt behaviour, will comprise just a small subset of those resources. But to the extent that cognitive integration is achieved, the whole brain will have participated in an open competition to nd that subset.
5. BRAIN CONNECTIVITY When one computer exchanges information with another through the Internet, the information in question is chopped up into packets, and each packet is sent separately to its destination. Packets can arrive in a different order from that in which they were sent (to be re-assembled in the correct order on arrival), and different packets can go via different routes. So the physical connections that underlie this process give away very little about the ow of information in the network they support. But the brain of an animal works very differently. If one brain region is to exercise a direct inuence on another, it must be directly connected to that region. So the pattern of physical connections between brain regions is good evidence of how information and inuence ow around the brain. It makes sense, therefore, to study brain connectivity with a view to understanding how arbitrarily
2708
(a)
(b)
modules (c) (d )
connector hubs
connective core
Figure 2. (a) A small-world network built using the WattsStrogatz procedure [43]. This network lacks modularity. (b) A modular small-world network that more closely resembles the large-scale structural connectivity of an animals brain (but lacks connector hubs). (c) A modular small-world network with connector hubs. A node is designated a connector hub if it plays a signicant role in communication between modules. (d ) The connector hubs, along with their dense interconnections, are designated the connective core of the network.
constituted coalitions of anatomically distributed brain processes form. The large-scale network organization of an animals brain is revealed in several steps [32,33]. First, anatomical investigations yield data about the existence of pathways between brain regions, either using traditional tracing methods or using diffusion-based imaging in vivo. These data are then compiled into a connectivity matrix covering a major subdivision of the animals forebrainthe cerebral cortex, say, or perhaps the entire telencephalon. This is the animals (structural) connectome. The resulting connectivity matrix is then subjected to analysis using the mathematical theory of complex networks (or graphs), which reveals its large-scale organization. The animal that has been most thoroughly investigated in this way is the human [34 37]. But connectivity matrices and network analyses have been carried out for a number of other species, including the fruity [38], the cat [3941], the macaque [40,42] and the pigeon.6 A number of topological features consistently arise in these studies. These include small-world organization and modularity. A sparse network (all brain networks are sparse) is a small world if it has a mean path length comparable to, but a clustering coefcient higher than, an equivalent randomly wired network [43] (gure 2a). A networks mean path length is the average length of the shortest path between any two nodes, while its clustering coefcient is the proportion of a nodes neighbours that neighbour each other averaged over the whole network. A small world network retains a capacity for locally concentrated processing,
Phil. Trans. R. Soc. B (2012)
but facilitates the rapid ow of information around the network as a whole. A network is modular if it can be partitioned into subsets of nodes (modules) that are densely connected internally but only sparsely connected to other subsets [44] (gure 2b). Modules are suggestive of functional specialization. Another consistent feature of brain networks is the presence of topologically signicant hub nodes [40]. Although different authors use slightly different criteria for designating a node as a hub, the key attribute in all cases is high connectivity. A hub node is one that has an unusually large number of connections, or one that lies on an unusually large number of short paths. Hub nodes can be further classied in the context of a networks modularity. A connector hub is one that plays an important role in communication between modules (gure 2c), while a provincial hub is one that plays an important role in communication within a module. The human, macaque and cat connectomes have all been shown to possess a pronounced connective core of topologically central hub nodes [34,37,41,42]. The addition of the pigeon to this list would be signicant because it serves as a prototypical avian species, and it would be surprising indeed if the same network properties were not present in the (larger) forebrains of corvids. We might expect the set of hub nodes that comprises the connective core in the pigeon to be constituted by brain regions that are either homologous to, or functionally analogous to, connective core regions in humans and macaques, such as the hippocampal formation, which is implicated in long-term memory, and the
The brains connective core M. Shanahan nidopallium caudolaterale, a pre-frontal-like region which is implicated in goal-directed action and working memory. The presence in all these species of a connective core that includes a prefrontal functional analogue as well as homologous hippocampal structures would support the hypothesis of convergent evolution of intelligence in apes and birds [45,46], suggesting a common neural substrate for cognition.
2709
6. THE CONNECTIVE CORE HYPOTHESIS What are the specic cognitive implications of the fact that an animals forebrain has the network properties just listed? The contention of this paper is that a brains capacity for cognitive integration depends on the presence within the telencephalon of a pronounced connective core, that is to say a small set of hub nodes that are topologically central to the whole network and richly connected to each other [47, ch. 4]. Its topologically central situation entails that information and inuence can funnel into and fan out from the connective core, which makes it ideally suited to full three functional roles that together promote the availability of the brains full repertoire of process combinations. According to the hypothesis, the connective core is a locus of broadcast, a medium of coupling, and an arena for competition. The connective core is a potential locus of broadcast because the ongoing pattern of activity in the regions comprising it can exercise direct inuence on a large number of other forebrain regions, and indirect inuence on an even larger number via just a few hops in the network. Suppose an animal is presented with an unexpected object (such as a straight piece of wire in a tool manufacture paradigm). The resulting visual stimulus excites its visual areas, which in turn can inuence activity in its connective core. Thanks to the cores rich connections to other regions, this inuence can then be disseminated throughout the forebrain, arousing the widespread activation of multiple neuronal groups [48,49], including those associated with the visible affordances of the object (it can be pushed, picked up, pecked at, used as a prod and so on). To the extent that the connective core fulls this broadcast role, it addresses one aspect of the coalition formation problem, namely how the set of brain processes relevant to the ongoing situation becomes active and is thereby given the opportunity to join a coalition that might eventually take over the animals motor apparatus, while those that are irrelevant are excluded [22]. The relevance or otherwise of a process is not decided centrally, in any sense. Thanks to the connective core, this responsibility is delegated to the processes themselves, which carry it out individually and in parallel. But cutting down the set of active processes to only those that are relevant is only half the problem. Even the pool of relevant brain processes may be large, and the set of possible coalitions constituted by members of that pool is combinatorially larger.
Phil. Trans. R. Soc. B (2012)
One consequence of the combinatorics is the prohibitive quantity of wiring that would be needed if each group of neurons was directly connected to every other group of neurons with which it might enter into a coalition. This is where the second purported role of the connective core comes into play. In much the same way that activity within the connective core can inuence any part of the forebrain, so any part of the forebrain can inuence the connective core and, via the connective core, any other part of the forebrain. So the connective core has the potential to act as a communications infrastructure capable of routing information and inuence between any two brain processes, allowing them to become coupled [47, chs 4 and 5; 50]. Thanks to its convergentdivergent wiring pattern, the connective core allows arbitrary combinations of processes to become coupled without recourse to combinatorial wiring. This addresses the most prominent aspect of the coalition formation problem from the standpoint of cognitive integration, as it allows novel coalitions to arise. However, like any channel of communication, the connective core has only limited bandwidth. Rival coalitions must compete for this bandwidth. In a telephone network, contention for limited bandwidth is resolved equitably. In the early days of telephony, customers on a busy trunk line had to wait until a channel became available, while nowadays they might have to tolerate a reduction in their quality of service. Either way, everyone gets a fair share of the bandwidth. By contrast, contention for access to the connective core is not resolved equitably according to the present hypothesis. Rather, it is resolved through winnertakes-all competition. The coalition that ends up driving activity in the connective core will shut out its competitors by preventing their constituent processes from exchanging inuence and information through it. In this sense, the competition plays out in the connective core itself. It thereby addresses a further aspect of the coalition formation problem, namely the need for a single coalition at a time to dominate the dynamics of the brain and dictate the animals behaviour.7 So far we have been imagining a single transition in the state of the brain. We have examined the putative role played by the connective core in producing the globally integrated brain state that determines an animals response to its current situation. So the emphasis up to now has been on parallel processing. But the distinctive blend of serial and parallel processing supported by the connective core involves many such transitions. Insofar as the connective core has only a limited capacity, it constitutes a central bottleneck in the brain [54,55]. Apart from helping to enforce the sort of winner-takes-all competition that is necessary for an animal to commit to a coherent course of action, this central bottleneck facilitates serial processing, which is necessary for the chaining of mental operations [17]. Thanks to its limited capacity, a serial procession of states emerges via the connective core. Yet, each stateto-state transition in this series is the product of competition and co-operation among massively numerous parallel processes (gure 3). This unconventional model of computation can be characterized in dynamical systems terms. The brain
2710
serial procession of broadcast states punctuated by competition Figure 3. The connective core supports a distinctive blend of serial and parallel processing. The transition from one broadcast state to another results from competition (differently coloured circles) and co-operation (similarly coloured circles) among parallel brain processes (here organized into ve modules). In this case, the broadcast inuence of the green coalition gives rise to a competition among three potential successors (yellow, cyan and magenta). The magenta coalition triumphs, giving rise to the next broadcast state.
settles into global, attractor-like states, mediated by the connective core. But these states, though attractor-like, are only temporary. That is to say, they are not stable but metastable [56 59]. A coalition is held together by its own complex dynamics, which will eventually precipitate its break-up. (A likely feature of this complex dynamics is synchronous oscillation [60 62].) Moreover, in a behaving animal, the brain is subject to external perturbation due to incoming sensory stimulation, which can similarly upset the stability of a coalition. So the overall brain-wide pattern of activity is one of coalition formation, followed by break-up, followed by the formation of a new coalition and so on, yielding a serial procession of global metastable states punctuated by transients. One way to think of the blend of sequential and parallel processing envisioned here is in terms of a serial virtual machine realized on a massively parallel neural substrate [63, pp. 209 226; 64]. However, the processing carried out by this serial virtual machine is unlike that of a conventional von Neumann computer. In the present case, as in a conventional computer, we have states, but those states possess an ongoing internal dynamics that is absent in conventional computation. In the present case, as in a conventional computer, we have transitions from one state to another, but these transitions are here mediated by a rich temporal dynamics that is quite different from the discrete state-to-state transitions found in conventional computation. Inherent in the way the present system operates is the possibility of dealing with a cognitive challenge either through a series of sequential internal steps, as in a conventional AI planning algorithm, or through a form of parallel search resembling the convergence of an attractor network, or indeed through some combination of both methods. However, continuing the comparison with a von Neumann architecture, we might ask what, in the envisaged serial virtual
Phil. Trans. R. Soc. B (2012)
machine, is the analogue of memory in a conventional computer. For the serial machine to carry out anything resembling the sequential processing of an AI problem solving algorithm, it must be possible for certain aspects of the present state to endure and inuence future states. This is accomplished in the framework of this paper in the following way. When the presently dominant coalition of processes breaks up and gives way to a successor, certain members of the outgoing coalition can remain active, retaining information, possibly to join a later dominant coalition. There is ample evidencefrom lesion studies with delayed response tasks, for examplethat mnemonic brain processes of this sort are supported by the prefrontal cortex in mammals, a member of the connective core, and by its analogue in birds (the nidopallium caudolaterale) [65,66]. To summarize, the presence of a small set of topologically central forebrain regions has been established empirically. The potential implications of this nding for understanding animal cognition, and perhaps for understanding animal consciousness, are profound. They are encapsulated by the following slogan: unity from multiplicity, serial from parallel. Unity arises from multiplicity because the connective core ensures that a singular, coherent response to the ongoing situation is integrated from the brains full resources. Serial emerges from parallel because brain-wide information and inuence are channelled through the connective core, which acts as a limited bandwidth processing bottleneck, allowing mental operations to be chained together.
7. SYNTHETIC METHODOLOGY The connective core hypothesis purports to explain how the brains connectivity and dynamics enable certain kinds of sophisticated cognition that have been demonstrated experimentally in non-human animals. How can
2711
(a)
192
W1 W2
W3
128
C1
64
cortical modules
C2
C3
50
100
250
300
Figure 4. (a) The architecture of the spiking neuron model of a global neuronal workspace presented in Shanahan [76]. The workspace nodes and their interconnections are functional analogues of the connective core, according to the hypothesis of the present paper. (Compare gure 2c.) The blue portions were included in the computer simulation. (b) A raster plot of the output from the model, showing the sequential production of reverberating spatial patterns in the connective core (global workspace).
we test this hypothesis and, insofar as it is true, build a prole of the serial and parallel activity involved in the successful performance of different tasks? The behavioural paradigm described earlier is one approach. But to really get at the underlying mechanisms, such behavioural methods need to be supplemented. Although neuroanatomy, imaging and electrophysiological recording are vital sources of relevant data, the focus of this nal section is a complementary but less well-established methodology, namely the construction of computer and robot models. As Braitenberg [67] noted, in his law of uphill analysis and downhill invention, it is often more difcult to infer principles of internal operation from experimental observation than it is to invent mechanisms that produce the same outward behaviour. Invention cannot be a substitute for observation, because more than one mechanism is typically capable of generating a given behaviour. What we seek are computer and robot models that not only generate the required behaviour, but are also consistent with the empirical ndings of neuroscience. A computer program is not a scientic theory, of course, and nor is a working robot. But properly engineered computer programs and robots are built according to principles and specications, and it is these principles and specications, rather than any artefact conforming to them, that contribute to our scientic understanding. A computational model of a neuroscientic hypothesis is a demonstration that the hypothesis is sufciently clear to be implementable. Moreover, the rigours of implementation often reveal unexpected nuances of a hypothesis, as well as extensions and alternatives that might otherwise have been missed. This is the rationale behind the eld of computational neuroscience. But here we are advocating the extension of this synthetic methodology to the embodied case, to the use of robots. This approach has been applied to various brain structures, including the hippocampus [68], the basal ganglia [69] and the amygdala [70], as well as to low-level behaviours
Phil. Trans. R. Soc. B (2012)
such as phonotaxis in invertebrates [71]. (For an overview of recent work, see Krichmar & Wagatsuma [72].) In lieu of detailed experimental data, can the synthetic approach be used to bolster the connective core hypothesis, and the claim that it can explain aspects of animal cognition? Some steps in this direction have already been taken. In Shanahan [18], a cognitive architecture is presented whose key component is a network of articial neurons (the global workspace [48]) that fulls the same role as that imputed to the connective core. It is an arena for competition among rival groups of neurons, each of which attempts to make it settle into a particular pattern of reverberating activation. The winning pattern is broadcast back to the full cohort of competitors, giving rise to a cycle of competition followed by broadcast. Thus, a serial procession of states is produced, but each serial step is the outcome of parallel processing in tens of thousands of articial neurons. This serial process is used to rehearse the consequences of actions before they are carried out. The resulting predictions modulate the salience of possible actions in the context of a winner-takes-all action selection mechanism, which controls a simple simulated robot. The resulting model was used to emulate a classic experiment of Tolman & Gleitman [73] that appeared to show a capacity for look-ahead in rats (see electronic supplementary material, gure S3). The work just described showed that a global neuronal workspacea structure analogous to a connective corecould form the basis of serial computation in a massively parallel setting, and demonstrated its deployment in a control architecture for a robot. However, the model used a form of articial neuron that lacks biological delity. A more realistic computer model of competitive access to a global neuronal workspace was developed by Deheane and colleagues using spiking neurons [74,75]. A complementary spiking model of global neuronal workspace broadcast was later reported in Shanahan [76] (gure 4), and then elaborated by Connor & Shanahan [77]. A related
2712
M. Shanahan The brains connective core unifying the distributed activity of the brains massively parallel resources, the connective core promotes cognitive integration by allowing the formation of novel coalitions of brain processes, and facilitates serial mental operations. However, possession of a connective core is no guarantee of cognitive sophistication. Indeed, the evolutionary pressures that led to this network organization may at rst have been less to do with cognitive prowess than to do with the space and energy costs of connecting large numbers of neurons to each other. It is typically taken for granted that the most parsimonious explanation of an animals behaviour is to be preferred in the absence of evidence ruling it out. Yet it is not always obvious what constitutes parsimony [80]. A simple explanation of a simple behaviour may look like the right choice at the level of a single species. But from a wider scientic perspective, a unied theory that encompassed the full spectrum of animal behaviour would be preferable. The goal of supplying such a theory is likely to be furthered by a better understanding of the underlying neural mechanisms. It may be the case, for example, that behaviours that can be given an associative account and behaviours that seem to demand a cognitive account both emerge from the same underlying network organization. Perhaps the neural underpinnings of insight, far from being mysterious, are prevalent in large centralized nervous systems, thanks to massive parallelism, rich dynamics and a particular network organization.
Thanks to Jackie Chappell, Lucy Cheke, Nicky Clayton, Uri Grodzinski, Onur Gu ntu rku n, Richard Newcombe and Alex Taylor.
model showcasing the possible role of the brains connective core in inter-process communication was described by Zylberberg et al. [50]. While these models contain many of the elements required to instantiate the connective core hypothesis, they are also lacking certain crucial features. In particular, they do not address the issue of coalition formation, which is central to the present paper. On the assumption that elevated levels of synchronous oscillation in spatially separated brain regions are a sign that those regions are operating as a coalition, the model described Shanahan [61] is a step in the right direction. The model is based on oscillators, and is thus dened at a higher level than spiking neurons. Moreover, although its connectivity is modular and small-world, it lacks a connective core. Nevertheless, the model shows how metastable coalitions can form and break up in a modular, small-world network of dynamical elements. Cabral et al. [78] have shown that a similar setup can be used to model human resting state fMRI data, including the default mode network in which the regions of the human connective core are prominent. A more complete implementation of the connective core hypothesis, one that could actually exhibit cognitive integration, would combine the important features of each of the models described. The rst step would be to reverse engineer the basic blueprint of the vertebrate brain, something akin to what Striedter [79] calls the vertebrate brain archetype (after the nineteenth century biologist Richard Owen). This should describe the simplest viable brain that incorporates every structure common to all vertebrates (whether as homologues or functional analogues), and retain those connectivity properties that are similarly shared. The next step would be to construct a model brain according to this blueprint that conforms to the connective core hypothesis. The key features of this model would be: (i) biologically realistic network topology, specically a hierarchically modular small-world network with a connective core of hub nodes, (ii) functional differentiation to include at least sensory, motor, basal gangliar, amygdaloid, hippocampal and prefrontal regions, and (iii) embodiment in a behaving robot endowed with a basic motivational system and associative learning mechanisms. To be consistent with the framework of the present paper, the connective core would be expected to exhibit certain dynamical properties, namely (i) episodes of (metastable) broadcast punctuated by competition and (ii) the formation of coalitions of brain processes drawn from a large, open-ended repertoire. If the proposals of the present paper are correct, then, by scaling up the model in different ways, it ought be possible to replicate the cognitive feats of a range of species, including those that seemingly demonstrate insight.
ENDNOTES
1
8. CONCLUDING REMARKS There is growing evidence that the brains of cognitively sophisticated animals are hierarchically modular, small-world networks with connective cores comprising sets of connector hubs. According to the hypothesis of this paper, possession of a connective core is a prerequisite for sophisticated animal cognition. As well as
Phil. Trans. R. Soc. B (2012)
Four-year-old children consistently fail at the wire-bending task while 8-year-old children typically succeed [4], a result that lends credibility to this paradigm as a benchmark for a distinctive capacity for innovation. 2 These experiments are used here as motivating examples. The aim is not to explain their results per se. Attempting to do this without further development of the theory would most likely result in justso stories. The paper uses examples from avian cognition. For a wider overview of animal tool use, see Seed & Byrne [7]. 3 The term planning is sometimes used in the animal cognition literature to mean planning for the future needs [10]. Here we are using the term in the sense prevalent in articial intelligence and robotics. 4 Accomplishments of this sort evoke the frame problem, in the philosophers sense of that term [19 21]. This is the challenge of explaining how all of an agents beliefs that are relevant to a problem or situation can be brought to bear without incurring an impractical cost to explicitly discriminate relevant from non-relevant beliefs. This is obviously closely related to the issue of cognitive integration. As a solution to the frame problem, the present paper builds on the proposal of Shanahan & Baars [22]. 5 We are again reminded of the frame problem here, as well as Mithens [30] notion of cognitive uidity. 6 The pigeon connectome is the subject of ongoing collaborative work by the present author, Vern Bingman, Toru Shimizu, Martin Wild, and Onur Gu ntu rku n. 7 Competitive mechanisms have long been thought to play an important part in brain dynamics [51]. The need for the brain to settle into an integrated global state that determines a coherent behavioural outcome when faced with competing opportunities or threats has been dubbed the action selection problem [52]. Various structures, notably the basal ganglia, have been hypothesized to address
2713
22
REFERENCES
1 Chittka, L. & Niven, J. 2009 Are bigger brains better? Curr. Biol. 19, R995 R1008. (doi:10.1016/j.cub. 2009.08.023) 2 Weir, A. A. S., Chappell, J. & Kacelnik, A. 2002 Shaping of hooks in New Caledonian crows. Science 297, 981. (doi:10.1126/science.1073433) 3 Bird, C. D. & Emery, N. J. 2009 Insightful problem solving and creative tool modication by captive nontool-using rooks. Proc. Natl Acad. Sci. USA 106, 10 370 10 375. (doi:10.1073/pnas.0901008106) 4 Beck, S. R., Apperly, I. A., Chappell, J., Guthrie, C. & Cutting, N. 2011 Making tools isnt childs play. Cognition 119, 301306. (doi:10.1016/j.cognition.2011.01.003) 5 Wimpenny, J. H., Weir, A. A. S., Clayton, L., Rutz, C. & Kacelnik, A. 2009 Cognitive processes associated with sequential tool use in New Caledonian crows. PLoS ONE 4, e6471. (doi:10.1371/journal.pone.0006471) 6 Taylor, A. H., Eliffe, D., Hunt, G. R. & Gray, R. D. 2010 Complex cognition and behavioural innovation in New Caledonian crows. Proc. R. Soc. B 277, 2637 2643. (doi:10.1098/rspb.2010.0285) 7 Seed, A. M. & Byrne, R. 2011 Animal tool-use. Curr. Biol. 20, R1032R1039. (doi:10.1016/j.cub.2010.09.042) 8 Epstein, R. 1985 Animal cognition as the praxist views it. Neurosci. Biobehav. Rev. 9, 623 630. (doi:10.1016/01497634(85)90009-0) 9 Shettleworth, S. J. 2010 Cognition, evolution, and behavior, 2nd edn. Oxford, UK: Oxford University Press. 10 Raby, C. R., Alexis, D. M., Dickinson, A. & Clayton, N. S. 2007 Planning for the future by western scrubjays. Nature 445, 919 921. (doi:10.1038/nature05575) 11 Hansell, M. 2007 Built by animals: the natural history of animal architecture. Oxford, UK: Oxford University Press. 12 Russell, S. & Norvig, P. 2010 Articial intelligence: a modern approach, 3rd edn. Boston, MA: Pearson. 13 Blum, A. L. & Furst, M. L. 1997 Fast planning through planning graph analysis. Artif. Intell. 90, 281300. (doi:10.1016/S0004-3702(96)00047-1) 14 Langley, P., Laird, J. E. & Rogers, S. 2009 Cognitive architectures: research issues and challenges. Cogn. Syst. Res. 10, 141160. (doi:10.1016/j.cogsys.2006.07.004) 15 Duncan, J. 2010 The multiple-demand (MD) system of the primate brain: mental programs for intelligent behaviour. Trends Cogn. Sci. 14, 172 179. (doi:10.1016/j.tics. 2010.01.004) 16 Zylberberg, A., Dehaene, S., Roelfsema, P. R. & Sigman, M. 2011 The human Turing machine: a neural framework for mental programs. Trends Cogn. Sci. 15, 293 300. 17 Sackur, J. & Dehaene, S. 2009 The cognitive architecture for chaining of two mental operations. Cognition 111, 187 211. (doi:10.1016/j.cognition.2009.01.010) 18 Shanahan, M. P. 2006 A cognitive architecture that combines internal simulation with a global workspace. Conscious. Cogn. 19, 433 449. (doi:10.1016/j.concog. 2005.11.005) 19 Fodor, J. A. 1983 The modularity of mind: an essay on faculty psychology. Cambridge, MA: MIT Press. 20 Dennett, D. 1984 Cognitive wheels: the frame problem in articial intelligence. In Minds, machines and evolution (ed. C. Hookway), pp. 129 151. Cambridge, UK: Cambridge University Press. 21 Shanahan, M. P. 2009 The frame problem. In The Stanford encyclopedia of philosophy (winter 2009 edn),
Phil. Trans. R. Soc. B (2012)
23
24
25
26 27 28
29 30 31
32
33 34
35
36
37
38
39
40
41
(ed. E. N. Zalta). (http://plato.stanford.edu/archives/ win2009/entries/frame-problem/). Shanahan, M. P. & Baars, B. 2005 Applying global workspace theory to the frame problem. Cognition 98, 157 176. (doi:10.1016/j.cognition.2004.11.007) Hopeld, J. 1982 Neural networks and physical systems with emergent collective computational abilities. Proc. Natl Acad. Sci. USA 79, 2554 2558. (doi:10.1073/ pnas.79.8.2554) Amit, D. J. 1989 Modeling brain function: the world of attractor neural networks. Cambridge, UK: Cambridge University Press. Smith, K. 1999 Neural networks for combinatorial optimization: a review of more than a decade of research. INFORMS J. Comput. 11, 1534. (doi:10.1287/ijoc.11.1.15) Freeman, W. J. 1999 How brains make up their minds. London, UK: Weidenfeld & Nicolson. Rolls, E. T. 2009 Attractor networks. Wiley Interdiscip. Rev. Cogn. Sci. 1, 119 134. (doi:10.1002/wcs.1) Lansner, A. 2009 Associative memory models: from the cell-assembly theory to biophysically detailed cortex simulations. Trends Neurosci. 32, 178 186. (doi:10. 1016/j.tins.2008.12.002) Sutton, R. S. & Barto, A. G. 1998 Reinforcement learning: an introduction. Cambridge, MA: MIT Press. Mithen, S. 1996 The prehistory of the mind. London, UK: Thames & Hudson. Seed, A. M., Tebbich, S., Emery, N. J. & Clayton, N. S. 2006 Investigating physical cognition in rooks, Corvus frugilegus. Curr. Biol. 16, 697 701. (doi:10.1016/j.cub. 2006.02.066) Bullmore, E. & Sporns, O. 2009 Complex brain networks: graph theoretical analysis of structural and functional systems. Nat. Rev. Neurosci. 10, 186 198. (doi:10.1038/nrn2575) Sporns, O. 2010 Networks of the brain. Cambridge, MA: MIT Press. Hagmann, P., Cammoun, L., Gigandet, X., Meuli, R., Honey, C. J., Wedeen, C. J. & Sporns, O. 2008 Mapping the structural core of human cerebral cortex. PLoS Biol. 6, e159. (doi:10.1371/journal.pbio.0060159) guez, Iturria-Medina, Y., Sotero, R. C., Canales-Rodr n-Go mez, Y. & Melie-Garc a, L. 2008 E. J., Alema Studying the human brain anatomical network via diffusion-weighted MRI and graph theory. NeuroImage 40, 1064 1076. (doi:10.1016/j.neuroimage.2007.10.060) Gong, G., He, Y., Concha, L., Lebel, C., Gross, D. W., Evans, A. C. & Beaulieu, C. 2009 Mapping anatomical connectivity patterns of human cerebral cortex using in vivo diffusion tensor imaging tractography. Cereb. Cortex 19, 524 536. (doi:10.1093/cercor/bhn102) van den Heuvel, M. P. & Sporns, O. 2011 Rich-club organization of the human connectome. J. Neurosci. 31, 15 77515 786. (doi:10.1523/JNEUROSCI.3539-11.2011) Chiang, A.-S. et al. 2011 Three-dimensional reconstruction of brain-wide wiring networks in Drosophila at single-cell resolution. Curr. Biol. 21, 111. (doi:10.1016/ j.cub.2010.11.056) Scannell, J. W., Burns, G. A. P. C., Hilgetag, G. C., ONeil, M. A. & Young, M. P. 1999 The connectional organization of the cortico-thalamic system of the cat. Cereb. Cortex 9, 277 299. (doi:10.1093/cercor/ 9.3.277) Sporns, O., Honey, C. J. & Ko tter, R. 2007 Identication and classication of hubs in brain networks. PLoS ONE 10, e1049. (doi:10.1371/journal.pone.0001049) pez, G., Zhou, C. & Kurths, J. 2010 Cortical Zamora-Lo hubs form a module for multisensory integration on top of the hierarchy of cortical networks. Front. Neuroinform. 4, 1. (doi:10.3389/neuro.11.001.2010)
2714
42 Modha, D. S. & Singh, R. 2010 Network architecture of the long-distance pathways in the macaque brain. Proc. Natl Acad. Sci. USA 107, 13 48513 490. (doi:10. 1073/pnas.1008054107) 43 Watts, D. J. & Strogatz, S. H. 1998 Collective dynamic of small-world networks. Nature 393, 440 442. (doi:10. 1038/30918) 44 Girvan, M. & Newman, M. E. J. 2002 Community structure in social and biological networks. Proc. Natl Acad. Sci. USA 99, 7821 7826. (doi:10.1073/pnas. 122653799) 45 Emery, N. J. & Clayton, N. S. 2005 Evolution of the avian brain and intelligence. Curr. Biol. 15, R946R950. (doi:10.1016/j.cub.2005.11.029) 46 Seed, A., Emery, N. & Clayton, N. 2009 Intelligence in corvids and apes: a case of convergent evolution? Ethology 115, 401 420. (doi:10.1111/j.1439-0310. 2009.01644.x) 47 Shanahan, M. P. 2010 Embodiment and the inner life: cognition and consciousness in the space of possible minds. Oxford, UK: Oxford University Press. 48 Baars, B. J. 1988 A cognitive theory of consciousness. Cambridge, UK: Cambridge University Pres. 49 Dehaene, S., Changeux, J.-P., Naccache, L., Sackur, J. & Sergent, C. 2006 Conscious, preconscious, and subliminal processing: a testable taxonomy. Trends Cogn. Sci. 10, 204 211. (doi:10.1016/j.tics.2006.03.007) ndez Slezak, D., Roelfsema, P. R., 50 Zylberberg, A., Ferna Dehaene, S. & Sigman, M. 2010 The brains router: a cortical network model of serial processing in the primate brain. PLoS Comput. Biol. 6, e1000765. (doi:10.1371/ journal.pcbi.1000765) 51 Desimone, R. & Duncan, J. 1995 Neural mechanisms of selective visual attention. Annu. Rev. Neurosci. 18, 193 222. 52 Prescott, T. J. 2007 Forced moves or good tricks in design space? Landmarks in the evolution of neural mechanisms for action selection. Adapt. Behav. 15, 931. (doi:10.1177/1059712306076252) 53 Redgrave, P., Prescott, T. J. & Gurney, K. 1999 The basal ganglia: a vertebrate solution to the selection problem. Neuroscience 89, 10091023. (doi:10.1016/S03064522(98)00319-4) 54 Pashler, H. 1984 Processing stages in overlapping tasks: evidence for a central bottleneck. J. Exp. Psychol. Hum. Percept. Perform. 10, 358377. (doi:10.1037/0096-1523. 10.3.358) 55 Marois, R. & Ivanoff, J. 2005 Capacity limits of information processing in the brain. Trends Cogn. Sci. 9, 296 305. (doi:10.1016/j.tics.2005.04.010) 56 Kelso, J. A. S. 1995 Dynamic patterns: the self-organization of brain and behavior. Cambridge, MA: MIT Press. 57 Bressler, S. L. & Kelso, J. A. S. 2001 Cortical coordination dynamics and cognition. Trends Cogn. Sci. 5, 2636. (doi:10.1016/S1364-6613(00)01564-3) 58 Werner, G. 2007 Metastability, criticality, and phase transitions in brain and its models. BioSystems 90, 496 508. (doi:10.1016/j.biosystems.2006.12.001) 59 Deco, G., Jirsa, V. K. & McIntosh, A. R. 2011 Emerging concepts for the dynamical organization of resting-state activity in the brain. Nat. Rev. Neurosci. 12, 43 56. (doi:10.1038/nrn2961) 60 Fries, P. 2009 Neuronal gamma-band synchronization as a fundamental process in cortical communication. Annu. Rev. Neurosci. 32, 209224. (doi:10.1146/annurev.neuro. 051508.135603)