Emergent Intelligence: The Brain as a Complex Adaptive System December 4, 2013 1 Complex Systems At various points in our young lives, we come to the realization that neither our parents nor our teachers know everything, that people in charge make all kinds of mistakes, and that many powerful entities decidedly do not have the best interest of the people nor the planet at heart. In short, we learn that the world is not a well-oiled machine. For the most part, things normally just work out, though no laws can guarantee it. For sure, there are plenty of laws, but there is no master plan to maximize much of anything: happiness, economic growth, traffic flow, environmental sustainability, etc. In reality, the world works due to a complex combination of imposed constraints (e.g. traffic laws) and individual interactions (e.g. drivers swerving to avoid those who occasionally forget them). From all of this, large-scale patterns and systems emerge, ones that are rarely considered optimal or well-oiled, but that generally work well for at least a sizeable fraction of the people. The eye-opening fact that the world works without excessive global manipulation, and without many guarantees, may initially disappoint us, but the understanding of how predominantly local interactions produce sophisticated global results can turn that childhood dismay into adult fascination. The field of complex systems or complex adaptive systems (CAS) specializes in these intriguing feats of self-organization in a variety of systems as diverse as the planet itself: rainforests, termite mounds, traffic jams, epidemics, global economies, climate, the internet, immune systems, politics, migration patterns, stampedes, stock markets, bacterial colonies, retail distribution chains, brains, languages, metabolic pathways, and ancient civilizations. As described by various authors [11, 7, 8], complex systems typically consist of many interacting components (or agents) and exhibit several of the following properties: 1. Distributed control - nothing oversees or manages the components and their interactions. 2. Synergistic - The sophistication or complexity of the gestalt greatly exceeds the summed complexities of the components. 3. Emergent (Self-Organizing) - The global structure arises solely from the local interactions, often via positive feedback. 4. Autopoietic - The global structure is maintained by local interactions, often via negative feedback. 5. Dissipative - The system is far from equilibrium (and thus in a low entropy state), yet stable. 6. Adaptive - The system can modify its own structure and/or behavior in order to maintain other aspects of itself. 1 Precise definitions of complexity abound, but have little formal agreement. Within particular disciplines, such as computer science or physics, formalizations of the term aid in intra-disciplinary assessments; but over-arching characterizations to, for example, compare the complexities of a sorting algorithm to that of a steam engine do not exist, and seem somewhat ludicrous. However, a few cross-disciplinary generalizations are in order. Complexity rarely depends upon the number of components, nor their individual intricasies, but on their interaction protocols. Though individual humans seem infinitely complex on their own, the mind-boggling complexity of society stems not from the many billion individuals, but from the billions (or trillions) of interactions between them, many of which are (even in complete isolation) highly unpredictable. Typically, complex systems defy the prediction of global behavior from cursory analyses of components and their interactions. Surprises are commonplace, and in many cases, the only way to avoid them is to observe the system itself or run very large-scale simulations. In fact, Bedau [2] defines a weakly emergent system in terms of explanatory incompressibility: the only way to ascertain the system’s behavior is to run a complete simulation of it; no shortcuts, such as analytically model solutions or slick heuristics will do the trick. Finally, emergence and self-organization are the two words most synonymous with complex systems, while many of the other adjectives, such as nonlinear and autopoietic, have straightforward ties to these anchor terms. For example, emergence and its inherent surprise often result from synergistic interactions, where the global sophistication greatly exceeds the sum of the individual agent complexities; and this super-additivity is the hallmark of a non-linear system. Similarly, self-organization and autopoiesis are often co-occuring complements: self-assembly and self-maintenance typically require one another. A common element to discussions of emergence is that the self-organizing global structure provides constraints upon the local structure, thus biasing component interactions in ways not immediately predicted by the local behaviors. This leads to a useful working definition of emergence: The transition of a multi-component system into a state containing high-level properties and relationships: 1. whose causal origins involve a non-linear (and often unpredictable) combination of other component’ properties and relationships , and 2. that often add further constraint to the low-level interactions. For example, convection cooling of water on a pond’s surface produces vertical flows wherein warm water masses rise, are cooled, and then sink. These large-scale flows emerge from the local properties of water molecules that lose energy and are simply displaced from the surface by higher-energy molecules from warmer water. As small vertical streams form, their molecules may collide with nearby molecules, imparting vertical momentum and gradually broadening the stream. This constitutes positive feedback and leads to the rapid cooling of water. Since water is most dense at 4◦ C, any cooling below that temperature produces lower density surface water, which then halts the convection stream and reduces cooling to a (much slower) conductive process. Below 0◦ C, ice formation begins and further reduces the cooling rate of the remaining liquid. Thus, the local properties of water (i.e., it’s maximum-density and freezing points) support the emergence of the light upper layer and then ice, both of which have a negative feedback upon the vertical streams and thus upon water temperature, which they help stabilize. In sum, one emergent structure, a convection current, accelerates the cooling rate via positive feedback. The other emergent structure, ice, provides a negative feedback to maintain water temperatures (near freezing) despite much colder air temperatures. An often-cited example of emergence comes from the renowned neuroscientist, Roger Sperry[15]. As shown in Figure 2 1, the round wheel pattern enables a rolling motion, which imparts circular momentum upon the individual wheel molecules. Though an elegant example of top-down constraint, this fails to illustrate the actual self-organization of the global wheel form. Tumbleweed more convincingly exemplifies emergence. As the plant dries out and dies, it disengages from the root and is easily blown around. Being dry but still malleable, the dead plant attains a rounder form via repeated contact with the ground and the bending and breaking of individual stems. Eventually, the accumulated deformations round the weed to the point of being able to roll and impart top-down rotational momentum upon its molecules. Upon reaching a wetter area, the tumbleweed absorbs water and mechanically opens up, thus imparting another top-down momentum upon its molecules. Figure 1: Roger Sperry’s example of a wheel and the top-down effects of global structure upon the motions of individual molecules. As a more practical example of emergence, consider the work of Spears et. al. [14] in which purely local information guides the movements of agents (representing military vehicles), which then self-organize into target global patterns, such as hexagonal and square lattices. As shown in Figure 2, a hexagonal pattern emerges when all agents follow the simple rules of a) repelling all agents located closer than R units to themselves, while b) attracting all agents outside of that circle. For instance, on the left of the figure, agent B is pushed to the right by A and pulled upward by C, yielding a resultant force vector that moves it in a northeasterly direction, while agent C is pulled to the southwest, and agent A to the north. Together, these forces drive the agents into a stable triangular pattern. With many agents, this becomes a hexagonal grid. C C A A B B Figure 2: Emergence of a hexagonal lattice from simple attraction-repulsion behavior among agents (dots). Dashed circles denote borders between attraction and repulsion. (Left) All interactions between 3 agents are shown. These lead to the formation of a triangle, in which the resultant force vector on each agent vanishes. (Right) Sketch of some interaction forces among a larger group of agents, which eventually lead to a hexagonal lattice. To form square lattices, the authors introduce two types of √ agents differentiated by their spin values, either up or down. The preferred radius for dissimilar agents is R, and 2R for similar. Squares then emerge among groups of 4 agents, two of each type, with√alternating types along the perimeter (with edges of length R) and like types in opposite corners (a diagonal distance 2R apart). In addition, by allowing agents to occasionally change spin, they facilitate 3 an ingenious method of self-repair, as detailed in [14]. 2 Chaos and Complexity Complex systems are frequently characterized as those exhibiting a mixture of order and disorder; and the complex regime of a dynamical system is often defined as the edge of chaos: an area in state space between order and disorder/chaos. So an understanding of complexity requires one of chaos as well. The trademark of a chaotic system is Sensitive Dependence upon Initial Conditions (SDIC). Assume two separate runs of a system, one in which it begins in state s0 and another in s0 ∗, where the two states are distinct, but extremely (even infinitesimally) close in state space. The two runs of the system cause a transition through T states, ending in sT and sT ∗, respectively. SDIC entails that these two final states are not close: the two trajectories of the system diverge in state space. Figure 3 shows SDIC for a fictitious 2-variable system. s*T X2 s0 s*0 sT X1 Figure 3: Simple illustration of SDIC in a system composed of two variables: X1 and X2 . Though initial states s0 and s0 ∗ are very close, the trajectories from each quickly diverge. The problem with chaotic systems is that they are nearly unpredictable, since the accuracy of the initial state can rarely be assessed to more than a few significant digits. The mathematical beauty of chaos is that very simple, completely deterministic, non-linear systems can exhibit it. One such model is the logistic map of equation 1, one of a few classic examples of chaotic systems, and one that has applications in a wide range of fields, such as ecology, medicine, economics and neural computation. xt+1 = rxt (1 − xt ) (1) The gist of the model is simple: the next value of x is the product of a rate constant, r, the current value of x, and one minus that value. x0 is assumed to be between 0 and 1, and the model insures that x never goes outside those bounds. The choice of r strongly influences the trajectory of x in the following manner: • r < 3 - the system quickly transitions to a value that remains stable from timestep to timestep, a.k.a. a point attractor. • 3 ≤ r ≤ 3.57 - the system transitions to an infinite loop between k different values, a.k.a. a periodic attractor of 4 period k. When r = 3, k = 2; as r increases, the system goes through various bifurcation points, where k doubles. The frequency of occurrence of these points increases as r approaches 3.57. • r > 3.57 - the system never settles into a discernible attractor and SDIC arises. In short, small values of r produce an ordered state, while those above 3.57 produce chaos. To illustrate SDIC in this state, Figure 4 shows two trajectories that begin with a mere one millionth separation but gradually diverge. Also note that both trajectories are non-repeating. Rate=3.9000 1.0 0.6 0.6 X 0.8 X 0.8 0.4 0.4 0.2 0.2 0.00 Rate=3.9000 1.0 20 40 Time 60 80 0.00 100 x0 = 0.7 20 40 Time 60 80 100 x0 = 0.700001 Figure 4: Two applications of the logistic map (with r = 3.9) using initial values differing by only one millionth. The trajectories diverge after approximately 30 timesteps. Values of r near 3.57 represent the complex regime or edge of chaos. Many studies have shown that within this regime, a system is best able to perform computation: the storage, transmission and modification of information. Others indicate that systems often transition (or evolve) naturally to this regime. This natural transition is best exemplified by Bak’s sandpile [1]. Imagine dropping single sand grains onto a pile, always from the same point above the pile, and always just one grain at a time: the external perturbation to the pile is both small and constant. Initially, each grain will just fall to the table, but eventually the pile will develop a peak with steep slopes. At this point, the addition of each new grain can potentially cause avalanches, of various sizes (i.e. displacing varying numbers of sand grains). But when the sizes of those avalanches are plotted, it turns out that the vast majority of perturbations cause small avalanches, while an exponentially decreasing number of larger avalanches occur. Still, some very large avalanches do occur. This exponentially declining frequency curve is known as a power law distribution, and it is an often-cited signature of a system in a critical state. Figure 5 gives the plot of a typical power-law curve (left) along with a cartoon example of a distribution of avalanche sizes for a sandpile as described by Per Bak. For this and many other systems formally deemed complex, the frequency distribution of values for some important system variable is the basis for the power-law distribution expressed in equation 2. f (x) = cxk (2) As long as the perturbations are small and constant, then the occurrence of a power-law distribution over the internal effects of the disturbances is often grounds for the following type of claim: Most perturbations lead to small changes in the system, indicating that it is predominantly stable or ordered. But occasionally, a small perturbation can produce a large change, indicating a more unordered 5 Power-Law 1.0 0.8 Frequency of Occurrence f(X) 0.6 0.4 0.2 0.00 10 20 X 30 40 50 Avalanche Size Figure 5: (Left) Plot of the function f (x) = cxk with c = 1 and k = -2, a standard example of a power-law curve. (Right) Sketch of a power-law distribution for the avalanches of a typical sandpile. or chaotic situation. The mixture of the two, order and chaos, with the former dominating, is a strong indicator of complexity in the system. In addition to formalizing the critical state, Bak argued that many systems will naturally transition to it of their own accord, with only minor external assistance (such as the dropping of sand grains onto a pile): they are emergent in the same way that a convection current emerges from a) the internal physicochemical dynamics of water, combined with b) an external temperature perturbation. The term self-organized criticality (SOC) [1] expresses both the emergent and edge-of-chaos aspects of this important phenomenon. An important mathematical construct for assessing order, chaos and complexity is the Lyapunov exponent, λ in equation 3, where 4x0 and 4xt are the initial and final separations, respectively, between two trajectories of a system. 4xt = 4x0 eλt (3) Intuitively, when λ < 0, the trajectories converge; the system is ordered. Similarly, when λ > 0, the systems diverge, indicating SDIC and chaos. The complex regime is that (often narrow) band of parameter space where λ = 0 and trajectories neither diverge nor converge. For systems with large state-space vectors, this would seem to be a nontrivial band to find, although influential complexity theorists such as Stuart Kauffman [8] support Bak’s view that many systems will almost inevitably approach this edge-of-chaos area. Throughout this book, you will see examples of systems that appear to inhabit the complex regime as evidenced by a mixture of orderly and chaotic behavior, where order dominates but chaos has just enough presence to support adaptation. 3 Levels of Emergence Some of the more thorough philosophical investigations into emergence come from Terrence Deacon [4, 5], who describes three different levels, ranging from basic multi-scalarity to those invoking notions of self and memory/representation. Related to these levels are synchronic and diachronic aspects of emergence, where the latter involves global pattern formation over time, while the former involves changes of spatial scale without temporal progression. For example, temperature emerges synchronically at the macro level from energetic molecular interactions at the micro level; 6 this emergence does not unfold over time but merely becomes evident to the outside observer in the shift of spatial perspective and equipment: from the microscope to the thermometer. Conversely, pond freezing exhibits diachronic emergence through the temporal sequence of conduction, convection, conduction and ice formation. Deacon’s 3 levels or orders of emergence are more elaborate. First-order emergence is primarily synchronic and involves aggregates of low-level interactions that produce higher-level properties. Temperature from molecular interactions, color from molecular light-absorption properties, and weight from atomic mass are all examples of first-order emergence. This type of emergence is often devoid of surprise, with high-level properties being easily predicted from a statistical analysis of the underlying micro properties and events. Stochastic perturbations of first-order systems tend to average out at the macro level, as opposed to being magnified by positive feedbacks; and the global patterns exert no obvious constraints upon local dynamics. Although this is a fairly mundane type of emergence - essentially a base case - it exhibits some degree of explanatory complexity in that global patterns do not follow from local interactions by simple causal arguments: tools such as statistical mechanics, which average over the micro events, are needed for the logical derivation of global effects from aggregated local causes. With 2nd-order emergence, the fun begins! In these scenarios (as sketched in Figure 6), stochastic micro-level perturbations (a.k.a. asymmetries) can become amplified into global patterns in a diachronic manner. A recurrent causal architecture also comes into play as these global structures bias local behaviors. These systems are both self-organizing and autopoietic, and they involve a significant mix of bottom-up and top-down causality. These systems are dissipative in that material and energy may flow through them (i.e., they are far from equilibrium), but the global structure (i.e. the low entropy formation) is recirculated, amplified, and often eventually stabilized. This stability of form engenders notions of self : some hallmark of the system or system class that is created and maintained. Most examples of emergence in the popular complex-systems literature are of this type. At the top of Figure 6, notice that individual components begin with many degrees of freedom. This entails a large search space of possible actions, a space that gradually constricts as the global pattern emerges. Although activities such as the Brownian movement of molecules rarely evoke notions of search, when system components are cellular or multicellular, the metaphor begins to fit. In 3rd-order systems, the emergent high-level structures are imprinted into another structure: the memory or representation, which can function as a seed for another round of 2nd-order emergence, producing another, similar, high-level structure. Essentially, this facilitates reproduction and evolution of the self. This seeding is a re-presentation of the memory to the lower-level dynamics, which catalyzes 2nd-order emergence. Third-order emergence involves a synergy between self-organization and representation. The latter implicitly houses a memory of the self-organizational process, which, in turn, can create new representations. These representations may be disseminated and subjected to differential selection pressures (via different micro-level environments) and the resulting heterogeneous replication frequencies. Thus, 3rd-order emergence facilitates evolution by natural selection, where the representation is obviously analogous to DNA. Furthermore, any adaptations of the system to the environment will involve changes to the dynamics, some of which may be reflected in the emergent global pattern and thus encoded in the representation. In this way, the representation becomes an implicit memory of a history of environments (via the adaptations necessary to survive in them). For example, any animal’s genome serves as a history of its adaptations, a reflection of a long chain of salient environmental factors - ambiguous, noisy, and woefully incomplete, but a history nonetheless. Figure 7 illustrates the basic process, wherein the representation is built in parallel with the global structure. Then, when transferred to an unordered environment, i.e. one where the local interactions have few constraints, the memory provides a bias leading to both a) the rejuvenation of a high-order pattern, and b) the biasing of the low-level dynamics. Since this new situation may not be identical to the original, the representation may also change to more accurately 7 Time Symmetry of Local Interactions Global Structure Global Structure Asymmetry Amplified Asymmetry Figure 6: The essence of 2nd-order emergence. From predominantly unbiased (a.k.a. symmetric) local interactions, a small amount of bias arises randomly. Positive feedback accentuates this bias, forming a global pattern that provides significant constraint upon the low-level behaviors (i.e., increasing asymmetry). Negative feedbacks eventually kick in to stabilize the local dynamics and global structure. Arrows denote degrees of behavioral freedom of low-level components 8 reflect the new scenario. The representation acts as information in the formal, information-theoretic sense that, when added to a system of relatively unconstrained (i.e., high entropy) local interactions, it provides a strong bias, thereby promoting structure and reducing entropy; information content equates with induced entropy reduction (as detailed in a later chapter). Global Structure Memory-1 Memory-1 Amplified Asymmetry Global Structure Memory-2 Memory updated to include recent history Amplified Asymmetry Figure 7: Essence of the re-presentation process underlying Deacon’s 3rd-order emergence, where a memory structure arises as an additional result of 2nd-order emergence. It then serves as a seed for regenerating the global pattern in a new environment. Deacon illustrates 2nd- and 3rd-order emergence with an abstract biochemical system called an autogen, which is probably inspired by Varela et. al.’s earlier model of autopoiesis [6]. As shown in Figure 8, an autogen consists of an autocatalytic loop (AL) surrounded by a membrane. The loop and membrane are symbiotic: a by-product of the loop are membrane building blocks (M). The membrane prevents chemical constituents of the loop from dissipating into the environment. Though many components of the AL are produced from other AL constituents, a few raw material chemicals (e.g. X and Y) are also required to perpetuate operation. The autogen’s structure emerges through a 2nd-order process in which various compounds come into contact and begin producing other compounds, including M and other molecules of the AL. Positive feedbacks include a) the ability of M molecules to partially restrict the movement of the AL compounds, thus helping to prevent their immediate dispersal, and b) the eventual formation of the complete AL loop, allowing compounds A-D to produce one another (in the presence of X) and M (using Y). Once the complete membrane and AL have formed, negative feedbacks help to stabilize the structure. For example, since X and Y are only consumed (not produced) by the AL, their internal concentrations will wane without periodic resupply, which turns out to be an emergent property of the autogen. As the AL loop begins to slow down, production of M cannot keep pace with natural deterioration of the membrane, and gaps appear. These allow X and Y to flow into the cell, thereby restimulating the AL, producing more M, and mending the membrane. In short, the positive feedbacks involving AL and membrane lead to self-organizing autogen formation, while the 9 Y X M M M A X X X D Auto-catalytic Loop Y B Y C M M X X Y Figure 8: The basic structure of Deacon’s autogen, with an autocatalytic loop (AL) that requires raw materials (X and Y), and is surrounded by a membrane, whose building blocks (M) are produced by AL. negative feedbacks insure autopoiesis (self-maintenance). Furthermore, if the membrane splits in two places, creating pieces M1 and M2, then each piece may usurp enough of A-D, X and Y to have a private AL, which will then generate a complete new membrane for each, thus producing 2 copies of the original autogen. So reproduction can also be a 2nd-order emergent phenomenon!! Conceivably, an AL could become modified - for example by replacing compound D with E. If this produced a more efficient AL, possibly one requiring less raw materials to produce the same amount of membrane, then it might have a survival advantage in terms of having less frequent membrane tears. Of course, more tears could lead to faster reproduction, so an inefficient AL might actually be a more prolific copier. Either way, the roots of evolution are already evident in this 2nd-order emergent system. Y X M M A X X M X D Y B Y C M M X Y X Figure 9: Re-supply of X and Y via gaps in the deteriorating membrane. Deacon’s autogen exhibits 3rd-order emergence with the addition of a representation in the form of a chemical template. In general, chemical reactions are enhanced by mechanisms that increase the proximity of their reactants. For example, by elevating the odds that A and X molecules meet in Figure 10, production of B should increase. If molecules A* and X* exist, and bind to A and X, respectively, and if A* normally binds to X*, then the presence of an A*-X* compound should promote A-X proximity and B production. Thus, A* and X* serve as building blocks for a template that a) promotes AL formation, maintenance and productivity, and b) embodies a memory of the AL such that the infusion of a well-mixed chemical soup with the template can greatly accelerate autogen emergence. The information embodied by the template greatly biases the low-level dynamics of the chemical soup, producing structures (i.e., autogens) much faster than they would normally emerge. 10 Third-order emergence requires that the representation emerges as well. Although many of the building blocks may arise naturally in the chemical milieu, there is a decided selective advantage to any autogen whose AL produces some or all of them. For example, the AL of Figure 10 produces 4 of the 7 template blocks. Clearly then, autogens that produce templates should have an increasing frequency in the population compared to purely 2nd-order species, and each template modification that improves AL efficiency should give a selective advantage to these representations, whose relationship to DNA should now be obvious. Hence, 3rd-order emergence introduces the genotype-phenotype distinction. Templates, like other representations, are structural embodiments of the dynamic processes that underlie self. They play the informational role of catalyzing activities that reduce entropy. W* B* Z M A X D Auto-catalytic Loop Y B W C X* A* Template X* A* B* W* Z* D* Y* X A B W Z D Y Figure 10: Use of an auto-generated template to promote spatial proximity of the reactants of the autocatalytic loop. These three levels of Deacon will serve us well as we investigate emerging intelligence across the three levels of adaptivity: evolution, development and learning. As shown in later chapters, the two systems turn out to be more orthogonal than they may appear on the surface. For example, Deacon’s third level not only applies to evolution itself, but, in some cases, to learning. 4 Emergence and Search Have you ever wondered why migrating birds often fly in a V formation? The explanation is classic emergence. We begin with an important fact of physics: when a bird flaps its wings downward, it pushes air downward, creating a low pressure region above one of higher pressure (due to the added air). A split second after this happens (and the bird moves forward), this pressure difference produces a small updraft. A second bird, trailing behind and to the side, can use this updraft to lift its own wing, thus saving some energy. In addition, if birds prefer seeing the horizon ahead of them, instead of the backside of another bird, then they will naturally seek out locations in the moving flock that both save energy and provide an unobstructed view. When all birds have these motivations, the group self-organizes into a flying wedge. As illustrated in Figure 11, from the perspective of a single bird looking for an optimal location, the initial, unorganized phase affords many options (starred locations). These comprise a broad, unconstrained, private search space for the bird. However, as the global V pattern emerges, it embodies a top-down constraint, operationalized as a restriction of 11 the bird’s search space. So whereas the bird can (quite literally) wing it and randomly choose among many options in the unorganized state, its moves become forced as the wedge materializes. ? ? Figure 11: The emergence of a V pattern in flying birds (arrows), each of which creates pockets of uplifting air behind and to each side. Stars denote unused (by a trailing bird) pockets with an unobstructed view, while circles represent unused pockets with a poor view. (Left) Prior to V formation, many unobstructed pockets exist, which entails many options for the focal bird (drawn with an open arrowhead). (Right) In the V, the only optimal options for the focal bird are at the edges, and these promote extension of the V, a positive feedback. As shown in Figure 12, this transition from an open to a restricted search space is even more obvious for schooling fish, which conserve energy by swimming directly behind another and appear to follow a few local behavioral guidelines: head in the same direction as your neighbors while avoiding collisions. The same basic idea pertains to human scenarios as well, such as swimming in crowded pools, driving in congested freeway traffic, and finding your favorite concession stand at halftime of a football game. The important point is that the decisions, which may appear quite intelligent, involve no sophisticated knowledge or planning; they are merely reactions to an emerging set of largescale constraints, which often make the correct movement choice patently obvious. In short, as situations accumulate constraints, even the most intelligent of beings exhibits little more problem-solving wizardry than Simon’s [13] ant on the beach. Figure 12: Emergence of schooling in fish. (Left) In the unorganized state, a focal fish (open arrowhead) has many movement options (stars). (Right) The organized state offers few degrees of freedom. Emergence in systems of decision-making agents can therefore be viewed as an interaction among parallel search processes. The initial trial-and-error dynamics of each such process enables the system to jiggle into configurations (e.g. short diagonal lines and small wedges) where positive feedbacks commence (e.g. forcing the moves of birds near these patterns) and allow the small patterns to grow into large-scale constraints, which then exert natural restrictions on every individual search process. Without the trial and error, there is no jiggle, and the system can quickly stagnate in a highly ineffective state. Imagine what happens if every bird, fish or human simply choosse a movement sequence ahead of time and sticks to it, come hell or high water! 12 From an observer’s global viewpoint, the group as a whole might appear to be searching for an efficient pattern, though no over-arching goals nor instructions guide such a transition. Still, the system gives the impression of trying different alternatives and eventually becoming attracted toward particular stable configurations. A similar interpretation pertains when we observe populations of non-cognitive agents, such as neurons. Each does have a repertoire of possible behaviors, which become constrained by the behaviors of other neurons, but the neuron obviously has no awareness of these options. Despite the lack of cognizant deliberation by molecules or neurons, and by the aggregates that they form, the search perspective provides useful insights and predictions when analyzing such systems. Of course, at the highest levels, the brain itself performs cognitive acts, but the choices that organisms make are not explicitly about neural patterns; i.e. there are not specific target patterns of neural firing toward which organisms explicitly strive. Yet, from the perspective of an outside observer examining patterns of neural firing, the system can appear to naturally transition to certain attractive states, cycles or semi-recurring patterns. Just as in physics, where systems are often described as seeking a low-energy state, the search metaphor provides useful explanatory leverage in any population of interacting entities, each of which has several degrees of freedom. Then, in cases of emergence, these degrees may change as global patterns form and evolve, giving the outward appearance of a system seeking a particular ordered state. In emergent systems, individual agents make local choices, thus affecting a small subset of the global situation. However, unlike many localized design decisions in AI search, those in emergent systems are not tested in isolation but in vivo: in direct interaction with other system components. Emergent systems do not deal in partial states with expected completions, but in complete states with expected transitions. Through an understanding of the interaction topologies and behavioral repertoires of the agents, as well as the changing constraints, one can begin to attach probabilities to these expectations and to differentiate between situations in which strong constraints force agent choices and those in which agents possess a legitimate freedom to explore. It turns out that this exploratory freedom of individual agents is the foundation of natural adaptivity - a foundation composed of an extremely malleable substrate, but one that tends to bend without breaking. As we will see, the ability of living systems to modify themselves to accommodate external change relies heavily on the ability of their components to exhibit fairly explicit trial-and-error behavior. As the basis of adaptivity, parallel search among multiple agents is pivotal in the emergent complexity of life and intelligence. For example, the development of the nervous system is a highly adaptive process in which brain hooks up to body, not by carrying out a genetically-specified topology-building plan (i.e. connect neuron A to neuron B, then B to C and D, etc.), but by individual neurons performing various search processes, such as sending out axons in search of dendritic targets. Those that find a dendrite to synapse onto may survive, while failure to connect can spell death. Similarly, during cell mitosis, a pair of centromeres send out exploratory microtubules searching for chromosomes with which to attach. Those that hook up will form part of the mitotic spindle; the rest will wither away. This is exploration, trial-and-error search within individual cells, that enables the centrosomes to find all the chromosomes, wherever they lie within the cell. The cell adapts to random chromosome locations because the centrosomes can search/explore, in the same way that the nervous system adapts to varied locations of muscle cells via the exploratory axons of motor neurons. 5 Emergent Intelligence Although many different properties of complex systems have already been discussed, Melanie Mitchell’s formulation [10], motivated by a network perspective, seems very relevant for the emergence of intelligence. She provides four basic characteristics of complexity: 1. Global information is encoded as patterns over the component states. 13 2. Stochasticity abounds, and is often magnified by positive feedback. 3. Parallel exploration and exploitation at the level of individual agents is ubiquitous. 4. Top-down and bottom-up interactions co-occur continuously. The brain exhibits these in many different ways, though one example of each suffices for now. First, neural representations are commonly believed to involve population coding: concepts such as tumbleweed are encoded by large combinations of active neurons, not the individual cells alone. Second, random neural activations occur continuously in the brain, and these often employ positive feedback to illicit large-scale firing patterns, such as the synchronous activity common to brain oscillations, such as theta and gamma waves, which are critical for memory formation and retrieval [3]. Third, during development, neurons grow axons that independently explore the environment for other neurons with which to connect. In this search for efferents, links to already well-connected neurons have a selective advantage, since this exploitation of proven networkers enhances the new neuron’s chances of participating in active circuits and thus surviving. Finally, the brain’s interpretation of sensory information relies on a healthy mix of bottom-up (sensory-driven) signals that interact with top-down (expectation-driven) signals from higher cortical regions. Most popular accounts of complex systems, such as Mitchell’s [11], include the brain in their list of examples, with intelligence touted as an emergent result of billions of interacting neurons. Unfortunately, intelligence is a rather vague concept, and certainly not one easily described by a particular global pattern - although, as discussed in later chapters, scientists have attempted to define concepts such as consciousness in terms of information-theoretic metrics and the neural activation patterns that satisfy them. However, a detailed analysis of the many neural structures and mechanisms (along with their origins) that underlie intelligent behavior reveals a parade of self-organizing systems. These span the three primary levels of adaptation, and at each level, emergence relies heavily upon a search process. From an evolutionary perspective, brains have emerged along a million-year timescale via the gradual expansion of a very distant ancestral genotype (by duplication and modification of successful genetic components). Evolution has devised ingenious developmental recipes for converting genotypes to phenotypes. Although evolution is often characterized as random by laymen who either lack a solid biological background or have various religious agendas, the concept of selection pressure and its ability to steer evolution via genotypic adaptations normally suffices to convince open-minded people that evolution has the ability to craft complex organisms that purely random processes could never achieve. In short, evolutionary search is highly biased, favoring some genetic and phenotypic variations over others. However, even many biologists view the processes by which variations arise (prior to selection) as random, with selection then separating the wheat from the chaff. Conversely, the popular new theory of facilitated variation [9] indicates quite the opposite: the processes that generate variations are also highly biased. Essentially, millions of years of evolution have hit upon complex combinations of genes, and versatile genotype-phenotype relationships that greatly improve the odds that random mutations will produce viable phenotypes, thus giving selection a very high-quality pool of applicants. It seems that the DNA representations of natural 3rd-order emergence are both extremely informative and robust. Clarifying the source of this robustness can both a) enhance our general understanding and appreciation for evolution as a designer, and b) suggest feasible strategies for artificial evolutionary searches for intelligent algorithms. This permits an interpretation of the evolutionary search process as, in fact, quite well informed and strongly biased to choose routes that history has pre-selected by virtue of the constraints gradually molded into the genotype-phenotype mapping. Developmental processes themselves exhibit considerable self-organization, as a multitude of embryonic cells send and receive chemical signals, many of which strongly affect behavior and lead to the formation of global structures. As discussed later, exploratory search, involving real physical molecules and cells, along with their connective tubing and wiring, plays out in many parts of the body and brain. Without this developmental search and the grow to fit assembly paradigm that it engenders, the genome would become more a blueprint than a recipe, and lose most of its robustness along the way. An examination of the 2nd-order emergence in developmental search helps explain how 14 instances of apparently improbable simultaneous coevolution of many traits are actually the modification of one trait followed by a host of phenotypic adaptations to that perturbation, all made possible by the flexible search processes implicitly encoded by the genome. Development is one of the most impressive emergent phenomena in existence, as the complex global structure of the organism arises from an intricate interplay among a wide variety of cells. Postnatal modifications are often considered examples of phenotypic plasticity, with classic examples being the increase of muscle mass via weight lifting, the increase of red-blood-cell count by moving to higher altitude, and the modification of synaptic strengths in the brain (often equated with learning). Whereas development lays down the main components and their connections, plasticity/learning involves smaller, more local, tweaks to that infrastructure. In the brain, development involves the formation of large populations of neurons, along with the extensive axonal and dendritic networks linking them, whereas learning typically involves synaptic formation and/or tuning but may include the addition of new neurons to established populations (and their linking up to others), although only a few areas of the mammalian brain appear to support post-natal neurogenesis [12]. Learning also has many emergent aspects, again resulting from processes often characterized as search. For example, synaptic tuning in supervised-learning scenarios is commonly modeled as a gradient-descent search for a vector of synaptic weights that reduces the total error of a neural network’s input-output mapping. Classic AI supervisedlearning algorithms are not particularly emergent, due to the presence of externally-provided control signals and complex algorithms with a highly non-local flavor. However, some brain circuitry exhibits more decentralized supervised learning from which the mature, fully-functioning network gradually arises. Other brain networks display emergent reinforcement learning involving global chemical broadcasts but certainly nothing ressembling global hands of god steering the process (as in classic AI methods of reinforcement learning). Finally, unsupervised learning is a common brain mechanism capable of crafting complex topological maps from purely local interactions. Here, both natural and artificial versions of the algorithm exhibit considerable self-organization. In addition, the dynamic progression of networks tuned by unsupervised learning is often modeled as a search process in the space of neural-activation patterns. Compared to development, the self-organizing global structure of learning is less impressive for its physical differences from the original (simpler, less-organized) state but more recognized for its accentuated informational content, behavioral intricacy or functional ability. In many cases, to fully appreciate the emergence within a learning system, one needs mathematics (e.g. information theory) instead of a microscope or phylogenetic tree. Indeed, intelligence is emergent, and just as imperfect as Mom, Dad, tax laws and the United Nations. Still, the neural machinery is oiled well enough to perform many impressive feats. The full extent of this phenomena goes well beyond examples of sophisticated coordination arising from individual decisions, such as the ability of 5 basketball players to weave intricate patterns from only individual movement principles, or of jazz musicians to produce lively entertainment with band members whom they’ve never met before, or of fireflies to blink in synch. The true beauty of emergent intelligence is spread across the three adaptive mechanisms and Deacon’s three levels. And it’s all about search and neurons. References [1] P. BAK , C. TANG , AND K. W IESENFELD, Self-organized criticality, Physica Review A, 38 (1988). [2] M. A. B EDAU, Is weak emergence just in the mind?, Minds and Machines, 18 (2008), pp. 443–459. [3] G. B UZSAKI, Rhythms of the Brain, Oxford University Press, New York, NY, 2006. [4] T. D EACON, The hierarchic logic of emergence: untangling the interdependence of evolution and selforganization, in Evolution and Learning: The Baldwin Effect Reconsidered, B. Weber and D. Depew, eds., Cambridge, MA, 2003, The MIT Press, pp. 273–308. [5] , Incomplete Nature: How Mind Emerged from Matter, W.W. Norton and Company, New York, 2012. 15 [6] H. M. F. G. VARELA AND R. U RIBE, Autopoiesis: the organization of living systems, its characterization and a model, Biosystems, 5 (1974), pp. 187–196. [7] J. H. H OLLAND, Hidden Order: How Adaptation Builds Complexity, Addison-Wesley, Reading, Massachusetts, 1 ed., 1995. [8] S. K AUFFMAN, The Origins of Order, Oxford University Press, New York, 1993. [9] M. W. K IRSCHNER AND J. C. G ERHART, The Plausibility of Life: Resolving Darwin’s Dilemma, Yale University Press, New Haven, CN, 2005. [10] M. M ITCHELL, Complex systems: Network thinking, Artificial Intelligence, 170 (2006), pp. 1194–1212. [11] M. M ITCHELL, Complexity: A Guided Tour, Oxford University Press, New York, NY, 2009. [12] T. J. S HORS, Saving new brain cells, Scientific American, 300 (2009), pp. 40–48. [13] H. A. S IMON, The Sciences of the Artificial, MIT Press, Boston, MA, 1996. [14] W. S PEARS , D. S PEARS , J. H AMANN , AND R. H EIL, Distributed, physics-based control of swarm vehicles, Autonomous Robots, 17 (2004), pp. 137–162. [15] R. S PERRY, In defense of mentalism and emergent interaction, Journal of Mind and Behavior, 300 (1991), pp. 221–245. 16
© Copyright 2026 Paperzz