Building Aunt Hillary: Creating Artificial Minds with ‘Neural Nests’ Mike Reddy and Stuart Lewis School of Computing, University of Glamorgan Trefforest, Pontypridd,, RCT. CF37 1DL Wales, UK {mreddy,sflewis}@glam.ac.uk http://www.comp.glam.ac.uk/staff/mreddy/ Abstract. This paper describes work in progress to enable a real robot to recreate trail following of ants engaged in pheromone-reinforced recruitment to food gathering. Specifically, it is proposed that development of a set of macrobehaviours for creating and following a trail can be achieved by use of microbehaviours in a simulated environment to develop a novel neural architecture – the Neural Nest - for learning without explicit representation. A simulated ‘neural nest’ has been tested to determine the feasibility for ant colonies to encode higher-level behaviours for controlling a physical robot. In our experiments, the emergent behaviour from reinforcement of interactions between unsupervised simple agents, can allow a robot to sense and react to external stimuli in an appropriate way, under the control of a non-deterministic pheromone trail following program. Future work will be to implement the architecture entirely on the physical robot in real time. 1 Introduction Swarm Intelligence (SI) research has been useful for applying models of adaptive behaviours of complex biological structures, consisting of simple entities (e.g. termites, ant nests, and beehives), to computational problems that require prohibitively expensive search algorithms [1]. One of the key strengths of SI is that elements have flexibility to interact in unpredictable ways, to facilitate optimisation. Specifically, Ant Colony Optimisation (ACO) [2] and Ant Colony Routing (ACR) [3] have been proposed, which use the analogy of pheromones in stylised environments to govern search and routing strategies. However, these techniques have, so far, been mostly applied to specific topological/geographical problems (e.g. the Travelling Salesman Problem [2], Communications Networks [3-5], etc). It should be noted that other approaches, such as novel neural networks [6] or evolutionary algorithms [7] may be more successful, where potential solutions are numerous, or uniformly distributed [1]. SI has also been applied to the problem of coordinating groups of autonomous robots to collaborate on a task [8]; the concept of simulated ants being mapped directly onto sets of robots situated in a simulated or real environment. The power of SI has, therefore, been in coordinating these situated agents at the macro-level without the D. Polani et al. (Eds.): RoboCup 2003, LNAI 3020, pp. 584–591, 2004. Springer-Verlag Berlin Heidelberg 2004 Building Aunt Hillary: Creating Artificial Minds with ‘Neural Nests’ 585 need for planning or explicit communications. However, most existing ant colony simulations do not attempt to address the complex emergent control of individual physical ants (robotic or biological). We believe that the strength of SI – its use of micro-components to optimise a complex macro-system – may also contribute to the control of individual micro-components, such as a single physical robot, rather than just simulating them as part of a larger macro environment. Many adaptive robot control architectures have been proposed, to control agents at the micro-level, that is, within the individual robot; these include traditional and recurrent neural networks and evolutionary programming, etc. (A good summary can be found in [9, 10]). Ethology, in the form of Tinbergen’s instinct centres [11] has inspired a learning classifier approach using Genetic Algorithms (GAs) [12]. However, there has been no attempt to make direct use of SI to control robots at the micro-level; namely to implement an autonomous learning robot using an Ant Colony System (ACS). Simulation of ant behaviour may yet prove useful for understanding the emergent properties of complex biological structures within the human brain. The neural nest, proposed in this paper, is an ant colony inspired implementation of a real-time neural network, without a pre-defined symbolic representation or the need for supervised learning. To allow emergent intelligence, it is, vital that a physical platform be available to reproduce the conditions for evolutionary development, and to enable evaluation of the ant-powered brain. Finally, we describe integrating the symbolic ant brain with the physical robot. 2 Previous Work 2.1 Aunt Hillary – A Reprise In his seminal work “Gödel, Escher, Bach – an Eternal Golden Braid” [13], Douglas R. Hofstadter describes Aunt Hillary, a conscious ant colony, who consists of signal propagating and symbol manipulation activities using ants as base units interacting with the colony environment to shape the activities of ants into higher level concepts. This is not at odds with more symbolic representations. However, it is difficult to perceive how behaviours could evolve, when we only consider large grained views of intelligent systems. While Aunt Hillary serves principally as a clever analogy to explain how neurons can collaborate to produce symbol manipulation and, ultimately, human consciousness, the concept is an inspiring one, raising many interesting questions, especially when the brain can be considered a network of real-time processes (or neurons), with a flow of control from sensors to motors. Symbolic and subsymbolic techniques make no attempt to describe how higher-level concepts could be implemented in the structures of the brain, and draw a sharp distinction with the physical nervous system. It must also be remembered that it is not just brains, but also bodies (sensors, actuators, etc.) that have co-evolved. Biological systems have not only learned to adapt, but also adapted to learn. 586 Mike Reddy and Stuart Lewis 2.2 AntNest – A Reprise Real ants have limited senses and communication, but manage to evolve optimal solutions to search and routing problems, using pheromones trails, which result in a positive feedback effect. Several simulations of this activity exist, but most have a simple rule-based approach for controlling ant behaviour. Furthermore, these simulations are limited, especially where the strength, degradation and effectiveness of scents can be critical for success. Our previous work, AntNest, has been particularly successful due to using two scents: one for food-bearing ants (food scent), and one for foraging ants (nest scent) [14]. The revised rules in AntNest are as follows: 1) 2) 3) 4) 5) Ants leave the nest in a random direction, leaving a nest scent in the process. Random search influenced by the food scent level, if any, leads the ant to food. Some of the food is acquired and the ant begins to lay food scent towards home. Random movements, influenced by the nest scent level, lead the ant homeward. The ant then returns to step 1. AntNest was originally conceived as a means for visualising Economics problems, by modeling a biological simulation of an ant colony; ants had several roles in addition to food gathering: nursing babies, clearing the dead from the nest, attacking enemies and sleeping. However, we were inspired by the similarity of the simulations to the architectures of neurons: axons and dendrites. Several features of current neural network architectures have been abstracted from the biological reality: real neurons react to rates of fire, and do so in a continuous time-frame; they are also excitatory or inhibitory in nature, whereas artificial neurones can generate positive and negative values, mediated by weightings (again positive or negative) at the receiving neuron. The research question was whether we could use competing ant colonies to implement a neural architecture that was more closely inspired by the qualities of real neurons: ant trails being analogous to neural links, while the rate of flow of ants from source to nest representing the rate of fire of a neuron. 3 Neural Nests – A Proposal In this section, a connectionist reinforcement learning architecture is described, which may enable an autonomous robot to acquire motor-sensory navigation strategies without explicit representation. The use of an embedded ACS to implement a novel neural architecture has several advantages: Firstly, it allows us to experiment with combining the strengths of both SI and neural networks; Secondly, it allows the possibility of applying SI techniques in a direct manner, within a single entity, namely a physical robot ant in a real setting; Thirdly, it satisfies an aesthetic need to explore the ability for complex systems, such as the human brain, to evolve and manipulate meaningful symbols using simple components, without explicit prior representation as is the case for many other solutions to robot control. Building Aunt Hillary: Creating Artificial Minds with ‘Neural Nests’ ??? 587 Touch Food 1 Nest Scent 1 ??? Motor Nest 1 Motor Nest 2 Nest Scent 2 Food Scent Touch Food 2 Motor 1 Motor Control: Red = Fwd Blue = Bwd Motor 2 Touch 1 Touch Effect: Red = Free Blue = Blocked Touch 2 Fig. 1. Neural Nest Internals: Food Sources and Motor Nests are passive entities inside the environment, which represent sensor inputs and motor outputs respectively. Ants collecting food and returning to the nest will cause the external robot to act accordingly depending on their behavioural properties Furthermore, a tentative cognitive model of unsupervised neural learning is described – the Neural Nest – which uses a set of simple micro-agents to recreate the linkage and firing behaviours of biological neurons to demonstrate the feasibility of our approach to real autonomous robot control. It was recognised that a physical implementation in a real robot was needed as early as possible, rather than working purely with simulation, and then attempting to port to a robot afterwards. Simulation does not throw up the real-world errors and inconsistencies that are present with even the most simplest physical environments; this is especially important as we consider the body evolution to be key to brain evolution. However, the neural nest architecture is a slow one at present – certainly not capable of working in real-time – so a simulated robot was modelled in a closed environment. Initially, this robot had just two touch sensors and two motors. 588 Mike Reddy and Stuart Lewis The basic currency of the Neural Nest architecture is the collection of food from food sources linked to external sensors on a real or simulated robot. Nests are associated with external actuators (in this case, motors on the robot) and the collection of food and its return to the nest provides a signal processing capability that will control the robot’s behaviour. The first experiment in the Neural Nest simulator was to see whether the robot would more sensibly with two motor nests and two sensor food sources (See Figure 1). This first experiment showed the gradual grafting of the two touch sensors onto their respective motors. The early effect of this was to drive the robot into obstacles rather than away from them. However, over time a cross-link occurred that allowed the robot to be a little more effective at avoiding walls, when approaching them with one of the sensors. The second experiment involved adding light sensors to the robot, in order for it to engage in line following. We added food sources related to position of the sensor array on the front of the vehicle to the previous simulation arrangement, placing the new food sources in comparable positions to their placement on the physical robot described in Section 4. Again in simulation the architecture generated by the simulation did not show the bipolar symmetry that most animals display. This may be a fault in the model, such as scent evaporation, or faults in the simulated environment1 but it may also be due to a very limited time available for the system to evolve a solution before testing. The model generated is shown in Figure 2. From observation of the simulated robot, what appeared to be happening was as follows: 1) The right motor stayed on continuously, so directions and steering came from the activities of the left motor, unless the right touch sensor was activated; 2) The left motor stayed on continuously, but reacted to both the left touch sensor and the left and centre light sensors. 3) When a line was detected by either of these light sensors the left motor would stall. If both detected a line the left motor would reverse, turning the robot left. This gave a limited ability to sense and follow lines, while also maintaining the ability to react partially to the touch sensors. The effect of this architecture on the physical environment is detailed in Figure 4. Fig. 2. Touch and Light Simulator: (a) The two motor nests search for food; (b) Neural architecture is complete with the left-hand nest also using the left light sensor 1 This could be the random number generation skewing movement. Building Aunt Hillary: Creating Artificial Minds with ‘Neural Nests’ 589 2 4 Robot AUNT – The Macro-level Robot AUNT is a tracked robot with two touch sensors on the front corners and three light sensors mounted facing down at the front of the canopy. The experiments ran on a white board placed on the ground and fenced off with walls, which the touch sensors would detect. A retractable white-board marker allowed the robot to lay a trail for the light sensors to follow. The initial problem to be faces was whether such a pen-based trail on a whiteboard could act as a suitable environment for testing pheromone following with a robot. Initial experiments were performed using a handcoded program to test the feasibility of a real physical implementation of pheromone following. Note: This was done for various reasons, but the most significant was to test whether I could get the tracks to perform a rudimentary form of evaporation – the pen marks will not fade the way that a scent trail would, and there is only one robot, rather than thousands. In fact, the robot tracks provided an excellent mechanism for reproducing the effect of scent evaporation, even if not in the way it would in reality3. Fig. 3. Robot AUNT: (a) Touch sensors mounted on the corner; (b) Three light sensors facing down; (c) A retractable pen mounted at the pivot point to mark trails The key issue was then to determine whether the neural nest architecture could be used to control a physical robot. In order to do this effectively, an abstraction of the simulated network was implemented as a set of probabilistic choices in the control of robot motors, dependent upon sensor values. We simulated the arrival of ants as decisions to go forwards and backwards, with the delay between sensing and acting proportional to the average time for an ant to travel between the food source and the nest. Clearly, this is not a perfect reproduction of the actual nest architecture, but allowed a simple implementation on the robot. As was true for the simulation, the robot tended to veer to one side and occasionally trapped itself in corners. The next stage of the experiment was to implement the architecture described in Figure 2, using light and touch sensors. In order to do this, Robot AUNT would need to be able to lay trails autonomously. Modifications were made to the environment to allow the robot to detect whether it was at a food source (shiny surfaces) or the nest (black surface). The 2 3 AUNT stands for Autonomous Unsupervised Neural Turtle – in honour of Aunt Hillary. The robot is a turtle rather than an ant because of its previous incarnation as a drawing robot. It is a little known fact that white-board marker is easier to wipe off once it is completely dry. The efficacy of the tracks to wipe the board clean was proved beyond doubt when we forgot to change the pen for one experiment and returned to find the board eerily clean. 590 Mike Reddy and Stuart Lewis program was modified so that the pen would be lowered once the robot had found food, and raised when it had returned successfully to the nest. Given that there was no help from a number of ants leaving trails, Robot AUNT did mark out trails to and from the nest (see Figure 4). It also performed slightly better at following lines than the simulation did4. Fig. 4. Robot AUNT II: (a) Searching for food by following lines from previous journeys; (b) Carrying food back to the nest, leaving a scent trail in white-board pen 5 Conclusions A cognitive model of unsupervised neural learning has been described – the Neural Nest – which uses a set of simple micro-agents to control the macro-behaviours of a physical robot. We have seen a neural nest link touch and light sensors to motors in a way that is effective in both a simulated and real robot, with a degree of sensor fusion. Ant trails and food foraging have been used as the basis for a computational model of how neurons forge links and activate by rate of firing and excitatory and inhibitory behaviours. Results are encouraging, but tentative, with more quantifiable analysis needed to ensure that this is a valid technique for evolving robotic control systems. However, the current scent model needs some revision, as nests switch between different sensors, with only one good link at a time. Sensitivity analysis of the range of producible behaviours is also needed. Furthermore, the resulting robot search pattern is prone to the lack of positive reinforcement, where real ants would be surrounded by recruited colleagues in scent trail generation; our studies only used one macro-robot. Further research into the use of multiple robots may show more clearly the effectiveness of this macro and micro approach to robot control. While this may not produce the most efficient solution to the problem, it is our hope that such an evolutionary approach may also eventually lead to a better abstract representation of the link between low-level neural functioning and higher order behaviours of a robot. 4 This is most probably due to the need to use eight fixed directions in the simulation, where a physical robot would be changing direction more uniformly using real motors. Building Aunt Hillary: Creating Artificial Minds with ‘Neural Nests’ 591 References 1. Bonabeau, E., Dorigo, M., Theraulaz, G.: Swarm Intelligence: From Natural to Artificial Systems. Oxford Univ. Press, New York, (1999). 2. Dorigo, M., Gambardella, L. M.: Ant colonies for the traveling salesman problem. BioSystems 43, 73-81 (1997). 3. Schoonderwoerd, R., Holland, O., Bruten, J., Rothkrantz, L.: Ant-based load balancing in telecommunications networks. Adapt. Behav. 5, 169-207 (1997). 4. Heusse, M., Guérin, S., Snyers, D., Kuntz, P.: Adaptive agent-driven routing and load balancing in communication networks. Adv. Compl. Syst. 1, 237-254 (1998). 5. Di Caro, G., Dorigo, M.: AntNet: Distributed stigmergetic control for communications networks. J. Artif. Intell. Res. 9, 317-365 (1998). 6. Chen, K.: A simple learning algorithm for the traveling salesman problem. Phys. Rev. E 55, 7809-7812 (1997). 7. Baluja, S., Caruana, R. In Prieditis, A., Russell, S.: Proc. 12th Intl Conf. Machine Learning. Morgan Kaufmann, Palo Alto. (1995). 38-46. 8. Kube, C. R., Zhang, H.: Collective robotics: from social insects to robots. Adaptive Behavior 2, (1994). 189-218. 9. Dorigo, M. (Guest ed): Special Issue on Learning Autonomous Robots. IEEE Transactions on Systems, Man, and Cybernetics-Part B Vol.26, No.3, (1996). 10. Dorigo, M., Colombetti, M.: Robot Shaping: An Experiment in Behavior Engineering. MIT Press. (1997). 11. Tinbergen, N.: The Study of Instincts. Oxford University Press. (1966). 12. Dorigo, M., Schnepf, U.: Genetics-based Machine Learning and Behiour Based Robotics: A New Synthesis. IEEE Transactions on Systems, Man, and Cybernetics. Vol. 23, No. 1, (1993) 141-154. 13. Hofstadter, D.: Gödel, Escher, Bach: An Eternal Golden Braid. 14. Reddy, M., Kedward, P.: Ant Economics: What simple agents can teach us about supply, demand and eating your Dead. Procs. of ANTS'2000 - From Ant Colonies to Artificial Ants: 2nd International Workshop on Ant Colony Optimization. Brussels. Belgium. 8-9th September 2000.
© Copyright 2025 Paperzz