Transforming Robotics with Biologically Inspired Learning Models
Aisha Sohail, Heather Ames and Jasmin Leveille run simulations of the artificial visual system in Cog.
CREDIT: Massimiliano Versace, CELEST, Boston University
This Behind the Scenes article was provided to LiveScience in partnership with the National Science Foundation.
I walked into the building and there was a human-sized robot waiting to greet me.
It shook my hand, took my coat, and brought me to sit in the room where my interview was going to be held. It asked me whether I needed a drink, and then proceeded to clean the countertops and water the plants. When I asked whether there was a reason it was working so hard, it simply said, "I am putting myself to the fullest possible use, which is all I think that any conscious entity can ever hope to do."
If you have ever seen Stanley Kubrick's tribute to humanoid computers "2001: A Space Odyssey," then you already know I was merely making an allusion....
What actually happened during my first visit to the Neuromorphics Lab at Boston University was a slightly different, though no less entertaining, scenario.
I walked into an office and there was a Roomba-like robot approaching and avoiding multicolored objects. It made its decisions based on a reward history ("bad robot" vs. "good robot").
On a desk, I noticed a dismembered RC helicopter with half its parts missing. Peeking into an additional room, I couldn't help but notice a toy car with a camera installed at the helm, and EEG electrodes hanging off on all sides. All around me, researchers were creating and refining artificial brain systems in virtual environments before deploying them in robots.
Even before sitting down to talk with anyone about job opportunities, I knew this was the place for me.
The Neuromorphics Lab is researching innovative robot learning-algorithms. Imagine having a cleaning robot that did what no other cleaning robot is currently able to do: Learn. It could learn the one place in your house where your dog always loves to wipe his grubby little paws when he comes inside. It could learn that Tuesdays are softball practice, which means a certain trail of dirt leading up to your room.
The keyword here, obviously, is learning. The problem with the conventional approach to robotics is that it requires explicit programming for robots to carry out specific tasks, leading to a lack of autonomous, general-purpose artificial intelligence, or AI.
Working in collaboration with Hewlett-Packard laboratories, the Neuromorphics Lab, part of the NSF-sponsored Center of Excellence for Learning in Education, Science and Technology (CELEST), has undertaken the ambitious project of creating a brain on a chip — a fundamental predecessor to the design of autonomous robotics and general intelligence.
Researchers in the Neuromorphics Lab are closer than ever to being able to accomplish the goal of creating a general mammalian-type intelligence. Most people have never even heard of the term "neuromorphic" — which is a technology with a specific form ("morphic") that is based on brain ("neuro") architecture. The neural models being developed by the Neuromorphics Lab implement "whole brain systems," or large-scale brain models that allow virtual and robotic agents to learn on their own to interact with new environments.
Like any intelligent biological system, artificial-autonomous and adaptive systems need three things: A mind, a brain and a body. The CELEST models run on a software platform called Cog, which serves as the operating system within which the artificial "brain" is developed.
Along with the hardware — currently general-purpose processors to be augmented by innovative nanotechnologies under development at HP — Cog offers an ideal environment for the design and testing of whole-brain simulation. The work of the Neuromorphics Lab focuses primarily on engineering the mind of the adaptive system. Once complete, a virtual animat, equipped with the artificial brain, will be able to learn how to navigate in its environment based on its inherent capabilities for responding to motivations, evaluating sensory data, and making intelligent decisions that are transformed into motor outputs.
As a new employee of the Neuromorphics Lab, I recently participated in a demonstration of the adaptive robot. I watched as it was able to learn to distinguish and develop a preference for a set of multicolored blocks. Although this may seem like a trivial task, one that comes naturally to humans, the immensity of this task lies in the fact that the animat is not explicitly programmed to approach certain colored blocks, but rather to learn which objects to approach and avoid based on rewards and punishments associated with them. The process is similar to how animals learn by trial and error to interact with a world they were not "pre-programmed" to act upon.
Whole-brain systems are difficult to engineer and test. The Neuromorphics Lab accelerates these processes by training the animat brain in virtual environments. Not being bounded by a physical substrate such as a robot, researchers are able to test thousands of different brains in parallel on high-performance computing resources, such as NSF's TeraGrid, and use the best versions on the robot. The platform the developers selected is the iRobot Create, a robot that looks a lot like the Roomba vacuum-cleaning robot.
Since the animat is not explicitly programmed to solve specific tasks, there is greater flexibility for the robot's prospective functions. Eventually, it will function on an autonomous level and be able to take on more complex adaptive tasks such as intelligently interacting and caring for the elderly, autonomously exploring and collecting samples on an alien planet, and generally employing more humanoid behavior.
This is a challenge for any artificial intelligence program under development: it is simply impossible to program a lifetime's set of knowledge into a robot! That is why it is so important for the next generation of artificial intelligence to be able to learn throughout a lifetime without needing constant reprogramming.
Science fiction is rife with examples of learning robots and HAL 9000 from Kubrick's Space Odyssey will forever come to mind as the media's favorite malfunctioning robot. Although confident about the advent of general intelligence machines in the near future, researchers at the Neuromorphics Lab are optimistic that misbehaving robots like HAL will live only in science fiction movies. Future robots will not be programmed, but will be trained. The key is to educate them well!
This work was partially funded by the Center of Excellence for Learning in Education, Science and Technology (CELEST), a National Science Foundation Science of Learning Center (NSF OMA-0835976) and by the DARPA SyNAPSE program, contract HR0011-09-3-0001. The views, opinions, and/or findings contained in this article are those of the authors and should not be interpreted as representing the official views or policies, either expressed or implied, of the Defense Advanced Research Projects Agency, the Department of Defense, or the National Science Foundation.
Editor's Note: This research was supported by the National Science Foundation (NSF), the federal agency charged with funding basic research and education across all fields of science and engineering. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the author and do not necessarily reflect the views of the National Science Foundation. See the Behind the Scenes Archive.
MORE FROM LiveScience.com