Robots and Artificial Intelligence
Artificial intelligence (AI) is arguably the most exciting field in robotics.
It's certainly the most controversial: Everybody agrees that a robot can
work in an assembly line, but there's no consensus on whether a robot can
ever be intelligent.
Like the term "robot" itself, artificial intelligence is hard to define.
Ultimate AI would be a recreation of the human thought process -- a man-made
machine with our intellectual abilities. This would include the ability to
learn just about anything, the ability to reason, the ability to use
language and the ability to formulate original ideas. Roboticists are
nowhere near achieving this level of artificial intelligence, but they have
had made a lot of progress with more limited AI. Today's AI machines can
replicate some specific elements of intellectual ability.
Computers can already solve problems in limited realms. The basic idea of AI
problem-solving is very simple, though its execution is complicated. First,
the AI robot or computer gathers facts about a situation through sensors or
human input. The computer compares this information to stored data and
decides what the information signifies. The computer runs through various
possible actions and predicts which action will be most successful based on
the collected information. Of course, the computer can only solve problems
it's programmed to solve -- it doesn't have any generalized analytical
ability. Chess computers are one example of this sort of machine.
Some modern robots also have the ability to learn in a limited capacity.
Learning robots recognize if a certain action (moving its legs in a certain
way, for instance) achieved a desired result (navigating an obstacle). The
robot stores this information and attempts the successful action the next
time it encounters the same situation. Again, modern computers can only do
this in very limited situations. They can't absorb any sort of information
like a human can. Some robots can learn by mimicking human actions. In Japan,
roboticists have taught a robot to dance by demonstrating the moves
themselves.
Some robots can interact socially. Kismet, a robot at M.I.T's Artificial
Intelligence Lab, recognizes human body language and voice inflection and
responds appropriately. Kismet's creators are interested in how humans and
babies interact, based only on tone of speech and visual cue. This low-level
interaction could be the foundation of a human-like learning system.
Kismet and other humanoid robots at the M.I.T. AI Lab operate using an
unconventional control structure. Instead of directing every action using a
central computer, the robots control lower-level actions with lower-level
computers. The program's director, Rodney Brooks, believes this is a more
accurate model of human intelligence. We do most things automatically; we
don't decide to do them at the highest level of consciousness.
The real challenge of AI is to understand how natural intelligence works.
Developing AI isn't like building an artificial heart -- scientists don't
have a simple, concrete model to work from. We do know that the brain
contains billions and billions of neurons, and that we think and learn by
establishing electrical connections between different neurons. But we don't
know exactly how all of these connections add up to higher reasoning, or
even low-level operations. The complex circuitry seems incomprehensible.
Because of this, AI research is largely theoretical. Scientists hypothesize
on how and why we learn and think, and they experiment with their ideas
using robots. Brooks and his team focus on humanoid robots because they feel
that being able to experience the world like a human is essential to
developing human-like intelligence. It also makes it easier for people to
interact with the robots, which potentially makes it easier for the robot to
learn.