Combining engineering and linguistics leads to new discoveries
1:20 p.m., April 12, 2011--Babies learn to talk by listening intently to those around them. They observe which sounds come naturally together, intuitively building rules as to what fits and what does not.
Herbert Tanner believes similar rules exist in robotics, when it comes to deciding the sequence of actions a robot can execute.
Self-assembled materials, InSPACE
Tanner, an assistant professor of mechanical engineering at the University of Delaware, and fellow researcher Jeffrey Heinz, are applying formal language theory and linguistics algorithms to design robots that can “think for themselves.”
“Amazingly, we can use insights from how children learn language to design robots which can likewise learn from their experience,” said Heinz, an assistant professor of linguistics and cognitive science.
This is accomplished through a theoretical modeling and analysis framework that uses a combination of logic (the language of computer science) and differential equations (the language of traditional engineering). The new algorithms will give the robot the tools to devise plans on its own, based on its environment.
Currently, human designers pre-program robots to complete actions in a certain sequence. This is because while planning is easy for humans, it is analytically and computationally challenging for robots.
In one example, a robot is asked to pick up and deliver a computer printout. The robot uses high frequency wireless communications and a network of eight cameras to triangulate its location and its destination. The cameras offer pinpoint accuracy – within one millimeter –measuring the robots’ location one hundred times per second using reflective markers. The sequence of maneuvers involved is currently pre-programmed, but Tanner has a plan to enable the robot to “program” itself.
“We want our robots to learn both from a human instructor and from their environment how to plan their actions,” said Tanner.
In the future, he hopes to include multiple robots, potentially using a combination of both ground and aerial machines.
“A team of robots that exploits the strengths of each member is more valuable than adding new capabilities to a single robot,” he added.
If successful, Tanner’s research will allow robots to observe the behaviors around them and adapt their actions according to their surroundings. He believes this could be useful in emergency response to automatically generate an action plan for how first responders can quickly coordinate relief efforts.
Medics, for example, cannot reach victims unless engineers have cleared the rubble; meanwhile, engineers need to wait for firefighters to put out the flames before they can work.
“It becomes a problem of planning, scheduling and resource allocation,” explained Tanner. “Having an automated planning tool generated by the robots can eliminate uncertainty and confusion and get things going.”
The research is funded by the National Science Foundation’s Cyber-Physical Systems program. Tanner is the principal investigator on the grant, which totals $1 million. Heinz and Calin Belta, a mechanical engineer at Boston University, serve as co-principal investigators on the project.
Article by Karen B. Roberts