Squirrels, Birds Teach Robots To Deceive
Military robots learn from animals how to trick their adversaries.
Deception is something that people do all the time - and it plays an important role in military strategy. Now some researchers are trying to figure out how to get robots to do it, by looking at the behavior of squirrels and birds.
At Georgia Tech, a team led by Ronald Arkin, a professor at the School of Interactive Computing, studied the literature on squirrels hiding their caches of acorns. Squirrels will hide their food in a certain place, but when they see other squirrels trying to steal from them, they attempt to fool the thieves by running to a false location.
Ronald Arkin, a professor in Georgia Tech's School of Interactive Computing, and his Ph.D. student Jaeeun Shim, used that as a model for robot behavior. They programmed the robot into tricking a "predator" machine by doing what a squirrel does: showing the adversary a false location for an important resource.
The team also looked at how other animals -– in this case, a species of bird called an Arabian babbler –- drive off predators. Babblers will make an alarm call when they see a predator and other babblers will join the bird and make more calls. They then mob the predator, all the while flapping wings and making noise. The babblers don't ever actually fight the animal they want to drive off; they just make enough noise and flap around enough that attacking a babbler seems like it isn't worth it.
Arkin and and Ph.D. student Justin Davis found that the deception works when the group reaches a certain size - essentially, when enough backup arrives to convince the adversary that it's best to back off. Davis modeled that behavior in software using a military scenario and found that it worked even if the group didn't have the firepower to confront the enemy directly.
The military is interested in this because a robot that can fool an opponent is a valuable tool. It could lead an enemy down a false trail or make itself look more dangerous than it actually is.
The work is an extension of similar
algorithms using a kind of hide-and-seek game.
If robots can fool other robots – or people – that does raise interesting ethical problems. When does fooling people become dangerous?
How do you tell the robot when the right time to do that is? We won't be seeing anything like the Terminator anytime soon, but we already have drones, and the military has explored the use of autonomous
concern over robots that can make targeting decisions - the ability to deceive would complicate that.
via Georgia Tech Credit: Tetra Images/Corbis