You might think a robot would be more likely to win people over if it were good at its job. But according to a recent study, people find imperfect robots more likable.
In previous studies, researchers noticed that human subjects reacted differently to robots that made unplanned errors in their tasks. For their new investigation, the study authors programmed a small, humanoid robot to deliberately make mistakes so the scientists could learn more about how that fallibility affected the way people responded to the bots. They also wanted to see how these social cues might provide opportunities for robots to learn from their experiences. [Super-Intelligent Machines: 7 Robotic Futures]
The researchers found that people liked the error-prone robot more than the error-free one, and that they responded to the robot's mistakes with social signals that robots could possibly be trained to recognize, in order to modify future behavior.
For the study, 45 human subjects — 25 men and 20 women — were paired with a robot that was programmed to perform two tasks: ask interview questions, and direct several simple Lego brick assemblies.
For 24 of the users, the robot behaved flawlessly. It posed questions and waited for their responses, and then instructed them to sort the Lego bricks and build towers, bridges and "something creative," ending the exercise by having the person arrange Legos into a facial expression to show a current emotional state, according to the study.
But for 21 people in the study, the robot's performance was less than stellar. Some of the mistakes were technical glitches, such as failing to grasp Lego bricks or repeating a question six times. And some of the mistakes were so-called "social norm violations," such as interrupting while their human partner was answering a question or telling them to throw the Lego bricks on the floor.
The scientists observed the interactions from a nearby station. They tracked how people reacted when the robots made a mistake, gauging their head and body movements, their expressions, the angle of their gaze, and whether they laughed, smiled or said something in response to the error. After the tasks were done, they gave participants a questionnaire to rate how much they liked the robot, and how smart and human-like they thought it was, on a scale from 1 to 5.
The researchers found that the participants responded more positively to the bumbling robot in their behavior and body language, and they said they liked it "significantly more" than the people liked the robot that made no mistakes at all.
However, the subjects who found the error-prone robot more likable didn't see it as more intelligent or more human-like than the robot that made fewer mistakes, the researchers found.
Their results suggest that robots in social settings would probably benefit from small imperfections; if that makes the bots more likable, the robots could possibly be more successful in tasks meant to serve people, the study authors wrote.
And by understanding how people respond when robots make mistakes, programmers can develop ways for robots to read those social cues and learn from them, and thereby avoid making problematic mistakes in the future, the scientists added.
"Future research should be targeted at making a robot understand the signals and make sense of them," the researchers wrote in the study.
"A robot that can understand its human interaction partner's social signals will be a better interaction partner itself, and the overall user experience will improve," they concluded.
The findings were published online May 31 in the journal Frontiers in Robotics and AI.
Original article on Live Science.