Is my robot happy to see me?

Oct 19, 2009
Graduate student Jenay Beer tested adults ability to read a robots expressions.

(PhysOrg.com) -- People are social creatures. Robots... not so much. When we think of robots, we think of cold, metallic computers without emotion. If science fiction has taught us anything, though, it's that we crave emotion, even in our robots - think C-3PO or Star Trek's Data. So it stands to reason that if robots are ever going to become a fixture in our society, even becoming integrated into our households, we need to be able to read their faces. But how good are we at reading robot faces?

Scientists at Georgia Tech decided to test our ability to interpret a robot's "emotion" by reading its expression to see if there were any differences between the ages. They found that older adults showed some unexpected differences in the way they read a robot's face from the way younger adults performed. The research is being presented this week at the Human Factors and Ergonomics Society Annual Meeting in San Antonio.

"Home-based assistive robots have the potential to help older adults age in place. They have the potential to keep older adults independent longer, reduce healthcare needs and provide everyday assistance," said Jenay Beer, graduate student in Georgia Tech's School of Psychology.

Beer, along with Wendy Rogers and Arthur Fisk, professors of Engineering Psychology at Georgia Tech and directors of the Human Factors and Aging Laboratory, used a virtual version of the iCat, called appropriately enough the virtual iCat, to test the difference among adults between the ages of 65 and 75 and 18 to 27. They had the virtual iCat exhibit seven emotions at various levels of intensity: happiness, sadness, anger, fear, surprise, disgust and neutral. They tested how well each participant could read the emotions of the virtual iCat.

Existing research on how well adults can recognize emotions on human faces has found that older adults are less accurate in recognizing anger, fear and sadness. But the robotic study found that older adults were less accurate in recognizing anger and fear, as expected, but had difficulty recognizing happiness, not sadness. In fact, they most often confused the happy expression with the neutral expression of the robot.

Beer reasoned that the similar success both younger and older adults had in recognizing sadness could be due to the difference in the way a human actually expresses an emotion and the way it's exaggerated in art.

"It may be due to the 'cartoon' look of the iCat, with the mouth turned down being very prominent," she said.

As for why the older adults had trouble recognizing the happy robot compared with their success in recognizing happy people, Beer suspected that the robot just didn't do a good enough job of expressing its emotion.

"It may be that older adults were not as cognizant of the facial features differentiating happy from neutral," she explained.

Researchers also found that neither the young nor old could easily distinguish the emotion disgust on the virtual iCat. Beer explained that this could be due to the difficulty in programming a robot to show this emotion.

"When humans express disgust, the nose is wrinkled and the lips are drawn back, creating creases on each side of the mouth," said Beer. "Manipulating these wrinkles is difficult for a made with a plastic face."

Beer is continuing her work by studying whether other virtual versions of robots show the same differences when compared to the virtual iCat and the human face. What seems clear already, though, is if robots are going to be accepted by older adults in any social situations, they need to be designed with emotion displays that are easy to recognize, with some of them potentially being exaggerated to overcome any trouble may have in reading that emotion among human faces.

Source: Georgia Institute of Technology

Explore further: In funk music, rhythmic complexity influences dancing desire

add to favorites email to friend print save as pdf

Related Stories

Mood Affects Young and Old Differently, Study Finds

Mar 15, 2006

The effect of mood on how people process information changes greatly as they age, suggests new research from the Georgia Institute of Technology. The study, which offers a window into the changing nature of the aging mind ...

Linguistics may be clue to emotions

Jan 20, 2005

Words may be a clue to how people, regardless of their language, think about and process emotions, according to a Penn State researcher. "It has been suggested in the past that all cultures have in common a small number of ...

Recommended for you

Screenagers face troubling addictions from an early age

14 hours ago

In 1997, Douglas Rushkoff boldly predicted the emergence a new caste of tech-literate adolescents. He argued that the children of his day would soon blossom into "screenagers", endowed with effortless advantages over their parents, ...

Better memory at ideal temperature

14 hours ago

People's working memory functions better if they are working in an ambient temperature where they feel most comfortable. That is what Leiden psychologists Lorenza Colzato and Roberta Sellaro conclude after having conducted ...

User comments : 0

More news stories

Down's chromosome cause genome-wide disruption

The extra copy of Chromosome 21 that causes Down's syndrome throws a spanner into the workings of all the other chromosomes as well, said a study published Wednesday that surprised its authors.

Progress in the fight against quantum dissipation

(Phys.org) —Scientists at Yale have confirmed a 50-year-old, previously untested theoretical prediction in physics and improved the energy storage time of a quantum switch by several orders of magnitude. ...

Meteorites yield clues to Martian early atmosphere

(Phys.org) —Geologists who analyzed 40 meteorites that fell to Earth from Mars unlocked secrets of the Martian atmosphere hidden in the chemical signatures of these ancient rocks. Their study, published ...