Avatars make the Internet sign to deaf people

Aug 29, 2014
Computer scientists from Saarbrücken in Germany have developed an approach where animated online characters display content in sign language. They collaborate with Peter Schaar (picture) who is deaf and is a lecturer for sign language. Credit: AG Heloir

It is challenging for deaf people to learn a sound-based language, since they are physically not able to hear those sounds. Hence, most of them struggle with written language as well as with text reading and comprehension. Therefore, most website content remains inaccessible for them. Computer scientists from Saarbrücken, Germany, want to change the situation by means of a method they developed: animated online characters display content in sign language. In the long term, deaf people would be able to use the technique to communicate on online platforms via sign language. To realize the technique, users would only need readily available devices.

Sign language was first acknowledged as a separate language in the sixties, which is quite recent. Similar to , it evolved from different cultural backgrounds. Every country has its own with various dialects, which are based on different rules than the spoken language. For the deaf, sign language is their native language. Therefore, it is not easy for them to learn spoken language, which is why they may struggle with text reading and comprehension even after their graduation. Although several websites provide video clips in which sign language interpreters translate the text, much Internet content remains cryptic for the deaf community. To inform quickly in cases where there is no interpreter on hand, researchers are working on a novel approach to provide content. Their idea: avatars. These animated characters could be used in the context of announcements at train stations, or on websites.

"We have already gained initial experiences with avatars", explains Alexis Heloir, who is the leader of the research group "Sign Language Synthesis and Interaction" at the Cluster of Excellence "Multimodal Computing and Interaction" and also a researcher at the German Center for Artificial Intelligence. "If we try to animate them like human beings, deaf people have issues with understanding the avatars." The researcher assumes that this is caused by the greater variety of emotional expressions of humans compared to avatars. To deal with that problem, Alexis Heloir and Fabrizio Nunnari create avatars that make more accentuated movements. The researchers are closely cooperating with Peter Schaar, who is deaf and is a lecturer for German sign language at the Saarland University Language Center and the College of Engineering and Commerce in Saarbrücken.

"Our method should be inexpensive and easy to use so that every member of the deaf community will be able to use it", says Fabrizio Nunnari. To capture the motions of deaf people, the scientists make use of affordable cameras and sensors that are typically used by teenagers for computer games. A computing method transfers the movements of the entire body onto the avatar. In the long term, the researchers want to create a collection of short sign language sequences that can be used by the deaf to interact on the web.

Explore further: Hand gestures improve learning in both signers and speakers

add to favorites email to friend print save as pdf

Related Stories

Early exposure to language for deaf children

Jun 05, 2012

(Medical Xpress) -- Most agree that the earlier you expose a child to a language, the easier it is for that child to pick it up. The same rules apply for deaf children.

Recommended for you

Oculus unveils new prototype VR headset

Sep 20, 2014

Oculus has unveiled a new prototype of its virtual reality headset. However, the VR company still isn't ready to release a consumer edition.

Who drives Alibaba's Taobao traffic—buyers or sellers?

Sep 18, 2014

As Chinese e-commerce firm Alibaba prepares for what could be the biggest IPO in history, University of Michigan professor Puneet Manchanda dug into its Taobao website data to help solve a lingering chicken-and-egg question.

Computerized emotion detector

Sep 16, 2014

Face recognition software measures various parameters in a mug shot, such as the distance between the person's eyes, the height from lip to top of their nose and various other metrics and then compares it with photos of people ...

User comments : 0