Lip-reading computers can detect different languages

Scientists at the University of East Anglia (UEA) have created lip-reading computers that can distinguish between different languages.

Computers that can read lips are already in development but this is the first time they have been 'taught' to recognise different languages. The discovery could have practical uses for deaf people, for law enforcement agencies, and in noisy environments.

Led by Stephen Cox and Jake Newman of UEA's School of Computing Sciences, the groundbreaking research will be presented at a major conference in Taiwan on Wednesday April 22.

The technology was developed by statistical modelling of the lip motions made by a group of 23 bilingual and trilingual . The system was able to identify which was spoken by an individual speaker with very high accuracy. These languages included English, French, German, Arabic, Mandarin, Cantonese, Italian, Polish and Russian.

"This is an exciting advance in automatic lip-reading technology and the first scientific confirmation of something we already intuitively suspected -¬ that when people speak different languages, they use different mouth shapes in different sequences," said Prof Cox.

"For example, we found frequent 'lip-rounding' among French speakers and more prominent tongue movements among Arabic speakers."

Funded by the EPSRC, the research is part of a wider UEA project on automatic lip-reading. The next step will be to make the system more robust to an individual's physiology and his or her way of speaking.

Source: University of East Anglia

Citation: Lip-reading computers can detect different languages (2009, April 22) retrieved 19 April 2024 from https://phys.org/news/2009-04-lip-reading-languages.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Crime fighting potential for computerised lip-reading

0 shares

Feedback to editors