Gesture-driven computers

Feb 29, 2008

It isn’t always easy to communicate with a computer. Two Fraunhofer Institutes will be presenting new possibilities of man-machine interaction at CeBIT in Hanover (Germany) on March 4 through 9. They will demonstrate how computers can be operated simply by gesturing or pointing a finger.

A man stands in front of a large screen gesticulating in a seemingly hectic manner. As if by magic, images suddenly appear on the display. Their movements follow the actor’s gestures, rotate at the slightest turn of a finger, and become larger or smaller as desired. This scene will look familiar to anyone who has watched the science fiction film ‘Minority Report’. Paul Chojecki, scientist and project manager at the Fraunhofer Institute for Telecommunications, Heinrich-Hertz-Institut, HHI in Berlin, explains the iPoint Presenter in a manner reminiscent of the John Anderton character played by Tom Cruise.

At the heart of the system is a set of cameras which enable the computer to observe the person standing in front of the projection screen. The moment this person moves their hands, the computer reacts without being touched at all. “It begins by determining the position of the user’s index finger, then follows its movements,” Chojecki explains. The user can point to buttons or use gestures to move virtual objects. Through ‘multipointing interaction’, i.e. commands using multiple fingers, he can rotate, enlarge or minimize objects. This requires neither special gloves nor any particular markings. Anyone can intuitively operate the device with their bare hands without any preparation whatsoever.

The iPoint Presenter will be demonstrated for the first time at CeBIT 2008, using the example of an interactive game and a photo viewer. But these are just two of the many possible applications it can be used for. It could replace touch screens at info terminals, for example, or help to edit and organize photos. “What is special about it is that the human-computer cummunication is entirely contact-free. The system is therefore ideal for scenarios in which contact between the user and the computer is not allowed or not possible, for example in an operating theater,” says Chojecki.

The system would also be ideal for presentations in large auditoriums. The speaker would no longer need a mouse or a laser pointer, and could click through the presentation and highlight important information simply by pointing. A particularly useful feature for situations like these is that the system can be extended to as many as nine cameras. This immensely increases the user’s operating range and enables them to interact with very large screens, for example at trade fairs or advertising events.

Identifying gestures

Gestures enable people of different nationalities to communicate without the need for spoken words. How useful would it be if this type of communication were also possible between humans and technical devices? This form of giving commands would make many situations safer and more pleasant than they are today. Drivers, for example, could operate their car radios and navigators more easily, and TV viewers at home in their armchairs would no longer need a remote control to flick through the channels.

A whole new generation of video games could be created if the technology involved were able to identify and interpret human gestures. Even machines, household appliances or video conference systems could be controlled by mere hand signals. The system could also be of help to physically disabled people, enabling them to interact with a computer without the need for a mouse and keyboard.

To translate these scenarios into reality as soon as possible, researchers at the Fraunhofer Institute for Digital Media Technology IDMT in Ilmenau are now teaching computers to understand human gestures, and are developing a method of automatically recognizing different hand signals. “Our work is based on optical pattern recognition,” explains IDMT project manager Valiantsin Hardzeyeu. “This technique mimics the way in which humans see things. To this end, we modeled the processes taking place in the human visual apparatus – from the point where the photons hit the retina to the stage in which they are processed in the visual cortex – in a computer simulation.”

A first prototype, which comprises an ‘intelligent’ camera connected to a computer with this new type of pattern recognition software, will be presented at the Fraunhofer stand. The camera will record visitors’ gestures, and the software behind it will analyze them and convert the hand signals into machine commands.

Source: Fraunhofer-Gesellschaft

Explore further: Coping with floods—of water and data

add to favorites email to friend print save as pdf

Related Stories

Skin icons can tap into promise of smartwatch

Oct 21, 2014

You have heard it before: smartwatches are cool wearables but critics remind us of the fact that their small size makes many actions cumbersome and they question how many people will really have them on their ...

Recommended for you

Coping with floods—of water and data

Dec 19, 2014

Halloween 2013 brought real terror to an Austin, Texas, neighborhood, when a flash flood killed four residents and damaged roughly 1,200 homes. Following torrential rains, Onion Creek swept over its banks and inundated the ...

Cloud computing helps make sense of cloud forests

Dec 17, 2014

The forests that surround Campos do Jordao are among the foggiest places on Earth. With a canopy shrouded in mist much of time, these are the renowned cloud forests of the Brazilian state of São Paulo. It is here that researchers ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.