Robotic perception, on purpose

Oct 26, 2009
Robotic perception, on purpose

(PhysOrg.com) -- European researchers developed technology that enables a robot to combine data from both sound and vision to create combined, purposeful perception. In the process, they have taken the field to a new level.

Currently, computer vision is good at recognising objects in images and videos and has been successfully employed in several specialised industrial applications, such as quality control during microchip fabrication.

But robotic perception is much weaker in less defined situations, like understanding and responding to human behaviour and even conversations. Yet, it is precisely this sort of interaction which promises the most compelling applications for future humanoid technology, where people-like robots can act as guides, or mix with people, or use perception to infer appropriate actions.

More importantly, these broad robotic applications will deliver insights into other disciplines, like cognition and neuroscience.

A truly perceptive robot, capable of acting independently and appropriately in complex situations remains a distant goal, but European researchers brought it much closer with their Perception-on-Purpose (POP) project.

Original, by design

“The originality of our project was our attempt to integrate two different sensory modalities, namely sound and vision,” explains Radu Horaud, POP’s coordinator.

“This was very difficult to do, because you are integrating two completely different physical phenomena,” he adds.

Vision works from the reflection of light waves from an object, and it allows the observer to infer certain properties, like size, shape, density and texture. But with sound you are interested in locating the direction of the source, and trying to identify the type of sound it is.

Tricky issue

On its own, sound is difficult to pinpoint, because it needs to be located in a 3D space. Then there is the problem of background noise, such as an open window letting in sounds from next door.

But it turned out that integrating two different senses helped the researchers in their bid to locate and tune into relevant sounds.

“It is not that easy to decide what is foreground and what is background using sound alone, but by combining the two modalities - sound and vision - it becomes much easier,” reveals Horaud.

“If you are able to locate ten sound sources in ten different directions, but if in one of these directions you see a face, then you can much more easily concentrate on that sound and throw out the other ones.”

This video is not supported by your browser at this time.

Integrated technology

This was one approach that the team took and, with the algorithms they developed, their robot, called Popeye, was able to identify the speaker with a fair degree of reliability.

“There is more work to be done on that aspect of the work, it is not completely robust yet,” warns Horaud.

Still, it was a very strong result, and what makes it even more impressive is that the team managed to integrate all the technology into a neat and compact robotic platform.

“Most often, sound research is conducted in specialised labs, with arrays of microphones and a very controlled acoustic environment. But we integrated our two microphones and two cameras onto the head of our Popeye. The idea is to have an agent-centred system,” Horaud stresses.

Powerful technology

The Popeye packs a lot of powerful technology into a small space and offers purposeful robotic perception. This is important because Horaud argues persuasively that, in evolutionary terms, multi-sensory perception and cognition are linked.

By perceiving a hand-held object with their two eyes, for example, monkeys - and the first hominids after them - developed stereo vision and hence were able to learn many properties of an object from combined tactile and visual data. Over time, they developed new skills, including building tools, from this information.

Horaud feels, too, that some modern uses of artificial intelligence (AI), like chess applications, are limited because they do not learn from their environment. They are programmed with abstract data - say, chess moves - and they process that.

“They cannot infer predicates from natural images; they cannot draw abstract information from physical observations,” he stresses.

For now, POP has achieved many of its aims and developed very promising approaches. Commercial applications for this type of technology are not out of the question, and the researchers also hope to continue their work in a further project.

That project would look at extending some of POP’s results into a functioning humanoid robot. In the meantime, POP’s work means that the purposefully perceptive has become a not-so-distant future technology.

More information: perception.inrialpes.fr/POP/

Provided by ICT Results (news : web)

Explore further: Using social media for behavioral studies is cheap, fast, but fraught with biases

add to favorites email to friend print save as pdf

Related Stories

New robot skier takes to the slopes (w/ Video)

Oct 23, 2009

(PhysOrg.com) -- A new robot skier has been invented that can be fitted with off-the-shelf skis. This is not the first skiing robot, since Japanese scientists have produced their own (see PhysOrg.com article here), but is bigger and heavie ...

Two Robot Chefs Make Omelets

Dec 04, 2008

(PhysOrg.com) -- No "house of the future" is complete without a household robot to do the cooking and cleaning. Although today´s robots still have a ways to go before substituting for a real live-in maid, ...

BlackBerry Storm 2 coming soon (w/ Video)

Oct 20, 2009

(PhysOrg.com) -- RIM are soon to release their updated BlackBerry, the Storm 2 smart phone, with a more streamlined design and touch-sensitive buttons instead of the hardware buttons of the first version.

iRobot Unveils Morphing Blob Robot (w/ Video)

Oct 15, 2009

(PhysOrg.com) -- iRobot's latest robot is unique on many levels. The doughy blob moves by inflating and deflating - a new technique its developers call "jamming." As the researchers explain in the video below, ...

Recommended for you

Brain inspired data engineering

19 hours ago

What if next-generation ICT systems could be based on the brain's structure and its cognitive and adaptive processes? A groundbreaking paradigm of brain-inspired intelligent ICT architectures is being born.

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.