The Next Level in Robots: Monkey See, Monkey Do, Monkey Create

Oct 29, 2007 by Mary Anne Simpson weblog
The  Next Level in Robots: Monkey See, Monkey Do, Monkey Create
Ancient Japanese Three Wise Monkeys - Photo Credit: Wikipedia

The next level of robot is currently in the research and development stage in Japan's National Institute of Information and Communication Technology. The next level of robot untethered by human omnipresence allows it to take cues from gestures and make immediate and appropriate responses.

The Japanese, National Institute of Information and Communications Technology is working on a project wherein machines can learn and teach themselves what to do. Presently, robots are tethered to human commands or guided by programs in advance that operate in real time. The new level of robot will take cues from gestures and operate more autonomously through a learning process.

The Institutes's Spoken Language Division is in the development stage of creating a robot that measure 155 cm and weighs 85 kg that learns through gestures, thereby creating a more autonomous robot. The Spoken Language Group´s main focus is to develop an information communication system that understands when people talk correctly and automatically takes appropriate actions to people and other machines. The actions are based on the knowledge they receive from the talk by people in their presence.

According to the Institute, the current research is involved in producing stress-free unambiguous communication that a machine understands immediately and tells its understanding immediately to a person or another machine. Its primary goal is to establish a technology to give messages to network terminals by people's natural expressions, such as gestures, hand signals and body language that transcend language differences and allow for approximations.

As of this writing, the prototype of this next level of robot has not made its public debut. There are reports of its development. According to Digital World Tokyo, the work in progress robot can understand the gesture of pointing a finger at an object. It can possibly understand the traditional Japanese bow indicating a respectful greeting.

In addition, the new robot can repeat the same gestures in the appropriate circumstances. Specifically it can pointing out a direction and then move in that direction. This indicates the robot has formed its own learning process without being programmed to do so or by a formal teaching command.

Explore further: Socially-assistive robots help kids with autism learn by providing personalized prompts

add to favorites email to friend print save as pdf

Related Stories

Dogs' behavior could help to design social robots

Sep 12, 2013

Designers of social robots, take note. Bring your dog to the lab next time you test a prototype, and watch how your pet interacts with it. You might just learn a thing or two that could help you fine-tune ...

Makr Shakr uses three arms for drink-recipe collabs

May 15, 2013

(Phys.org) —We're told it's the wave of the future. Design, make, enjoy. Beyond home-based 3-D printers, there will be new machines and display screens and apps that will invite you to have day to day products ...

Recommended for you

Ride-sharing could cut cabs' road time by 30 percent

12 hours ago

Cellphone apps that find users car rides in real time are exploding in popularity: The car-service company Uber was recently valued at $18 billion, and even as it faces legal wrangles, a number of companies ...

Jumping into streaming TV

13 hours ago

More TV viewers are picking up so-called streaming media boxes in the hope of fulfilling a simple wish: Let me watch what I want when I want.

Job listing service ZipRecruiter raises $63 million

14 hours ago

ZipRecruiter, a California start-up that tries to simplify tasks for recruiters, has raised $63 million in initial venture capital funding as the 4-year-old service races to keep up with growing demand.

User comments : 0