Generating 'oohs' and 'aahs': Vocal Joystick uses voice to surf the Internet

Oct 09, 2007
Vocal Joystick uses voice to surf the Internet
Doctoral student Brandi House uses Vocal Joystick to control the movement of a robotic arm. The screen on the lower right shows how the software analyzes her vocalizations to create instructions for the arm's movement. Credit: University of Washington

The Internet offers wide appeal to people with disabilities. But many of those same people find it frustrating or impossible to use a handheld mouse. Software developed at the University of Washington provides an alternative using one of the oldest and most versatile modes of communication: the human voice.

"There are many people who have perfect use of their voice who don't have use of their hands and arms," said Jeffrey Bilmes, a UW associate professor of electrical engineering. "I think there are several reasons why Vocal Joystick might be a better approach, or at least a viable alternative, to brain-computer interfaces." The tool's latest developments will be presented this month in Tempe, Ariz. at the Assets Conference on Computers and Accessibility.

Vocal Joystick detects sounds 100 times a second and instantaneously turns that sound into movement on the screen. Different vowel sounds dictate the direction: "ah," "ee," "aw" and "oo" and other sounds move the cursor one of eight directions. Users can transition smoothly from one vowel to another, and louder sounds make the cursor move faster. The sounds "k" and "ch" simulate clicking and releasing the mouse buttons.

Versions of Vocal Joystick exist for browsing the Web, drawing on a screen, controlling a cursor and playing a video game. A version also exists for operating a robotic arm, and Bilmes believes the technology could be used to control an electronic wheelchair.

Existing substitutes for the handheld mouse include eye trackers, sip-and-puff devices, head-tracking systems and other tools. Each technology has drawbacks. Eye-tracking devices are expensive and require that the eye simultaneously take in information and control the cursor, which can cause confusion. Sip-and-puff joysticks held in the mouth must be spit out if the user wants to speak, and can be tiring. Head-tracking devices require neck movement and expensive hardware.

Vocal Joystick requires only a microphone, a computer with a standard sound card and a user who can produce vocal sounds.

"A lot of people ask: 'Why don't you just use speech recognition"'" Bilmes said. "It would be very slow to move a cursor using discrete commands like 'move right' or 'go faster.' The voice, however, is able to do continuous commands quickly and easily." Early tests suggest that an experienced user of Vocal Joystick would have as much control as someone using a handheld device.

In the laboratory, doctoral student Jonathan Malkin, who helped develop the tool, uses Vocal Joystick to play a game called Fish Tale. It takes two minutes to train the program for Malkin's voice. He then moves the fish character easily around the screen, raising his voice slightly to speed up and avoid being eaten by a predator fish.

The newest development, which will be presented at the October meeting in Tempe, uses Vocal Joystick to control a robotic arm. The pitch of the tone moves the arm up and down; other commands are unchanged. This is the first time that vocal commands have been used to control a three-dimensional object, Bilmes said.

One initial concern, he said, was whether people would feel self-conscious using the tool.

"But once you try it you immediately forget what you're saying," Bilmes said. "I usually go to the New York Times' Web site to test the system and then I get distracted and start reading the news. I forget that I'm using it."

To test the device, the group has been working with about eight spinal-cord injury patients at the UW Medical Center since March.

"It's a really exciting idea. I think it has tremendous potential," said Kurt Johnson, a professor of rehabilitation medicine who is helping with the tests.

Bilmes said he hopes people will become more adept at using the system over time. Future research will incorporate more advanced controls that use more aspects of the human voice, such as repeated vocalizations, vibrato, degree of nasality and trills.

"While people use their voices to communicate with just words and phrases," Bilmes said, "the human voice is an incredibly flexible instrument, and can do so much more."

Source: University of Washington

Explore further: Researchers develop new program to evaluate prominent individuals' personalities

add to favorites email to friend print save as pdf

Related Stories

Recommended for you

Watching others play video games is the new spectator sport

Aug 29, 2014

As the UK's largest gaming festival, Insomnia, wrapped up its latest event on August 25, I watched a short piece of BBC Breakfast news reporting from the festival. The reporter and some of the interviewees appeared baff ...

SHORE facial analysis spots emotions on Google Glass

Aug 28, 2014

One of the key concerns about facial recognition software has been over privacy. The very idea of having tracking mechanisms as part of an Internet-connected wearable would be likely to upset many privacy ...

User comments : 0