Researchers develop new image-recognition software

May 21, 2008
What do you see?
Question: What do you see in the red circles? A bottle, a cell phone, a person, a shoe? The answer: They're all the same. Professor Antonio Torralba created these low-resolution images, in which the circled shapes were inserted and are all identical, to demonstrate how context affects our recognition of objects. Even the 'car' in the lower left image is the same object. Photo / Antonio Torralba

It takes surprisingly few pixels of information to be able to identify the subject of an image, a team led by an MIT researcher has found. The discovery could lead to great advances in the automated identification of online images and, ultimately, provide a basis for computers to see like humans do.

Antonio Torralba, assistant professor in MIT's Computer Science and Artificial Intelligence Laboratory, and colleagues have been trying to find out what is the smallest amount of information--that is, the shortest numerical representation--that can be derived from an image that will provide a useful indication of its content.

Deriving such a short representation would be an important step toward making it possible to catalog the billions of images on the Internet automatically. At present, the only ways to search for images are based on text captions that people have entered by hand for each picture, and many images lack such information. Automatic identification would also provide a way to index pictures people download from digital cameras onto their computers, without having to go through and caption each one by hand. And ultimately it could lead to true machine vision, which could someday allow robots to make sense of the data coming from their cameras and figure out where they are.

"We're trying to find very short codes for images," says Torralba, "so that if two images have a similar sequence [of numbers], they are probably similar--composed of roughly the same object, in roughly the same configuration." If one image has been identified with a caption or title, then other images that match its numerical code would likely show the same object (such as a car, tree, or person) and so the name associated with one picture can be transferred to the others.

"With very large amounts of images, even relatively simple algorithms are able to perform fairly well" in identifying images this way, says Torralba. He will be presenting his latest findings this June in Alaska at a conference on Computer Vision and Pattern Recognition. The work was done in collaboration with Rob Fergus at the Courant Institute in New York University and Yair Weiss of Hebrew University in Jerusalem.

To find out how little image information is needed for people to recognize the subject of a picture, Torralba and his co-authors tried reducing images to lower and lower resolution, and seeing how many images at each level people could identify.

"We are able to recognize what is in images, even if the resolution is very low, because we know so much about images," he says. "The amount of information you need to identify most images is about 32 by 32." By contrast, even the small "thumbnail" images shown in a Google search are typically 100 by 100.

Even an inexpensive current digital camera produces images consisting of several megapixels of data--and each pixel typically consists of 24 bits (zero or one) of data. But Torralba and his collaborators devised a mathematical system that can reduce the data from each picture even further, and it turns out that many images are recognizable even when coded into a numerical representation containing as little as 256 to 1024 bits of data.

Using such small amounts of data per image makes it possible to search for similar pictures through millions of images in a database, using an ordinary PC, in less than a second, Torralba says. And unlike other methods that require first breaking down an image into sections containing different objects, this method uses the entire image, making it simple to apply to large datasets without human intervention.

For example, using the coding system they developed, Torralba and his colleagues were able to represent a set of 12.9 million images from the Internet with just 600 megabytes of data--small enough to fit in the RAM memory of most current PCs, and to be stored on a memory stick. The image database and software to enable searches of the database, are being made publicly available on the web.

Of course, a system using drastically reduced amounts of information can't come close to perfect identification. At present, the matching works for the most common kinds of images. "Not all images are created equal," he says. The more complex or unusual an image is, the less likely it is to be correctly matched. But for the most common objects in pictures--people, cars, flowers, buildings--the results are quite impressive.

The work is part of research being carried out by hundreds of teams around the world, aimed at analyzing the content of visual information. Torralba has also collaborated on related work with other MIT researchers including William Freeman, a professor in the Department of Electrical Engineering and Computer Science; Aude Oliva, professor in the Department of Brain and Cognitive Sciences; and graduate students Bryan Russell and Ce Liu, in CSAIL. Torralba's work is supported in part by a grant from the National Science Foundation.

Torralba stresses that the research is still preliminary and that there will always be problems with identifying the more-unusual subjects. It's similar to the way we recognize language, Torralba says. "There are many words you hear very often, but no matter how long you have been living, there will always be one that you haven't heard before. You always need to be able to understand [something new] from one example."

Source: MIT

Explore further: Researcher develops method for monitoring whether private information is sufficiently protected

add to favorites email to friend print save as pdf

Related Stories

Gesture-based computing on the cheap (w/ Video)

May 20, 2010

Ever since Steven Spielberg’s 2002 sci-fi movie Minority Report, in which a black-clad Tom Cruise stands in front of a transparent screen manipulating a host of video images simply by waving his hands, the ...

Recommended for you

Tackling urban problems with Big Data

16 hours ago

Paul Waddell, a city planning professor at the University of California, Berkeley, with a penchant for conducting research with what he calls his "big urban data," is putting his work to a real-world test ...

Computer-assisted accelerator design

Apr 22, 2014

Stephen Brooks uses his own custom software tool to fire electron beams into a virtual model of proposed accelerator designs for eRHIC. The goal: Keep the cost down and be sure the beams will circulate in ...

User comments : 3

Adjust slider to filter visible comments by rank

Display comments: newest first

Maynard_G
1.7 / 5 (3) May 21, 2008
Got to be kidding. The circles are obviously not the same. For example, only the upper left hand circle has yellow in it.
Valentiinro
4.5 / 5 (2) May 21, 2008
No, not the circles fillings, but the object in the center of the circle. The thing in the center is the same object at different angles and with different stuff around it.
Sophos
4 / 5 (1) May 22, 2008
Maynard, I think they mean they are all rectangles of similar shading and aspect ratio

I agree with you its poorly worded

More news stories

Genetic code of the deadly tsetse fly unraveled

Mining the genome of the disease-transmitting tsetse fly, researchers have revealed the genetic adaptions that allow it to have such unique biology and transmit disease to both humans and animals.