Filtering web images effectively

Jun 25, 2012 By Rob Knies

You’re looking for a photo of a flower. Not just any photo—it needs to be horizontal in shape. And not just any flower—it needs to be a purple flower.

What do you do? You could perform a conventional image search on the web. There are lots of flowers out there—lots of shapes, lots of colors. Poke around for a while, and you just might find what you need.

Alternatively, you can use the filter bar in Bing Image Search, which has been augmented by work from Microsoft Research Asia. You type in a textual query: “flower” and filter for “purple,” “photograph,” and “wide,” and voilà, a collection of horizontal shots of purple flowers pops up.

The color filter is thanks, in large part, to research by Jingdong Wang and Shipeng Li. They are in Providence, R.I., from June 16 to 21, attending the Institute of Electrical and Electronics Engineers’ 2012 Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2012), during which they are presenting their paper Salient Object Detection for Searched Web Images via Global Saliency, written in collaboration with Peng Wang, Gang Zeng, Jie Feng, and Hongbin Zha of the Key Laboratory on Machine Perception at Peking University.

That paper is part of a substantial conference contribution from Microsoft Research, which had no fewer than 41 peer-reviewed papers accepted for CVPR 2012. Andrew Blake, Microsoft distinguished scientist and managing director of Microsoft Research Cambridge, is program chair for the event, and Jian Sun of Microsoft Research Asia and Richard Szeliski of Microsoft Research Redmond are serving as area chairs.

A “salient object” is the primary component in an image. Imagine a photo of a horse in a field under a blue sky. The grass might be green, and the sky might be gorgeous, but in all likelihood, the salient object in the shot is the horse.

“Our goal is to develop an effective and efficient technique to locate the salient object,” Wang explains. “Particularly, we also predict if an image contains a salient object, which has been rarely studied before.”

The detection of salient objects in images has been—for a long time. Such an ability has broad applications—image cropping, adaptive image displays on mobile devices, extracting dominant colors within images, removing images lacking an object of interest. But localizing salient objects remains a challenge.

The problem is that objects have a variety of visual characteristics, making it difficult to differentiate salient objects from an image background simply by appearance. And while low-resolution thumbnail images, which proliferate on the web, are recognizable by humans, previous detection methods make it difficult to get the reliable image segmentation those methods require for success.

The Microsoft Research Asia researchers, though, use a learning approach called a “random forest,” an algorithm that clusters data points into groups, to predict the existence and the position of a salient object in an image.

“The key,” Wang explains, “is to describe an image using a global saliency description and to conduct a classification stage, to check the existence of the salient object, and a regression stage, to check the location.”

If a salient object can be identified, using the researchers’ techniques, its dominant color is extracted. If a salient object can’t be identified—which could be possible, for instance, in a landscape shot—the dominant color of the entire image is extracted.

To validate their results, Li and Wang constructed a large image database consisting of hundreds of thousands of manually labeled web images from Bing image search and deployed their algorithm to identify the existence, predict the location, and identify the color of the salient object in thumbnail .

Subsequent tests determined that the researchers’ algorithm significantly outperforms existing state-of-the-art methods.

“We are pleased to see that the results justified our hypothesis that salient objects essentially share common patterns,” Wang says, “even though the objects are generally different. That motivated us to investigate machine-learning tools to solve the problems.”

Explore further: Researchers use Twitter to predict crime

More information: research.microsoft.com/en-us/um/people/jingdw/pubs/CVPR12-Saliency.pdf

Provided by Microsoft Corporation

not rated yet
add to favorites email to friend print save as pdf

Related Stories

Information processing: Adding a touch of color

Jun 08, 2012

Creating a high-quality realistic color image from a grayscale picture can be challenging. Conventional methods typically require the user’s input, either by using a scribbling tool to color the image ...

Faster computer graphics

Jun 13, 2011

Photographs of moving objects are almost always a little blurry — or a lot blurry, if the objects are moving rapidly enough. To make their work look as much like conventional film as possible, game and ...

2-D photos spring to 3-D life

Jun 16, 2011

You’re interested in purchasing a car you’ve seen on the web. It’s the right make, model and vintage. It seems to be in great shape, and it’s just the right color. The price seems reasonable. So what’s ...

Recommended for you

User comments : 0

More news stories

Finnish inventor rethinks design of the axe

(Phys.org) —Finnish inventor Heikki Kärnä is the man behind the Vipukirves Leveraxe, which is a precision tool for splitting firewood. He designed the tool to make the job easier and more efficient, with ...

Making graphene in your kitchen

Graphene has been touted as a wonder material—the world's thinnest substance, but super-strong. Now scientists say it is so easy to make you could produce some in your kitchen.

Poll: Big Bang a big question for most Americans

Few Americans question that smoking causes cancer. But they have more skepticism than confidence in global warming, the age of the Earth and evolution and have the most trouble believing a Big Bang created the universe 13.8 ...