Filtering web images effectively

Jun 25, 2012 By Rob Knies

You’re looking for a photo of a flower. Not just any photo—it needs to be horizontal in shape. And not just any flower—it needs to be a purple flower.

What do you do? You could perform a conventional image search on the web. There are lots of flowers out there—lots of shapes, lots of colors. Poke around for a while, and you just might find what you need.

Alternatively, you can use the filter bar in Bing Image Search, which has been augmented by work from Microsoft Research Asia. You type in a textual query: “flower” and filter for “purple,” “photograph,” and “wide,” and voilà, a collection of horizontal shots of purple flowers pops up.

The color filter is thanks, in large part, to research by Jingdong Wang and Shipeng Li. They are in Providence, R.I., from June 16 to 21, attending the Institute of Electrical and Electronics Engineers’ 2012 Computer Society Conference on Computer Vision and Pattern Recognition (CVPR 2012), during which they are presenting their paper Salient Object Detection for Searched Web Images via Global Saliency, written in collaboration with Peng Wang, Gang Zeng, Jie Feng, and Hongbin Zha of the Key Laboratory on Machine Perception at Peking University.

That paper is part of a substantial conference contribution from Microsoft Research, which had no fewer than 41 peer-reviewed papers accepted for CVPR 2012. Andrew Blake, Microsoft distinguished scientist and managing director of Microsoft Research Cambridge, is program chair for the event, and Jian Sun of Microsoft Research Asia and Richard Szeliski of Microsoft Research Redmond are serving as area chairs.

A “salient object” is the primary component in an image. Imagine a photo of a horse in a field under a blue sky. The grass might be green, and the sky might be gorgeous, but in all likelihood, the salient object in the shot is the horse.

“Our goal is to develop an effective and efficient technique to locate the salient object,” Wang explains. “Particularly, we also predict if an image contains a salient object, which has been rarely studied before.”

The detection of salient objects in images has been—for a long time. Such an ability has broad applications—image cropping, adaptive image displays on mobile devices, extracting dominant colors within images, removing images lacking an object of interest. But localizing salient objects remains a challenge.

The problem is that objects have a variety of visual characteristics, making it difficult to differentiate salient objects from an image background simply by appearance. And while low-resolution thumbnail images, which proliferate on the web, are recognizable by humans, previous detection methods make it difficult to get the reliable image segmentation those methods require for success.

The Microsoft Research Asia researchers, though, use a learning approach called a “random forest,” an algorithm that clusters data points into groups, to predict the existence and the position of a salient object in an image.

“The key,” Wang explains, “is to describe an image using a global saliency description and to conduct a classification stage, to check the existence of the salient object, and a regression stage, to check the location.”

If a salient object can be identified, using the researchers’ techniques, its dominant color is extracted. If a salient object can’t be identified—which could be possible, for instance, in a landscape shot—the dominant color of the entire image is extracted.

To validate their results, Li and Wang constructed a large image database consisting of hundreds of thousands of manually labeled web images from Bing image search and deployed their algorithm to identify the existence, predict the location, and identify the color of the salient object in thumbnail .

Subsequent tests determined that the researchers’ algorithm significantly outperforms existing state-of-the-art methods.

“We are pleased to see that the results justified our hypothesis that salient objects essentially share common patterns,” Wang says, “even though the objects are generally different. That motivated us to investigate machine-learning tools to solve the problems.”

Explore further: Coping with floods—of water and data

More information: research.microsoft.com/en-us/u… /CVPR12-Saliency.pdf

Provided by Microsoft Corporation

not rated yet
add to favorites email to friend print save as pdf

Related Stories

Information processing: Adding a touch of color

Jun 08, 2012

Creating a high-quality realistic color image from a grayscale picture can be challenging. Conventional methods typically require the user’s input, either by using a scribbling tool to color the image ...

Faster computer graphics

Jun 13, 2011

Photographs of moving objects are almost always a little blurry — or a lot blurry, if the objects are moving rapidly enough. To make their work look as much like conventional film as possible, game and ...

2-D photos spring to 3-D life

Jun 16, 2011

You’re interested in purchasing a car you’ve seen on the web. It’s the right make, model and vintage. It seems to be in great shape, and it’s just the right color. The price seems reasonable. So what’s ...

Recommended for you

Coping with floods—of water and data

22 hours ago

Halloween 2013 brought real terror to an Austin, Texas, neighborhood, when a flash flood killed four residents and damaged roughly 1,200 homes. Following torrential rains, Onion Creek swept over its banks and inundated the ...

Cloud computing helps make sense of cloud forests

Dec 17, 2014

The forests that surround Campos do Jordao are among the foggiest places on Earth. With a canopy shrouded in mist much of time, these are the renowned cloud forests of the Brazilian state of São Paulo. It is here that researchers ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.