Researcher sees new angles in visual search

Researcher sees new angles in visual search
Engineering professor Shih-Fu Chang is developing technology to make visual search as effortless as typing keywords into a search engine. Image credit: Eileen Barroso/Columbia University

Engineering professor Shih-Fu Chang is trying to make visual search technology as effortless as typing a keyword like “Morningside restaurants” into Google.

In Chang’s ideal world, you could sift through a season’s worth of Major League Baseball games to find every double play or sort through your digital archives to find all the pictures you’ve taken of your kid blowing out her birthday candles.

Chang, the Richard Dicker Professor of Telecommunications, who has a joint appointment in electrical engineering and computer science, develops algorithms to identify and index data, as well as new techniques and software systems to help users manage large amounts of multimedia information. The director of Columbia Engineering’s Digital Video and Multimedia Lab, Chang was recently honored with a lifetime achievement award from the Association for Computing Machinery’s Special Interest Group on Multimedia.

Chang credits his associates for the prestigious award. “It really should be attributed to all the wonderful students and collaborators I have had the good fortune to work with during my career,” he said.

A native of Taiwan, he was fascinated by technology from a young age, he remembers getting his first computer, an Apple II, as a high school student in 1981 and envying classmates who were fortunate enough to have an Atari computer system.

“I grew up in a time when the information technology industry was rapidly expanding, and it caught the imagination of young people,” Chang recalls. “My parents didn’t know anything about computers, but we as students caught the wave and jumped on it.” He received his Ph.D. from the University of California, Berkeley, and joined Columbia in 1993.

In 1998, he developed one of the first video systems, VideoQ, and in the 1990s he pioneered search by sketch technology, in which a user can draw what he or she is looking for.

His work has been broadly funded by government and industry, including Eastman Kodak, and many video indexing technologies developed by his group have been licensed to companies. With the support of the National Science Foundation, he is working on technology to determine whether images and videos have been tampered with.

Recently, he’s been working on technology that lets users adjust the importance of multiple characteristics in a search. If you’re looking for an image of a sun setting over a mountain near a body of water, for instance, you would enter the search words “sunset,” “mountain” and “water” into a grid next to a database of images. Then you could home in on a specific feature—the mountain, for instance—by using your mouse to nudge the cursor closer to that word. All the images in the database would be indexed automatically without time-consuming human tagging.

Other research projects let users upload a specific image they want to match, where a photo can be matched with images all over the Internet. Chang is working on refining technology that allows people to specify the exact portion of an image to search for—a certain sculpture or building, for example—so that searching is more accurate and efficient, even on mobile devices like iPhones. His research could also be applied in the medical field, allowing technicians to find a moment in an ultrasound film that could help make a diagnosis.

Another aspect of his research involves using the human brain as a search tool. In collaboration with biomedical engineering professor Paul Sajda, Chang designed a device that monitors brain activity as a subject looks at pictures. First, an EEG machine records the “aha” moments of recognition, then a computer analyzes the recognition patterns to identify similarities in other photographs in large databases that elicited the same strong reaction.

“The machine does what it’s best at, and the human does what he or she is best at in the most natural way,” Chang says.

Citation: Researcher sees new angles in visual search (2011, October 26) retrieved 23 April 2024 from https://phys.org/news/2011-10-angles-visual.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.

Explore further

Professor develops mobile app to identify plant species

0 shares

Feedback to editors