Research streamlines data processing to solve problems more efficiently

Mar 10, 2010

Researchers at North Carolina State University have developed a new analytical method that opens the door to faster processing of large amounts of information, with applications in fields as diverse as the military, medical diagnostics and homeland security.

"The problem we address here is this: When faced with a large amount of data, how do you determine which pieces of that information are relevant for solving a specific problem," says Dr. Joel Trussell, a professor of electrical and at NC State and co-author of a paper describing the research. "For example, how would you select the smallest number of features that would allow a to differentiate between water and solid ground, based on visual data collected by video?"

This is important, because the more data you need to solve a problem, the more expensive it is to collect the data and the longer it will take to process the data. "The work we've done here allows for a more efficient collection of data by targeting exactly what information is most important to the decision-making process," Trussell says. "Basically, we've created a new that can be used to determine how much data is needed to make a decision with a minimal rate of error."

One application for the new algorithm, discussed in the paper, is for the development of programs that can analyze hyperspectral data from military cameras in order to identify potential targets. Hyperspectral technology allows for finer resolution of the wavelengths of light that are visible to the human eye, though it can also collect information from the infrared spectrum - which can be used to identify specific materials, among other things. The algorithm could be used to ensure that such a program would operate efficiently, minimizing needs and allowing the data to be processed more quickly.

But Trussell notes that "there are plenty of problems out there where people are faced with a vast amount of data, visual or otherwise, - such as medical situations, where doctors may have the results from multiple imaging tests. For example, the algorithm would allow the development of a more efficient screening process for evaluating medical images - such as mammograms - from a large group of people."

Another potential application would be for biometrics, such as homeland security efforts to identify terrorists and others on the Department of Homeland Security watchlist based on video and camera images.

Explore further: Coping with floods—of water and data

More information: The work is published in the March issue of IEEE Transactions on Knowledge and Data Engineering.

Related Stories

New tool enables powerful data analysis

Jan 08, 2009

(PhysOrg.com) -- A powerful computing tool that allows scientists to extract features and patterns from enormously large and complex sets of raw data has been developed by scientists at University of California, ...

New research advances voice security technology

Mar 08, 2010

Most people are familiar with security technology that scans a person's handprint or eye for identification purposes. Now, thanks in part to research from North Carolina State University, we are closer to practical technology ...

Recommended for you

Coping with floods—of water and data

Dec 19, 2014

Halloween 2013 brought real terror to an Austin, Texas, neighborhood, when a flash flood killed four residents and damaged roughly 1,200 homes. Following torrential rains, Onion Creek swept over its banks and inundated the ...

Cloud computing helps make sense of cloud forests

Dec 17, 2014

The forests that surround Campos do Jordao are among the foggiest places on Earth. With a canopy shrouded in mist much of time, these are the renowned cloud forests of the Brazilian state of São Paulo. It is here that researchers ...

User comments : 0

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.