Research streamlines data processing to solve problems more efficiently

Mar 10, 2010

Researchers at North Carolina State University have developed a new analytical method that opens the door to faster processing of large amounts of information, with applications in fields as diverse as the military, medical diagnostics and homeland security.

"The problem we address here is this: When faced with a large amount of data, how do you determine which pieces of that information are relevant for solving a specific problem," says Dr. Joel Trussell, a professor of electrical and at NC State and co-author of a paper describing the research. "For example, how would you select the smallest number of features that would allow a to differentiate between water and solid ground, based on visual data collected by video?"

This is important, because the more data you need to solve a problem, the more expensive it is to collect the data and the longer it will take to process the data. "The work we've done here allows for a more efficient collection of data by targeting exactly what information is most important to the decision-making process," Trussell says. "Basically, we've created a new that can be used to determine how much data is needed to make a decision with a minimal rate of error."

One application for the new algorithm, discussed in the paper, is for the development of programs that can analyze hyperspectral data from military cameras in order to identify potential targets. Hyperspectral technology allows for finer resolution of the wavelengths of light that are visible to the human eye, though it can also collect information from the infrared spectrum - which can be used to identify specific materials, among other things. The algorithm could be used to ensure that such a program would operate efficiently, minimizing needs and allowing the data to be processed more quickly.

But Trussell notes that "there are plenty of problems out there where people are faced with a vast amount of data, visual or otherwise, - such as medical situations, where doctors may have the results from multiple imaging tests. For example, the algorithm would allow the development of a more efficient screening process for evaluating medical images - such as mammograms - from a large group of people."

Another potential application would be for biometrics, such as homeland security efforts to identify terrorists and others on the Department of Homeland Security watchlist based on video and camera images.

Explore further: New frontier in error-correcting codes

More information: The work is published in the March issue of IEEE Transactions on Knowledge and Data Engineering.

Related Stories

New tool enables powerful data analysis

Jan 08, 2009

(PhysOrg.com) -- A powerful computing tool that allows scientists to extract features and patterns from enormously large and complex sets of raw data has been developed by scientists at University of California, ...

New research advances voice security technology

Mar 08, 2010

Most people are familiar with security technology that scans a person's handprint or eye for identification purposes. Now, thanks in part to research from North Carolina State University, we are closer to practical technology ...

Recommended for you

New frontier in error-correcting codes

20 hours ago

Error-correcting codes are one of the glories of the information age: They're what guarantee the flawless transmission of digital information over the airwaves or through copper wire, even in the presence ...

Five ways the superintelligence revolution might happen

Sep 26, 2014

Biological brains are unlikely to be the final stage of intelligence. Machines already have superhuman strength, speed and stamina – and one day they will have superhuman intelligence. This is of course ...

User comments : 0