Virginia Tech to tackle the 'Big Data' challenges of next-generation sequencing with HokieSpeed

October 3, 2012
Wu Feng, associate professor of computer science in the College of Engineering at Virginia Tech, will engage in Big Data research with promising advances for genomics. Credit: Virginia Tech

The National Science Foundation (NSF) and the National Institutes of Health (NIH) today announced nearly $15 million in new big data fundamental research projects. These awards aim to develop new tools and methods to extract and use knowledge from collections of large data sets to accelerate progress in science and engineering research.

Among the awards is a $2 million grant to Iowa State, Virginia Tech, and Stanford University to develop high-performance computing techniques on massively parallel heterogeneous computing resources for large-scale data analytics.

Such heterogeneous computing resources include the NSF Major (MRI) funded HokieSpeed supercomputing instrument with in-situ visualization. HokieSpeed was the highest-ranked commodity supercomputer in the U.S. on the Green500 when it debuted in November 2011.

Specifically, the three-university team intends to develop techniques that would enable researchers to innovatively leverage high-performance computing to analyze the data deluge of high-throughput DNA sequencing, also known as next generation sequencing (NGS).

The research will be conducted in the context of grand challenge problems in and metagenomics or the study of metagenomes, the genetic material received directly from environmental samples.

On this grant, working together are Srinivas Aluru, a chaired professor of computer engineering at Iowa State University and principal investigator; Patrick S. Schnable, a chaired professor of agronomy, also at Iowa State; Oyekunle A. Olukotun, a professor of electrical engineering and computer science at Stanford University; and Wu Feng, who holds the Turner Fellowship and who is an associate professor of computer science at Virginia Tech. Olukotun and Feng are co-.

In previous research Aluru has advanced the assembly of , comparative genomics, deep-sequencing data analysis, and parallel bioinformatics methods and tools. Aluru and Schnable previously worked together on generating a reference genome for the complex stalk of corn genome that will help speed efforts to develop better crop varieties.

Feng's relevant prior work lies at the synergistic intersection of life sciences and high-performance computing, particularly in the context of big data. For example, in 2007, Feng and his colleagues created an ad-hoc environment called ParaMEDIC, short for Parallel Metadata Environment for Distributed I/O and Computing, to conduct a massive sequence search over a distributed ephemeral supercomputer that enabled bioinformaticists to "identify missing genes in genomes."

Feng said, "With apologies to the movie Willy Wonka and the Chocolate Factory, one can view ParaMEDIC as WonkaVision for Scientific Data – a way to intelligently teleport data using semantic-based cues. "

Feng also recently studied how heterogeneous at the small scale and large scale, e.g., HokieSpeed, could be used for short-read mapping and alignment of genetic sequences in support of the philanthropic award that he received from NVIDIA Foundation as part of its "Compute the Cure" program.

With this award, Feng, the principal investigator, and his colleagues created a framework for faster genome analysis to make it easier for genomics researchers to identify mutations that are relevant to cancer.

In all, NSF and NIH announced a total of eight new projects in response to a call for proposals on "Core Techniques and Technologies for Advancing Big Data Science & Engineering," or "Big Data," in March of 2012. They run the gamut of scientific approaches with possible future applications in scientific disciplines, such as physics, psychology, economics, and medicine.

"I am delighted to provide such a positive progress report just six months after fellow federal agency heads joined the White House in launching the Big Data Initiative," said NSF Director Subra Suresh. "By funding new types of collaborations—multi-disciplinary teams and communities enabled by new data access policies—and with the start of an exciting competition, we are realizing plans to advance the complex, science and engineering grand challenges of today and to fortify U.S. competitiveness for decades to come."

"To get the most value from the massive biological data sets we are now able to collect, we need better ways of managing and analyzing the information they contain," said NIH Director Francis S. Collins. "The new awards that NIH is funding will help address these technological challenges—and ultimately help accelerate research to improve health—by developing methods for extracting important, biomedically relevant information from large amounts of complex data."

The TopCoder Open Innovation platform and process allows U.S. government agencies to conduct high risk/high reward challenges in an open and transparent environment with predictable cost, measurable outcomes-based results and the potential to move quickly into unanticipated directions and new areas of software technology.

Big data is a term applied to data sets whose size is beyond the ability of commonly used software tools to capture, manage and process the data within a tolerable elapsed time. Big data sizes are a constantly moving targets currently ranging from a few dozen terabytes to many petabytes of data in a single data set.

Explore further: Open Science Grid receives $30 million award

Related Stories

Open Science Grid receives $30 million award

September 25, 2006

Scientists on the track to discovery got good news this month when a powerful computing tool received critical government funding. A five-year, $30 million award to the Open Science Grid Consortium, announced by the National ...

Microsoft and NSF enable research in the cloud

February 4, 2010

Microsoft Corp. and the National Science Foundation (NSF) today announced an agreement that will offer individual researchers and research groups selected through NSF's merit review process free access to advanced cloud computing ...

High-performance computing reveals missing genes

April 13, 2010

Scientists at the Virginia Bioinformatics Institute (VBI) and the Department of Computer Science at Virginia Tech have used high-performance computing to locate small genes that have been missed by scientists in their quest ...

Virginia Tech to build new supercomputer

August 5, 2010

Two million dollars is coming to Virginia Tech to create HokieSpeed, a versatile new supercomputing instrument for accelerating and transforming discovery and innovation across a myriad of disciplines.

Recommended for you

AT&T testing superfast 5G mobile network

February 12, 2016

AT&T announced plans Friday to begin testing a superfast fifth generation, or 5G, mobile network with speeds up to 100 times faster than the most rapid existing connections.

Using stories to teach human values to artificial agents

February 12, 2016

The rapid pace of artificial intelligence (AI) has raised fears about whether robots could act unethically or soon choose to harm humans. Some are calling for bans on robotics research; others are calling for more research ...


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.