Nautilus, the powerful computer for visualizing and analyzing large datasets at the Remote Data Analysis and Visualization Center (RDAV), goes into full production on September 20. Managed by the University of Tennessee (UT) and funded by a grant from the National Science Foundation (NSF), RDAV and Nautilus provide scientists with cutting-edge capabilities for analyzing and visualizing massive quantities of data. As Nautilus goes into service, RDAV will serve researchers in a wide range of fields.
Louis Gross, professor of ecology and evolutionary biology and mathematics at the University of Tennessee and director of the National Institute for Mathematical and Biological Synthesis (NIMBioS) says, "NIMBioS and RDAV have already initiated new collaborations to enhance the use of visualization for large datasets arising from field observations and from mathematical model results. Our objective is to increase the ability of biologists to interpret and analyze complex, multivariate data to address fundamental and applied questions in the life sciences."
"For a scientist, visualization is more than just generating pretty pictures," said astrophysicist Bronson Messer of the Oak Ridge National Lab (ORNL) and UT. "As our simulations grow larger and larger, visualization and the associated data analysis are absolutely essential in producing scientific insight from computation." Messer notes, "Nautilus, and the way it is being integrated into the computational ecosystem at NICS, looks like a very promising avenue for us to increase the amount of scientific insight we obtain from simulations on Kraken. The large, shared memory also allows us to translate analysis workflows directly from earlier, smaller versions."
In addition to addressing scientific problems in the life sciences and astrophysics, Nautilus will be used to process data spanning many other research areas. These include: visualizing data results from computer simulations with many complex variables, such as weather or climate modeling; analyzing large amounts of data coming from experimental facilities like ORNL's Spallation Neutron Source; and aggregating and interpreting input from a large number of sensors distributed over a wide geographic region. The computer will also have the capability to study large bodies of text and aggregations of documents.
With 1024 cores and four terabytes of memory, Nautilus can process large volumes of data and analyze it in ways unlike those used by previous systems. Manufactured by SGI as part of their UltraViolet product line, Nautilus features a shared-memory architecture with a single system image. This configuration allows researchers great flexibility in taking advantage of computing power to analyze larger amounts of data in ways that are impossible on many other high performance computing systems. RDAV has installed a 1.1 petabyte filesystem on Nautilus in anticipation of these vast amounts of data.
Founded in the second half of 2009, the RDAV center is now fully staffed and ready to support scientists and their data and visualization challenges. In addition to configuring and fine-tuning the Nautilus hardware, the team has been developing and refining new and existing software to address the latest scientific problems and to engage the scientific community.
RDAV Director Sean Ahern says, "I'm very excited about standing up this machine for the NSF TeraGrid, as it's going to provide much needed capability for understanding complex datasets from Kraken and other sources. And as datasets continue to grow, the shared memory nature of the SGI is a fertile ground for new analysis and visualization research for very large datasets."
Explore further: Coping with floods—of water and data