The universe is a vast and mysterious place, but thanks to high-performance computing technology scientists around the world are beginning to understand it better. They are using supercomputers to simulate how the Big Bang generated the seeds that led to the formation of galaxies such as the Milky Way.
A new project involving DOE's Argonne Lab, Fermilab and Berkeley Lab will allow scientists to study this vastness in greater detail with a new cosmological simulation analysis toolbox.
Modeling the universe with a computer is very difficult, and the output of those simulations is typically very large. By anyone's standards, this is "big data," as each of these data sets can require hundreds of terabytes of storage space. Efficient storage and sharing of these huge data sets among scientists is paramount. Many different scientific analyses and processing sequences are carried out with each data set, making it impractical to rerun the simulations for each new study.
This past year Argonne Lab, Fermilab and Berkeley Lab began a unique partnership on an ambitious advanced-computing project. Together the three labs are developing a new, state-of-the-art cosmological simulation analysis toolbox that takes advantage of DOE's investments in supercomputers and specialized high-performance computing codes. Argonne's team is led by Salman Habib, principal investigator, and Ravi Madduri, system designer. Jim Kowalkowski and Richard Gerber are the team leaders at Fermilab and Berkeley Lab.
The three labs have embarked on an innovative project to develop an open platform with a web-based front end that will allow the scientific community to download, transfer, manipulate, search and record simulation data. The system will allow scientists to upload and share applications as well as carry out complex computational analyses using the resources available to and assigned by the system.
To achieve these objectives, the team uses and enhances existing high-performance computing, high-energy physics and cosmology-specific software systems. As they modify the existing software so that it can handle the large datasets of galaxy-formation simulations, team members take advantage of the expertise they have gained by working on the big data challenges posed by particle physics experiments at the Large Hadron Collider.
This is an exciting project for the three labs. Large-scale simulations of cosmological structure formation are key discovery tools in the Cosmic Frontier program of DOE's Office of High Energy Physics. Not only will this new project provide an important toolbox for Cosmic Frontier scientists and the many institutions involved in this research, but it will also serve as a prototype for a successful big-data software project spanning many groups and communities.
The commercial world has taken notice, too. This month, Rob Roser, head of Fermilab's Scientific Computing Division, will present this project as part of his keynote speech at the Big Data Conference in Chicago.
Explore further: Coping with floods—of water and data