Sequoia supercomputer transitions to classified work

April 18, 2013 by Donald B Johnston
The Sequoia supercomputer.

The National Nuclear Security Administration (NNSA) today announced that its Sequoia supercomputer at Lawrence Livermore National Laboratory (LLNL) has completed its transition to classified computing in support of the Stockpile Stewardship Program, which helps the United States ensure the safety, security and effectiveness of its aging nuclear weapons stockpile without the use of underground testing.

The 20 /s (quadrillion floating point operations per second) IBM /Q system is now dedicated exclusively to NNSA's Advanced Simulation and Computing (ASC) program. ASC is a tri-lab effort drawing on the computational engineering and scientific computing expertise resident at Los Alamos, Sandia and Lawrence Livermore national laboratories.

"The success of early science runs on Sequoia have prepared the system to take on the complex calculations necessary to continue certifying the nation's aging nuclear stockpile," said NNSA Assistant Deputy Administrator for Stockpile Stewardship Chris Deeney. "Sequoia's mammoth computing power will provide scientists and engineers with a more complete understanding of weapons' performance, notably hydrodynamics and the properties of materials at and temperatures. These capabilities provide confidence in the U.S. deterrent as it is reduced under treaty agreements and represent the nation's continued leadership in high performance computing."

Bob Meisner, director of the ASC program, says that among the critical enhanced capabilities Sequoia provides is uncertainty quantification, or UQ, the quantitative characterization and reduction of uncertainty in computer applications made possible by the running of large suites of calculations designed to assess the effects of minor differences in the systems. Sources of uncertainty are rife in the natural science and engineering. UQ uses statistical methods to determine likely outcomes.

"The work we've done to date on Sequoia gives every indication that we will be able to run suites of highly resolved uncertainty quantification calculations in support of ASC's effort to extend the life of aging weapons systems such as the B61 and W78—what we call the life extension program," Meisner said. "By reducing the time required for the studies that underlie life extension, total costs also are reduced. The machine will also provide a means to do an assessment of its entry-level 3-D UQ capabilities. 3-D UQ will become increasingly important as the stockpile ages."

Additionally, NNSA expects the machine to enhance the program's ability to: sustain the stockpile by resolving any significant findings in weapons systems; bring greater computing power to all aspects of the annual assessment of the stockpile; and anticipate and avoid future problems that inevitably result from aging. These capabilities help ensure that the nation will never have to return to nuclear testing. Supercomputers such as Sequoia have allowed the U.S. to have confidence in its nuclear weapons stockpile over the 20 years since nuclear testing ended in 1992. The insights that come from supercomputing simulations also are vital to addressing nonproliferation and counterterrorism issues as well as informing other national security decisions such as nuclear weapon policy and treaty agreements.

Delivered and deployed in early 2012, the 96-rack Sequoia machine not only took the No. 1 ranking on the June 2012 Top500 list of the world's most powerful supercomputers, it was also rated as the world's most energy efficient system and earned top honors on the Graph500 list for its ability to solve big data problems - finding the proverbial needle in the haystack. While Sequoia dropped to No. 2 on the November 2012 Top500 list, it remains one of the most energy efficient HPC systems and retained its No. 1 Graph500 ranking.

Early unclassified work on the machine allowed NNSA researchers and IBM computer scientists to work out the bugs and optimize the system before it transitioned to classified work. Los Alamos National Laboratory researchers ran turbulence simulations and Sandia National Laboratories scientists explored the properties of tantalum on Sequoia.

LLNL researchers performed record simulations using all 1,572,864 cores of Sequoia to study the interaction of ultra-powerful lasers with dense plasmas in a proposed method to produce fusion energy, the energy source that powers the sun. Sequoia is the first machine to exceed one million computational cores. The simulations are the largest particle-in-cell (PIC) code simulations by number of cores ever performed and are important to laser fusion experiments in LLNL's National Ignition Facility (NIF). PIC simulations are used extensively in plasma physics to model the motion of the charged particles, and the electromagnetic interactions between them, that make up ionized matter. For more, see the March 19, 2013, announcement.

In addition, LLNL scientists investigated burn in doped plasmas, exploiting the full capability of Sequoia and the code developed for this purpose. Following a benchmark exploration of the density and temperature dependence of burn in undoped hydrogen plasma, researchers began a series of extreme-scale simulations of burn in the presence of small fractions of a percent of high-Z dopants. These studies are being used to deepen scientists' understanding of the effect of dopants on burn, physics that is vital to capsule design for NIF, a facility critical to NNSA's stockpile stewardship program.

Early efforts by LLNL scientists also included a QBox first principles molecular dynamics code examination of the electronic structure of heavy metals, research of interest to stockpile stewardship. QBox was developed at LLNL to perform large-scale simulations of materials directly from first-principles, allowing scientists to predict the properties of complex systems without first having to carry out experiments.

Sequoia also demonstrated its great scalability with a 3-D simulation of the human heart's electrophysiology. Using a code created in a partnership between LLNL and IBM scientists called Cardioid, researchers are modeling the electrical signals moving throughout the heart. Cardioid has the potential to be used to test drugs and medical devices, paving the way for tests on humans. Techniques employed by the code are useful to Sequoia's national security applications. Development of Cardioid is continuing on a smaller version of Sequoia called Vulcan, a five petaflop/s BlueGene/Q system used for unclassified research collaborations.

Sequoia was recognized with a Breakthrough Award from Popular Mechanics as one of the top technology innovations of 2012.

Explore further: Record simulations conducted on Lawrence Livermore supercomputer

Related Stories

Predictive simulation successes on Dawn supercomputer

September 30, 2009

( -- The 500-teraFLOPS Advanced Simulation and Computing program's Sequoia Initial Delivery System (Dawn), an IBM machine of the same lineage as BlueGene/L, has immediately proved itself useful as several initial ...

IBM To Build Supercomputer For U.S. Government

February 3, 2009

( -- The U.S. Government has contracted out IBM to build a massive supercomputer bigger than any supercomputer out there. The supercomputer system, called Sequoia, will be capable of delivering 20 petaflops (1,000 ...

Bug repellent for supercomputers proves effective

November 14, 2012

(—Lawrence Livermore National Laboratory (LLNL) researchers have used the Stack Trace Analysis Tool (STAT), a highly scalable, lightweight tool to debug a program running more than one million MPI processes on ...

Recommended for you

Microsoft aims at Apple with high-end PCs, 3D software

October 26, 2016

Microsoft launched a new consumer offensive Wednesday, unveiling a high-end computer that challenges the Apple iMac along with an updated Windows operating system that showcases three-dimensional content and "mixed reality."

Making it easier to collaborate on code

October 26, 2016

Git is an open-source system with a polarizing reputation among programmers. It's a powerful tool to help developers track changes to code, but many view it as prohibitively difficult to use.

Dutch unveil giant vacuum to clean outside air

October 25, 2016

Dutch inventors Tuesday unveiled what they called the world's first giant outside air vacuum cleaner—a large purifying system intended to filter out toxic tiny particles from the atmosphere surrounding the machine.


Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.