(PhysOrg.com) -- Scientists have made many discoveries about the origins of our 13 billion-year-old universe. But many scientific mysteries remain. What exactly happened during the Big Bang, when rapidly evolving physical processes set the stage for gases to form stars, planets and galaxies?

Now astrophysicists using supercomputers to simulate the Big Bang have a new mathematical tool to unravel those mysteries, says Daniel Reynolds, assistant professor of mathematics at Southern Methodist University in Dallas.

Reynolds collaborated with astrophysicists at the University of California at San Diego as part of a National Science Foundation project to simulate cosmic reionization, the time from 380,000 years to 400 million years after the universe was born.

Together the scientists built a computer model of events during the "Dark Ages" when the first stars emitted radiation that altered the surrounding matter, enabling light to pass through. The team tested its model on two of the largest existing NSF supercomputers, "Ranger" at the University of Texas at Austin and "Kraken" at the University of Tennessee.

The new mathematical model tightly couples a myriad of physical processes present during cosmic reionization, such as gas motion, radiation transport, chemical kinetics and gravitational acceleration due to star clustering and dark matter dynamics, Reynolds says. The key characteristic of the model that differentiates it from competing work is that the researchers focused on enforcing a very tight coupling in the model between the different physical processes.

"By forcing the computational methods to tightly bind these processes together, our new model allows us to generate simulations that are highly accurate, numerically stable and computationally scalable to the largest supercomputers available," Reynolds says.

Reynolds' mathematical research was published as "Self-Consistent Solution of Cosmological Radiation-Hydrodynamics and Chemical Ionization" in the *Journal of Computational Physics*.

Simulation models typically consist of a complex bundle of mathematical equations representing physical processes. The equations are integrated to reflect interaction of the physical processes. Only supercomputers can simultaneously solve the equations. Scientific intuition and creativity come into play by developing the base model with equations with the best parameters, Reynolds says. Variables can be altered to describe different scenarios that might have occurred. The objective is to develop a simulation model with results that most closely resemble telescope observations and that predict a universe that looks like what we have. If that happens, scientists have discovered the set of physical processes that existed at the birth of the universe as it was evolving from one instant to the next.

Physical processes include the heating of various gases, gravity, the conservation of mass, the conservation of momentum, the conservation of energy, expansion of the universe, the transport of radiation, and the chemical ionization of different species such as Hydrogen and Helium, the primary elements present at the beginning of the universe. An additional equation running in the background describes and models the dynamics of dark matter — the majority of the matter in the universe — which gives rise to gravity and is attributed with helping the universe form stars, planets and galaxies.

"Supercomputers are so big, they hold so much data, you can build models that work with many processes at one time," Reynolds says. "A lot of these processes behave nonlinearly. When they are put together, they inhibit each other, feed off each other, so you end up with many different processes when they are put together."

A direct consequence of the tight coupling that the researchers enforce in their model is that the resulting system of equations is much more complex than those that must be solved by other models, Reynolds says.

"This paper describes both how we form the coupled model, as well as the mathematical methods that enable us to solve the systems of equations that result. These include methods that accurately track the different time scales of each process, which often occur at rates that vary by orders of magnitude," he says. "However, perhaps the most important contribution of this paper is our description of how we pose the complex interaction of different models as a nonlinear problem with potentially billions of equations and unknowns, and solve that problem using new algorithms designed for next-generation supercomputers. We conclude by demonstrating that the new model lives up to the ideal, providing an approach that allows high accuracy, stability and scalability on a suite of difficult test problems."

Only recently have mathematics algorithms been invented to solve basic problems — like diffusion of heat — using resources as large as those available on modern supercomputers, Reynolds says. There have been simple analytical solutions to many problems from mathematical physics for hundreds of years. However, those analytical solutions only work when scientists simplify the problem in some way or another. For example, he says, they may approximate the shape of a planet as a sphere, instead of an ellipsoid, or may assume that ocean water is incompressible, which only works for very shallow water, or assume the Earth is homogeneous, instead of formed using widely differing layers of rock.

"Scientists have been able to approximate a great many physical processes in such idealized situations. But the true frontier nowadays is to let go of these simplifying approximations and treat the problems as they really are, by modeling all of the geometric structure and the in-homogeneity," Reynolds says. "To do that, you need to solve harder equations with lots of data, which is ideally suited to using supercomputers. The numerical methods that can allow us to use larger and larger computers have only just come out. The problems are getting more challenging and harder to solve, but the numerical methods are reaching greater capability, so you can really start moving them forward. These new computers make everything a new frontier."

**Explore further:**
New Look At Microwave Background May Cast Doubts On Big Bang Theory

## Parsec

Completely untrue. The equations could be solved with a handheld calculator. Of course it would take a wee bit more time and the ol fingers would likely snap from old age before you got a billionth of the way through...

## dutchman

Sequentially perhaps, but simultaneously? Rather unlikely, methinks. :)

## frajo

There's no supercomputer with more than a million CPUs/cores. Thus the supercomputers have do do it sequentially, too.

## SincerelyTwo

## de_la_meu

Jan 06, 2010## PPihkala

I think that you must very tighly describe what you mean with your word see. In strictest sense there was nothing to 'see' (happening at visible light energies) at actual big bang as it is currently described. An analogy would be to put your hands into exploding fission bomb and to try to feel how it feels. Only to notice that you don't have any hands nor any body to feel with. After big bang it took lot's of time and expansion to cool things enough to let even protons and neutrons to form. Before those it was just magnitudes brighter than anything imaginable.

## frajo

## Noumenon

## frajo