Nanostructure complex materials modeling  

March 25, 2015 by Karen Mcnulty Walsh  
Scientists have demonstrated a way to increase the chances of solving material structures by feeding data from multiple experiments and theory into a "global optimizer" that uses mathematical algorithms to narrow the range of possible solutions based on its analysis of the complementary data sets.

Materials with chemical, optical, and electronic properties driven by structures measuring billionths of a meter could lead to improved energy technologies—from more efficient solar cells to longer-lasting energy-dense batteries. Scientific instruments such as those at Brookhaven Lab's Center for Functional Nanomaterials (CFN) and the just-opened National Synchrotron Light Source II (NSLS-II), both DOE Office of Science User Facilities, offer new ways to study materials at this nanometer length scale—including as they operate in actual devices. 

These experiments produce huge amounts of data, revealing important details about . But right now scientists don't have the computational tools they need to use that data for rational materials design—a step that's essential to accelerate the discovery of materials with the performance characteristics required for real-world, large-scale implementation. To achieve that goal, what's needed is a new way to combine the data from a range of experiments along with theoretical descriptions of materials' behavior into valid predictive models that scientists can use to develop powerful new materials.

"There is a very good chance that expertise in and applied mathematical algorithms developed with the support of DOE's Office of Advanced Scientific Computing Research (ASCR) can help us make progress on this frontier," said Simon Billinge, a physicist at Brookhaven National Laboratory and Columbia University's School of Engineering and Applied Science. At the SC14 conference, Simon presented a potential solution that makes use of mathematical theory and to extract the information essential for strengthening models of material performance.

"A lot of the models we have for these materials are not robust," he said. "Ideally, we'd like to be able to program in the properties we want—say, efficient solar energy conversion, superconductivity, massive electrical storage capacity—and have the model spit out the design for a new material that will have that property, but this is clearly impossible with unreliable models." 

Finding possible solutions for a material's structure by global optimization of a model incorporating data from multiple experimental techniques is computationally expensive/time consuming, but must be done multiple times to check whether there is more than one solution compatible with the data. Each dot in these triangle diagrams represents a different way of combing inputs from the three experiments and the color indicates the number of structure solutions found for that combination (red being the most unique solution). Only when it is done for many dots can scientists zero in on the most likely common solution that accounts for all the data. It takes high performance computers to handle the scale of computation needed to give clear results.

Powerful experimental tools such as NSLS-II make possible more intricate experiments.  But ironically, some of the new techniques make the discovery process more difficult.

Billinge explains: "Real materials and real applications depend on fine details of materials' structure such as defects, surfaces, and morphology, so experiments that help reveal fine structural details are essential. But some of the most interesting materials are very complex, and then they are made into even more complex, multi-component devices. When we place these complex devices in an x-ray or neutron-scattering beamline, interactions of the beams with all the complex components produce overlapping results. You are looking for small signals from defects and surfaces hidden in a huge background of other information from the extra components, all of which degrades the useful information."

At the same time, the complexity of the models the scientists are trying to build to understand these materials is also increasing, so they need more information about those essential details to feed into the model, not less. Solving these problems reliability, given all the uncertainties, requires advanced mathematical approaches and high-performance computing, so Simon and his collaborators are working with ASCR on a two-pronged approach to improve the process.

On the input side, they combine results from multiple experiments—x-ray scattering, neutron scattering, and also theory. On the output side, the scientists try to reduce what's called the dimensionality of the model. Billinge explained this as being similar to the compression that creates an mp3 music file by shedding unessential information most people wouldn't notice is missing. 

"If we reduce the complexity to minimize the input needed to make the problem solvable, we can run it through powerful high-performance computers that use advanced mathematical methods derived from information theory, uncertainty quantification, and other data analytic techniques to sort through all the details," he said. The can put back together the complementary information from the different experiments—sort of like the parable about the blind men exploring different parts of an elephant, but now sharing and combining their results—and use it to predict complex materials structure.

"Similar to an mp3 file, there is some missing information in these models. But with the right representation it can be good enough to have predictive value and allow us to design new materials," Billinge said.

Explore further: First scientific publication from data collected at NSLS-II

Related Stories

First scientific publication from data collected at NSLS-II

March 3, 2015

Just weeks after the National Synchrotron Light Source II (NSLS-II), a U.S. Department of Energy Office of Science User Facility at Brookhaven National Laboratory, achieved first light, a team of scientists at the X-Ray Powder ...

A new mathematics for experimental science

April 1, 2014

Mathematics is the ultimate scientific tool. For centuries it has been used to describe the forces of nature, from planetary motion to fluid dynamics. It helped unlock the secrets of DNA and unleashed the Digital Revolution. ...

Solar cell polymers with multiplied electrical output

January 12, 2015

One challenge in improving the efficiency of solar cells is that some of the absorbed light energy is lost as heat. So scientists have been looking to design materials that can convert more of that energy into useful electricity. ...

Recommended for you

Atomic blasting creates new devices to measure nanoparticles

December 14, 2017

Like sandblasting at the nanometer scale, focused beams of ions ablate hard materials to form intricate three-dimensional patterns. The beams can create tiny features in the lateral dimensions—length and width, but to create ...

Engineers create plants that glow

December 13, 2017

Imagine that instead of switching on a lamp when it gets dark, you could read by the light of a glowing plant on your desk.

Faster, more accurate cancer detection using nanoparticles

December 12, 2017

Using light-emitting nanoparticles, Rutgers University-New Brunswick scientists have invented a highly effective method to detect tiny tumors and track their spread, potentially leading to earlier cancer detection and more ...

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

swordsman
not rated yet Mar 26, 2015
Sort of a part experiment, part mathematical model technique.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.