Subscribe free to our newsletters via your
. 24/7 Space News .




NUKEWARS
Nuclear weapon simulations show performance in molecular detail
by Emil Venere
West Lafayette IN (SPX) Jun 11, 2012


Employees at Lawrence Livermore National Laboratory work on a high-performance computer. Purdue researchers have collaborated with the national laboratory, using a similar high-performance computer to improve simulations that show a nuclear weapon's performance in precise molecular detail. (Photo courtesy of Lawrence Livermore National Laboratory).

U.S. researchers are perfecting simulations that show a nuclear weapon's performance in precise molecular detail, tools that are becoming critical for national defense because international treaties forbid the detonation of nuclear test weapons. The simulations must be operated on supercomputers containing thousands of processors, but doing so has posed reliability and accuracy problems, said Saurabh Bagchi, an associate professor in Purdue University's School of Electrical and Computer Engineering.

Now researchers at Purdue and high-performance computing experts at the National Nuclear Security Administration's (NNSA) Lawrence Livermore National Laboratory have solved several problems hindering the use of the ultra-precise simulations. NNSA is the quasi-independent agency within the U.S. Department of Energy that oversees the nation's nuclear security activities.

The simulations, which are needed to more efficiently certify nuclear weapons, may require 100,000 machines, a level of complexity that is essential to accurately show molecular-scale reactions taking place over milliseconds, or thousandths of a second. The same types of simulations also could be used in areas such as climate modeling and studying the dynamic changes in a protein's shape.

Such highly complex jobs must be split into many processes that execute in parallel on separate machines in large computer clusters, Bagchi said.

"Due to natural faults in the execution environment there is a high likelihood that some processing element will have an error during the application's execution, resulting in corrupted memory or failed communication between machines," Bagchi said. "There are bottlenecks in terms of communication and computation."

These errors are compounded as long as the simulation continues to run before the glitch is detected and may cause simulations to stall or crash altogether.

"We are particularly concerned with errors that corrupt data silently, possibly generating incorrect results with no indication that the error has occurred," said Bronis R. de Supinski, co-leader of the ASC Application Development Environment Performance Team at Lawrence Livermore. "Errors that significantly reduce system performance are also a major concern since the systems on which the simulations run are very expensive."

Advanced Simulation and Computing is the computational arm of NNSA's Stockpile Stewardship Program, which ensures the safety, security and reliability of the nation's nuclear deterrent without underground testing.

New findings will be detailed in a paper to be presented during the Annual IEEE/IFIP International Conference on Dependable Systems and Networks from June 25-28 in Boston. Recent research findings were detailed in two papers last year, one presented during the IEEE Supercomputing Conference and the other during the International Symposium on High-Performance Parallel and Distributed Computing.

The researchers have developed automated methods to detect a glitch soon after it occurs.

"You want the system to automatically pinpoint when and in what machine the error took place and also the part of the code that was involved," Bagchi said. "Then, a developer can come in, look at it and fix the problem."

One bottleneck arises from the fact that data are streaming to a central server.

"Streaming data to a central server works fine for a hundred machines, but it can't keep up when you are streaming data from a thousand machines," said Purdue doctoral student Ignacio Laguna, who worked with Lawrence Livermore computer scientists. "We've eliminated this central brain, so we no longer have that bottleneck."

Each machine in the supercomputer cluster contains several cores, or processors, and each core might run one "process" during simulations. The researchers created an automated method for "clustering," or grouping the large number of processes into a smaller number of "equivalence classes" with similar traits. Grouping the processes into equivalence classes makes it possible to quickly detect and pinpoint problems.

"The recent breakthrough was to be able to scale up the clustering so that it works with a large supercomputer," Bagchi said.

Lawrence Livermore computer scientist Todd Gamblin came up with the scalable clustering approach.

A lingering bottleneck in using the simulations is related to a procedure called checkpointing, or periodically storing data to prevent its loss in case a machine or application crashes. The information is saved in a file called a checkpoint and stored in a parallel system distant from the machines on which the application runs.

"The problem is that when you scale up to 10,000 machines, this parallel file system bogs down," Bagchi said. "It's about 10 times too much activity for the system to handle, and this mismatch will just become worse because we are continuing to create faster and faster computers."

Doctoral student Tanzima Zerin and Rudolf Eigenmann, a professor of electrical and computer engineering, along with Bagchi, led work to develop a method for compressing the checkpoints, similar to the compression of data for images.

"We're beginning to solve the checkpointing problem," Bagchi said. "It's not completely solved, but we are getting there."

The checkpointing bottleneck must be solved in order for researchers to create supercomputers capable of "exascale computing," or 1,000 quadrillion operations per second.

"It's the Holy Grail of supercomputing," Bagchi said.

The research has been funded by Lawrence Livermore and the National Science Foundation. The work also involves Lawrence Livermore scientists Greg Bronevetsky, Dong H. Ahn, Martin Schulz and IBM Austin researcher Mootaz Elnozahy.

Laguna was awarded a George Michael Memorial High Performance Computing Fellowship during the Supercomputing 2011 Conference in recognition of his work on the project. The Supercomputing Conference paper can be accessed online at IEEExplore or ACM Digital Library or from the research group's home page here.

Purdue researchers did not work with the actual classified nuclear weapons software code, but instead used generic benchmarks, a set of programs designed to help evaluate the performance of parallel supercomputers.

.


Related Links
Purdue University
Learn about nuclear weapons doctrine and defense at SpaceWar.com
Learn about missile defense at SpaceWar.com
All about missiles at SpaceWar.com
Learn about the Superpowers of the 21st Century at SpaceWar.com






Comment on this article via your Facebook, Yahoo, AOL, Hotmail login.

Share this article via these popular social media networks
del.icio.usdel.icio.us DiggDigg RedditReddit GoogleGoogle








NUKEWARS
Obama hits back at 'offensive' leak allegations
Washington (AFP) June 8, 2012
US President Barack Obama hit back Friday at accusations that the White House was somehow involved in leaking classified national security secrets, labeling the allegations "offensive." Lawmakers have fumed this week over a series of news reports, some quoting anonymous administration officials, that reveal details of covert US operations. Some Republicans have suggested the information was ... read more


NUKEWARS
UA Lunar-Mining Team Wins National Contest

NASA Lunar Spacecraft Complete Prime Mission Ahead of Schedule

NASA Offers Guidelines To Protect Historic Sites On The Moon

Neil Armstrong gives rare interview - to accountant

NUKEWARS
NASA's Mars rover zeroes in on August landing

Russia May Join Mars Orbiter Project in Nov. - ESA

Robotic Arm Gets to Work on Veins of Gypsum

Odyssey Orbiter Puts Itself into Standby Safe Mode

NUKEWARS
US scientists host 'bake sale for NASA'

XCOR Appoints Space Expedition Corp As General Sales Agent For Space Tourism Flights

European Union launches latest space regulation efforts

CU-Boulder students to help NASA develop plant food production for deep space

NUKEWARS
Two Women For Tiangong

Shenzhou 9 Ready For Manned Mission To Tiangong-1

China to launch manned spacecraft this month

What will China's Taikonauts do aboard Tiangong 1?

NUKEWARS
Strange Geometry - Yes, It's All About the Math

Capillarity in Space - Then and Now, 1962-2012

Dragon on board

SpaceX Launches Falcon 9 Dragon on Historic Mission

NUKEWARS
NuSTAR Arrives at Island Launch Site

Another Ariane 5 begins its initial build-up at the Spaceport

Boeing Receives DARPA Airborne Satellite Launch Study Contract

Sea Launch Delivers the Intelsat 19 Spacecraft into Orbit

NUKEWARS
Tiny Planet-Finding Mirrors Borrow from Webb Telescope Playbook

Astronomers Probe 'Evaporating' Planet Around Nearby Star with Hobby-Eberly Telescope

Venus transit may boost hunt for other worlds

NSO To Use Venus Transit To Fine-Tune Search For Other Worlds

NUKEWARS
Lawrence Livermore research identifies precise measurement of radiation damage

Hologram developers raise real cash for virtual stars

Smooth moves: how space animates Hollywood

Skeleton key




The content herein, unless otherwise known to be public domain, are Copyright 1995-2014 - Space Media Network. AFP, UPI and IANS news wire stories are copyright Agence France-Presse, United Press International and Indo-Asia News Service. ESA Portal Reports are copyright European Space Agency. All NASA sourced material is public domain. Additional copyrights may apply in whole or part to other bona fide parties. Advertising does not imply endorsement,agreement or approval of any opinions, statements or information provided by Space Media Network on any Web page published or hosted by Space Media Network. Privacy Statement