In this video from The Good Stuff program, Katherine Riley and Argonne National Laboratory describes what a supercomputer really is and how Argonne uses high performance computing to solve the world’s most challenging problems. “This is a great program for explaining HPC to your friends, neighbors, and even your kids.”
Argonne has selected 10 computational science and engineering research projects for its Aurora Early Science Program starting this month. Aurora, a massively parallel, manycore Intel-Cray supercomputer, will be ALCF’s next leadership-class computing resource and is expected to arrive in 2018. The Early Science Program helps lay the path for hundreds of other users by doing actual science, using real scientific applications, to ready a future machine. “As with any bleeding edge resource, there’s testing and debugging that has to be done,” said ALCF Director of Science Katherine Riley.
“Argonne National Labs has created a process to assist in moving large applications to a new system. Their current HPC system, Mira will give way to the next generation system, Aurora, which is part of the collaboration of Oak Ridge, Argonne, and Livermore (CORAL) joint procurement. Since Aurora contains technology that was not available in Mira, the challenge is to give scientists and developers access to some of the new technology, well before the new system goes online. This allows for a more productive environment once the full scale new system is up.”
Today the U.S. Department of Energy announced that it will invest $16 million over the next four years to accelerate the design of new materials through use of supercomputers. “Our simulations will rely on current petascale and future exascale capabilities at DOE supercomputing centers. To validate the predictions about material behavior, we’ll conduct experiments and use the facilities of the Advanced Photon Source, Spallation Neutron Source and the Nanoscale Science Research Centers.”
“High performance computing has transformed how science and engineering research is conducted. Answering a question in 30 minutes that used to take 6 months can quickly change the way one asks questions. Large computing facilities provide access to some of the world’s largest computing, data, and network resources in the world. Indeed, the DOE complex has the highest concentration of supercomputing capability in the world. However, by nature of their existence, making use of the largest computers in the world can be a challenging and unique task. This talk will discuss how supercomputers are unique and explain how that impacts their use.”
The Argonne Leadership Computing Facility (ALCF) is now accepting proposals for its Aurora Early Science Program (ESP) through September 2, 2016. The program will award computing time to 10 science teams to pursue innovative research as part of pre-production testing on the facility’s next-generation system. Aurora is a massively parallel, many-core Intel-Cray supercomputer that will deliver 18 times the computational performance of Mira, ALCF’s current production system.
The ASCR Leadership Computing Challenge (ALCC) has awarded 26 projects a total of 1.7 billion core-hours at the Argonne Leadership Computing Facility (ALCF), a DOE Office of Science User Facility. The one-year awards began July 1.
In this WBEZ podcast, Katherine Riley from Argonne explains what a supercomputer does. “Argonne National Laboratory is celebrating its 70th anniversary. The science and engineering research center has had a prominent place in historical research developments. They’re currently in the process of installing their newewst supercomputer, Theta.”
Argonne Distinguished Fellow Paul Messina has been tapped to lead the Exascale Computing Project, heading a team with representation from the six major participating DOE national laboratories: Argonne, Los Alamos, Lawrence Berkeley, Lawrence Livermore, Oak Ridge and Sandia. The project will focus its efforts on four areas: Applications, Software, Hardware, and Exascale Systems.
Researchers from Michigan State University are using the Mira supercomputer to perform large-scale 3-D simulations of the final moments of a supernova’s life cycle. While the 3-D simulation approach is still in its infancy, early results indicate that the models are providing a clearer picture than ever before of the mechanisms that drive supernova explosions.