Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Future HPC Leaders Gather at Argonne Training Program on Extreme-Scale Computing

Over at ALCF, Andrea Manning writes that the recent Argonne Training Program on Extreme-Scale Computing brought together HPC practitioners from around the world. “You can’t get this material out of a textbook,” said Eric Nielsen, a research scientist at NASA’s Langley Research Center. Added Johann Dahm of IBM Research, “I haven’t had this material presented to me in this sort of way ever.”

Supercomputing Jet Noise for a Quieter World

Researchers at the University of Minnesota are using Argonne supercomputers to to look for new ways to reduce the noise produced by jet engines. Among the loudest sources of human-made noise that exist, jet engines can produce sound in excess of 130 decibels. “The University of Minnesota team developed a new method based on input-output analysis that can predict both the downstream noise and the sideline noise. While it was thought that the sideline noise was random, the input-output modes show coherent structure in the jet that is connected to the sideline noise, such that it can be predicted and controlled.”

Illinois Supercomputers Tag Team for Big Bang Simulation

Researchers are tapping Argonne and NCSA supercomputers to tackle the unprecedented amounts of data involved with simulating the Big Bang. “Researchers performed cosmological simulations on the ALCF’s Mira supercomputer, and then sent huge quantities of data to UI’s Blue Waters, which is better suited to perform the required data analysis tasks because of its processing power and memory balance.”

Is Aurora Morphing into an Exascale AI Supercomputer?

The recently published Department of Energy FY 2018 Congressional Budget Request has raised a lot of questions about the Aurora supercomputer that was scheduled to be deployed at Argonne ALCF next year. “As we covered in our Radio Free HPC podcast, Aurora appears to be morphing into a very different kind of machine.”

Argonne Seeking Proposals to Advance Big Data in Science

The Argonne Leadership Computing Facility Data Science Program (ADSP) is now accepting proposals for projects hoping to gain insight into very large datasets produced by experimental, simulation, or observational methods. The larger the data, in fact, the better. Applications are due by June 15, 2017.

Video: How Smart is a Supercomputer?

In this video from The Good Stuff program, Katherine Riley and Argonne National Laboratory describes what a supercomputer really is and how Argonne uses high performance computing to solve the world’s most challenging problems. “This is a great program for explaining HPC to your friends, neighbors, and even your kids.”

Argonne Annouces Early Science Projects for Aurora Supercomputer

Argonne has selected 10 computational science and engineering research projects for its Aurora Early Science Program starting this month. Aurora, a massively parallel, manycore Intel-Cray supercomputer, will be ALCF’s next leadership-class computing resource and is expected to arrive in 2018. The Early Science Program helps lay the path for hundreds of other users by doing actual science, using real scientific applications, to ready a future machine. “As with any bleeding edge resource, there’s testing and debugging that has to be done,” said ALCF Director of Science Katherine Riley.

Building for the Future Aurora Supercomputer at Argonne

“Argonne National Labs has created a process to assist in moving large applications to a new system. Their current HPC system, Mira will give way to the next generation system, Aurora, which is part of the collaboration of Oak Ridge, Argonne, and Livermore (CORAL) joint procurement. Since Aurora contains technology that was not available in Mira, the challenge is to give scientists and developers access to some of the new technology, well before the new system goes online. This allows for a more productive environment once the full scale new system is up.”

DOE to Invest $16 Million in Supercomputing Materials

Today the U.S. Department of Energy announced that it will invest $16 million over the next four years to accelerate the design of new materials through use of supercomputers. “Our simulations will rely on current petascale and future exascale capabilities at DOE supercomputing centers. To validate the predictions about material behavior, we’ll conduct experiments and use the facilities of the Advanced Photon Source, Spallation Neutron Source and the Nanoscale Science Research Centers.”

Video: How the HPC Environment is Different from the Desktop (and Why)

“High performance computing has transformed how science and engineering research is conducted. Answering a question in 30 minutes that used to take 6 months can quickly change the way one asks questions. Large computing facilities provide access to some of the world’s largest computing, data, and network resources in the world. Indeed, the DOE complex has the highest concentration of supercomputing capability in the world. However, by nature of their existence, making use of the largest computers in the world can be a challenging and unique task. This talk will discuss how supercomputers are unique and explain how that impacts their use.”