MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Supercomputing Better Engines at Argonne

Researchers are using Argonne supercomputers to jump-start internal-combustion engine designs in the name of conservation. “Improving engine efficiencies by even a few percentage points can take a big chunk out of our carbon footprint. We are working on a proof-of-concept to demonstrate how simulating several thousand engine configurations simultaneously can really help engineers zero in on the optimum engine designs and operating strategies to maximize efficiency while minimizing harmful emissions.”

Call for Proposals: Aurora Early Science Program at Argonne

The Argonne Leadership Computing Facility (ALCF) is now accepting proposals for its Aurora Early Science Program (ESP) through September 2, 2016. The program will award computing time to 10 science teams to pursue innovative research as part of pre-production testing on the facility’s next-generation system. Aurora is a massively parallel, many-core Intel-Cray supercomputer that will deliver 18 times the computational performance of Mira, ALCF’s current production system.

DOE Awards 1.7 Billion Core-hours on Argonne Supercomputers

The ASCR Leadership Computing Challenge (ALCC) has awarded 26 projects a total of 1.7 billion core-hours at the Argonne Leadership Computing Facility (ALCF), a DOE Office of Science User Facility. The one-year awards began July 1.

ESnet Timeline Celebrates 30 Years of Servicing Science

The Department of Energy’s Energy Sciences Network (ESnet) has published a 3D timeline celebrating thirty years of service. With the launch of an interactive timeline, viewers can explore ESnet’s history and contributions.

Supercomputers Joining the Fight Against Cancer

“Supercomputers are key to the Cancer Moonshot. These exceptionally high-powered machines have the potential to greatly accelerate the development of cancer therapies by finding patterns in massive datasets too large for human analysis. Supercomputers can help us better understand the complexity of cancer development, identify novel and effective treatments, and help elucidate patterns in vast and complex data sets that advance our understanding of cancer.”

Paul Messina on the New ECP Exascale Computing Project

Argonne Distinguished Fellow Paul Messina has been tapped to lead the Exascale Computing Project, heading a team with representation from the six major participating DOE national laboratories: Argonne, Los Alamos, Lawrence Berkeley, Lawrence Livermore, Oak Ridge and Sandia. The project will focus its efforts on four areas: Applications, Software, Hardware, and Exascale Systems.

Slimming Down Supercomputer Power Bills

Any performance improvements that could be wrung out of supercomputers by adding more power have long been exhausted. New supercomputers demand new options that will give scientists a sleek, efficient partner in making new discoveries such as the new supercomputer called Summit that’s being developed and is to arrive at Oak Ridge National Lab in the next couple of years. “If necessity is the mother of invention, we’ll have some inventions happening soon,” says deputy division director of Argonne Leadership Computing Facility Susan Coghlan.

X-Stack PI Meeting Showcases Exascale Code

Berkeley Lab recently hosted the fourth annual X-Stack PI event, where X-Stack researchers, facilities teams, application scientists, and developers from national labs, universities, and industry met to share the latest developments in X-Stack application codes. “X-Stack was launched in 2012 by the U.S. Department of Energy’s Advanced Scientific Computing Research program to support the development of exascale software tools, including programming languages and libraries, compilers and runtime systems, that will help programmers handle massive parallelism, data movement, heterogeneity and failures as the scientific community transitions to the next generation of extreme-scale supercomputers.”

NSCI Update from the HPC User Forum

In this video from the HPC User Forum in Tucson, Saul Gonzalez Martirena from NSF provides an update on the NSCI initiative. “As a coordinated research, development, and deployment strategy, NSCI will draw on the strengths of departments and agencies to move the Federal government into a position that sharpens, develops, and streamlines a wide range of new 21st century applications. It is designed to advance core technologies to solve difficult computational problems and foster increased use of the new capabilities in the public and private sectors.”

INCITE Seeking Proposals to Advance Science with Leadership Computing

The DoE INCITE program is now accepting proposals for high-impact, computationally intensive research campaigns in a broad array of science, engineering and computer science domains.