Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


The U.S. D.O.E. Exascale Computing Project – Goals and Challenges

Paul Messina from Argonne gave this Invited Talk at SC17. “Balancing evolution with innovation is challenging, especially since the ecosystem must be ready to support critical mission needs of DOE, other Federal agencies, and industry, when the first DOE exascale systems are delivered in 2021. The software ecosystem needs to evolve both to support new functionality demanded by applications and to use new hardware features efficiently. We are utilizing a co-design approach that uses over two dozen applications to guide the development of supporting software and R&D on hardware technologies as well as feedback from the latter to influence application development.

Speeding Data Transfer with ESnet’s Petascale DTN Project

Researchers at the DOE are looking to dramatically increase their data transfer capabilities with the Petascale DTN project. “The collaboration, named the Petascale DTN project, also includes the National Center for Supercomputing Applications (NCSA) at the University of Illinois in Urbana-Champaign, a leading center funded by the National Science Foundation (NSF). Together, the collaboration aims to achieve regular disk-to-disk, end-to-end transfer rates of one petabyte per week between major facilities, which translates to achievable throughput rates of about 15 Gbps on real world science data sets.”

Video: Comanche Collaboration Moves ARM HPC forward at National Labs

In this video from SC17 in Denver, Rick Stevens from Argonne leads a discussion about the Comanche Advanced Technology Collaboration. By initiating the Comanche collaboration, HPE brought together industry partners and leadership sites like Argonne National Laboratory to work in a joint development effort,” said HPE’s Chief Strategist for HPC and Technical Lead for the Advanced Development Team Nic Dubé. “This program represents one of the largest customer-driven prototyping efforts focused on the enablement of the HPC software stack for ARM. We look forward to further collaboration on the path to an open hardware and software ecosystem.”

Apply now for Argonne Training Program on Extreme-Scale Computing 2018

Computational scientists now have the opportunity to apply for the upcoming Argonne Training Program on Extreme-Scale Computing (ATPESC). The event takes place from July 29-August 10, 2018 in greater Chicago. “With the challenges posed by the architecture and software environments of today’s most powerful supercomputers, and even greater complexity on the horizon from next-generation and exascale systems, there is a critical need for specialized, in-depth training for the computational scientists poised to facilitate breakthrough science and engineering using these amazing resources.”

Argonne to Install Comanche System to Explore ARM Technology for HPC

Argonne National Laboratory is collaborating with HPE to provide system software expertise and a development ecosystem for a future high-performance computing system based on 64-bit ARM processors. “Argonne is working with HPE to evaluate early versions of chipmaker Cavium ARM ThunderX2 64-bit processors for the ARM ecosystem. Argonne is interested in evaluating the ARM ecosystem as a cost-effective and power-effective alternative to x86 architectures based on Intel CPUs, which currently dominate the high-performance computing market.”

Video: The Legion Programming Model

“Developed by Stanford University, Legion is a data-centric programming model for writing high-performance applications for distributed heterogeneous architectures. Legion provides a common framework for implementing applications which can achieve portable performance across a range of architectures. The target class of users dictates that productivity in Legion will always be a second-class design constraint behind performance. Instead Legion is designed to be extensible and to support higher-level productivity languages and libraries.”

Video: System Interconnects for HPC

In this video from the 2017 Argonne Training Program on Extreme-Scale Computing, Pavan Balaji from Argonne presents an overview of system interconnects for HPC. “The Argonne Training Program on Extreme-Scale Computing (ATPESC) provides intensive, two weeks of training on the key skills, approaches, and tools to design, implement, and execute computational science and engineering applications on current high-end computing systems and the leadership-class computing systems of the future.”

HPC Connects with Smart Cities at SC17

In this video from the SC17 HPC Connects series, Pete Beckman and Charlie Catlett from Argonne describe how the Smart Cities initiative to improve the quality of life for residents using HPC, urban informatics, and other technologies to improve the efficiency of services. “Smart Cities will be the topic of the SC17 plenary session, which kicks off the conference at 5:30pm on Monday, Nov. 13 in the Colorado Convention Center.”

Video: Evolution of MATLAB

Cleve Moler from MathWorks gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. “MATLAB is a high-performance language for technical computing. It integrates computation, visualization, and programming in an easy-to-use environment where problems and solutions are expressed in familiar mathematical notation. Typical uses include: Data analysis, exploration, and visualization.”

Video: Silicon Photonics for Extreme Computing

Keren Bergman from Columbia University gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. “Exaflop machines would represent a thousand-fold improvement over the current standard, the petaflop machines that first came on line in 2008. But while exaflop computers already appear on funders’ technology roadmaps, making the exaflop leap on the short timescales of those roadmaps constitutes a formidable challenge.”