Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Videos, Slides from MUG ’20 Now Available

The MVAPICH User Group Meeting (MUG ’20), built around an implementation of the MPI standard developed by Ohio State University, has posted videos and slides from presentations on a variety of topics at its recent annual conference. The program included keynote talks from Brian van Essen from Lawrence Livermore National Labs and Michael Norman from San Diego Supercomputing […]


In this special guest feature, Gilad Shainer from Mellanox Technologies writes that the new GPCNeT benchmark is actually a measure of relative performance under load rather than a measure of absolute performance. “When it comes to evaluating high-performance computing systems or interconnects, there are much better benchmarks available for use. Moreover, the ability to benchmark real workloads is obviously a better approach for determining system or interconnect performance and capabilities. The drawbacks of GPCNeT benchmarks can be much more than its benefits.”

Distributed HPC Applications with Unprivileged Containers

Felix Abecassis and Jonathan Calmels from NVIDIA gave this talk at FOSDEM 2020. “We will present the challenges in doing distributed deep learning training at scale on shared heterogeneous infrastructure. At NVIDIA, we use containers extensively in our GPU clusters for both HPC and deep learning applications. We love containers for how they simplify software packaging and enable reproducibility without sacrificing performance.”

Geoffrey C. Fox to receive Ken Kennedy Award at SC19

Today ACM/IEEE named Geoffrey C. Fox of Indiana University Bloomington as the recipient of the 2019 ACM-IEEE CS Ken Kennedy Award. “Fox was cited for foundational contributions to parallel computing methodology, algorithms and software, and data analysis, and their interfaces with broad classes of applications. The award will be presented at SC19 in Denver.”

Checkpointing the Un-checkpointable: MANA and the Split-Process Approach

Gene Cooperman from Northeastern University gave this talk at the MVAPICH User Group. “This talk presents an efficient, new software architecture: split processes. The “MANA for MPI” software demonstrates this split-process architecture. The MPI application code resides in “upper-half memory”, and the MPI/network libraries reside in “lower-half memory”.

Video: Three Perspectives on Message Passing

Robert Harrison from Brookhaven gave this talk at the MVAPICH User Group. “MADNESS, TESSE/EPEXA, and MolSSI are three quite different large and long-lived projects that provide different perspectives and driving needs for the future of message passing. All three of these projects employ MPI and have a vested interest in computation at all scales, spanning the classroom to future exascale systems.”

Benchmarking MPI Applications in Singularity Containers on Traditional HPC and Cloud Infrastructures

Andrei Plamada from ETH Zurich gave this talk at the hpc-ch forum on Cloud and Containers. “Singularity is a container solution that promises to both integrate MPI applications seamlessly and run containers without privilege escalation. These benefits make Singularity a potentially good candidate for the scientific high-performance computing community. However, the performance overhead introduced by Singularity is unclear. In this work we will analyze the overhead and the user experience on both traditional HPC and cloud infrastructures.”

Converging Workflows Pushing Converged Software onto HPC Platforms

Are we witnessing the convergence of HPC, big data analytics, and AI? Once, these were separate domains, each with its own system architecture and software stack, but the data deluge is driving their convergence. Traditional big science HPC is looking more like big data analytics and AI, while analytics and AI are taking on the flavor of HPC.

Scalable and Distributed DNN Training on Modern HPC Systems

DK Panda from Ohio State University gave this talk at the Swiss HPC Conference. “We will provide an overview of interesting trends in DNN design and how cutting-edge hardware architectures are playing a key role in moving the field forward. We will also present an overview of different DNN architectures and DL frameworks. Most DL frameworks started with a single-node/single-GPU design.”

Call for Papers: EuroMPI Conference in Zurich

The EuroMPI conference has issued its Call for Papers. The event takes place September 10-13 in Zurich, Switzerland. “The EuroMPI conference is since 1994 the preeminent meeting for users, developers and researchers to interact and discuss new developments and applications of message-passing parallel computing, in particular in and related to the Message Passing Interface (MPI). This includes parallel programming interfaces, libraries and langauges, architectures, networks, algorithms, tools, applications, and High Performance Computing with particular focus on quality, portability, performance and scalability.”