MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Altair Launches PBS Pro 13

Altair Logo Stacked

Today Altair announced the general availability of PBS Professional 13.0, the latest version of the market-leading software product for high-performance computing workload management and job scheduling on clusters and supercomputers.

Computing With MPI in Heterogeneous Environments


Designating the appropriate provider for large MPI applications is critical to taking advantage of all of the compute power available. “A modern HPC system with multiple host cpus and multiple coprocessors such as the Intel Xeon Phi coprocessor housed in numerous racks can be optimized for maximum application performance with intelligent thread placement.”

Concurrent Kernel Offloading


“The combination of using a host cpu such as an Intel Xeon combined with a dedicated coprocessor such as the Intel Xeon Phi coprocessor has been shown in many cases to improve the performance of an application by significant amounts. When the datasets are large enough, it makes sense to offload as much of the workload as possible. But is this the case when the potential offload data sets are not as large?”

Numerical Optimization for Deep Learning


“With the advent of massively parallel computing coprocessors, numerical optimization for deep-learning disciplines is now possible. Complex real-time pattern recognition, for example, that can be used for self driving cars and augmented reality can be developed and high performance achieved with the use of specialized, highly tuned libraries. By just using the Message Passing Interface (MPI) API, very high performance can be attained on hundreds to thousands of Intel Xeon Phi processors.”

RCE Podcast: Jonathan Dursi on “HPC is dying, and MPI is killing it”

Jonathan Dursi,

In this RCE Podcast, Brock Palen and Jeff Squyres speak with Jonathan Dursi about his recent editorial entitled HPC is dying, and MPI is killing it. The article that spawned a lot of attention in good discussion for our community.

Interview: AutoTune – Automated Optimization and Tuning

Michael Gerndt, Technische Universität München

The main goal of AutoTune is the automatic optimization of applications in the area of HPC, targeting both performance optimization and energy efficiency. In this interview, Michael Gerndt from the Technische Universitaet Muenchen tells us more about the project.

Come to Portland! MPI 3.1 is Just Around the Corner

Jeff Squyres

Over at Cisco’s High Performance Computing Networking Blog, Jeff Squyres writes that MPI 3.1 is coming soon.

Using Advanced MPI: Modern Features of the Message-Passing Interface

We need a reviewer for this book!

“These authors are experts in MPI, but more importantly, they are experts at teaching MPI. If you want to master MPI, there no better guides than this book and its companion.”

Bill Gropp Presents: MPI for Scalable Computing

Bill Gropp

In this video from the 2014 Argonne Training Program on Extreme-Scale Computing, Bill Gropp from NCSA presents: Cost of Unintended Synchronization. “At ATPESC 2014, we captured 67 hours of lectures in 86 videos of presentations by pioneers and elites in the HPC community on topics ranging from programming techniques and numerical algorithms best suited for leading-edge HPC systems to trends in HPC architectures and software most likely to provide performance portability through the next decade and beyond.”

HPC Thought Leaders Publish New Book on Using Advanced MPI


A new MPI book is available for pre-order on Amazon. Written by William Gropp, Torsten Hoefler, Ewing Lusk, and Rajeev Thakur, Using Advanced MPI: Modern Features of the Message-Passing Interface offers a practical guide to the advanced features of the MPI (Message-Passing Interface) standard library for writing programs for parallel computers. It covers new features added in MPI-3, the latest version of the MPI standard, and updates from MPI-2.