While we’re always on the lookout for HPC news, not everything makes it to the front page. Here are some notable items from this week.
“Rapid growth in the use cases and demands for extreme computing and huge data processing is leading to convergence of the two infrastructures. The trend towards convergence is not only strategic however but rather inevitable as the Moore’s law ends such that sustained growth in data capabilities, not compute, will advance the capacity and thus the overall capacities towards accelerating research and ultimately the industry.”
“In this session we describe how GPUs can be used within virtual environments with near-native performance. We begin by showing GPU performance across four hypervisors: VMWare ESXi, KVM, Xen, and LXC. After showing that performance characteristics of each platform, we extend the results to the multi-node case with nodes interconnected by QDR InfiniBand. We demonstrate multi-node GPU performance using GPUDirect-enabled MPI, achieving efficiencies of 97-99% of a non-virtualized system.”
“OpenACC was applied to the a global high-resolution atmosphere model named NICAM. We executed the dynamical core test without re-writing any specific kernel subroutines for GPU execution. Only 5% of the lines of source code were modified, demonstrating good portability. The results showed that the kernels generated by OpenACC achieved good performance, which was appropriate to the memory performance of GPU, as well as weak scalability. A large-scale simulation was carried out using 2560 GPUs, which achieved 60 TFLOPS.”
In this podcast from the 2015 NCSA Blue Waters Symposium, Arden L. Bement discusses the Blue Waters supercomputer and the future of HPC. Formerly Director of the NSF, Bement keynoted the symposium and is currently the Davis A. Ross Distinguished Professor Emeritus and Adjunct Professor of the College of Technology at Purdue University.
Today CD-adapco announced a significant scalability milestone for its STAR-CCM+ CFD software. Optimized over the course of a year in collaboration with HLRS and SICOS BW, STAR-CCM+ was run on the entirety of the 1.045 PetaFlop Hermit cluster, managing to maintain perfect scalability beyond 55,000 cores.