It can be tough to keep up with all that’s happening in HPC these days. Here are some notable items that may have not made it to the front page.
“The drive toward exascale computing, a renewed emphasis on data-centric processing, energy efficiency concerns, and the limitations of memory and I/O performance are all working to reshape High Performance Computing platforms. Many-core accelerators, flash storage, 3D memory, integrated networking, and optical interconnects are just some of the technologies propelling these future architectures. In concert with those developments, the HPC vendor landscape has been churning in response to broader market forces, and these events are going to drive some interesting changes in the coming year.”
“This talk will focus on programming models and their designs for upcoming exascale systems with millions of processors and accelerators. Current status and future trends of MPI and PGAS (UPC and OpenSHMEM) programming models will be presented. We will discuss challenges in designing runtime environments for these programming models by taking into account support for multi-core, high-performance networks, GPGPUs, Intel MIC, scalable collectives (multi-core-aware, topology-aware, and power-aware), non-blocking collectives using Offload framework, one-sided RMA operations, schemes and architectures for fault-tolerance/fault-resilience.”
CORAL (Collaboration of Oak Ridge, Argonne and Lawrence Livermore National Labs) is a project that was launched in 2013 to develop the technology and meet the Department of Energy’s 2017-2018 leadership computing needs with supercomputers. The collaboration between Mellanox, IBM and NVIDIA was selected by the CORAL project team after a comprehensive evaluation of future technologies from a variety of vendors. The development of these supercomputers is well underway with installation expected in 2017.
“Written by one of the foremost experts in high-performance computing and the inventor of Gustafson’s Law, The End of Error: Unum Computing explains a new approach to computer arithmetic: the universal number (unum). The unum encompasses all IEEE floating-point formats as well as fixed-point and exact integer arithmetic. This new number type obtains more accurate answers than floating-point arithmetic yet uses fewer bits in many cases, saving memory, bandwidth, energy, and power.”