Texas A&M University’s High Performance Research Computing (HPRC) center is the latest Intel® Parallel Computing Center. “HPRC is proud to be recognized as an Intel Parallel Computing Center,” said Honggao Liu, director of High Performance Research Computing. “At HPRC we use high-performance computing to unite experts in numerous fields of study. This grant and multi-disciplinary project will allow us to better understand and solve issues within this critical software.”
Today Allinea announced plans to showcase its software tools for developing and optimizing high performance code at the GPU Technology Conference April 4-7 in San Jose. The company will highlight the best practices required to unleash the potential performance within the latest generation of NVIDIA GPUs for a wide range of software applications.
Kenneth Hoste from the University Ghent presented this tutorial at the Switzerland HPC Conference. “One unnecessarily time-consuming task for HPC user support teams is installing software for users. Due to the advanced nature of a supercomputing system (think: multiple multi-core modern microprocessors (possibly next to co-processors like GPUs), the availability of a high performance network interconnect, bleeding edge compilers & libraries, etc.), compiling the software from source on the actual operating system and system architecture that it is going to be used on is typically highly preferred over using readily available binary packages that were built in a generic way.
“Containers wrap up software with all its dependencies in packages that can be executed anywhere. This can be specially useful in HPC environments where, often, getting the right combination of software tools to build applications is a daunting task. However, typical container solutions such as Docker are not a perfect fit for HPC environments. Instead, Shifter is a better fit as it has been built from the ground up with HPC in mind. In this talk, we show you what Shifter is and how to leverage from the current Docker environment to run your ap- plications with Shifter.”
Today the OpenACC standards group announced a set of additional hackathons and a broad range of learning opportunities taking place during the upcoming GPU Technology Conference being held in San Jose, CA April 4-7, 2016. OpenACC is a mature and performance-portable path for developing scalable parallel programs across multi-core CPUs, GPU accelerators or many-core processors.
DK Panda from Ohio State University presented this talk at the Switzerland HPC Conference. “This talk will focus on challenges in designing runtime environments for Exascale systems with millions of processors and accelerators to support various programming models. We will focus on MPI, PGAS (OpenSHMEM, CAF, UPC and UPC++) and Hybrid MPI+PGAS programming models by taking into account support for multi-core, high-performance networks, accelerators (GPUs and Intel MIC) and energy-awareness. Features and sample performance numbers from the MVAPICH2 libraries will be presented.”
“EasyBuild, a software build and installation framework, can be used to automatically install software and generate environment modules. By using a hierarchical module naming scheme to offer environment modules to users in a more structured way, and providing Lmod, a modern tool for working with environment modules, we help typical users avoid common mistakes while giving power users the flexibility they demand. EasyBuild is developed by the High-Performance Computing team at Ghent University together with the members of the EasyBuild community, and is made available under the GNU General Public License (GPL) version 2.”
Dr. Rosa Badia from BSC/CNS presented this Invited Talk at SC15. “StarSs (Star superscalar) is a task-based family of programming models that is based on the idea of writing sequential code which is executed in parallel at run-time taking into account the data dependencies between tasks. The talk will describe the evolution of this programming model and the different challenges that have been addressed in order to consider different underlying platforms from heterogeneous platforms used in HPC to distributed environments, such as federated clouds and mobile systems.”
In this slidecast, Jeff Squyres from Cisco Systems presents: How to make MPI Awesome – MPI Sessions. As a proposal for future versions of the MPI Standard, MPI Sessions could become a powerful tool tool to improve system resiliency as we move towards exascale. “Now that we have brought these ideas to a larger audience, my hope is that we (the Forum) start refining these ideas to fit them into a future release of the MPI standard. Meaning: please don’t assume that exactly what is proposed in these slides are going to make it into the MPI standard.”