With 69 contributors from academia and industry, a new book shows how to leverage parallelism on processors and coprocessors with the same programming, providing detailed illustrations of effective ways to combine Intel Xeon Phi coprocessors with multicore processors.
In this video from the 2014 Argonne Training Program on Extreme-Scale Computing, James Reinders presents: Computer Architecture and Structured Parallel Programming. “At ATPESC 2014, we captured 67 hours of lectures in 86 videos of presentations by pioneers and elites in the HPC community on topics ranging from programming techniques and numerical algorithms best suited for leading-edge HPC systems to trends in HPC architectures and software most likely to provide performance portability through the next decade and beyond.”
“A crazy idea was born at ISC14 while answering questions about the new energy metrics in Allinea Performance Reports and Allinea MAP – could the information in these reports show us an easy way to increase energy efficiency without having to change the program? The idea was to use CPU frequency scaling to run memory-bound jobs at a lower clock frequency. In Lazy Energy Efficiency Challenge One, I found that on a synthetic benchmark I could increase the iterations per watt by 19% on a memory-bound code.”
“Scalable parallel programmers (and desktop parallel programmers, for that matter) would benefit greatly from a productive new language supporting better syntax, features and optimizations than today’s parallel notations offer. HPC has a rich set of lower-level technologies upon which to build and numerous failed parallel languages from which to learn.”
A new MPI book is available for pre-order on Amazon. Written by William Gropp, Torsten Hoefler, Ewing Lusk, and Rajeev Thakur, Using Advanced MPI: Modern Features of the Message-Passing Interface offers a practical guide to the advanced features of the MPI (Message-Passing Interface) standard library for writing programs for parallel computers. It covers new features added in MPI-3, the latest version of the MPI standard, and updates from MPI-2.
Researchers at the San Diego Supercomputer Center (SDSC) at the University of California, San Diego have received a three-year, $1.3 million award from the National Science Foundation (NSF) to develop a web-based resource that lets scientists seamlessly share and access preliminary results and transient data from research on a variety of platforms, including mobile devices. […]