“As we see Moore’s Law alive and well, more and more parallelism is introduced into all computing platforms and on all levels of integration and programming to achieve higher performance and energy efficiency. We will discuss Multi- and Many-Core solutions for highly parallel workloads with general purpose and energy efficient technologies. We will also touch on the challenges and opportunities for parallel programming models, methodologies and software tools to achieve highly efficient and highly productive parallel applications. At the end we will take a brief look towards Exascale computing.”
The Open Compute Project, initiated by Facebook as a way to increase computing power while lowering associated costs with hyper-scale computing, has gained a significant industry following. This guide to Open Computing is design to help organizations optimize their HPC environment to achieve higher performance at a lower operating cost.
In this episode, the Radio Free HPC team wraps up the GPU Technology Conference. The theme of the show this year was Deep Learning, a topic that is heating up the market for GPUs with challenges like image recognition and self-driving cars. As a sister conference, the OpenPOWER Summit this week in San Jose showcased the first OpenPower hardware, including a prototype HPC server from IBM that will pave the way to the two IBM/Nvidia/Mellanox Coral supercomputers expected in 2017.
“Does it matter if women in HPC are clustered in certain areas of work? I think that it does. It is vital in an industry where multi-skilled, interdisciplinary teams are seen as the model of the future that women are not there just to make up the numbers, but are given every support and encouragement in achieving their full potential in whatever roles in HPC teams they feel themselves most suited.”
“I came to IBM via the acquisition of Platform Computing. There’s also been other IBM assets around HPC, namely GPFS. What’s been the evolution of those items as well and how they really come together under this concept of software-defined infrastructure, and how we’re now taking these capabilities and expanding them into other initiatives that have sort of bled into the HPC space.”
For about 40 years, developers and users could count on an increase in CPU performance that would make applications run faster. However, with the slowdown in constant clock rate increases being replaced by additional core counts and even more new instructions, rethinking algorithms, their use of the latest APIs, and using the latest compilers has become critical for the next generation of application performance enhancements.