Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


A Recap of the 2017 OpenFabrics Workshop

The 13th Annual OpenFabrics Alliance (OFA) Workshop wrapped at the end of March with a look toward the future. The annual gathering, held this year in Austin, Texas, was devoted to advancing cutting edge networking technology through the ongoing collaborative efforts of OpenFabrics Software (OFS) producers and users. With a record 130+ attendees, the 2017 Workshop expanded on the OFA’s commitment to being an open organization by hosting an engaging Town Hall discussion and an At Large Board election, filling two newly added director seats for current members.

Mellanox InfiniBand to Power Science at University of Waterloo

Today Mellanox announced that the University of Waterloo selected Mellanox EDR 100G InfiniBand solutions to accelerate their new supercomputer. The new supercomputer will support a broad and diverse range of academic and scientific research in mathematics, astronomy, science, the environment and more. “The growing demands for research and supporting more complex simulations led us to look for the most advanced, efficient, and scalable HPC platforms,” said John Morton, technical manager for SHARCNET. “We have selected the Mellanox InfiniBand solutions because their smart acceleration engines enable high performance, efficiency and robustness for our applications.”

InfiniBand Roadmap Foretells a World Where Server Connectivity is at 1000 Gb/sec

The InfiniBand Trade Association (IBTA) has updated their InfiniBand Roadmap. With HDR 200 Gb/sec technolgies shipping this year, the roadmap looks out to an XDR world where server connectivity reaches 1000 Gb/sec. “The IBTA‘s InfiniBand roadmap is continuously developed as a collaborative effort from the various IBTA working groups. Members of the IBTA working groups include leading enterprise IT vendors who are actively contributing to the advancement of InfiniBand. The roadmap details 1x, 4x, and 12x port widths with bandwidths reaching 600Gb/s data rate HDR in 2017. The roadmap is intended to keep the rate of InfiniBand performance increase in line with systems-level performance gains.”

IBM’s New PowerAI Software Speeds Deep Learning

IBM PowerAI on Power servers with GPU accelerators provide at least twice the performance of our x86 platform; everything is faster and easier: adding memory, setting up new servers and so on,” said current PowerAI customer Ari Juntunen, CTO at Elinar Oy Ltd. “As a result, we can get new solutions to market very quickly, protecting our edge over the competition. We think that the combination of IBM Power and PowerAI is the best platform for AI developers in the market today. For AI, speed is everything —nothing else comes close in our opinion.”

Liqid Delivers Composable Infrastructure Solution for Dynamic GPU Resource Allocation

Liqid Inc. has fully integrated GPU support into the Liqid Composable Infrastructure (CI) Platform. “Liqid’s CI Platform is the first solution to support GPUs as a dynamic, assignable, bare-metal resource. With the addition of graphics processing, the Liqid CI Platform delivers the industry’s most fully realized approach to composable infrastructure architecture.”

Call for Papers: Hot Interconnects

The 25th International Symposium on High Performance Interconnects (HotI 2017) has issued its Call for Papers. The event takes place August 29-30 at the Ericsson Campus in Santa Clara, California. “Hot Interconnects is the premier international forum for researchers and developers of state-of-the-art hardware and software architectures and implementations for interconnection networks of all scales, ranging from multi-core on-chip interconnects to those within systems, clusters, data centers, and clouds. This yearly conference is attended by leaders in industry and academia. The atmosphere provides for a wealth of opportunities to interact with individuals at the forefront of this field.”

Slidecast: BXI – Bull eXascale Interconnect

“The interconnect is going to be a key enabling technology for exascale systems. This is why one of the cornerstones of Bull’s exascale program is the development of our own new-generation interconnect. The Bull eXascale Interconnect or BXI introduces a paradigm shift in terms of performance, scalability, efficiency, reliability and quality of service for extreme workloads.”

Interview: Peter Braam on How Campaign Storage Bridges the Small & Big, Fast & Slow

Peter Braam is well-known in the HPC Community for his early work with Lustre and other projects like the SKA telescope Science Data Processor. As one of the featured speakers at the upcoming MSST Mass Storage Conference, Braam will describe how his Campaign Storage Startup provides tools for massive parallel data movement between the new low cost, industry standard campaign storage tiers with premium storage for performance or availability.

Rock Stars of HPC: DK Panda

As our newest Rock Star of HPC, DK Panda sat down with us to discuss his passion for teaching High Performance Computing. “During the last several years, HPC systems have been going through rapid changes to incorporate accelerators. The main software challenges for such systems have been to provide efficient support for programming models with high performance and high productivity. For NVIDIA-GPU based systems, seven years back, my team introduced a novel `CUDA-aware MPI’ concept. This paradigm allows complete freedom to application developers for not using CUDA calls to perform data movement.”

Jülich to Build 5 Petaflop Supercomputing Booster with Dell

Today Intel and the Jülich Supercomputing Centre together with ParTec and Dell today announced plans to develop and deploy a next-generation modular supercomputing system. Leveraging the experience and results gained in the EU-funded DEEP and DEEP-ER projects, in which three of the partners have been strongly engaged, the group will develop the necessary mechanisms required to augment JSC’s JURECA cluster with a highly-scalable component named “Booster” and being based on Intel’s Scalable Systems Framework (Intel SSF).