IBTA Celebrates 20 Years of Growth and Industry Success

“This year, the IBTA is celebrating 20 years of growth and success in delivering these widely used and valued technologies to the high-performance networking industry. Over the past two decades, the IBTA has provided the industry with technical specifications and educational resources that have advanced a wide range of high-performance platforms. InfiniBand and RoCE interconnects are deployed in the world’s fastest supercomputers and continue to significantly impact future-facing applications such as Machine Learning and AI.”

A Performance Comparison of Different MPI Implementations on an ARM HPC System

Nicholas Brown from EPCC gave this talk at the MVAPICH User Group. “In this talk I will describe work we have done in exploring the performance properties of MVAPICH, OpenMPI and MPT on one of these systems, Fulhame, which is an HPE Apollo 70-based system with 64 nodes of Cavium ThunderX2 ARM processors and Mellanox InfiniBand interconnect. In order to take advantage of these systems most effectively, it is very important to understand the performance that different MPI implementations can provide and any further opportunities to optimize these.”

Overview of the MVAPICH Project and Future Roadmap

DK Panda gave this talk at the MVAPICH User Group. “This talk will provide an overview of the MVAPICH project (past, present, and future). Future roadmap and features for upcoming releases of the MVAPICH2 software family (including MVAPICH2-X and MVAPICH2-GDR) for HPC and Deep Learning will be presented. Features and releases for Microsoft Azure and Amazon AWS will also be presented. Current status and future plans for OSU INAM, OMB, and Best Practices Page will also be presented.”

Video: InfiniBand In-Network Computing Technology and Roadmap

Gilad Shainer from Mellanox gave this talk at the MVAPICH User Group. “In-Network Computing transforms the data center interconnect to become a “distributed CPU”, and “distributed memory”, enables to overcome performance barriers and to enable faster and more scalable data analysis. These technologies are in use at some of the recent large scale supercomputers around the world, including the top TOP500 platforms. The session will discuss the InfiniBand In-Network Computing technology and performance results, as well as view to future roadmap.”

Mellanox HDR 200G InfiniBand is powering next-gen supercomputers

Today Mellanox announced that HDR 200G InfiniBand is powering the next generation of supercomputers world-wide, enabling higher levels of research and scientific discovery. HDR 200G InfiniBand solutions include the ConnectX-6 adapters, Mellanox Quantum switches, LinkX cables and transceivers and software packages. With its highest data throughput, extremely low latency, and smart In-Network Computing acceleration engines, HDR InfiniBand provides world leading performance and scalability for the most demanding compute and data applications.

Mellanox Rocks the TOP500 with Ethernet and InfiniBand

Today Mellanox announced that the company’s InfiniBand solutions accelerate six of the top ten HPC and AI Supercomputers on the June TOP500 list. The six systems Mellanox accelerates include the top three, and four of the top five: The fastest supercomputer in the world at Oak Ridge National Laboratory, #2 at Lawrence Livermore National Laboratory, #3 at Wuxi Supercomputing Center in China, #5 at Texas Advanced Computing Center, #8 at Japan’s Advanced Industrial Science and Technology, and #10 at Lawrence Livermore National Laboratory. “HDR 200G InfiniBand, the fastest and most advanced interconnect technology, makes its debut on the list, accelerating four supercomputers worldwide, including the fifth top-ranked supercomputer in the world located at the Texas Advanced Computing Center, which also represents the fastest supercomputer built in 2019.”

Agenda Posted: Exacomm 2019 Workshop at ISC High Performance

“The goal of this workshop is to bring together researchers and software/hardware designers from academia, industry and national laboratories who are involved in creating network-based computing solutions for extreme scale architectures. The objectives of this workshop will be to share the experiences of the members of this community and to learn the opportunities and challenges in the design trends for exascale communication architectures.”

Radio Free HPC Ponders Amdahl’s Law, GPUs, and the ASC19 Student Cluster Competition

In this podcast, the Radio Free HPC team reviews the results of the ASC19 Student Cluster Competition and HPC workshop in China. “For the first time, a non-mainland-Chinese team wins the top spot. Taiwan takes the gold in part by their stellar performance in HPCG benchmark where they achieved 2 TFlops, some 25% better than the 2nd best team. The system was a 5-node cluster with Infiniband FDR interconnect.”

Scalable and Distributed DNN Training on Modern HPC Systems

DK Panda from Ohio State University gave this talk at the Swiss HPC Conference. “We will provide an overview of interesting trends in DNN design and how cutting-edge hardware architectures are playing a key role in moving the field forward. We will also present an overview of different DNN architectures and DL frameworks. Most DL frameworks started with a single-node/single-GPU design.”

Excelero Powers AI as a Service with Shared NVMe at InstaDeep

“InstaDeep offers a pioneering AI as a Service solution enabling organizations of any size to leverage the benefits of AI and Machine Learning without the time, costs and expertise required to run their own AI stacks. Excelero’s NVMesh, in turn, allows InstaDeep to access the low-latency, high-bandwidth performance that is essential for running customer AI and ML workloads efficiently – and gain the scalability vital to InstaDeep’s own rapid growth.”