“Cavium ThunderX has significant differentiation in the 64-bit ARM market as Cavium is the first ARMv8 vendor to deliver dual socket support with full ARMv8.1 implementation and significant advantage in CPU cores with 48 cores per socket. In addition, ThunderX supports large memory capacity (512GB per socket, 1TB in a 2S system) with excellent memory bandwidth and low memory latency. In addition, ThunderX includes multiple 10 GbE / 40GbE network interfaces delivering excellent IO throughput. These features enable ThunderX to deliver the core performance & scale out capability that the HPC market requires.”
In this video from the HPC User Forum in Tucson, Earl Joseph from IDC presents: 2016 IDC HPC Market Update. “The HPC User Forum was established in 1999 to promote the health of the global HPC industry and address issues of common concern to users. The organization has since grown to 150 members.”
Today ISC 2016 announced the awards-winning research papers for the PRACE ISC Award and the Gauss Award.
If you are in the Northwest and you happen to like surf and turf, have I got a deal for you! Dell is hosting a series of Big Data lunch events in Seattle and Portland at the end of April. On April 26, Dell brings the event to Blueacre Seafood in Seattle. In Portland, lunch is on April 27 at the mighty Fogo de Chao, a Brazilian steak house for the Where’s the Beef? crowd. They’re also coming to Flemings in Salt Lake City on April 28.
In this video from the 2016 GPU Technology Conference, Jason Pai from Supermicro describes the new 1028GQ-TRT SuperServer. With support for up to four Nvidia Tesla K80 GPUs, the 1U superserver offers extreme compute density in 1U of rack space. “From HPC to Deep Learning and Big Data Analytics, denser, more powerful GPU solutions have become a necessity in order to service the next generation of GPU-accelerated applications. At GTC, Supermicro demonstrated how these applications have progressed, and how its GPU solutions are influencing this evolution.”
In this video from the 2016 GPU Technology Conference, Jaan Mannik from One Stop Systems describes the GPUltima system. Delivering up to 1 Petaflop in a rack, the GPUltima is a single 19″ rack comprised of 8 OSS High Density Compute Accelerators (HDCA) each with 16 NIVIDA Dual GPUs (128 total), 16 dual-socket servers, an Infiniband Switch and an Ethernet Switch.”
A design for virtual Ethernet networks over IB is described. The virtual Ethernet networks are implemented as overlays on the IB network. They are managed in a flexible manner using software. Virtual networks can be created, removed, and assigned to servers dynamically using this software. A virtual network can exist entirely on the IB fabric, or it can have an uplink connecting it to physical Ethernet using a gateway. The virtual networks are represented on the servers by virtual network interfaces which can be used with para-virtualized I/O, SRIOV,and non-virtualized I/O. This technology has many uses: communication between applications which are not IB-aware, communication between IB-connected servers and Ethernet-connected servers, and multi-tenancy for cloud environments. It can be used in conjunction with OpenStack, such as for tenant networks.”
NVMe over Fabrics is an enhancement / expansion of the NVMe specification to support remote access to NVMe storage resources via RDMA. This presentation will review the major components of the NVMe over Fabrics specification, describe the host and target stack, discuss the plan for Linux open source drivers, and preview some performance characteristics. The discussion will also touch on possible deployment models.
Over at the SC16 Blog, Dan Stanzione from TACC writes that the conference is starting a multi-year emphasis designed to advance the state of the practice in the HPC community by providing a track for the professionals driving innovation and development in designing, building, and operating the world’s largest supercomputers, along with all of the system and application software that make them run effectively.
“Research computational workflows consist of several pieces of third party software and, because of their experimental nature, frequent changes and updates are commonly necessary thus raising serious deployment and reproducibility issues. Docker containers are emerging as a possible solution for many of these problems, as they allow the packaging of pipelines in an isolated and self-contained manner. This presentation will introduce our experience deploying genomic pipelines with Docker containers at the Center for Genomic Regulation (CRG). I will discuss how we implemented it, the main issues we faced, the pros and cons of using Docker in an HPC environment including a benchmark of the impact of containers technology on the performance of the executed applications.”