Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

XGC Fusion Code Selected for all 3 Pre-exascale Supercomputers

When the DOE’s pre-exascale supercomputers come online soon, all three will be running an optimized version of the XGC dynamic fusion code. Developed by a team at the DOE’s Princeton Plasma Physics Laboratory (PPPL), the XGC code was one of only three codes out of more than 30 science and engineering programs selected to participate in Early Science programs on all three new supercomputers, which will serve as forerunners for even more powerful exascale machines that are to begin operating in the United States in the early 2020s.

Adrian Cockcroft Presents: Shrinking Microservices to Functions

In this fascinating talk, Cockcroft describes how hardware networking has reshaped how services like Machine Learning are being developed rapidly in the cloud with AWS Lamda. “We’ve seen the same service oriented architecture principles track advancements in technology from the coarse grain services of SOA a decade ago, through microservices that are usually scoped to a more fine grain single area of responsibility, and now functions as a service, serverless architectures where each function is a separately deployed and invoked unit.”

DDN and Lustre to Power TSUBAME3.0 Supercomputer

“The IO infrastructure of TSUBAME3.0 combines fast in-node NVMe SSDs and a large, fast, Lustre-based system from DDN. The 15.9PB Lustre* parallel file system, composed of three of DDN’s high-end ES14KX storage appliances, is rated at a peak performance of 150GB/s. The TSUBAME collaboration represents an evolutionary branch of HPC that could well develop into the dominant HPC paradigm at about the time the most advanced supercomputing nations and consortia achieve Exascale computing.”

Ingram Micro Artizen Soutions for HPC

In this video from the GPU Technology Conference, Rick Young from Ingram Micro describes the company’s Artizen HPC solutions. “Available now to channel partners in the U.S., the distributor’s new and exclusive line of Artizen High Performance Computing (HPC) offerings include turnkey high performance servers, ultimate workstations, and customizable supercomputing clusters, as well as computing integration and software installation services.”

Monitoring and Management Interfaces for GPU Devices in a Cluster Environment

“This presentation will provide an overview of the Nvidia Tesla Deployment Kit (TDK) from a user and a system administrator point of view. TDL contains Nvidia Management Library (NVML) and nvidia-healthmon–a tool for detecting and troubleshooting known GPU issues in a cluster environment. Usage models within a cluster environment will be presented along with a discussion on how existing resource management tools can be extended to improve allocation and accounting of GPU resources.”

Video: ClusterStor Update

“HPC storage solutions and futures continue to evolve as growth and performance requirements permeate every HPC market segment. Torben discusses these challenges and how the company’s storage solutions are addressing these shifting needs with new developments around disk drives, RAID, CIFS, security, small file handling, and other related technologies.”

Managing the GPUs of Your Cluster in a Flexible Way with rCUDA

In this talk, we introduce the rCUDA remote GPU virtualization framework, which has been shown to be the only one that supports the most recent CUDA versions, in addition to leverage the InfiniBand interconnect for the sake of performance. Furthermore, we also present the last developments within this framework, related with the use of low-power processors, enhanced job schedulers, and virtual machine environments.”

Troy Baer from NICS Wins Lifetime Achievement Adaptie Award

“Congratulations go out to Troy Baer, HPC system administrator at the National Institute for Computational Sciences (NICS), University of Tennessee. Troy Baer’s contributions in scheduling and resource management using Moab have helped Kraken—NICS’ flagship computing resource and the first academic computer to break the petaflop barrier—achieve outstanding 90-95% utilization rates since 2010. Baer’s administrative capabilities enable researchers in numerous scientific arenas, from climate to materials science to astrophysics, to achieve breakthroughs not yet possible on other resources. In November 2012, Baer helped NICS’ Beacon system secure a No. 1 ranking on the Green500 list of energy-efficient supercomputers.”

DK Panda Presents: Programming Models for Exascale Systems

“This talk will focus on programming models and their designs for upcoming exascale systems with millions of processors and accelerators. Current status and future trends of MPI and PGAS (UPC and OpenSHMEM) programming models will be presented. We will discuss challenges in designing runtime environments for these programming models by taking into account support for multi-core, high-performance networks, GPGPUs, Intel MIC, scalable collectives (multi-core-aware, topology-aware, and power-aware), non-blocking collectives using Offload framework, one-sided RMA operations, schemes and architectures for fault-tolerance/fault-resilience.”

High Performance Computing at CSCS

“With around 3.2 billion computer operations (3.2 gigaflops) per watt, the combination of GPUs CPUs makes “Piz Daint” one of the world’s most energy-­efficient supercomputers in the petaflop performance class.”