Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


UK to Launch Six HPC Centers

The UK is launching six HPC centers this month. Funded by £20 million from the Engineering and Physical Sciences Research Council (EPSRC) the centres are located around the UK, at the universities of Cambridge, Edinburgh, Exeter, and Oxford, Loughborough University, and UCL. “These centres will enable new discoveries, drive innovation and allow new insights into today’s scientific challenges. They are important because they address an existing gulf in capability between local university systems and the UK National Supercomputing Service ARCHER,” said Professor Philip Nelson, EPSRC’s Chief Executive.

Video: Computational Fluid Dynamics for Surgical Planning

“The current versions of the codes use MPI and depend on finer and finer meshes for higher accuracy which are computationally demanding. To overcome the demands, the team has gained access to their state-of-the-art cluster equipped with POWER CPUs and Tesla P100 GPUs — and turning to OpenACC and machine learning to accelerate their science. This has allowed them to spend the least resources on programming, and effectively utilize available compute resources.”

Berkeley Lab Tunes NWChem for Intel Xeon Phi Processor

A team of researchers at Berkeley Lab, PNNL, and Intel are working hard to make sure that computational chemists are prepared to compute efficiently on next-generation exascale machines. Recently, they achieved a milestone, successfully adding thread-level parallelism on top of MPI-level parallelism in the planewave density functional theory method within the popular software suite NWChem. “Planewave codes are useful for solution chemistry and materials science; they allow us to look at the structure, coordination, reactions and thermodynamics of complex dynamical chemical processes in solutions and on surfaces.”

Ingram Micro Artizen Soutions for HPC

In this video from the GPU Technology Conference, Rick Young from Ingram Micro describes the company’s Artizen HPC solutions. “Available now to channel partners in the U.S., the distributor’s new and exclusive line of Artizen High Performance Computing (HPC) offerings include turnkey high performance servers, ultimate workstations, and customizable supercomputing clusters, as well as computing integration and software installation services.”

Monitoring and Management Interfaces for GPU Devices in a Cluster Environment

“This presentation will provide an overview of the Nvidia Tesla Deployment Kit (TDK) from a user and a system administrator point of view. TDL contains Nvidia Management Library (NVML) and nvidia-healthmon–a tool for detecting and troubleshooting known GPU issues in a cluster environment. Usage models within a cluster environment will be presented along with a discussion on how existing resource management tools can be extended to improve allocation and accounting of GPU resources.”

Video: ClusterStor Update

“HPC storage solutions and futures continue to evolve as growth and performance requirements permeate every HPC market segment. Torben discusses these challenges and how the company’s storage solutions are addressing these shifting needs with new developments around disk drives, RAID, CIFS, security, small file handling, and other related technologies.”

Managing the GPUs of Your Cluster in a Flexible Way with rCUDA

In this talk, we introduce the rCUDA remote GPU virtualization framework, which has been shown to be the only one that supports the most recent CUDA versions, in addition to leverage the InfiniBand interconnect for the sake of performance. Furthermore, we also present the last developments within this framework, related with the use of low-power processors, enhanced job schedulers, and virtual machine environments.”

Troy Baer from NICS Wins Lifetime Achievement Adaptie Award

“Congratulations go out to Troy Baer, HPC system administrator at the National Institute for Computational Sciences (NICS), University of Tennessee. Troy Baer’s contributions in scheduling and resource management using Moab have helped Kraken—NICS’ flagship computing resource and the first academic computer to break the petaflop barrier—achieve outstanding 90-95% utilization rates since 2010. Baer’s administrative capabilities enable researchers in numerous scientific arenas, from climate to materials science to astrophysics, to achieve breakthroughs not yet possible on other resources. In November 2012, Baer helped NICS’ Beacon system secure a No. 1 ranking on the Green500 list of energy-efficient supercomputers.”

DK Panda Presents: Programming Models for Exascale Systems

“This talk will focus on programming models and their designs for upcoming exascale systems with millions of processors and accelerators. Current status and future trends of MPI and PGAS (UPC and OpenSHMEM) programming models will be presented. We will discuss challenges in designing runtime environments for these programming models by taking into account support for multi-core, high-performance networks, GPGPUs, Intel MIC, scalable collectives (multi-core-aware, topology-aware, and power-aware), non-blocking collectives using Offload framework, one-sided RMA operations, schemes and architectures for fault-tolerance/fault-resilience.”

High Performance Computing at CSCS

“With around 3.2 billion computer operations (3.2 gigaflops) per watt, the combination of GPUs CPUs makes “Piz Daint” one of the world’s most energy-­efficient supercomputers in the petaflop performance class.”