High Performance Computing integrator OCF is supporting scientific research at the UK Atomic Weapons Establishment (AWE), with the design, testing and implementation of a new HPC cluster and a separate big data storage system. “The new HPC system is built on IBM’s POWER8 architecture and a separate parallel file system, called Cedar 3, built on IBM Spectrum Scale. In early benchmark testing, Cedar 3 is operating 10 times faster than the previous high-performance storage system at AWE. Both server and storage systems use IBM Spectrum Protect for data backup and recovery.”
Over at the Altair Blog, Jochen Krebs writes that the new HPC cluster at BASF will run PBS Works workload management software. “What does it take to go from months to mere days in gaining results when conducting research? Supercomputing now plays a vital role in the advancement of systems efficiency across industries. On March 17th, BASF and HPE announced in a press release that BASF has chosen HPE to build a new supercomputer for chemical research projects. HPE’s Apollo System supercomputer will help BASF to reduce computer simulation and modeling times from months to days and will drive the digitalization of BASF’s worldwide research activities.”
“This talk will present the motivating factors for considering OpenStack for the management of research computing infrastructure. Stig Telfer will give an overview of the differences in design criteria between cloud, HPC and data analytics, and how these differences can be mitigated through architectural and configuration choices of an OpenStack private cloud. Some real-world examples will be given that demonstrate the potential for using OpenStack for managing HPC infrastructure. This talk will present ways that the HPC community can gain the benefits of using software-defined infrastructure without paying the performance overhead.”
Ingolf Wittmann from IBM presented this talk for the Switzerland HPC Conference. “This presentation will point out based on real examples how HPC environments can benefit from such solutions and technologies to drive cognitive solutions, machine/deep learning where we can ask ourselves, ‘What will be possible in the near future – can the future computers be smarter than humans?”
“This is the first in a series of short videos to introduce you to parallel programming with OpenACC and the PGI compilers, using C++ or Fortran. You will learn by example how to build a simple example program, how to add OpenACC directives, and to rebuild the program for parallel execution on a multicore system. To get the most out of this video, you should download the example programs and follow along on your workstation.”
“In this talk we will discuss a workflow for building and testing Docker containers and their deployment on an HPC system using Shifter. Docker is widely used by developers as a powerful tool for standardizing the packaging of applications across multiple environments, which greatly eases the porting efforts. On the other hand, Shifter provides a container runtime that has been specifically built to fit the needs of HPC. We will briefly introduce these tools while discussing the advantages of using these technologies to fulfill the needs of specific workflows for HPC, e.g., security, high-performance, portability and parallel scalability.”
Over at the Google Blog, Alex Barrett writes that an MIT math professor recently broke the record for the largest-ever Compute Engine cluster, with 220,000 cores on Preemptible VMs. According to Google, this is the largest known HPC cluster to ever run in the public cloud.
Dr. Eng Lim Goh from Hewlett Packard Enterprise gave this talk at the HPC User Forum. “SGI’s highly complementary portfolio, including its in-memory high-performance data analytics technology and leading high-performance computing solutions will extend and strengthen HPE’s current leadership position in the growing mission critical and high-performance computing segments of the server market.”
“Just as developers need tools to understand the performance of a CPU intensive application in order to modify the code for higher performance, so do those that develop interactive 3D computer graphics applications. An excellent tool for t this purpose is the Intel Graphics Performance Analyzer set. This tool, which is free to download, can help the developer understand at a very low level how the application is performing, from a number of aspects.”
“Starting today, Intel will contribute all Lustre features and enhancements to the open source community. This will mean that we will no longer provide Intel-branded releases of Lustre, and instead align our efforts and support around the community release.” In related news, former Whamcloud CEO and Lustre team leader Brent Gorda has left the Intel Lustre team for another management position.