Over at the SUSE Blog, Jay Kruemcke writes that the High-Performance Computing Module (HPC Module) for SUSE Linux Enterprise (SLES) is now available for 64-bit ARM (AArch64) systems. The HPC Module is delivered as an add-on product to SUSE Linux Enterprise Server. “In summary, the HPC module allows us to keep the content closer to what’s happening in the HPC community upstream, providing more leading-edge tools in a more manageable fashion, leveraging a different lifecycle than the base SUSE Linux Enterprise Server. The new HPC module contains packages to optimize and manage HPC systems, and build HPC applications – building a bridge between the base server and an HPC stack (such as the stack provided by OpenHPC). This journey has started – some packages have already been made public and we have much more in the works and in our release queue.”
In this video, researchers describe how the Tianhe-1 supercomputer supports scientific research. “Currently #43 on the TOP500, the 2.56 Petaflop Tianhe-1A carries out 1,400 computing tasks per day. It is mainly used to serve universities, research institutions, small and medium-sized enterprises, and provide scientific computing services.”
Southern Methodist University is seeking a Senior Linux System Administrator for High Performance Computing (HPC) to build, maintain, operate and manage HPC systems. The individual in this position will have shared support responsibility for university HPC as member of a two-person team.
The University of Melbourne has launched a new HPC service called Spartan that combines traditional HPC with a flexible cloud computing component. “Many research projects demand high speed interconnect,” said Bernard Meade, Head of Research Computer Services at the University of Melbourne. “Spartan can quickly scale into cloud based virtual machines as needed, and expand the HPC system as user needs evolve. Traditional HPC systems are typically tailored for a few specific use cases, but in practice are used for a much wider variety of applications, resulting in less than optimal usage.”
“Based on a containerized HPC environment this talk shows of a state-of-the-art stack including performance monitoring, log event handling and GraphDB based inventory to provide insights into what is going on within a SLURM cluster. The framework used is QNIBTerminal incorporating the ELK stack, a graphite backend and neo4j as a GraphDB.”