“PushToCompute is the easiest and most advanced DevOps pipeline for high performance applications available today”, said Nimbix CTO Leo Reiter. “It seamlessly enables serverless computing of even the most complex workflows, greatly simplifying application deployment at scale, and eliminating the need for any platform orchestration or user interface work. Developers simply focus on their specific functionality, rather than on building cloud capabilities into their applications.”
In this special guest feature from Scientific Computing World, Wolfgang Gentzsch explains the role of HPC container technology in providing ubiquitous access to HPC. “The advent of lightweight pervasive, packageable, portable, scalable, interactive, easy to access and use HPC application containers based on Docker technology running seamlessly on workstations, servers, and clouds, is bringing us ever closer to the democratization of HPC.”
Based on the “Barreleye” platform design pioneered by Rackspace and promoted by the OpenPOWER Foundation and the Open Compute Project (OCP) Foundation, Penguin Magna 1015 targets memory and I/O intensive workloads, including high density virtualization and data analytics. The Magna 1015 system uses the Open Rack physical infrastructure defined by the OCP Foundation and adopted by the largest hyperscale data centers, providing operational cost savings from the shared power infrastructure and improved serviceability.
Today the University of Alabama at Birmingham unveiled a new supercomputer powered by Dell. With a peak performance of 110 Teraflops, the system is 10 times faster than its predecessor. “With their new Dell EMC HPC cluster, UAB researchers will have the compute and storage they need to aggressively research, uncover and apply knowledge that changes the lives of individuals and communities in many areas, including genomics and personalized medicine.”
In this video from the 2016 HPC User Forum in Austin, a select panel of HPC vendors describe their disruptive technologies for high performance computing. Vendors include: Altair, SUSE, ARM, AMD, Ryft, Red Hat, Cray, and Hewlett Packard Enterprise. “A disruptive innovation is an innovation that creates a new market and value network and eventually disrupts an existing market and value network, displacing established market leading firms, products and alliances.”
“This project will make a substantial contribution to advancing wind energy,” said Steve Hammond, NREL’s Director of Computational Science and the principal investigator on the project. “It will advance our fundamental understanding of the complex flow physics of whole wind plants, which will help further reduce the cost of electricity derived from wind energy.”
Engineers of the Hikari HVDC power feeding system predict it will save 15 percent compared to conventional systems. “The 380 volt design reduces the number of power conversions when compared to AC voltage systems,” said James Stark, director of Engineering and Construction at the Electronic Environments Corporation (EEC), a Division of NTT FACILITIES. “What’s interesting about that,” Stark added, “is the computers themselves – the supercomputer, the blade servers, cooling units, and lighting – are really all designed to run on DC voltage. By supplying 380 volts DC to Hikari instead of having an AC supply with conversion steps, it just makes a lot more sense. That’s really the largest technical innovation.”
Loyola University Maryland has been awarded a $280,120 grant from the National Science Foundation (NSF) to build an HPC cluster that will exponentially expand research opportunities for faculty and students across disciplines.
“The drive towards Exascale computing requires cooling the next generation of extremely hot CPUs, while staying within a manageable power envelope,” said Bob Bolz, HPC and Data Center Business development at Aquila. “Liquid cooling holds the key. Aquarius is designed from the ground up to meet reliability and the feature-specific demands of high performance and high density computing. Our design goal was to reduce the cost of cooling server resources to well under 5% of overall data center usage.”
Vectorization and threading are critical to using such innovative hardware product such as the Intel Xeon Phi processor. Using tools early in the design and development processor that identify where vectorization can be used or improved will lead to increased performance of the overall application. Modern tools can be used to determine what might be blocking compiler vectorization and the potential gain from the work involved.