Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Podcast: Extreme Power and Cooling Efficiency at DownUnder Geosolutions

In this podcast, the Radio Free HPC team looks at an interesting presentation on Power and Cooling by Stuart Midgley from DownUnder Geosolutions. Their 250 Petaflop “Bubba” supercomputer uses immersive cooling for extreme density and power efficiency.

Deep Learning at scale for the construction of galaxy catalogs

A team of scientists is now applying the power of artificial intelligence (AI) and high-performance supercomputers to accelerate efforts to analyze the increasingly massive datasets produced by ongoing and future cosmological surveys. “Deep learning research has rapidly become a booming enterprise across multiple disciplines. Our findings show that the convergence of deep learning and HPC can address big-data challenges of large-scale electromagnetic surveys.”

NVIDIA Powers Rosie Supercomputer at MSOE

An NVIDIA GPU-powered supercomputer named “Rosie” is at the heart of a new computational science facility at the Milwaukee School of Engineering. “Housed in a glass-walled area within the newly constructed four-story Diercks Hall, the new NVIDIA-powered AI supercomputer includes three NVIDIA DGX-1 pods, each with eight NVIDIA V100 Tensor Core GPUs, and 20 servers each with four NVIDIA T4 GPUs. The nodes are joined together by Mellanox networking fabric and share 200TB of network-attached storage. Rare among supercomputers in higher education, the system —which provides 8.2 petaflops of deep learning performance — will be used for teaching undergrad classes.”

Supermicro Enhances SuperWorkstations for Desktop HPC

Today Supermicro Computer launched a new server-grade high-end workstation to its broad portfolio of fully configurable SuperWorkstation systems. The SYS-5049A-T joins a robust range of solutions supporting dynamic computational workloads for demanding requirements found in scientific research, deep learning (DL), artificial intelligence (AI), augmented reality (AR), and 3-D modeling with real-time simulation.

Applying Cloud Techniques to Address Complexity in HPC System Integrations

Arno Kolster from Providentia Worldwide gave this talk at the HPC User Forum. “OLCF and technology consulting company Providentia Worldwide recently collaborated to develop an intelligence system that combines real-time updates from the IBM AC922 Summit supercomputer with local weather and operational data from its adjacent cooling plant, with the goal of optimizing Summit’s energy efficiency. The OLCF proposed the idea and provided facility data, and Providentia developed a scalable platform to integrate and analyze the data.”

Exxact Releases New eBook “Getting Started with Deep Learning”

Today Exxact Corporation, a leading provider of High Performance Computing (HPC) and GPU-accelerated deep learning solutions, announced that their new, comprehensive eBook: “Getting Started with Deep Learning” is now available globally for free digital download. “We’ve condensed everything we’ve learned from building state of the art deep learning systems for some of the top research institutes and fortune 500 companies, so that whoever reads this can plan strategically and avoid costly errors.”

AI on XSEDE Systems Promises Early Prediction of Breast Cancer

Researchers are using XSEDE supercomputers and artificial intelligence to predict breast cancer. “Our local GPUs did not have enough memory to accommodate such a scale of data for AI modeling. It could take weeks to run one experiment without the support of powerful GPUs. Using the GPUs from XSEDE, with larger memory, reduced that to a couple of hours.”

Checkpointing the Un-checkpointable: MANA and the Split-Process Approach

Gene Cooperman from Northeastern University gave this talk at the MVAPICH User Group. “This talk presents an efficient, new software architecture: split processes. The “MANA for MPI” software demonstrates this split-process architecture. The MPI application code resides in “upper-half memory”, and the MPI/network libraries reside in “lower-half memory”.

AMD: Delivering the Future of High-Performance Computing

Dr. Lisa Su from AMD gave this talk at the recent DARPA Electronics Resurgence Initiative Summit. “Optimum system performance requires co-design of silicon chips, system architecture, and software. She presented the example of the Frontier exascale computer system being developed for Oak Ridge National Lab, which should exhibit 1.5 exaflops by 2021. While the highest-performance chips and systems will initially be limited to the most expensive machines, it is expected that similar technology will become available within a few years in data centers, edge computers, and even mobile devices.”

The ABCI Supercomputer: World’s First Open AI Computing Infrastructure

Shinichiro Takizawa from AIST gave this talk at the MVAPICH User Group. “ABCI is the world’s first large-scale Open AI Computing Infrastructure, constructed and operated by AIST, Japan. It delivers 19.9 petaflops of HPL performance and world’ fastest training time of 1.17 minutes in ResNet-50 training on ImageNet datasets as of July 2019. In this talk, we focus on ABCI’s network architecture and communication libraries available on ABCI and shows their performance and recent research achievements.”