Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Oakforest-PACS: Overview of the Fastest Supercomputer in Japan

Prof. Taisuke Boku from the University of Tsukuba & JCAHPC presented this talk at the DDN User Group at SC16. “Thanks to DDN’s IME Burst Buffer, researchers using Oakforest-PACS at the Joint Center for Advanced High Performance Computing (JCAHPC) are able to improve modeling of fundamental physical systems and advance understanding of requirements for Exascale-level systems architectures. With DDN’s advanced technology, JCAHPC has achieved effective I/O performance exceeding 1TB/s in writing tens of thousands of processes to the same file.”

Altair Moves Forward with Open Source PBS Pro and PBS Cloud at SC16

In this video from SC16, Bill Nitzberg provides an update on the Open Source release of PBS Professional workload management software. After that, Jeremie Bourdoncle announces the new Altair PBS Cloud. “Altair is excited to announce the upcoming availability of Altair PBS Cloud, its latest appliance solution to further cloud computing for organizations. Altair PBS Cloud is the solution to build and run high-performance computing (HPC) appliances for both public clouds, private clouds, and bare-metal infrastructure. Altair will release Altair PBS Cloud in the first quarter of 2017 following conclusion of a private preview.”

GPUs & Deep Learning in the Spotlight for Nvidia at SC16

In this video from SC16, Roy Kim from Nvidia describes how the company is bringing in a new age of AI with accelerated computing for Deep Learning applications. “Deep learning is the fastest-growing field in artificial intelligence, helping computers make sense of infinite amounts of data in the form of images, sound, and text. Using multiple levels of neural networks, computers now have the capacity to see, learn, and react to complex situations as well or better than humans. This is leading to a profoundly different way of thinking about your data, your technology, and the products and services you deliver.”

How Researchers Will Benefit from Canada’s National Data Cyberinfrastructure

“Individual institutions or organizations will have opportunities to deploy storage locally and can federate their local repository into the national system,” says Dr. Greg Newby, Compute Canada’s Chief Technology Officer. “This provides enhanced privacy and sharing capabilities on a robust, country-wide solution with improved data security and back-up. This is a great solution to address the data explosion we are currently experiencing in Canada and globally.”

HPE Apollo 6500 for Deep Learning

“With up to eight high performance NVIDIA GPUs designed for maximum transfer bandwidth, the HPE Apollo 6500 is purpose-built for HPC and deep learning applications. Its high ratio of GPUs to CPUs, dense 4U form factor and efficient design enable organizations to run deep learning recommendation algorithms faster and more efficiently, significantly reducing model training time and accelerating the delivery of real-time results, all while controlling costs.”

Kx Streaming Analytics Crunches 1.2 Billion NYC Taxi Data Points using Intel Xeon Phi

“The complexity and high costs of architecting and maintaining streaming analytics solutions often make it difficult to get new projects off the ground. That’s part of the reason Kx, a leading provider of high-volume, high-performance databases and real-time analytics solutions, is always interested in exploring how new technologies may help it push streaming analytics performance and efficiency boundaries. The Intel Xeon Phi processor is a case in point. At SC16 in Salt Lake City, Kx used a 1.2 billion record database of New York City taxi cab ride data to demonstrate what the Intel Xeon Phi processor could mean to distributed big data processing. And the potential cost/performance implications were quite promising.”

Building HPC Clusters as Code in the (Almost) Infinite Cloud

“Researchers can run one cluster for 10,000 hours or 10,000 clusters for one hour anytime, from anywhere, and both cost the same in the cloud. And with the availability of Public Data Sets in Amazon S3, petabyte scale data is instantly accessible in the cloud. Attend and learn how to build HPC clusters on the fly, leverage Amazon’s Spot market pricing to minimize the cost of HPC jobs, and scale HPC jobs on a small budget, using all the same tools you use today, and a few new ones too.”

Intel Xeon Phi with Software Defined Visualization at SC16

“Software Defined Visualization (SDVis) is an open source initiative from Intel and industry collaborators to improve the visual fidelity, performance and efficiency of prominent visualization solutions – with a particular emphasis on supporting the rapidly growing “Big Data” usage on workstations through HPC supercomputing clusters without the memory limitations and cost of GPU based solutions. Existing applications can be enhanced using the high performing parallel software rendering libraries OpenSWR, Embree, and OSPRay. At the Intel HPC Developer Conference, Amstutz provided an introduction to this initiative, its benefits, a brief descriptions of accomplishments in the past year and talk about the changes made to Intel provided libraries in the past year.”

Video: Behind the Machine at HPE

How is Hewlett Packard Enterprise reinventing the fundamental architecture on which all computers have been built for the past 60 years? In this video, HPC describes the evolution of The Machine research project – one of the largest and most complex research projects in the company’s history – and how HPE demonstrated the world’s first Memory-Driven Computing architecture.

HIP and CAFFE Porting and Profiling with AMD’s ROCm

In this video from SC16, Ben Sander from AMD presents: HIP and CAFFE Porting and Profiling with AMD’s ROCm. “We are excited to present ROCm, the first open-source HPC/Hyperscale-class platform for GPU computing that’s also programming-language independent. We are bringing the UNIX philosophy of choice, minimalism and modular software development to GPU computing. The new ROCm foundation lets you choose or even develop tools and a language run time for your application. ROCm is built for scale; it supports multi-GPU computing in and out of server-node communication through RDMA.”