Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Podcast: Optimizing Cosmos Code on Intel Xeon Phi

In this TACC podcast, Cosmos code developer Chris Fragile joins host Jorge Salazar for a discussion on how researchers are using supercomputers to simulate the inner workings of Black holes. “For this simulation, the manycore architecture of KNL presents new challenges for researchers trying to get the best compute performance. This is a computer chip that has lots of cores compared to some of the other chips one might have interacted with on other systems,” McDougall explained. “More attention needs to be paid to the design of software to run effectively on those types of chips.”

1000x Faster Deep-Learning at Petascale Using Intel Xeon Phi Processors

A cumulative effort over several years to scale the training of deep-learning neural networks has resulted in the first demonstration of petascale deep-learning training performance, and further to deliver this performance when solving real science problems. The result reflects the combined efforts of NERSC (National Energy Research Scientific Computing Center), Stanford and Intel to solve real world use cases rather than simply report on performance benchmarks.

Intel Parallel Studio XE 2018 For Demanding HPC Applications

“For those that develop HPC applications, there are usually two main areas that must be considered. The first is the translation of the algorithm, whether simulation based, physics based or pure research into the code that a modern computer system can run. A second challenge is how to move from the implementation of an algorithm to the performance that takes advantage of modern CPUs and accelerators.”

JAIST in Japan installs Cray XC40 Supercomputer

Today Cray announced the Japan Advanced Institute for Science and Technology (JAIST) has put a Cray XC40 supercomputer into production. The Cray XC40 supercomputers incorporate the Aries high performance network interconnect for low latency and scalable global bandwidth, as well as the latest Intel Xeon processors, Intel Xeon Phi processors, and NVIDIA Tesla GPU accelerators. “Our new Cray XC40 supercomputer will support our mission of becoming a premier center of excellence in education and research.”

OCF Builds Isca Supercomputer for Life Sciences at University of Exeter

Researchers from across the University of Exeter can now benefit from a new HPC machine – Isca – that was configured and integrated by OCF to give the university a larger capacity for computational research. “We’ve seen in the last few years a real growth in interest in High-Performance Computing from life sciences, particularly with the availability of new high-fidelity genome sequencers, which have heavy compute requirements, and that demand will keep going up.”

Intel Discontinues Xeon Phi 7200-Series Knights Landing Coprocessors

AnandTech reports that Intel is discontinuing it’s line of Knights Landing PCIe based Co-processors. The move is not a surprise given that the company has been producing bootable Xeon Phi processors since early in 2016. “At least for now, Intel does not want to compete against add-on PCIe compute accelerators with its Xeon Phi products. A big question is whether it actually needs to, given the stand-alone capabilities of Xeon Phi and its performance characteristics.”

Feed The Cores – Memory Bandwidth Usage

“Memory bandwidth to the CPUs has always been important. There were typically CPU cores that would wait for the data (if not in cache) from main memory. However, with the advanced capabilities of the Intel Xeon Phi processor, there are new concepts to understand and take advantage of.”

3X Performance Boost Using Intel Advisor and Intel Trace Analyzer in Astrophysics Simulations

On today’s processors, it is crucial to both vectorize (using AVX* or SIMD* instructions) and parallelize software to realize the full performance potential of the processor. By optimizing their MHD astrophysics applications with tools from Intel Parallel Studio XE, and running on the latest Intel hardware, the NSU team achieved a performance speed-up of 3X, cutting the standard time for calculating one problem from one week to just two days.

18 Petaflop Stampede2 Supercomputer Dedicated at TACC

Stampede2 is the newest strategic supercomputing resource for the nation’s research and education community, enabling scientists and engineers across the U.S., from multiple disciplines, to answer questions at the forefront of science and engineering. “Building on the success of the initial Stampede system, the Stampede team has partnered with other institutions as well as industry to bring the latest in forward-looking computing technologies combined with deep computational and data science expertise to take on some of the most challenging science and engineering frontiers,” said Irene Qualters, director of NSF’s Office of Advanced Cyberinfrastructure.

Podcast: A Retrospective on Great Science and the Stampede Supercomputer

TACC will soon deploy Phase 2 of the Stampede II supercomputer. In this podcast, they celebrate by looking back on some of the great science computed on the original Stampede machine. “In 2017, the Stampede supercomputer, funded by the NSF, completed its five-year mission to provide world-class computational resources and support staff to more than 11,000 U.S. users on over 3,000 projects in the open science community. But what made it special? Stampede was like a bridge that moved thousands of researchers off of soon-to-be decommissioned supercomputers, while at the same time building a framework that anticipated the eminent trends that came to dominate advanced computing.”