Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Video: Revolution in Computer and Data-enabled Science and Engineering

Ed Seidel from the University of Illinois gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. The theme of his talk centers around the need for interdisciplinary research. “Interdisciplinary research (IDR) is a mode of research by teams or individuals that integrates information, data, techniques, tools, perspectives, concepts, and/or theories from two or more disciplines or bodies of specialized knowledge to advance fundamental understanding or to solve problems whose solutions are beyond the scope of a single discipline or area of research practice.”

Video: Argonne’s Theta Supercomputer Architecture

Scott Parker gave this talk at the Argonne Training Program on Extreme-Scale Computing. “Designed in collaboration with Intel and Cray, Theta is a 9.65-petaflops system based on the second-generation Intel Xeon Phi processor and Cray’s high-performance computing software stack. Capable of nearly 10 quadrillion calculations per second, Theta will enable researchers to break new ground in scientific investigations that range from modeling the inner workings of the brain to developing new materials for renewable energy applications.”

A Vision for Exascale: Simulation, Data and Learning

Rick Stevens gave this talk at the recent ATPESC training program. “The ATPESC program provides intensive, two weeks of training on the key skills, approaches, and tools to design, implement, and execute computational science and engineering applications on current high-end computing systems and the leadership-class computing systems of the future. As a bridge to that future, this two-week program fills the gap that exists in the training computational scientists typically receive through formal education or other shorter courses.”

OpenHPC: Project Overview and Updates

Karl Schulz from Intel gave this talk at the MVAPICH User Group. “There is a growing sense within the HPC community for the need to have an open community effort to more efficiently build, test, and deliver integrated HPC software components and tools. To address this need, OpenHPC launched as a Linux Foundation collaborative project in 2016 with combined participation from academia, national labs, and industry. The project’s mission is to provide a reference collection of open-source HPC software components and best practices in order to lower barriers to deployment and advance the use of modern HPC methods and tools.”

Infinite Memory Engine: HPC in the FLASH Era

In this RichReport slidecast, James Coomer from DDN presents an overview of the Infinite Memory Engine IME. “IME is a scale-out, flash-native, software-defined, storage cache that streamlines the data path for application IO. IME interfaces directly to applications and secures IO via a data path that eliminates file system bottlenecks. With IME, architects can realize true flash-cache economics with a storage architecture that separates capacity from performance.”

Video: Omni-Path Status, Upstreaming, and Ongoing Work

Todd Rimmer from Intel presented this talk at the OpenFabrics Workshop. “Intel Omni-Path was first released in early 2016. Omni-Path host and management software is all open sourced. This session will provide an overview of Omni-Path including some of the technical capabilities and performance results as well as some recent industry results.”

Video: An Overview of Supercomputing

“Computers are an invaluable tool for most scientific fields. It is used to process measurement data and make simulation models of e.g. the climate or the universe. Brian Vinter talks about what makes a computer a supercomputer, and why it is so hard to build and program supercomputers.”