Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Video: Evolution of MATLAB

Cleve Moler from MathWorks gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. “MATLAB is a high-performance language for technical computing. It integrates computation, visualization, and programming in an easy-to-use environment where problems and solutions are expressed in familiar mathematical notation. Typical uses include: Data analysis, exploration, and visualization.”

HPC in Agriculture: NCSA and Syngenta’s Dynamic Partnership

In this video, Jim Mellon from Sygenta describes how the company’s partnership with NCSA is helping the company answer the agricultural challenges of the future. “Together, we’re solving some of the toughest issues in agriculture today, like how to feed our rapidly growing population knowing that the amount of land we have for growing crops is finite. NCSA Industry provides the HPC resources that Syngenta’s scientists need to solve these issues, as well as an industry focus on security, performance, and availability, with the consultancy to better understand how to maximize these resources.”

Video: Silicon Photonics for Extreme Computing

Keren Bergman from Columbia University gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. “Exaflop machines would represent a thousand-fold improvement over the current standard, the petaflop machines that first came on line in 2008. But while exaflop computers already appear on funders’ technology roadmaps, making the exaflop leap on the short timescales of those roadmaps constitutes a formidable challenge.”

Video: How R-Systems Helps Customers Move HPC to the Cloud

In this video from the HPC User Forum in Milwaukee, Brian Kucic from R-Systems describes how the company enables companies of all sizes to move their technical computing workloads to the Cloud. “R Systems provides High Performance Computer Cluster resources and technical expertise to commercial and institutional research clients through the R Systems brand and the Dell HPC Cloud Services Partnership. In addition to our industry standard solutions, R Systems Engineers assist clients in selecting the components of their optimal cluster configuration.”

HPC Connects: Mapping Global Ocean Currents

In this video from the SC17 HPC Connects series, Dimitris Menemenlis from NASA JPL/Caltech describes how supercomputing enables scientists to accurately map global ocean currents. The ocean is vast and there are still a lot of unknowns. We still can’t represent all the conditions and are pushing the boundaries of current supercomputer power,” said Menemenlis. “This is an exciting time to be an oceanographer who can use satellite observations and numerical simulations to push our understanding of ocean circulation forward.”

HPC I/O for Computational Scientists

Phil Carns from Argonne gave this talk at the 2017 Argonne Training Program on Extreme-Scale Computing. “Darshan is a scalable HPC I/O characterization tool. It captures an accurate but concise picture of application I/O behavior with minimum overhead. Darshan was originally developed on the IBM Blue Gene series of computers deployed at the Argonne Leadership Computing Facility, but it is portable across a wide variety of platforms include the Cray XE6, Cray XC30, and Linux clusters.  Darshan routinely instruments jobs using up to 786,432 compute cores on the Mira system at ALCF.”

Video: The AI Initiative at NIST

Michael Garris from NIST gave this talk at the HPC User Forum. “AI must be developed in a trustworthy manner to ensure reliability and safety. NIST cultivates trust in AI technology by developing and deploying standards, tests and metrics that make technology more secure, usable, interoperable and reliable, and by strengthening measurement science. This work is critically relevant to building the public trust of rapidly evolving AI technologies.”

Video: Supercomputing Models Enable Detection of a Cosmic Cataclysm

In this podcast, Peter Nugent from Berkeley Lab explains how scientists confirmed the first-ever measurement of the merger of two neutron stars and its explosive aftermath. “Simulations succeeded in modeling what would happen in an incredibly complex phenomenon like a neutron star merger. Without the models, we all probably all would have been mystified by exactly what we were seeing in the sky.”

Video: NASA Advanced Computing Environment for Science & Engineering

Rupak Biswas from NASA gave this talk at the Argonne Training Program on Extreme-Scale Computing. “High performance computing is now integral to NASA’s portfolio of missions to pioneer the future of space exploration, accelerate scientific discovery, and enable aeronautics research. Anchored by the Pleiades supercomputer at NASA Ames Research Center, the High End Computing Capability (HECC) Project provides a fully integrated environment to satisfy NASA’s diverse modeling, simulation, and analysis needs.”

John Gustafson to host BoF on Posit Arithmetic at SC17

John Gustafson from A*STAR will host a BoF on Posit arithmetic at SC17. Entitled, “Improving Numerical Computation with Practical Tools and Novel Computer Arithmetic,”  this BOF will be co-hosted by Mike Lam with discussions on tools for measuring floating point accuracy. “This approach obtains more accurate answers than floating-point arithmetic yet uses fewer bits in many cases, saving memory, bandwidth, energy, and power.”