John Turner from ORNL presented this talk at the HPC User Forum. “Fully exploiting future exascale architectures will require a rethinking of the algorithms used in the large scale applications that advance many science areas vital to DOE and NNSA, such as global climate modeling, turbulent combustion in internal combustion engines, nuclear reactor modeling, additive manufacturing, subsurface flow, and national security applications.”
The National Security Agency in Maryland is seeking an HPC Software Engineer in our Job of the Week. “NSA’s High Performance Computing team develops and integrates advanced architectures and unique technologies to sustain its world-class HPC inventory. Applicants have the opportunity to research, design, develop, program, integrate, and test HPCs and all related components. NSA stays abreast of, and utilizes, new and emerging HPC technologies to address NSA’s unique and critical mission.”
Today Intel and the Jülich Supercomputing Centre together with ParTec and Dell today announced plans to develop and deploy a next-generation modular supercomputing system. Leveraging the experience and results gained in the EU-funded DEEP and DEEP-ER projects, in which three of the partners have been strongly engaged, the group will develop the necessary mechanisms required to augment JSC’s JURECA cluster with a highly-scalable component named “Booster” and being based on Intel’s Scalable Systems Framework (Intel SSF).
Rich Brueckner from insideHPC presented this talk at the Switzerland HPC Conferene. “While High Performance Computing has gone through dramatic changes since Seymour Cray created the supercomputer industry in the 1970’s, misnomers, myths, and Alternative Facts have established themselves in the hive mind of the HPX community. In this session, Rich will turn the industry on its ear and reveal the whole truth in the service of outright parody.”
Intel Advisor, an integral part of Intel Parallel Studio XE 2017, can help identify portions of code that could be good candidates for parallelization (both vectorization and threading). It can also help determine when it might not be appropriate to parallelize a section of code, depending on the platform, processor, and configuration it’s running on. Intel Advisor Roofline Analysis reveals the gap between an application’s performance and its expected performance.
“Inside a jet engine, air flowing faster than a hurricane is combined with fuel to generate heat that powers the plane. Designers are turning to supercomputers to model these complex processes to make new engines that are cleaner, quieter and cheaper. Our ﬁrst aircraft engine transformed the aviation industry. It was the 410-horsepower, aircooled Wasp, which delivered unprecedented performance and reliability for the time. We have been leading change ever since.”
Today Cray announced it has signed a solutions provider agreement with Mark III Systems to develop, market and sell solutions that leverage Cray’s portfolio of supercomputing and big data analytics systems. “We’re very excited to be partnering with Cray to deliver unique platforms and data-driven solutions to our joint clients, especially around the key opportunities of data analytics, artificial intelligence, cognitive compute, and deep learning,” said Chris Bogan, Mark III’s director of business development and alliances. “Combined with Mark III’s full stack approach of helping clients capitalize on the big data and digital transformation opportunities, we think that this partnership offers enterprises and organizations the ability to differentiate and win in the marketplace in the digital era.”
In this video from Switzerland HPC Conference, Michael Feldman from TOP500.org presents an annual deep dive into the trends, technologies and usage models that will be propelling the HPC community through 2017 and beyond. “Emerging areas of focus and opportunities to expand will be explored along with insightful observations needed to support measurably positive decision making within your operations.”
“Like OpenHPC on which it is based, Intel HPC Orchestrator is a middleware stack for HPC environments that sits on top of the base operating system and provides everything needed to run a variety of HPC applications from independent software vendors together on the same HPC system. The product includes tools for administration, provisioning, resource and workload management, HPC runtimes, I/O services, scientific libraries, MPI libraries, compilers and other development tools.”
In this video from the HPC User Forum, Bob Sorensen from Hyperion Research moderates a panel discussion on Disruptive Technologies for HPC. “A disruptive innovation is an innovation that creates a new market and value network and eventually disrupts an existing market and value network, displacing established market leading firms, products and alliances. The term was defined and phenomenon analyzed by Clayton M. Christensen beginning in 1995.”