Today IBM along with Nvidia and two U.S. Department of Energy National Laboratories today announced a pair of Centers of Excellence for supercomputing – one at the Lawrence Livermore National Laboratory and the other at the Oak Ridge National Laboratory. The collaborations are in support of IBM’s supercomputing contract with the U.S. Department of Energy. They will enable advanced, large-scale scientific and engineering applications both for supporting DOE missions, and for the Summit and Sierra supercomputer systems to be delivered respectively to Oak Ridge and Lawrence Livermore in 2017 and to be operational in 2018.
In this podcast, the Radio Free HPC team looks at how the KatRisk startup is using GPUs on the Titan supercomputer to calculate global flood maps. “KatRisk develops event-based probabilistic models to quantify portfolio aggregate losses and exceeding probability curves. Their goal is to develop models that fully correlate all sources of flood loss including explicit consideration of tropical cyclone rainfall and storm surge.”
Today IBM Research announced that working with alliance partners at SUNY Polytechnic Institute’s Colleges of Nanoscale Science and Engineering it has produced the semiconductor industry’s first 7nm node test chips with functional transistors. According to IBM, the breakthrough underscores the company’s continued leadership and long-term commitment to semiconductor technology research.
Today JISC in the U.K. announced that Rolls-Royce is the first company to join its industrial supercomputing initiative. Designed to break down barriers between industry and academia, JISC will provide Rolls-Royce with easy access to supercomputing equipment at the Engineering and Physical Sciences Research Council (EPSRC) HPC Midlands.
Daniel Gutierrez, Managing Editor, of insideBIGDATA has put together a terrific Guide to Scientific Research. The goal of this paper is to provide a road map for scientific researchers wishing to capitalize on the rapid growth of big data technology for collecting, transforming, analyzing, and visualizing large scientific data sets.
“Supercomputing has reached a level of maturity and capability where many areas of science and engineering are not only advancing rapidly due to computing power, they cannot progress without it. I will illustrate examples from NCSA’s Blue Waters supercomputer, and from major data-intensive projects including the Large Synoptic Survey Telescope, and give thoughts on what will be needed going forward.”