Today ACM announced that Ron Perrott, an international leader in the development and promotion of parallel computing, will receive ACM’s prestigious Distinguished Service Award. The goal of ACM’s Awards and Recognition Program is to highlight outstanding technical and professional achievements and contributions in computer science and IT. Perrott will be formally honored at the ACM Awards Banquet on June 11 in San Francisco.
The Gauss Centre for Supercomputing (GCS) in Germany has allocated a record 1,648 million core hours of computing time to 21 scientifically outstanding national research projects as part of its Call for Large-Scale Projects. “GCS is excited to support simulation projects of these excelling scopes as they clearly underline our claim of Germany being a world leader in High Performance Computing. Beyond dispute, they produce proof of us being at eye level with the largest international research projects such as the INCITE Program supported by the Office of Science of the U.S. Department of Energy,” states Prof. Thomas Lippert of JSC, GCS Chairman of the Board.”
In its latest move to build a practical quantum computer, IBM Research for the first time ever is making quantum computing available in the cloud to anyone interested in hands-on access to the company’s advanced experimental quantum system. “The cloud-enabled quantum computing platform, called IBM Quantum Experience, will allow users to run algorithms and experiments on IBM’s quantum processor, work with the individual quantum bits (qubits), and explore tutorials and simulations around what might be possible with quantum computing.”
In this video from the 2016 MSST Conference, Ian Corner from CSIRO in Australia presents: A Journey to a Holistic Framework for Data-intensive Workflows. “At CSIRO, we are Australia’s national science organization and one of the largest and most diverse scientific research organizations in the world. Our research focuses on the biggest challenges facing the nation. We also manage national research infrastructure and collections.”
Over at ALCF, Katie Jones writes that researchers are using the Mira supercomputer to validate a new “wave-like” model of the van der Waals force—a weak attraction that has strong ties to function and stability in materials and biological systems.
Jeff Bonwick from EMC DSSD presented this talk at the MSST conference. “Jeff Bonwick is co-founder and CTO of DSSD, where he co-invented both the system hardware architecture and the Flood software stack. His talk will focus on extracting maximum performance from flash at scale. Jeff has a long history of developing at-scale storage starting with leading the team that developed the ZFS filesystem, which powers Oracle’s ZFS storage line as well as numerous startups including Nexenta, Delphix, Joyent, and Datto.”
Altair is making a big investment toward uniting the whole HPC community to accelerate the state of the art (and the state of actual production operations) for HPC scheduling. Altair is joining the OpenHPC project with PBS Pro. They are focused on longevity – creating a viable, sustainable community to focus on job scheduling software that can truly bridge the gap in the HPC world.
The MJO occurs on its own timetable—every 30 to 60 days—but its worldwide impact spurs scientists to unlock its secrets. The ultimate answer? Timely preparation for the precipitation havoc it brings—and insight into how it will behave when pressured by a warming climate.
In this video, technicians install a new supercomputer at UK Met Office. The Met Office is the National Weather Service for the UK, providing internationally-renowned weather and climate science and services to support the public, government and businesses.
In this special guest feature from Scientific Computing World, Darren Watkins from Virtus Data Centres explains the importance of building a data centre from the ground up to support the requirements of HPC users – while maximizing productivity, efficiency and energy usage. “The reality for many IT users is they want to run analytics that –with the growth of data – have become too complex and time critical for normal enterprise servers to handle efficiently.”