Accelerated computing continues to gain momentum. This year the GPU Technology Conference will feature 90 sessions on HPC and Supercomputing. “Sessions will focus on how computational and data science are used to solve traditional HPC problems in healthcare, weather, astronomy, and other domains. GPU developers can also connect with innovators and researchers as they share their groundbreaking work using GPU computing.”
“2017 will see the introduction of many technologies that will help shape the future of HPC systems. Production-scale ARM supercomputers, advancements in memory and storage technology such as DDN’s Infinite Memory Engine (IME), and much wider adoption of accelerator technologies and from Nvidia, Intel and FPGA manufacturers such as Xilinx and Altera, are all helping to define the supercomputers of tomorrow.”
Tokyo-based Startup XTREME DESIGN recently announced it has raised $700K of funding in its pre-series A round. Launched in early 2015, the Startup’s XTREME DNA software automates the process of configuring, deploying, and monitoring virtual supercomputers on public clouds. To learn more, we caught up with the company’s founder, Naoki Shibata.
Today IBM announced it has created the world’s smallest magnet using a single atom – and stored one bit of data on it. Currently, hard disk drives use about 100,000 atoms to store a single bit. The ability to read and write one bit on one atom creates new possibilities for developing significantly smaller and denser storage devices, that could someday, for example, enable storing the entire iTunes library of 35 million songs on a device the size of a credit card.
Today the integrated Rule-Oriented Data System (iRODS) Consortium announced that Intel Corporation has joined their membership-based foundation. “As a consortium member, Intel plans to improve integration between iRODS, the free open source software for data virtualization, data discovery, workflow automation, and secure collaboration, and Lustre, an open source parallel distributed file system used for computing on large-scale high performance computing clusters. Membership in the consortium is a first step in offering an integrated tiered solution to Lustre end-users that allows them to easily move data sets from HPC systems into less costly long-term storage systems, where the data can be managed, shared and kept secure using iRODS.”
Researchers at SDSC have developed a new seismic software package with Intel Corporation that has enabled the fastest seismic simulation to-date. SDSC’s ground-breaking performance of 10.4 Petaflops on earthquake simulations used 612,000 Intel Xeon Phi processor cores of the new Cori Phase II supercomputer at NERSC.
The use of simulation software for validation and testing is a well established process for many engineering companies – but advances in software, combined with increased pressure from competition and regulatory requirements, are driving simulation earlier into the design and concept stages of new product development. “If you can test lots of scenarios virtually, then you know that you have accounted for any possibility of it failing.”
“IBM has invested over decades to growing the field of quantum computing and we are committed to expanding access to quantum systems and their powerful capabilities for the science and business communities,” said Arvind Krishna, senior vice president of Hybrid Cloud and director for IBM Research. “Following Watson and blockchain, we believe that quantum computing will provide the next powerful set of services delivered via the IBM Cloud platform, and promises to be the next major technology that has the potential to drive a new era of innovation across industries.”
In this video, Dr. Marcelo Ponce from SciNet presents: Scientific Visualization with Python. “SciNet is Canada’s largest supercomputer centre, providing Canadian researchers with computational resources and expertise necessary to perform their research on scales not previously possible in Canada. We help power work from the biomedical sciences and aerospace engineering to astrophysics and climate science.”
“This talk will describe Monotasks, a new architecture for the core of Spark that makes performance easier to reason about. In Spark today, pervasive parallelism and pipelining make it difficult to answer even simple performance questions like “what is the bottleneck for this workload?” As a result, it’s difficult for developers to know what to optimize, and it’s even more difficult for users to understand what hardware to use and what configuration parameters to set to get the best performance.”