Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Agenda Posted for Next Week’s HPC Advisory Council Stanford Conference

“Over two days we’ll delve into a wide range of interests and best practices – in applications, tools and techniques and share new insights on the trends, technologies and collaborative partnerships that foster this robust ecosystem. Designed to be highly interactive, the open forum will feature industry notables in keynotes, technical sessions, workshops and tutorials. These highly regarded subject matter experts (SME’s) will share their works and wisdom covering everything from established HPC disciplines to emerging usage models from old-school architectures and breakthrough applications to pioneering research and provocative results. Plus a healthy smattering of conversation and controversy on endeavors in Exascale, Big Data, Artificial Intelligence, Machine Learning and much much more!”

Asetek Achieves 200 Million Hours of Fault-Free Pump Operations at DataCenters

Today Asetek announced that its server pump has achieved 200 million hours of reliable operation in real world use. Installed at end-user locations as diverse as Singapore and Norway, Asetek pumps have run fault-free for the equivalent of 22,000 years. “To date, we have not had a single server pump failure at any of our data center installations around the world,” said Mette Nørmølle, Vice President of Engineering. “Our low-pressure architecture is the key to enabling a cost-effective solution that is relied on by data centers demanding unrivaled performance and maximum uptime.”

Exascale Computing Project Announces ECP Industry Council

“I’m pleased to have the opportunity to lead this important Council,” said Dr. J. Michael McQuade of United Technologies Corporation, who will serve as the first Chair of the ECP Industry Council. “Exascale level computing will help industry address ever more complex, competitively important problems, ones which are beyond the reach of today’s leading edge computing systems. We compete globally for scientific, technological and engineering innovations. Maintaining our lead at the highest level of computational capability is essential for our continued success.”

Video: Rescale Night Showcases HPC in the Cloud

“Billed as an exposition into ‘The Future of Cloud HPC Simulation,’ the event brought together experts in high-performance computing and simulation, cloud computing technologists, startup founders, and VC investors across the technology landscape. In addition to product demonstrations with Rescale engineers, including the popular Deep Learning workshop led by Mark Whitney, Rescale Director of Algorithms, booths featuring ANSYS, Microsoft Azure, Data Collective, and Microsoft Ventures offered interactive sessions for attendees.”

HPC4Mfg Industry Day Comes to San Diego March 2-3

HPC4Mfg will host their first annual High Performance Computing for Manufacturing Industry Engagement Day on March 2-3 in San Diego. With a theme of “Spurring Innovation in U.S. Manufacturing Through Advanced Computing,” the conference will bring together representatives from U.S. manufacturing, national laboratories, universities, and consortiums to discuss the recent advancements in manufacturing realized through the application of HPC and how leveraging HPC expertise through public-private partnerships has lowered the risk of adoption.

Argonne Annouces Early Science Projects for Aurora Supercomputer

Argonne has selected 10 computational science and engineering research projects for its Aurora Early Science Program starting this month. Aurora, a massively parallel, manycore Intel-Cray supercomputer, will be ALCF’s next leadership-class computing resource and is expected to arrive in 2018. The Early Science Program helps lay the path for hundreds of other users by doing actual science, using real scientific applications, to ready a future machine. “As with any bleeding edge resource, there’s testing and debugging that has to be done,” said ALCF Director of Science Katherine Riley.

In Memory Computing Speeds Results

In-Memory Computing can accelerate traditional applications by using a memory first design. Applicable to a wide range of domains, In-Memory Computing and In-Memory Data Grids take advantage of the latest trends in computer systems technology. “In-memory computing is designed to address some of the most critical and real-time task requirements today. This include real-time fraud detection, biometrics and border security and financial risk analytics. All of these use cases require very low latency access to data from very large amounts of data, which results in faster and more accurate decisions.”

HiPINEB Workshop on High Performance Networks Posts Agenda for February 5 in Austin

The IEEE International Workshop on High-Performance Interconnection Networks in the Exascale and Big-Data Era (HiPINEB) has published its program agenda. Held in conjunction with the IEEE HPCA 2017 Conference, HiPINEB takes place Feb 5, in Austin, Texas.

Video: Modern Computing – Cloud, Distributed, & High Performance

Dr. Umit Catalyurek from Georgia Institute of Technology presented this talk as part of the USC Big Data to Knowledge series. “This lecture will be a brief crash course on computer architecture, high performance computing and parallel computing. We will, again very briefly, discuss how to classify computer architectures and applications, and what to look in applications to achieve best performance on different architectures.”

HPC User Forum Returns to Stuttgart Feb 28 – March 1

The IDC HPC User Forum, taking place at the HLRS premises in Stuttgart-Vaihingen on February 28 and March 1, will bring together the HPC community to hear experts from all of Europe and the USA. During the two-day event, one of the subjects will be the role of small and medium-sized enterprises (SME) in HPC.