MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Setting a Path for the Next-Generation of High-Performance Computing Architecture


At SC15, Intel talked about some transformational high-performance computing technologies and the architecture—Intel® Scalable System Framework (Intel® SSF). Intel describes Intel SSF as “an advanced architectural approach for simplifying the procurement, deployment, and management of HPC systems, while broadening the accessibility of HPC to more industries and workloads.” Intel SSF is designed to eliminate the traditional bottlenecks; the so called power, memory, storage, and I/O walls that system builders and operators have run into over the years.

Video: SGI Looks to Zero Copy Architecture for HPC and Big Data

Dr. Eng Lim Goh, CTO, SGI

In this video from SC15, Dr. Eng Lim Goh from SGI describes how the company is embracing new HPC technology trends such as new memory hierarchies. With the convergence of HPC and Big Data as a growing trend, SGI is envisions a “Zero Copy Architecture” that would bring together a traditional supercomputer with a Big Data analytics machine in a way that would not require users to move their data between systems.

Rise of Direct-to-Chip Liquid Cooling Highlighted at SC15

Fujitsu Cool-Central® Liquid Cooling

Of the varied approaches to liquid cooling, most remain technical curiosities and fail to show real-world adoption in any significant degree. In contrast, both Asetek RackCDU D2C™ (Direct-to-Chip) and Internal Loop Liquid Cooling are seeing accelerating adoption both by OEMs and end users.

Slidecast: Micron Persistent Memory & NVDIMM


“Micron is delivering on the promise of persistent memory with a solution that gives system architects a new approach for designing systems with better performance, reduced energy usage and improved total cost of ownership,” said Tom Eby, vice president for Micron’s compute and networking business unit. “With NVDIMM, we have a powerful solution that is available today. We’re also leading the way on future persistent memory development by spearheading R&D efforts on promising new technologies such as 3D XPoint memory, which will be available in 2016 and beyond.”

insideHPC Special Report on HPC Weather Forecasting and Climate Research


In the pantheon of HPC grand challenges, weather forecasting and long term climate simulation rank right up there with the most complex and computationally demanding problems in astrophysics, aeronautics, fusion power, exotic materials, and earthquake prediction, to name just a few. This special reports looks at how HPC takes on the challenge of global weather forecasting and climate research.

GPUs in Virtual Desktop Infrastructure Solutions


VDI or Virtual Desktop Infrastructure helps companies save money, time and resources. Instead of large bulky machines on every desk in the office, companies can connect multiple workstations to a single computer using thin clients. Instead of replacing individual desktops every year, companies only have to replace thin clients every 5 years. And when it comes time to do updates, the IT staff updates the one computer instead of spending time updating every individual workstation.

It’s Here: The Print ‘n Fly Guide to SC15 in Austin


At insideHPC, are very pleased to publish the Print ‘n Fly Guide to SC15 in Austin. We designed this Guide to be an in-flight magazine custom tailored for your journey to SC15 — the world’s largest gathering of high performance computing professionals.

Lustre* Accelerates the Convergence of Big Data and HPC in Financial Services

HPC BIGDATA Convergence

Across industries, companies are beginning to watch the convergence of High-performance Computing (HPC) and Big Data. Many organizations in the Financial Services Industry (FSI) are running their financial simulations on business analytics systems, some on HPC clusters. But they have a growing problem: integrating analytics of non-structured data from sources like social media with their internal data. Learn how Lustre can help solves these challenges.

High Performance Computing GPU Accelerators in Deep Learning

Katie Garrison, Marketing Commuincations, One Stop Solutions

Training the neural networks used in deep learning is an ideal task for GPUs because GPUs can perform many calculations at once (parallel calculations), meaning the training will take much less time than it used to take. More GPUs means more computational power so if a system has multiple GPUs, it can compute data much faster than a system with CPUs only, or a system with a CPU and a single GPU. One Stop System’s High Density Compute Accelerator is the densest GPU expansion system to date.

Pete Beckman Presents: Exascale Architecture Trends


“Argonne National Laboratory is one of the labs helping to lead the exascale push for the nation with the DOE. We lead in a numbers of areas with software and storage systems and applied math. And we’re really focusing, our expertise is focusing on those new ideas, those novel new things that will allow us to sort of leapfrog the standard slow evolution of technology and get something further out ahead, three years, five years out ahead. And that’s where our research is focused.”