Sign up for our newsletter and get the latest HPC news and analysis.


Direct N-Body Simulation

galaxy_s

In some domains, an N-Body simulation is key to solving for the movement and forces of a dynamic system of particles. At each time step, the force that one body exacts on each other, and then the velocity can be computed. The simulation can continue up to a desired number of time steps.

HPC Appliance Computing Goes Virtual

HPC Cloud

Altair’s HyperWorks Unlimited Virtual Appliance goes fully into the cloud with an Amazon-hosted option that lets users get started with HPC in just minutes.

Gather Scatter Operations

Gatter

Gather and scatter operations are used in many domains. However, to use these types of functions on an SIMD architecture creates some programming challenges.

How HPC is increasing speed and accuracy

Mark Gunn, Sr. VP, One Stop Systems

The overwhelming task of high performance computing today is the processing of huge amounts of data quickly and accurately. Just adding greater numbers of more intensive, sophisticated servers only partially solves the problem.

Fortran Still Going Strong

Fortran

Fortran still going strong. NERSC estimates that over half the hours on their systems are used by Fortran codes. This is quite amazing, given that Fortran first appeared about 60 years ago.

Numerical Optimization for Deep Learning

phi

“With the advent of massively parallel computing coprocessors, numerical optimization for deep-learning disciplines is now possible. Complex real-time pattern recognition, for example, that can be used for self driving cars and augmented reality can be developed and high performance achieved with the use of specialized, highly tuned libraries. By just using the Message Passing Interface (MPI) API, very high performance can be attained on hundreds to thousands of Intel Xeon Phi processors.”

Benefits of RackCDU D2C for High Performance Computing

DC2 Liquid Cooling

From bio-engineering and climate studies to big data and high frequency trading, HPC is playing an even greater role in today’s society. Without the power of HPC, the complex analysis and data driven decisions that are made as a result would be impossible. Because these super computers and HPC clusters are so powerful, they are expensive to cool, use massive amounts of energy, and can require a great deal of space.

Introduction to the Lustre File System

lustre-logo2

Although there are a number of truly huge implementations of Lustre today, the community is still far from reaching the maximum configurations that the Lustre architecture is designed for. Inside the Lustre File System describes the basics of how the Lustre File System operates with descriptions of the newest features.

The High Performance Data Analytics Market

HPDA

As data analytics becomes more mission critical, hardware and software need to evolve to handle both historical data (batch) and real time streaming data. This combined ability to manage different types of data is critical for a wide range of organizations.

Open Computing: Vendor Landscape

open-compute-project

The Open Compute Project, initiated by Facebook as a way to increase computing power while lowering associated costs with hyper-scale computing, has gained a significant industry following. Several vendors have released, or will soon release, servers based on Open Compute specifications.