MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:

Call for Benchmark Proposals: SC16 Student Cluster Competition


SC16 has issued a Call for Proposals for a new initiative that aims to integrate aspects of past technical papers into the Student Cluster Competition.

Video: Prologue O/S – Improving the Odds of Job Success


“When looking to buy a used car, you kick the tires, make sure the radio works, check underneath for leaks, etc. You should be just as careful when deciding which nodes to use to run job scripts. At the NASA Advanced Supercomputing Facility (NAS), our prologue and epilogue have grown almost into an extension of the O/S to make sure resources that are nominally capable of running jobs are, in fact, able to run the jobs. This presentation describes the issues and solutions used by the NAS for this purpose.”

Case Study: PBS Pro on a Large Scale Scientific GPU Cluster


Professor Taisuke Boku from the University of Tsukuba presented this talk at the PBS User Group. “We have been operating a large scale GPU cluster HA-PACS with 332 computation nodes equipped with 1,328 GPUs managed by PBS Professional scheduler. The users are spread out across a wide variety of computational science fields with widely distributed resource sizes from single node to full-scale parallel processing. There are also several categories of user groups with paid and free scientific projects. It is a challenging operation of such a large system keeping high system utilization rate as well as keeping fairness over these user groups. We have successfully been keeping over 85%-90% of job utilization under multiple constraints.”

Evolution of NASA Earth Science Data Systems in the Era of Big Data


Christopher Lynnes from NASA presented this talk at the HPC User Forum. “The Earth Observing System Data and Information System is a key core capability in NASA’s Earth Science Data Systems Program. It provides end-to-end capabilities for managing NASA’s Earth science data from various sources—satellites, aircraft, field measurements, and various other programs.”

Scientific Cloud Computing Lags Behind the Enterprise


“In business and commercial computing, momentum towards cloud and big data has already built up to the point where it is unstoppable. In technical computing, the growth of the Internet of Things is pressing towards convergence of technologies, but obstacles remain, in that HPC and big data have evolved different hardware and software systems while Open Stack, the Open Source cloud computing platform, does not work well with HPC.”

Video: DDN Infinite Memory Engine IME


Tommaso Cecchi from DDN presented this talk at the HPCAC Spain Conference. “IME unleashes a new I/O provisioning paradigm. This breakthrough, software defined storage application introduces a whole new new tier of transparent, extendable, non-volatile memory (NVM), that provides game-changing latency reduction and greater bandwidth and IOPS performance for the next generation of performance hungry scientific, analytic and big data applications – all while offering significantly greater economic and operational efficiency than today’s traditional disk-based and all flash array storage approaches that are currently used to scale performance.”

ISC 2016 Issues Call for BoFs


ISC 2016 has issued its Call for BoFs. “Like-minded ISC High Performance conference attendees come together in our informal Birds-of-a-Feather (BoF) sessions to discuss current HPC topics, network and share their thoughts and ideas. Each 60-minute BoF session addresses a different topic and is led by one or more individuals with expertise in the area. The ISC 2016 BoF sessions will be held from Monday, June 20 through Wednesday, June 22.”

Pete Beckman Presents: Exascale Architecture Trends


“Argonne National Laboratory is one of the labs helping to lead the exascale push for the nation with the DOE. We lead in a numbers of areas with software and storage systems and applied math. And we’re really focusing, our expertise is focusing on those new ideas, those novel new things that will allow us to sort of leapfrog the standard slow evolution of technology and get something further out ahead, three years, five years out ahead. And that’s where our research is focused.”

Maximizing HPC Compute Resources with Minimal Cost


“As HPC resource requirements continue to increase, the need for finding economical solutions to handle the rising requirements increases as well. There are numerous ways to approach this challenge. For example, leveraging existing equipment, adding new or used equipment, and handling uncommon peak usage dynamically through cloud solutions managed by a central job management system can prove to be highly available and resource rich, while remaining economical. In this presentation we will discuss how Wayne State University implemented a combination of these approaches to dramatically increase our compute resources for the equivalent cost of only a few new servers.”

Lustre Video: Robinhood v3 Policy Engine and Beyond


“The Robinhood Policy Engine is a versatile tool to manage contents of large file systems. It maintains a replicate of filesystem medatada in a database that can be queried at will. It makes it possible to schedule mass action on filesystem entries by defining attribute-based policies.”