The UberCloud has published their 3rd Compendium of HPC Cloud Case Studies. Like its predecessors in 2013 and 2014, this year’s edition draws from a select group of projects undertaken as part of the UberCloud Experiment. “Our efforts are paying off. Based on the experience gained over the past several years, we have now increased the success rate of the individual experiments to almost 100%, as compared to 40% in 2013 and 60% in 2014.”
In this video from ISC 2015, Wolfgang Gentzsch from The UberCloud presents: HPC Workloads Using Docker. “The UberCloud Project provides an environment to run HPC workloads on a variety of cloud providers. Wolfgang elaborates on the projects and provides his experience about the roadblocks towards a more flexible cloud-backed HPC future.”
“Univa is a workload optimization company. Our core product, Grid Engine software, creates a single virtual high throughput, high performance and hyper-scale compute pool out of distributed data center resources. Our customers efficiently run large quantities of mission-critical compute-intensive applications faster with lower overall costs.”
In this slidecast, Einar Rustad from Numascale describes how the company achieved a world-record on the McCalpin STREAM benchmark using their innovative scale-out to scale-up architecture. The benchmark measures sustainable memory bandwidth and the corresponding computation rate for simple vector kernels.
“100% Flash in the Datacenter? It won’t happen any time soon. Many (most?) tier one workloads will be moved to flash of course, but data is adding up so quickly that it’s highly unlikely you will be seeing a 100% datacenter any time soon. It will take a few years to have about 10/20% of data stored on flash and the rest will remain on huge hard disks (cheap 10+TB hard disks will soon be broadly available for example).”