MailChimp Developer

Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Video: Trends in HPC Cloud Computing

Tejas Karmarker from Microsoft Azure presented this talk at the HPC User Forum. “The performance and scalability of a world-class supercomputing center is now available to everyone, on demand in the cloud. Run your Windows and Linux HPC applications using high performance A8 and A9 compute instances on Azure, and take advantage of a backend network with MPI latency under 3 microseconds and non-blocking 32 Gbps throughput. This backend network includes remote direct memory access (RDMA) technology on Windows and Linux that enables parallel applications to scale to thousands of cores. Azure provides you with high memory and HPC-class CPUs to help you get results fast. Scale up and down based upon what you need and pay only for what you use to reduce costs.”

Innovations in Non-volatile Memory – 3D NAND and its Implications

Rob Peglar from Micron presented this talk at the 2016 MSST Conference. “The growing demands of mobile computing and data centers continue to drive the need for high-capacity, high-performance NAND flash technology. With planar NAND nearing its practical scaling limits, delivering to those requirements has become more difficult with each generation. Enter our 3D NAND technology, which uses an innovative process architecture to provide 3X the capacity of planar NAND technologies while providing better performance and reliability. System designers who build products like laptops, mobile devices and servers can take advantage of 3D NAND’s unprecedented performance to meet the rising data movement needs for businesses and consumers.”

Video: Quantum Computing and D-Wave

Bo Ewald from D-Wave Systems presented this talk at the HPC User Forum in Tucson. “D-Wave continues to advance the state-of-the-art of quantum computing at a rapid pace, with a number of impressive application results, and the release of their 1000 qubit D-Wave 2X system is another major milestone in the industry,” said Earl Joseph, IDC program vice president for HPC. “Complementing today’s high performance computing systems, quantum computers will likely become an important tool to solve important problems that can’t be solved today.”

Cowboy Supercomputer Powers Research at Oklahoma State

In this video, Oklahoma State Director of HPC Dana Brunson describes how the Cowboy supercomputer powers research. “High performance computing is often used for simulations that may be too big, too small, too fast, too slow, too dangerous or too costly, another thing it’s used for involves data. So you may remember the human genome project it took nearly a decade and cost a billion dollars, these sorts of things can now be done over the weekend for under a thousand dollars. Our current super computer is named Cowboy and it was funded by a 2011 National Science Foundation Grant and it has been serving us very well.”

Video: Enabling High-Performance Storage on OpenPOWER Systems

In this video from the 2016 OpenPOWER Summit, Stephen Bates of Microsemi presents: Enabling high-performance storage on OpenPOWER Systems. “Non-Volatile Memory (NVM), and the low latency access to storage it provides, is changing the compute stack. NVM Express is the de-facto protocol for communicating with local NVM attached over the PCIe interface. In this talk we will demonstrate performance data for extermely low-latency NVMe devices operating inside OpenPOWER systems. We will discuss the implications of this for applications like in-memory databases, analytics and cognitive computing. In addition we will present data on the emerging NVMe over Fabrics (NVMf) protocol running on OpenPOWER systems.”

MarsFS – A Near-POSIX Namespace Leveraging Scalable Object Storage

David Bonnie from LANL presented this talk at the 2016 MSST Conference. “As we continue to scale system memory footprint, it becomes more and more challenging to scale the long-term storage systems with it. Scaling tape access for bandwidth becomes increasingly challenging and expensive when single files are in the many terabytes to petabyte range. Object-based scale out systems can handle the bandwidth requirements we have, but are also not ideal to store very large files as objects. MarFS sidesteps this while still leveraging the large pool of object storage systems already in existence by striping large files across many objects.”

Storage Performance Modeling for Future Systems

In this video from the 2016 MSST Conference, Yoonho Park from IBM presents: Storage Performance Modeling for Future Systems. “The burst buffer is an intermediate, high-speed layer of storage that is positioned between the application and the parallel file system (PFS), absorbing the bulk data produced by the application at a rate a hundred times higher than the PFS, while seamlessly draining the data to the PFS in the background.”

Video: Versity HSM – Archiving to Objects

In this video from the 2016 MSST Conference, Harriet Coverston from Versity presents: Versity – Archiving to Objects. “Introducing Versity Storage Manager – an enterprise-class storage virtualization and archiving system that runs on Linux. Offering comprehensive data management for tiered storage environments and the ability to preserve and protect your data forever. Maximum protection at a minimum cost. Versity supports nearly unlimited volumes of storage and offers the most robust archive policy engine on the market.”

Video: Accelerating Code at the GPU Hackathon in Delaware

In this video from the GPU Hackathon at the University of Delaware, attendees tune their code to accelerate their application performance. The 5-day intensive GPU programming Hackathon was held in collaboration with Oak Ridge National Lab (ORNL). “Thanks to a partnership with NASA Langley Research Center, Oak Ridge National Laboratory, National Cancer Institute, National Institutes of Health (NIH), Brookhaven National Laboratory and the UD College of Engineering, UD students had access to the world’s second largest supercomputer — the Titan — to help solve real-world problems.”

Video: Brendan Gregg Looks at Tools & Methodologies for Linux System Performance

In this video from the 2016 Percona Data Performance Conference, Brendan Gregg, Senior Performance Architect from Netflix presents: Linux Systems Performance. “Systems performance provides a different perspective for analysis and tuning, and can help you find performance wins for your databases, applications, and the kernel. However, most of us are not performance or kernel engineers, […]