In this podcast, the Radio Free HPC team discusses the upcoming MSST Mass Storage Conference with Program Chair Matthew O’Keefe from Oracle. “Since the conference was founded by the leading national laboratories, MSST has been a venue for massive-scale storage system designers and implementers, storage architects, researchers, and vendors to share best practices and discuss building and securing the world’s largest storage systems for HPC, web-scale systems, and enterprises.”
The International Conference on Massive Storage Systems and Technology (MSST 2017) has issued its Call for Participation. The event takes place May 15-19, 2017 in Santa Clara, California. “MSST 2017 will dedicate five days to computer-storage technology, including a day of tutorials, two days of invited papers, two days of peer-reviewed research papers, and a vendor exposition.”
Gleb Budman from Backblaze presented this talk at the 2016 MSST Conference. “For Q1 2016 we are reporting on 61,590 operational hard drives used to store encrypted customer data in our data center. In Q1 2016, the hard drives in our data center, past and present, totaled over one billion hours in operation to date. That’s nearly 42 million days or 114,155 years worth of spinning hard drives. Let’s take a look at what these hard drives have been up to.”
In this video from the 2016 MSST Conference, Dave Anderson from Seagate presents: Whither Hard Disk Archives? The talk was part of a panel discussion on Data-intensive Workflows.
David Bonnie from LANL presented this talk at the 2016 MSST Conference. “As we continue to scale system memory footprint, it becomes more and more challenging to scale the long-term storage systems with it. Scaling tape access for bandwidth becomes increasingly challenging and expensive when single files are in the many terabytes to petabyte range. Object-based scale out systems can handle the bandwidth requirements we have, but are also not ideal to store very large files as objects. MarFS sidesteps this while still leveraging the large pool of object storage systems already in existence by striping large files across many objects.”
In this video from the 2016 MSST Conference, Yoonho Park from IBM presents: Storage Performance Modeling for Future Systems. “The burst buffer is an intermediate, high-speed layer of storage that is positioned between the application and the parallel file system (PFS), absorbing the bulk data produced by the application at a rate a hundred times higher than the PFS, while seamlessly draining the data to the PFS in the background.”
In this video from the 2016 MSST Conference, Harriet Coverston from Versity presents: Versity – Archiving to Objects. “Introducing Versity Storage Manager – an enterprise-class storage virtualization and archiving system that runs on Linux. Offering comprehensive data management for tiered storage environments and the ability to preserve and protect your data forever. Maximum protection at a minimum cost. Versity supports nearly unlimited volumes of storage and offers the most robust archive policy engine on the market.”