“Our annual ‘Budget Map’ report series looks at the relative spending between all of the products, components, and services that make up the HPC market. With six years of end user data, we get a strong grip on where the money is flowing, whether it’s on big items like clusters and storage, or on topical things like power consumption, programming, or compute cycles in public cloud. We also get a sense of future budget outlook and how the market is likely to evolve.”
“We have developed an active programme of industry engagement, typically consulting in niche areas and helping those partners exploit advanced computing within their organisations. This may be improving fraud detection algorithms, large-scale data management, tailored training courses or oil & gas related work. In fact it is very rarely about computer time! We also work with public sector bodies to support computing centric projects by the state and provide so-called platform technology support for research communities that need more than just raw cycles.”
“LLNL’s largest supercomputer is paired with a 55-petabyte file system, known as Grove, that stores vast quantities of simulation data. Grove must transfer information to and from Sequoia at a minimum of 500 gigabytes per second. To support Grove’s high storage capacity and bandwidth requirements, LC software developers have engaged in a multi-year project to replace much of Grove’s Lustre foundation with ZFS. “
In this episode of This Week in HPC, Michael Feldman and Addison Snell from Intersect360 Research discuss high frequency trading and the High Performance on Wall Street Conference. After that, they look at a Fujitsu’s Petascale upgrade at the Japan Aerospace Exploration Agency. The novel architecture for this Sparc-powered FX10 follow-on system will feature Micron’s Hybrid Memory Cube technology.
“The current time-horizon of HPC is not much more than five to ten years, with Exascale as the goal. But can we ensure that supercomputing embeds itself so deeply into the productive economy as to be indispensable a century from now? And will it be regarded as a thing of beauty, rather than just a technical artifact? And is Exascale the only destination?”
In this video from LUG 2014, Roger Ronald from System Fabric Works presents: Integrating Array Management into Lustre. “Intel Enterprise Edition for Lustre Plug-ins address a significant adoption barrier by improving ease of use. Now, System Fabric Works has implemented a NetApp plug-in for Intel EE Lustre and additional plug-ins for storage, networks, and servers are being encouraged.”