The recent release of a commercial version of the Lustre* parallel file system was big news for business data centers facing ever expanding data analysis and storage demands. Now, Lustre, the predominant high-performing file system installed in most of the supercomputer installations around the world, could be deployed to business customers in a hardened, tested, easy to manage and fully supported distribution.
The software defined data center is the underlying data center architecture that allows most IT infrastructure to be defined in software and to function as enterprise-wide resources. This approach enables ITaaS to be delivered in a virtualized environment with greater agility, speed and quality of service.
Recent announcements, analyst reports, conferences and anecdotal evidence point to a certain upswing for high performance computing in industry. Many industries have reaped the benefit of HPC for considerable time and are now stepping up a gear with their systems – some even on a par with national facilities, in order to maintain or extend their advantage. Whether in upstream exploration, engine design or aerodynamics – if you can scale up or scale out, you can derive advantage.
Over the past several years, virtualization has made major inroads into enterprise IT infrastructures. And now it is moving into the realm of high performance computing (HPC), especially for such compute intensive applications as electronic design automation (EDA), life sciences, financial services and digital media entertainment. This article is the first in a series that explores the benefits the HPC community can achieve by adopting proven virtualization and cloud technologies.
Everything from life sciences to the financial industry are relying on HPC clusters to perform complex and critical operations. Moving forward, there will be a lot more reliance on various HPC systems. So the all-important question comes in – How do you select, deploy and manage it all? Fortunately, IBM, Intel and NCAR have teamed up to explain their view on best practices selecting an HPC cluster using the process behind building the NCAR Wyoming Supercomputing Center.
High performance technical computing continues to transform the capabilities of organizations across a range of industries—helping them to tackle unprecedented big data analysis, generate competitive business advantage, and expand the limits of science and medicine. To keep pushing those boundaries, organizations are continually seeking ways to get more out of their technical computing systems.