The move to network offloading is the first step in co-designed systems. A large amount of overhead is required to service the huge number of packets required for modern data rates. This amount of overhead can significantly reduce network performance. Offloading network processing to the network interface card helped solve this bottleneck as well as some others.
“When the history of HPC is viewed in terms of technological approaches, three epochs emerge. The most recent epoch, that of co-design systems, is new and somewhat unfamiliar to many HPC practitioners. Each epoch is defined by a fundamental shift in design, new technologies, and the economics of the day. “A network co-design model allows data algorithms to be executed more efficiently using smart interface cards and switches. As co-design approaches become more mainstream, design resources will begin to focus on specific issues and move away from optimizing general performance.”
A single issue has always defined the history of HPC systems: performance. While offloading and co-design may seem like new approaches to computing, they actually have been used, to a lesser degree, in the past as a way to enhance performance. Current co-design methods are now going deeper into cluster components than was previously possible. These new capabilities extend from the local cluster nodes into the “computing network.”
This very interesting whitepaper explains how selecting a proper parallel file system for your application can increase the performance of complex simulations and reduce time to completion.
Cloud computing is growing and replacing many data centers for High Performance Computing (HPC) applications. However, the movement towards using a cloud infrastructure is not without challenges. This whitepaper discusses many of the challenges in moving from an on-premise HPC solution to using an HPC Cloud Solution.
Organizations that implement high-performance computing (HPC) technologies have a wide range of requirements. From small manufacturing suppliers to national research institutions, using significant computing technologies is critical to creating innovative products and leading-edge research. No two HPC installations are the same. For maximum return, budget, software requirements, performance and customization all must be considered before installing and operating a successful environment.
Properly managed HPC supercomputers can deliver sustained Return on Investment for production supercomputing. To ensure ongoing operations are at peak efficiency, system management capabilities should not only include system monitoring and configuration tools, but also workload management, automated power capping, and error and fault detection capabilities.
Organizations that implement high-performance computing (HPC) technologies have a wide range of requirements. From small manufacturing suppliers to national research institutions, using significant computing technologies is critical to creating innovative products and leading-edge research. No two HPC installations are the same. “For maximum return, budget, software requirements, performance and customization all must be considered before installing and operating a successful environment.”
This whitepaper is an excellent summary of how a next generation platform can be developed to bring a wide range of data to life, giving users the ability to take action when needed. Organizations that need to deal with massive amounts of data but are having challenges figuring out how to make sense of all of the data should read this whitepaper.
Today’s HPC supercomputers have significant power requirements that must be considered as part of their Total Cost of Ownership. In addition, efficient power management capabilities are critical to sustained return on investment.