Everything from life sciences to the financial industry are relying on HPC clusters to perform complex and critical operations. Moving forward, there will be a lot more reliance on various HPC systems. So the all-important question comes in – How do you select, deploy and manage it all? Fortunately, IBM, Intel and NCAR have teamed up to explain their view on best practices selecting an HPC cluster using the process behind building the NCAR Wyoming Supercomputing Center.
High performance technical computing continues to transform the capabilities of organizations across a range of industries—helping them to tackle unprecedented big data analysis, generate competitive business advantage, and expand the limits of science and medicine. To keep pushing those boundaries, organizations are continually seeking ways to get more out of their technical computing systems.
“How can capital markets firms handle the computational challenges presented by regulatory mandates and big data? Chances are the solution will involve high-performance computing powered by parallelism, or the ability to leverage multiple hardware resources to run code simultaneously. But while hardware architectures have been moving in that direction for years, many firms’ software isn’t written to take advantage of multiple threads of execution.”
“During the previous GTC, Murex has shown how the company had adapted their generic Monte-Carlo & PDE codes compatible with a payoff language. With one more year of experience with GPUs and OpenCL Murex will show how the company has broadened the usage of GPUs for other subjects like vanilla screening or model calibration and focus on their new challenge: use as many GPUs as possible for one single computation.”
“One of the hottest topics we see is remote visualization for post-processing simulation results. One big issue in traditional workflows in technical and scientific computing is the transfer of large amounts of data from where these have been created to where they are analyzed. Streamlining this workflow by processing where the data have been created in the first place is tantamount to shorten the wall-clock time it takes end users to get final results. At the same time, hardware utilization is greatly enhanced by using innovative technology for remote 3D visualization. For this, we have long since entered into a strategic partnership with NICE.”
“Terascala’s intelligent operating system, TeraOS, simplifies managing Lustre®-based storage and optimizes workflows, providing the high throughput storage HPC users need to solve bigger problems faster. For the HPC folks, this means that Terascala-powered storage appliances can reduce run times to hours instead of days or weeks.”