“This webinar replay discusses the use of high performance computing (HPC) in the design of aircraft jet engines and gas turbines used to generate electrical power. HPC is the critical enabler in this process, but applying HPC effectively in an industrial design setting requires an integrated hardware/software solution and a clear understanding of how the value outweighs the costs. This webinar will share GE’s perspective on the successful deployment and utilization of HPC, offer examples of HPC’s impact on GE products, and discuss future trends.”
“Although the use of GPUs has generalized nowadays, including GPUs in current HPC clusters presents several drawbacks mainly related with increased costs. In this talk we present how the use of remote GPU virtualization may overcome these drawbacks while noticeably increasing the overall cluster throughput. The talk presents real throughput measurements by making use of the rCUDA remote GPU virtualization middleware.”
Joseph Lombardo from UNLV presented this talk at the PBS Works User Group. “Lombardo will highlight results from an Alzheimer’s research project that benefited from using PBS Professional. He will then describe the NSCEE’s new system at the Supernap and how this system can be used to advance research for HPC users in both academia/R&D and commercial industry. Lombardo will also highlight two emerging projects; the New School of Medicine and new Technology park.”
In this video from the Disruptive Technologies Session at the 2015 HPC User Forum, Intel’s Ralph Biesemeyer presents: Intel 3D XPoint Technology.
“For decades, the industry has searched for ways to reduce the lag time between the processor and data to allow much faster analysis,” said Rob Crooke, senior vice president and general manager of Intel’s Non-Volatile Memory Solutions Group. “This new class of non-volatile memory achieves this goal and brings game-changing performance to memory and storage solutions.”
Hussein Harake from CSCS presented this talk at the HPC Advisory Council Spain Conference. “IME unleashes a new I/O provisioning paradigm. This breakthrough, software defined storage application introduces a whole new new tier of transparent, extendable, non-volatile memory (NVM), that provides game-changing latency reduction and greater bandwidth and IOPS performance for the next generation of performance hungry scientific, analytic and big data applications – all while offering significantly greater economic and operational efficiency than today’s traditional disk-based and all flash array storage approaches that are currently used to scale performance.”
“For High Performance Computing users who leverage open-source Lustre software, a good file system for big data is now getting even better. Building on its substantial contributions to the Lustre community, Intel is rolling out new features that will make the file system more scalable, easier to use, and more accessible to enterprise customers.”