“This tutorial, part of the SC16 State of the Practice, will guide attendees through the process of purchasing and deploying a HPC system. It will cover the whole process from engaging with stakeholders in securing funding, requirements capture, market survey, specification of the tender/request for proposal documents, engaging with suppliers, evaluating proposals, and managing the installation. Attendees will learn how to specify what they want, yet enable the suppliers to provide innovative solutions beyond their specification both in technology and in the price; how to demonstrate to stakeholders that the solution selected is best value for money; and the common risks, pitfalls and mitigation strategies essential to achieve an on-time and on-quality installation process.”
Today Mellanox announced the availability of standard Linux kernel driver for the company Open Ethernet, Spectrum switch platforms. Developed within the large Linux community, the new driver enables standard Linux Operating Systems and off-the-shelf Linux-based applications to operate on the switch, including L2 and L3 switching. Open Ethernet provides data centers with the flexibility to choose the best hardware platform and the best software platform, resulting in optimized data center performance and higher return on investment.
From Megaflops to Gigaflops to Teraflops to Petaflops and soon to be Exaflops, the march in HPC is always on and moving ahead. This whitepaper details some of the technical challenges that will need to be addressed in the coming years in order to get to exascale computing.
Registration is now open for the Dell HPC Community event at SC16. The event takes place Nov. 12 at the Radisson Hotel in Salt Lake City. “The Dell HPC Community events feature keynote presentations by HPC experts and working group sessions to discuss best practices in the use of Dell HPC Systems.”
“We are still in the first minutes of the first day of the Intelligence revolution. In this keynote, Dr. Joseph Sirosh will present 5 solutions (and their implementations) that the intelligent cloud delivers. Sirosh shares five cloud AI patterns that his team and presented at the Summit. These five patterns are really about ways to bring data and learning together in cloud services, to infuse intelligence.”
“It’s often a challenge to test the scalability of system software components before a large deployment, particularly if you need low level hardware access”, said Dan Stanzione, Executive Director at TACC and a Co-PI on the Chameleon project. “Chameleon was designed for just these sort of cases – when your local test hardware is inadequate, and you are testing something that would be difficult to test in the commercial cloud – like replacing the available file system. Projects like Slash2 can use Chameleon to make tomorrow’s cloud systems better than today’s.”
Adrian Jackson from EPCC at the University of Edinburgh presented this tutorial to ARCHER users. “We have been working for a number of years on porting computational simulation applications to the KNC, with varying successes. We were keen to test this new processor with its promise of 3x serial performance compared to the KNC and 5x memory bandwidth over normal processors (using the high-bandwidth, MCDRAM, memory attached to the chip).”
Sure, your code seems fast, but how do you know if you are leaving potential performance on the table? Recognized HPC experts Georg Hager and Gerhard Wellein will teach a tutorial on Node-Level Performance Engineering at SC16. The session will take place 8:30-5:00pm on Sunday, Nov. 13 in Salt Lake City.
“Our high-performance computing solutions enable deep learning, engineering, and scientific fields to scale out their compute clusters to accelerate their most demanding workloads and achieve fastest time-to-results with maximum performance per watt, per square foot, and per dollar,” said Charles Liang, President and CEO of Supermicro. “With our latest innovations incorporating the new NVIDIA P100 processors in a performance and density optimized 1U and 4U architectures with NVLink, our customers can accelerate their applications and innovations to address the most complex real world problems.”
In this video from the 2016 Argonne Training Program on Extreme-Scale Computing, Mark Miller from LLNL leads a panel discussion on Experiences in eXtreme Scale in HPC with FASTMATH team members. “The FASTMath SciDAC Institute is developing and deploying scalable mathematical algorithms and software tools for reliable simulation of complex physical phenomena and collaborating with U.S. Department of Energy (DOE) domain scientists to ensure the usefulness and applicability of our work. The focus of our work is strongly driven by the requirements of DOE application scientists who work extensively with mesh-based, continuum-level models or particle-based techniques.”