Sign up for our newsletter and get the latest HPC news and analysis.
Send me information from insideHPC:


Slidecast: Announcing the Nvidia Tesla P100 for PCIe Servers

Marc Hamilton, Nvidia

Marc Hamilton, Nvidia

In this slidecast, Marc Hamilton from describes the Nvidia Tesla P100 for PCIe Servers.

“The Tesla P100 for PCIe is available in a standard PCIe form factor and is compatible with today’s GPU-accelerated servers. It is optimized to power the most computationally intensive AI and HPC data center applications. A single Tesla P100-powered server delivers higher performance than 50 CPU-only server nodes when running the AMBER molecular dynamics code, and is faster than 32 CPU-only nodes when running the VASP material science applications.”

Later this year, Tesla P100 accelerators for PCIe will power an upgraded version of Europe’s fastest supercomputer, the Piz Daint system at the Swiss National Supercomputing Center in Lugano, Switzerland.

Tesla P100 accelerators deliver new levels of performance and efficiency to address some of the most important computational challenges of our time,” said Thomas Schulthess, professor of computational physics at ETH Zurich and director of the Swiss National Supercomputing Center. “The upgrade of 4,500 GPU accelerated nodes on Piz Daint to Tesla P100 GPUs will more than double the system’s performance, enabling researchers to achieve breakthroughs in a range of fields, including cosmology, materials science, seismology and climatology.”

The Tesla P100 for PCIe is the latest addition to the NVIDIA Tesla Accelerated Computing Platform. Key features include:

  • Unmatched application performance for mixed-HPC workloads — Delivering 4.7 teraflops and 9.3 teraflops of double-precision and single-precision peak performance, respectively, a single Pascal-based Tesla P100 node provides the equivalent performance of more than 32 commodity CPU-only servers.
  • CoWoS with HBM2 for unprecedented efficiency — The Tesla P100 unifies processor and data into a single package to deliver unprecedented compute efficiency. An innovative approach to memory design — chip on wafer on substrate (CoWoS) with HBM2 — provides a 3x boost in memory bandwidth performance, or 720GB/sec, compared to the NVIDIA Maxwell architecture.
  • Page Migration Engine for simplified parallel programming — Frees developers to focus on tuning for higher performance and less on managing data movement, and allows applications to scale beyond the GPU physical memory size with support for virtual memory paging. Unified memory technology dramatically improves productivity by enabling developers to see a single memory space for the entire node.
  • Unmatched application support — With 410 GPU-accelerated applications, including nine of the top 10 HPC applications, the Tesla platform is the world’s leading HPC computing platform.
  • Tesla P100 for PCIe Specifications
  • 4.7 teraflops double-precision performance, 9.3 teraflops single-precision performance and 18.7 teraflops half-precision performance with NVIDIA GPU BOOST technology
  • Support for PCIe Gen 3 interconnect (32GB/sec bi-directional bandwidth)
  • Enhanced programmability with Page Migration Engine and unified memory
  • ECC protection for increased reliability
  • Server-optimized for highest data center throughput and reliability
  • Available in two configurations:
    • 16GB of CoWoS HBM2 stacked memory, delivering 720GB/sec of memory bandwidth
    • 12GB of CoWoS HBM2 stacked memory, delivering 540GB/sec of memory bandwidth

Availability

NVIDIA Tesla P100 GPU accelerator for PCIe-based systems is expected to be available beginning in Q4 2016 from NVIDIA reseller partners and server manufacturers, including Cray, Dell, Hewlett Packard Enterprise, IBM and SGI.

View the Slides * Download the MP3 *  Subscribe on iTunes Subscribe to RSS 

Resource Links: