A NEW LEVEL OF APPLICATION PERFORMANCE

FEATURES AND BENEFITS

Tesla P100 is reimagined from silicon to software, crafted with innovation at every level. Each groundbreaking technology delivers a dramatic jump in performance to inspire the creation of the world's fastest compute node.

Exponential Performance Leap with Pascal Architecture

The NVIDIA Pascal architecture enables the Tesla P100 to deliver superior performance for HPC and hyperscale workloads. With more than 21 teraFLOPS of 16-bit floating-point (FP16) performance, Pascal is optimised to drive exciting new possibilities in deep learning applications. Pascal also delivers over 5 and 10 teraFLOPS of double- and single-precision performance for HPC workloads.

Unprecedented Efficiency with CoWoS with HBM2

Tesla P100 tightly integrates compute and data on the same package by adding chip-on-wafer-on-substrate (CoWoS) with HBM2 technology to deliver 3X more memory performance over the NVIDIA Maxwell™ architecture. This provides a generational leap in time to solution for data-intensive applications.

Simpler Programming with Page Migration Engine

Page Migration Engine frees developers to focus more on tuning for computing performance and less on managing data movement. Applications can now scale beyond the GPU's physical memory size to virtually limitless amounts of memory.

TESLA P100 PRODUCTS

NVIDIA Tesla P100 for Strong-Scale HPC

Tesla P100 with NVIDIA NVLink technology enables lightning-fast nodes to substantially accelerate time to solution for strong-scale applications. A server node with NVLink can interconnect up to eight Tesla P100s at 5X the bandwidth of PCIe. It's designed to help solve the world's most important challenges that have infinite compute needs in HPC and deep learning.

NVIDIA Tesla P100 for Strong-Scale HPC

Tesla P100 with NVIDIA NVLink technology enables lightning-fast nodes to substantially accelerate time to solution for strong-scale applications. A server node with NVLink can interconnect up to eight Tesla P100s at 5X the bandwidth of PCIe. It's designed to help solve the world's most important challenges that have infinite compute needs in HPC and deep learning.

NVIDIA Tesla P100 for Mixed-Workload HPC

Tesla P100 for PCIe enables mixed-workload HPC data centers to realise a dramatic jump in throughput while saving money. For example, a single GPU-accelerated node powered by four Tesla P100s interconnected with PCIe replaces up to 32 commodity CPU nodes for a variety of applications. Completing all the jobs with far fewer powerful nodes means that customers can save up to 70 percent in overall data center costs.

NVIDIA Tesla P100 for Mixed-Workload HPC

Tesla P100 for PCIe enables mixed-workload HPC data centers to realise a dramatic jump in throughput while saving money. For example, a single GPU-accelerated node powered by four Tesla P100s interconnected with PCIe replaces up to 32 commodity CPU nodes for a variety of applications. Completing all the jobs with far fewer powerful nodes means that customers can save up to 70 percent in overall data center costs.