Powering New Levels of User Engagement

AI is constantly challenged to keep up with exploding volumes of data and still deliver fast responses. Meet the challenges with NVIDIA® Tesla®, the world’s fastest, most efficient data center platform for inference. Tesla supports all deep learning workloads and provides the optimal inference solution—combining the highest throughput, best efficiency, and best flexibility to power AI-driven experiences.

TESLA P40For Inference-Throughput Servers

FEATURES AND BENEFITS

50X Higher Throughput to Keep Up with Expanding Workloads

Volta-powered Tesla V100 GPUs give data centers a dramatic boost in throughput for deep learning workloads to extract intelligence from today’s tsunami of data. A server with a single Tesla V100 can replace up to 50 CPU-only servers for deep learning inference workloads, so you get dramatically higher throughput with lower acquisition cost.

Unprecedented Efficiency for Low-Power, Scale-Out Servers

The ultra-efficient Tesla P4 GPU accelerates density-optimized, scale-out servers with a small form factor and a 50/75 W power footprint design. It delivers an incredible 52X better energy efficiency than CPUs for deep learning inference workloads, so hyperscale customers can scale within their existing infrastructure and service the exponential growth in demand for AI-based applications.

A Dedicated Decode Engine for New AI-Based Video Services

The Tesla P4 GPU can analyze up to 39 HD video streams in real time. Powered by a dedicated hardware-accelerated decode engine, it works in parallel with the NVIDIA CUDA® cores performing inference. By integrating deep learning into the pipeline, customers can offer new levels of smart, innovative functionality that facilitates video search and other video-related services.

Faster Deployment with NVIDIA TensorRT and DeepStream SDK

NVIDIA TensorRT™ is a high-performance, neural-network inference engine for production deployment of deep learning applications. With TensorRT, neural nets trained in 32-bit or 16-bit data can be optimized for reduced-precision INT8 operations on Tesla P4 or FP16 on Tesla V100. NVIDIA DeepStream SDK taps into the power of Tesla GPUs to simultaneously decode and analyze video streams.