Magnum IO

Meeting the Bandwidth Demands of Compute-Intensive Workloads

GPUs provide the horsepower required by compute-intensive workloads, but their data consumption increases the demand for input/output (IO) bandwidth. NVIDIA Magnum IO? is NVIDIA’s set of APIs that integrates computing, networking, file systems, and storage to maximize IO performance for multi-GPU, multi-node accelerated systems. It interfaces with CUDA-X? libraries to accelerate IO across a broad range of workloads, from AI to visualization.

System Balance and Utilization: It relieves CPU contention to create a more balanced GPU-accelerated system and delivers peak IO bandwidth with up to 10X fewer CPU cores.

Seamless Integration: It provides optimized implementation for current and future platforms, whether the data transfers are latency sensitive, bandwidth sensitive, or collectives.

IO Optimization Stack

Magnum IO includes innovative IO optimization technologies such as NCCL, NVIDIA? GPUDirect RDMA, and NVIDIA Fabric Manager. GPUDirect Storage is a key feature of the stack. It opens a direct data path between GPU memory and storage, avoiding the CPU altogether. This direct path can increase bandwidth, decrease latency, and decrease the utilization load on the CPU and GPU. Addressing the IO problem with Magnum IO drives toward a balanced system.