Today GTC China, NVIDIA made a series of announcements around Deep Learning, and GPU-accelerated computing for Hyperscale datacenters. “Demand is surging for technology that can accelerate the delivery of AI services of all kinds. And NVIDIA’s deep learning platform — which the company updated Tuesday with new inferencing software — promises to be the fastest, most efficient way to deliver these services.”

Cloud adoption is accelerating at the blink of an eye, easing the burden of managing data-rich workloads for enterprises big and small. Yet, common myths and misconceptions about the hybrid cloud are delaying enterprises from reaping the benefits. “In this article, we will debunk five of the top most commonly believed myths that keep companies from strengthening their infrastructure with a hybrid approach.”

Today Cray announced it has completed the previously announced transaction and strategic partnership with Seagate centered around the addition of the ClusterStor high-performance storage business. “As a pioneer in providing large-scale storage systems for supercomputers, it’s fitting that Cray will take over the ClusterStor line.”

In this video from the GPU Technology Conference, Rick Young from Ingram Micro describes the company’s Artizen HPC solutions. “Available now to channel partners in the U.S., the distributor’s new and exclusive line of Artizen High Performance Computing (HPC) offerings include turnkey high performance servers, ultimate workstations, and customizable supercomputing clusters, as well as computing integration and software installation services.”

“This presentation will provide an overview of the Nvidia Tesla Deployment Kit (TDK) from a user and a system administrator point of view. TDL contains Nvidia Management Library (NVML) and nvidia-healthmon–a tool for detecting and troubleshooting known GPU issues in a cluster environment. Usage models within a cluster environment will be presented along with a discussion on how existing resource management tools can be extended to improve allocation and accounting of GPU resources.”

“HPC storage solutions and futures continue to evolve as growth and performance requirements permeate every HPC market segment. Torben discusses these challenges and how the company’s storage solutions are addressing these shifting needs with new developments around disk drives, RAID, CIFS, security, small file handling, and other related technologies.”

In this talk, we introduce the rCUDA remote GPU virtualization framework, which has been shown to be the only one that supports the most recent CUDA versions, in addition to leverage the InfiniBand interconnect for the sake of performance. Furthermore, we also present the last developments within this framework, related with the use of low-power processors, enhanced job schedulers, and virtual machine environments.”

“Congratulations go out to Troy Baer, HPC system administrator at the National Institute for Computational Sciences (NICS), University of Tennessee. Troy Baer’s contributions in scheduling and resource management using Moab have helped Kraken—NICS’ flagship computing resource and the first academic computer to break the petaflop barrier—achieve outstanding 90-95% utilization rates since 2010. Baer’s administrative capabilities enable researchers in numerous scientific arenas, from climate to materials science to astrophysics, to achieve breakthroughs not yet possible on other resources. In November 2012, Baer helped NICS’ Beacon system secure a No. 1 ranking on the Green500 list of energy-efficient supercomputers.”

“This talk will focus on programming models and their designs for upcoming exascale systems with millions of processors and accelerators. Current status and future trends of MPI and PGAS (UPC and OpenSHMEM) programming models will be presented. We will discuss challenges in designing runtime environments for these programming models by taking into account support for multi-core, high-performance networks, GPGPUs, Intel MIC, scalable collectives (multi-core-aware, topology-aware, and power-aware), non-blocking collectives using Offload framework, one-sided RMA operations, schemes and architectures for fault-tolerance/fault-resilience.”

“With around 3.2 billion computer operations (3.2 gigaflops) per watt, the combination of GPUs CPUs makes “Piz Daint” one of the world’s most energy-­efficient supercomputers in the petaflop performance class.”

Latest Video

Industry Perspectives

In this podcast, the Radio Free HPC team looks at China’s massive upgrade of the Tianhe-2A supercomputer to 95 Petaflops peak performance. "As detailed in a new 21-page report by Jack Dongarra from the University of Tennessee, the upgrade should nearly double the performance of the system, which is currently ranked at #2 on TOP500." [Read More...]