Today DDN announced record performance on the Securities Technology Analysis Center (STAC) benchmark. Using the company’s EXAScaler storage solution, DDN set new public records for multiple workload types and sizes, including large and small workloads as well as I/O and compute-intensive workloads.

Today Cray announced it has strengthened its Lustre and overall storage and data management expertise with the addition of key individuals from Terascala, a Boston-based storage company highly regarded for developing easy-to-use and easy-to-manage storage solutions.

“Modern macroeconomic theory has provided many qualitative insights into the functioning of financial and labor markets in our complex modern economies. But determining the quantitative importance of frequently countervailing forces requires much more than the pencil and paper methods used by economic theorists. My talk will describe some new economic ideas brought by economic theory and how efforts to develop computational tools promise to help macroeconomists perform quantitative analyses that can inform policy choices.”

“BlueTides has successfully used essentially the entire set of XE6 nodes on the Blue Waters. It follows the evolution of 0.7 trillion particles in a large volume of the universe (600 co-moving Mpc on a side) over the first billion years of the universe’s evolution with a dynamic range of 6 (12) orders of magnitude in space (mass). This makes BlueTides by far the largest cosmological hydrodynamic simulation ever run.”

There is still time to enjoy early bird registration rates the 30th annual ISC High Performance conference. Reduced rates end June 10 for the conference, which takes place July 12-16 in Frankfurt, Germany.

This week Lawrence Livermore National Laboratory broke ground on a modular and sustainable supercomputing facility that will provide a flexible infrastructure able to accommodate the Laboratory’s growing demand for HPC.

The Gauss Centre for Supercomputing (GCS) reports that the 10th PRACE Regular Call for Proposals resulted in GCS supporting 14 outstanding scientific research projects from eight European countries. A total of 246.93 million core hours of computing time were awarded on GCS supercomputers, including Hornet at HLRS and SuperMUC at LRZ.

“This talk will give an overview of training and graduate education initiatives in Sweden in the eScience and high performance computing areas and show why these are important but also different. Lessons learned from giving courses at six super computer centers and current and future courses will be discussed.”

In this video from SC14, Mark O’Conner from Allinea demonstrates the company’s new Forge software development suite. “A shared, intuitive user interface between the debugger and profiler with a single, shallow learning curve ensures scientific developers and HPC experts alike get the maximum value from your tools investment.”

Latest Video

Industry Perspectives

In this Nvidia podcast, Bryan Catanzaro from Baidu describes how machines with Deep Learning capabilities are now better at recognizing objects in images than humans. “AI gets better and better until it kind of disappears into the background,” says Catanzaro — NVIDIA’s head of applied deep learning research — in conversation with host Michael Copeland on this week’s edition of the new AI Podcast. “Once you stop noticing that it’s there because it works so well — that’s when it’s really landed.” [Read More...]

White Papers

This white paper reviews common HPC-environment challenges and outlines solutions that can help IT professionals deliver best-in-class HPC cloud solutions—without undue stress and organizational chaos.