GlobusWorld 2019 has issued its Call for Presentations. The event takes place May 1-2 in Chicago. The event centers around Globus, a secure, reliable research data management service that enables researchers to move, share, publish & discover data via a single interface. “We’re seeking Globus users, managers and providers to present “lightning talks” at GlobusWorld 2019. Presenters will include research computing leaders, system administrators, and others who use Globus at universities, national supercomputing centers, and other research facilities.”

The battle against cancer is now being fought at the data level. A new Ai Startup called PAIGE hopes to revolutionize clinical diagnosis and treatment in pathology and oncology through the use of artificial intelligence. “Through its partnership with Igneous, PAIGE will be able to securely and efficiently manage 8 petabytes of unstructured data, including anonymized tumor scan images and clinical notes, as part of an integrated machine learning-based healthcare AI workflow anchored by an industry-leading Pure Storage FlashBlade and NVIDIA GPU compute cluster.”

Today the Google Cloud announced Public Beta availability of NVIDIA T4 GPUs for Machine Learning workloads. Starting today, NVIDIA T4 GPU instances are available in the U.S. and Europe as well as several other regions across the globe, including Brazil, India, Japan and Singapore. “The T4 is the best GPU in our product portfolio for running inference workloads. Its high-performance characteristics for FP16, INT8, and INT4 allow you to run high-scale inference with flexible accuracy/performance tradeoffs that are not available on any other accelerator.”

In this video from SC18, Dr. Daniel Reed moderates a panel discussion entitled: “If you can’t measure it, you can’t improve it” — Software Improvements from Power/Energy Measurement Capabilities.” “We have made major gains in improving the energy efficiency of the facility as well as computing hardware, but there are still large gains to be had with software- particularly application software. Just tuning code for performance isn’t enough; the same time to solution can have very different power profiles.”

The Partnership for Advanced Computing in Europe has published a new report: PRACE in the EuroHPC Era. PRACE is an international non-profit association with its seat in Brussels. “The EuroHPC Joint Undertaking will advance the European supercomputer landscape by completing the infrastructure pyramid at the top level with European leadership-class supercomputers. In a context of a strong international competition with USA, China and Japan, this development is highly expected by all stakeholders of HPC in Europe. For the European HPC users from science and industry, i.e. industrials and SMEs, the seamless integration of these new top-level systems and services into the existing European HPC-ecosystem is an issue of paramount importance.”

Today Quobyte announced a partnership with EUROstor to expand the reach of its next-generation file system throughout the German, Austrian and Swiss marketplace to customers seeking scale-out NAS solutions for enterprise application workloads. “With scale-out NAS solutions becoming increasingly important within our market, Quobyte is an exciting expansion of our storage product portfolio,” said Wolfgang Bauer, technical director of EUROstor. “With Quobyte’s flexibility, scalability and high performance, we can easily deliver storage solutions that are purpose-built to satisfy our customers’ needs. We trust there are a lot more interesting projects to come.”

In this special guest feature from Scientific Computing World, Robert Roe reports on new technology and 30 years of the US supercomputing conference at SC18 in Dallas. “From our volunteers to our exhibitors to our students and attendees – SC18 was inspirational,” said SC18 general chair Ralph McEldowney. “Whether it was in technical sessions or on the exhibit floor, SC18 inspired people with the best in research, technology, and information sharing.”

The upcoming Rice University Oil and Gas HPC Conference will focus on the computational challenges and needs in the Energy industry. The event takes place March 4-6, 2019 in Houston. “High-end computing and information technology continues to stand out across the industry as a critical business enabler and differentiator with a relatively well understood return on investment. However, challenges such as constantly changing technology landscape, increasing focus on software and software innovation, and escalating concerns around workforce development still remain. The agenda for the conference includes invited keynote and plenary speakers, parallel sessions made up of at least four presentations each and a student poster session.”

Scientists and Engineers at Berkeley Lab are busy preparing for Exascale supercomputing this week at the ECP Annual Meeting in Houston. With a full agenda running five days, LBL researchers will contribute Two Plenaries, Five Tutorials, 15 Breakouts and 20 Posters. “Sponsored by the Exascale Computing Project, the ECP Annual Meeting centers around the many technical accomplishments of our talented research teams, while providing a collaborative working forum that includes featured speakers, workshops, tutorials, and numerous planning and co-design meetings in support of integrated project understanding, team building and continued progress.”

In this video, Bill Jenkins from Intel presents and introduction to High Performance Computing. “This online training will provide a high-level introduction to high performance computing, the problem it solves and the vertical markets it solves it in. Rather than focusing on the step by step, this training will educate into the concepts and resources available to perform the data analytics process and even also discuss where accelerators can be used.”

Latest Video

Industry Perspectives

At SC18 in Dallas, I had a chance to catch up with Gary Grider from LANL. “So we’re forming a consortium to chase efficient computing. We see many of the HPC sites today seem to be headed down the path of buying machines that work really well with very dense linear algebra problems. The problem is: hardcore simulation can often not be a great fit on machines built for high Linpack numbers.” [READ MORE…]

White Papers

The use of Co-Design and offloading are important tools in achieving Exascale computing. Application developers and system designers can take advantage of network offload and emerging co-design protocols to accelerate their current applications. Adopting some basic co-design and offloading methods to smaller scale systems can achieve more performance on less hardware resulting in low cost and higher throughput. Learn more by downloading this guide.