Something for Everyone at GPU Technology Conference

By Tiffany Trader

March 23, 2015

Once relegated to the category of specialized gaming hardware, today’s graphics processors are solving some of the world’s toughest computing problems. During GTC15 last week in San Jose, the full breadth and depth of session topics provided even more evidence of how far graphics processors have come from their gaming roots.

And while this year deep learning and self-driving cars were front and center (literally, the entrance hall was brimming with automotive eye candy), with nearly 500 sessions to choose from, there were still plenty of interesting and relevant sessions targeting the big compute side of GPUs – not to mention the impromptu hallway moments.

If you missed the show this year or if you weren’t able to make it to every session on your agenda, fear not because NVIDIA has once again posted all of the keynotes and a majority of the sessions online.

To get you started, we’ve compiled our top twelve picks for the HPC set. Included in the list are standouts from the very first OpenPOWER summit, which shared the convention space with GTC. If there’s an event you enjoyed that you don’t see, drop me an email (tiffany@hpcwire.com) or mention it in the comments section.

#1 — The Keynotes

http://www.ustream.tv/recorded/60025825

In GTC’s opening keynote address, NVIDIA CEO and co-founder Jen-Hsun Huang discusses the latest breakthroughs in visual computing, including the company’s inroads into deep learning. On hand to discuss the future of autonomous vehicles was special guest, 21st century visionary Elon Musk.

And if you’re following the burgeoning deep learning space, check out the other two keynotes here and here, featuring Jeff Dean, Senior Fellow with Google’s Knowledge Group, and Andrew Ng, Chief Scientist at Baidu.

#2

Heterogeneous HPC, Architectural Optimization, and NVLink
Steve Oberlin, CTO, Accelerated Computing, NVIDIA

http://on-demand.gputechconf.com/gtc/2015/video/S5649.html

In this talk, Oberlin explores heterogeneous node design and architecture and describes the role of NVLink, NVIDIA’s scalable node integration channel set to debut with the Pascal GPU in 2016. The technology is a key piece of the DOE CORAL Summit and Sierra supercomputer systems.

#3

E4-ARKA: ARM64+GPU+IB is Now Here
Piero Altoè, Project Manager, E4 Computer Engineering

http://on-demand.gputechconf.com/gtc/2015/video/S5422.html

E4 Computer Engineering Project Manager Piero Altoè presents on ARKA, said to be the first server solution based on ARM 64 bit SoC dedicated to HPC. The compute node is boosted by discrete GPU NVIDIA K20 cards, and features both 10GbE and FDR InfiniBand fabrics. The hardware configuration of the compute node is described in detail to demonstrate the unique capabilities of the ARM+GPU+IB combination. Benchmarks are also discussed with particular attention paid to molecular dynamics software.

#4

A CUDA Implementation of the High Performance Conjugate Gradient (HPCG) Benchmark
Everett Phillips, HPC Software Engineer, NVIDIA

http://on-demand.gputechconf.com/gtc/2015/video/S5185.html

NVIDIA’s Everett Phillips describes a CUDA implementation of the HPCG benchmark, including key optimization strategies and performance results on a wide range of GPU systems: from the smallest CUDA capable platform to the largest GPU supercomputers. Comparing it with the long-standing LINPACK benchmark, Phillips points out that HPCG is more representative of common computation patterns found in modern applications, which rather than being solely compute-intensive are more memory-heavy and require more network communication.

#5

GPU Errors on HPC Systems: Characterization, Quantification and Implications for Architects and Operations
James Rogers, Director of Operations, Oak Ridge Leadership Computing Facility (OLCF), Oak Ridge National Laboratory (ORNL)

http://on-demand.gputechconf.com/gtc/2015/video/S5566.html

The fastest US supercomputer, Titan, installed at Oak Ridge National Laboratory, has more than 18,000 GPUs that are used for a broad range of scientific workloads. In this talk, Rogers points out that while the performance efficiency of GPUs is well understood, their resilience characteristics in a large-scale computing system have not been fully evaluated. He goes on to describe a study, drawn from 300,000,000 Titan node hours, that was undertaken to boost understanding of GPU errors on large-scale heterogenous machines. The work has implications for future GPU architects and HPC centers that use graphics processors.

#6

Attacking HIV with Petascale Molecular Dynamics Simulations on Titan and Blue Waters
James Phillips, Senior Research Programmer, University of Illinois at Urbana-Champaign

http://on-demand.gputechconf.com/gtc/2015/video/S5149.html

Starting with a description of the HIV infectious cycle, University of Illinois’ James Phillips discusses how the highly parallel molecular dynamics code NAMD is being used to perform petascale biomolecular simulations on the GPU-accelerated Cray XK7 Blue Waters and ORNL Titan machines.

#7

Power 8 Microprocessor
Satish Kumar Sadasivam, Senior Performance Engineer and Master Inventor, IBM STG

http://on-demand.gputechconf.com/gtc/2015/video/S5701.html

The first OpenPOWER Summit took place this year in tandem with GTC15. In addition to the main keynotes on Wednesday,  sessions spanned many topics, including this standing-room-only talk about IBM’s first OpenPOWER processor, POWER8. Starting with an overview of the microarchitecture, IBM’s Satish Kumar Sadasivam looks at performance monitoring features in POWER8 and other factors important to optimizing application performance, as well as the POWER8 pipeline.

#8

OpenACC for Fortran Programmers
Michael Wolfe, Compiler Engineer, NVIDIA

http://on-demand.gputechconf.com/gtc/2015/video/S5388.html

Michael Wolfe, who will be familiar to HPCwire readers for his popular “Compilers and More” column, describes how to program NVIDIA GPUs using Fortran with OpenACC directives. The first half of the talk covers the basic material necessary to start using GPUs for Fortran programs and then gets into more advanced material, with hints and tips for Fortran programmers with larger applications that they want to accelerate with a GPU.

#9

Achieving Near-Native GPU Performance in the Cloud
John Paul Walters, Project Leader, University of Southern California, Information Sciences Institute (ISI)

http://on-demand.gputechconf.com/gtc/2015/video/S5323.html

If you’re wondering about the current state of HPC in the cloud, this session is not to be missed. John Paul Walters, technical lead for ISI’s HPC OpenStack initiative and the PI of ISI’s NVIDIA CUDA Research Center, begins with the question of whether it is possible to provide HPC class performance in the GPU-equipped cloud. Approaches described include PCI passthrough, SR-IOV and GPUDirect in tandem with multi-node GPUs with results showing near-native performance.

#10

Porting Apps to Titan: Results from the Inaugural GPU Hackathon

Fernanda Foertter, HPC User Assistance Specialist, Oak Ridge National Laboratory
Seth Johnson, R&D Staff, Monte Carlo Methods, Oak Ridge National Laboratory

http://on-demand.gputechconf.com/gtc/2015/video/S5515.html

The results of the inaugural GPU Hackathon held at the Oak Ridge Leadership Computing Facility are showcased in this 50 minute session, along with lessons learned. The event hosted six teams paired with mentors over a week where applications were ported to GPUs using OpenACC directives. According to ORNL’s Fernanda Foertter, the experience was so successful, three more GPU-porting hackathons are underway.

#11

Using Docker in High Performance Computing in OpenPOWER Environment
Sam Sanjabi, Advisory Software Engineer, IBM Systems & Technology Lab (Canada)

http://on-demand.gputechconf.com/gtc/2015/video/S5680.html

Popular container technology Docker has been widely used in cloud, but increasingly we are hearing about its potential for high-performance computing. Are there benefits of Docker in HPC? There are according to IBM’s Sam Sanjabi. During an OpenPOWER session at GTC15, Sanjabi talks about an integration that was done between IBM’s Platform LSF job scheduler and Docker on the OpenPOWER platform. He gives a short overview of both of these technologies and then describes the results of combining them.

#12

GPU vs Xeon Phi: Performance of Bandwidth Bound Applications with a Lattice QCD Case Study
Mathias Wagner, Postdoc, Indiana University

http://on-demand.gputechconf.com/gtc/2015/video/S5447.html

A very popular session delivered by theoretical physicist Mathias Wagner addresses the performance benefits of accelerators, which continue to make inroads in HPC. More specifically, Wagner sheds light on whether GPUs or Phi coprocessors make the most sense for bandwidth bound applications. He compares their performance using a Lattice QCD application as a case study and describes what it takes to achieve great performance on both architectures.

Didn’t see your favorite session listed? For the full agenda, go to https://registration.gputechconf.com/form/session-listing and play with the keyword search to find what you are looking for. More information on the first OpenPOWER summit is available here.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Why HPC Storage Matters More Now Than Ever: Analyst Q&A

September 17, 2021

With soaring data volumes and insatiable computing driving nearly every facet of economic, social and scientific progress, data storage is seizing the spotlight. Hyperion Research analyst and noted storage expert Mark No Read more…

GigaIO Gets $14.7M in Series B Funding to Expand Its Composable Fabric Technology to Customers

September 16, 2021

Just before the COVID-19 pandemic began in March 2020, GigaIO introduced its Universal Composable Fabric technology, which allows enterprises to bring together any HPC and AI resources and integrate them with networking, Read more…

What’s New in HPC Research: Solar Power, ExaWorks, Optane & More

September 16, 2021

In this regular feature, HPCwire highlights newly published research in the high-performance computing community and related domains. From parallel programming to exascale to quantum computing, the details are here. Read more…

Cerebras Brings Its Wafer-Scale Engine AI System to the Cloud

September 16, 2021

Five months ago, when Cerebras Systems debuted its second-generation wafer-scale silicon system (CS-2), co-founder and CEO Andrew Feldman hinted of the company’s coming cloud plans, and now those plans have come to fruition. Today, Cerebras and Cirrascale Cloud Services are launching... Read more…

AI Hardware Summit: Panel on Memory Looks Forward

September 15, 2021

What will system memory look like in five years? Good question. While Monday's panel, Designing AI Super-Chips at the Speed of Memory, at the AI Hardware Summit, tackled several topics, the panelists also took a brief glimpse into the future. Unlike compute, storage and networking, which... Read more…

AWS Solution Channel

Supporting Climate Model Simulations to Accelerate Climate Science

The Amazon Sustainability Data Initiative (ASDI), AWS is donating cloud resources, technical support, and access to scalable infrastructure and fast networking providing high performance computing (HPC) solutions to support simulations of near-term climate using the National Center for Atmospheric Research (NCAR) Community Earth System Model Version 2 (CESM2) and its Whole Atmosphere Community Climate Model (WACCM). Read more…

ECMWF Opens Bologna Datacenter in Preparation for Atos Supercomputer

September 14, 2021

In January 2020, the European Centre for Medium-Range Weather Forecasts (ECMWF) – a juggernaut in the weather forecasting scene – signed a four-year, $89-million contract with European tech firm Atos to quintuple its supercomputing capacity. With the deal approaching the two-year mark, ECMWF... Read more…

Why HPC Storage Matters More Now Than Ever: Analyst Q&A

September 17, 2021

With soaring data volumes and insatiable computing driving nearly every facet of economic, social and scientific progress, data storage is seizing the spotlight Read more…

Cerebras Brings Its Wafer-Scale Engine AI System to the Cloud

September 16, 2021

Five months ago, when Cerebras Systems debuted its second-generation wafer-scale silicon system (CS-2), co-founder and CEO Andrew Feldman hinted of the company’s coming cloud plans, and now those plans have come to fruition. Today, Cerebras and Cirrascale Cloud Services are launching... Read more…

AI Hardware Summit: Panel on Memory Looks Forward

September 15, 2021

What will system memory look like in five years? Good question. While Monday's panel, Designing AI Super-Chips at the Speed of Memory, at the AI Hardware Summit, tackled several topics, the panelists also took a brief glimpse into the future. Unlike compute, storage and networking, which... Read more…

ECMWF Opens Bologna Datacenter in Preparation for Atos Supercomputer

September 14, 2021

In January 2020, the European Centre for Medium-Range Weather Forecasts (ECMWF) – a juggernaut in the weather forecasting scene – signed a four-year, $89-million contract with European tech firm Atos to quintuple its supercomputing capacity. With the deal approaching the two-year mark, ECMWF... Read more…

Quantum Computer Market Headed to $830M in 2024

September 13, 2021

What is one to make of the quantum computing market? Energized (lots of funding) but still chaotic and advancing in unpredictable ways (e.g. competing qubit tec Read more…

Amazon, NCAR, SilverLining Team for Unprecedented Cloud Climate Simulations

September 10, 2021

Earth’s climate is, to put it mildly, not in a good place. In the wake of a damning report from the Intergovernmental Panel on Climate Change (IPCC), scientis Read more…

After Roadblocks and Renewals, EuroHPC Targets a Bigger, Quantum Future

September 9, 2021

The EuroHPC Joint Undertaking (JU) was formalized in 2018, beginning a new era of European supercomputing that began to bear fruit this year with the launch of several of the first EuroHPC systems. The undertaking, however, has not been without its speed bumps, and the Union faces an uphill... Read more…

How Argonne Is Preparing for Exascale in 2022

September 8, 2021

Additional details came to light on Argonne National Laboratory’s preparation for the 2022 Aurora exascale-class supercomputer, during the HPC User Forum, held virtually this week on account of pandemic. Exascale Computing Project director Doug Kothe reviewed some of the 'early exascale hardware' at Argonne, Oak Ridge and NERSC (Perlmutter), while Ti Leggett, Deputy Project Director & Deputy Director... Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer called Dojo to process truly vast amounts of video data. It’s a beast! … A truly useful exaflop at de facto FP32.” Read more…

Berkeley Lab Debuts Perlmutter, World’s Fastest AI Supercomputer

May 27, 2021

A ribbon-cutting ceremony held virtually at Berkeley Lab's National Energy Research Scientific Computing Center (NERSC) today marked the official launch of Perlmutter – aka NERSC-9 – the GPU-accelerated supercomputer built by HPE in partnership with Nvidia and AMD. Read more…

Esperanto, Silicon in Hand, Champions the Efficiency of Its 1,092-Core RISC-V Chip

August 27, 2021

Esperanto Technologies made waves last December when it announced ET-SoC-1, a new RISC-V-based chip aimed at machine learning that packed nearly 1,100 cores onto a package small enough to fit six times over on a single PCIe card. Now, Esperanto is back, silicon in-hand and taking aim... Read more…

Enter Dojo: Tesla Reveals Design for Modular Supercomputer & D1 Chip

August 20, 2021

Two months ago, Tesla revealed a massive GPU cluster that it said was “roughly the number five supercomputer in the world,” and which was just a precursor to Tesla’s real supercomputing moonshot: the long-rumored, little-detailed Dojo system. “We’ve been scaling our neural network training compute dramatically over the last few years,” said Milan Kovac, Tesla’s director of autopilot engineering. Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months after Red Hat deprecated its support for the widely popular, free CentOS server operating system. The Rocky Linux development effort... Read more…

Google Launches TPU v4 AI Chips

May 20, 2021

Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I Read more…

Intel Completes LLVM Adoption; Will End Updates to Classic C/C++ Compilers in Future

August 10, 2021

Intel reported in a blog this week that its adoption of the open source LLVM architecture for Intel’s C/C++ compiler is complete. The transition is part of In Read more…

AMD-Xilinx Deal Gains UK, EU Approvals — China’s Decision Still Pending

July 1, 2021

AMD’s planned acquisition of FPGA maker Xilinx is now in the hands of Chinese regulators after needed antitrust approvals for the $35 billion deal were receiv Read more…

Leading Solution Providers

Contributors

Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

HPE Wins $2B GreenLake HPC-as-a-Service Deal with NSA

September 1, 2021

In the heated, oft-contentious, government IT space, HPE has won a massive $2 billion contract to provide HPC and AI services to the United States’ National Security Agency (NSA). Following on the heels of the now-canceled $10 billion JEDI contract (reissued as JWCC) and a $10 billion... Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Intel Unveils New Node Names; Sapphire Rapids Is Now an ‘Intel 7’ CPU

July 27, 2021

What's a preeminent chip company to do when its process node technology lags the competition by (roughly) one generation, but outmoded naming conventions make it seem like it's two nodes behind? For Intel, the response was to change how it refers to its nodes with the aim of better reflecting its positioning within the leadership semiconductor manufacturing space. Intel revealed its new node nomenclature, and... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire