Exascale Watch: El Capitan Will Use AMD CPUs & GPUs to Reach 2 Exaflops

By John Russell

March 4, 2020

HPE and its collaborators reported today that El Capitan, the forthcoming exascale supercomputer to be sited at Lawrence Livermore National Laboratory and serve the National Nuclear Security Administration (NNSA), will use AMD’s next-gen ‘Genoa’ Epyc CPUs and Radeon GPUs and deliver 2 exaflops (peak double-precision) performance, a 30 percent increase over the original spec. The new system, expected to be put into service in 2023, will be 10x faster than Summit, the fastest publicly-ranked supercomputer in the world today (Top500, November 2019).

The choice of AMD processor technology had not been made when the Department of Energy first announced the ~$600 million El Capitan procurement last August. Cray, now part of HPE, was announced as the prime contractor as was selection of its Shasta architecture. More detail on the CPU/GPU selections along with a few other system elements were presented in a media pre-briefing this week given by Bronis de Supinski, CTO, LLNL, Steve Scott, SVP, senior fellow, and CTO, HPE, and Forrest Norrod, SVP and GM, datacenter and embedded systems group, AMD.

HPE, through Cray, has been the big winner so far in the U.S. Exascale sweepstakes, obtaining contracts for all three systems – Aurora, with an Intel CPU/GPU pair; Frontier, with another AMD CPU/GPU pair, and El Capitan, which we now know will also feature AMD processors and AMD accelerators. After re-entering the HPC server market with its Epyc line of CPUs in 2017, AMD at first treaded lightly in pairing Epyc with Radeon GPUs in high-end servers. That clearly has changed.

Steve Scott, HPE/Cray

Talking about the delayed processor selections for El Capitan, Scott said, “The strategy that they’ve (DoE/LLNL) used – and increasingly others are using it as well – is to choose the system architecture, start getting that in place, and then make the processor choice as late as possible in the process. By doing that you can you get better visibility. Your headlights are illuminating farther in the future. The processor roadmaps have continued to improve and by making that decision later you tend to get a better result in the end, and that’s exactly what happened here.”

He declined to specify the process technology for the new processors but there’s been speculation Genoa will be fabbed on a 5nm process. We may know more soon, “We’ll be unpacking more the details in those parts as time goes by,” said AMD’s Norrod, “Our next disclosure on Genoa will, quite frankly…we’ll say a little bit about that at our financial analyst day which is this Thursday.” That’s tomorrow.

El Capitan’s primary mission is within NNSA’s advanced simulation and computing program, which uses simulations to certify the country’s nuclear stockpile is safe, secure and reliable. “To provide that certification we require complex simulations and as the nuclear stockpile ages, the complexity of the simulations only increases and need to be able to use larger and larger systems,” said de Supinski.

El Capitan will leverage HPE’s Shasta architecture which is at the core of the most recent refresh of the HPE/Cray advanced scale product line. Other core components include a new software stack, new Slingshot interconnect technology, and new storage system.

“[The] new software stack that provides a much more dynamic cloud like environment for hybrid workflows,” said Scott. “It has open documented APIs between the software components. It has a management system that’s built with redundant microservices and managed as a Kubernetes cluster, and has robust container support to allow users to take any workload that runs anyplace and run under this system as well.” A systems monitoring framework will run underneath the stack to optimize performance and help predict failures.”

Calling it a future-proof design, Scott said “We’ve designed it to accommodate a wide diversity of processors, different amounts of power, different types of processing, different physical sizes of the processor and memory system. And we’ve given it the power and the cooling headroom to handle processors that are headed again in the years ahead up to the kilowatt power levels.”

El Capitan will be liquid cooled and have an energy budget between 30-to-40 megawatts with expectations it will end up closer to 30MW than 40MW according to Scott. Slingshot and high performance Ethernet comprise the planned system interconnect. The planned storage system is HPE’s new ClusterStor E1000, which Scott said is, “a highly flexible tiered storage system using flash and hard drive partitions That allows you to individually optimize for performance as well as capacity and then does intelligent tiering of data between the partitions and this attaches directly to the slingshot interconnect, which helps take out cost and complexity and latency.”

Shasta Compute Blade

Specific performance specs were generally not disclosed. The new AMD CPU will use the Zen4 core which is reportedly on schedule for launch in 2021. The new CPU-GPU pairing (A-plus-A in AMD parlance) will leverage AMD’s Infinity fabric 3.0 to deliver memory coherency. The detailed node structure and number of nodes for El Capitan were not discussed in the pre-briefing but the official press release characterized the architecture as, “using accelerator-centric compute blades (in a 4:1 GPU to CPU ratio, connected by the 3rd Gen AMD Infinity Architecture for high-bandwidth, low latency connections) to increase performance for data-intensive AI, machine learning and analytics needs by offloading processing from the CPU to the GPU.”

As is the case generally in heterogeneous architectures, accelerators handle most of the work and require efficient IO. Norrod said, “We have next generation memory and IO subsystems that can provide non-blocking access to memory, non-blocking access to IO, and ensure that the full power of the Zen4 CPU engine and the Radeon Instinct GPU engines.”

He said the new GPU is optimized for high performance computing and machine intelligence applications. “It has extensive mix precision operations to optimize that deep learning performance, as well as [the ability] to provide peak single and dual precision performance with more traditional HPC applications. It does embody a next generation of high bandwidth memory (HBM) memory on package to provide the memory bandwidth and capacity that’s so critical to again feed the beast (GPU).”

While many data analytic workloads look quite different from high performance simulation, Scott said, “It turns out AI is one of the workloads that shares a lot in common with high performance simulation. Typically, the granularity or the precision that you use for the computations is quite a bit different. Most of AI is done at 16- or, or 32-bit precision, whereas, most of the scientific simulation is done at 64-bit precision. But modern processors like the AMD GPUs can take their function units and run them either in 64-bit mode or in 16-bit mode or 32-bit mode depending upon the particular computation. [To do that] you need a strong interconnect and need very high memory bandwidth which it shares in common with scientific workloads.

“We find the combination of the CPU and GPU with flexible precision, married with very high memory bandwidth and interconnect bandwidth and storage bandwidth to be well suited for both simulation and AI workloads and we can use all the compute nodes in the system to bring to bear to either those workloads,” Scott said.

Interestingly, AI is not currently a top priority at LLNL.

De Supinski said, “We’re doing a lot of research and development at Livermore exploring how we can bring [AI] to bear our simulations. Whereas we need a certain accuracy, deep learning models are probabilistic and so you can often be good enough with lower precision operations whereas we have to be able to understand where the errors are and where they are becoming larger because of the reduced precision and then be able to bring some mechanism in to increase precision and accuracy required.”

At the pre-briefing, a question was asked about El Capitan’s ability to use non-GPU accelerators. Scott said while GPUs are currently the AI accelerator of choice, many users are looking at alternatives and that El Capitan’s system architecture is “designed to accommodate that kind of heterogeneous mix.”

De Supinski noted LLNL is using an unclassified system, Lassen, a sister machine to the classified Sierra system, to learn more about emerging AI accelerators. “We’re actively exploring ways of adding purpose-built machine learning accelerators to that system. I would anticipate that the mechanism by which we’re doing that is available entirely in El Capitan; that is we can add additional nodes to the system that are designed specifically for that purpose. We will see how things go with our exploratory studies on Lassen. If they go well, we will be very likely to engage HPE in helping us figure out how we can exploit that.”

AMD, HPE and LLNL are collaborating on software tools for El Capitan. Part of the plan is to leverage AMD’s ROCm framework to take advantage of  “coherent acceleration in the OpenMP environment as well as other environments” according to Norrod.

Scott said, “As part of this procurement, the Department of Energy has provided additional funds beyond the purchase of the machine to fund non-recurring engineering efforts and one major piece of that is to work closely with AMD on enhancing the programming environment for their new CPU-GPU architecture.” Work is ongoing by all three partners to take the critical applications and workloads forward and optimize them to get the best performance in the machine when El Capitan is delivered.

De Supinski emphasized, “This is a collaborative process particularly for the software. Some of the software for the system is being developed at Lawrence Livermore in addition to the applications. For instance, we very much expect Spack, which is an open source management package [to able to run] on the new system.”

One interesting feature which was not discussed in the briefing but was mentioned in the official announcement is El Capitan’s planned use of optical data transmission.

According to the release, “HPE is expanding its partnership with LLNL to actively explore HPE optics technologies, a computing solution that uses light to transmit data, to feature in the DOE’s El Capitan. HPE’s optics technologies stem from R&D efforts related to PathForward, a program backed by U.S. DOE’s Exascale Computing Project. HPE developed and demonstrated breakthrough optics prototypes that integrate electrical-to-optical interfaces to enable broad use in future classes of system interconnects. Together, HPE and LLNL are exploring ways to integrate these optics technologies with HPE’s Cray Slingshot for DOE’s El Capitan to transmit more data, more efficiently. This approach aims to improve power efficiency, reliability and ability to cost-effectively increase global system bandwidth.”

Optical data transmission is a hotbed of research with many companies aggressively seeking practical implementation.

El Capitan will become NNSA’s fastest computer and greatly enhance NNSA’s ability to run 3D simulations quickly instead of 2D simulation. LLNL is managing the new system for the NNSA and has developed emerging techniques that allow researchers to create faster, more accurate models for primary missions across stockpile modernization and inertial confinement fusion (ICF), a key aspect of stockpile stewardship.

Link to the official announcement: https://www.llnl.gov/news/llnl-and-hpe-partner-amd-el-capitan-projected-worlds-fastest-supercomputer

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Nvidia Aims Clara Healthcare at Drug Discovery, Imaging via DGX

April 12, 2021

Nvidia Corp. continues to expand its Clara healthcare platform with the addition of computational drug discovery and medical imaging tools based on its DGX A100 platform, related InfiniBand networking and its AGX develop Read more…

Nvidia Serves Up Its First Arm Datacenter CPU ‘Grace’ During Kitchen Keynote

April 12, 2021

Today at Nvidia’s annual spring GPU technology conference, held virtually once more due to the ongoing pandemic, the company announced its first ever Arm-based CPU, called Grace in honor of the famous American programmer Grace Hopper. Read more…

Nvidia Debuts BlueField-3 – Its Next DPU with Big Plans for an Expanded Role

April 12, 2021

Nvidia today announced its next generation data processing unit (DPU) – BlueField-3 – adding more substance to its evolving concept of the DPU as a full-fledged partner to CPUs and GPUs in delivering advanced computi Read more…

Nvidia’s Newly DPU-Enabled SuperPOD Is a Multi-Tenant, Cloud-Native Supercomputer

April 12, 2021

At GTC 2021, Nvidia has announced an upgraded iteration of its DGX SuperPods, calling the new offering “the first cloud-native, multi-tenant supercomputer.” The newly announced SuperPods come just two years after the Read more…

Tune in to Watch Nvidia’s GTC21 Keynote with Jensen Huang – Recording Now Available

April 12, 2021

Join HPCwire right here on Monday, April 12, at 8:30 am PT to see the Nvidia GTC21 keynote from Nvidia’s CEO, Jensen Huang, livestreamed in its entirety. Hosted by HPCwire, you can click to join the Huang keynote on our livestream to hear Nvidia’s expected news and... Read more…

AWS Solution Channel

Volkswagen Passenger Cars Uses NICE DCV for High-Performance 3D Remote Visualization

 

Volkswagen Passenger Cars has been one of the world’s largest car manufacturers for over 70 years. The company delivers more than 6 million automobiles to global customers every year, from 50 production locations on five continents. Read more…

The US Places Seven Additional Chinese Supercomputing Entities on Blacklist

April 8, 2021

As tensions between the U.S. and China continue to simmer, the U.S. government today added seven Chinese supercomputing entities to an economic blacklist. The U.S. Entity List bars U.S. firms from supplying key technolog Read more…

Nvidia Serves Up Its First Arm Datacenter CPU ‘Grace’ During Kitchen Keynote

April 12, 2021

Today at Nvidia’s annual spring GPU technology conference, held virtually once more due to the ongoing pandemic, the company announced its first ever Arm-based CPU, called Grace in honor of the famous American programmer Grace Hopper. Read more…

Nvidia Debuts BlueField-3 – Its Next DPU with Big Plans for an Expanded Role

April 12, 2021

Nvidia today announced its next generation data processing unit (DPU) – BlueField-3 – adding more substance to its evolving concept of the DPU as a full-fle Read more…

Nvidia’s Newly DPU-Enabled SuperPOD Is a Multi-Tenant, Cloud-Native Supercomputer

April 12, 2021

At GTC 2021, Nvidia has announced an upgraded iteration of its DGX SuperPods, calling the new offering “the first cloud-native, multi-tenant supercomputer.” Read more…

Tune in to Watch Nvidia’s GTC21 Keynote with Jensen Huang – Recording Now Available

April 12, 2021

Join HPCwire right here on Monday, April 12, at 8:30 am PT to see the Nvidia GTC21 keynote from Nvidia’s CEO, Jensen Huang, livestreamed in its entirety. Hosted by HPCwire, you can click to join the Huang keynote on our livestream to hear Nvidia’s expected news and... Read more…

The US Places Seven Additional Chinese Supercomputing Entities on Blacklist

April 8, 2021

As tensions between the U.S. and China continue to simmer, the U.S. government today added seven Chinese supercomputing entities to an economic blacklist. The U Read more…

Habana’s AI Silicon Comes to San Diego Supercomputer Center

April 8, 2021

Habana Labs, an Intel-owned AI company, has partnered with server maker Supermicro to provide high-performance, high-efficiency AI computing in the form of new Read more…

Intel Partners Debut Latest Servers Based on the New Intel Gen 3 ‘Ice Lake’ Xeons

April 7, 2021

Fresh from Intel’s launch of the company’s latest third-generation Xeon Scalable “Ice Lake” processors on April 6 (Tuesday), Intel server partners Cisco, Dell EMC, HPE and Lenovo simultaneously unveiled their first server models built around the latest chips. And though arch-rival AMD may... Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

CERN Is Betting Big on Exascale

April 1, 2021

The European Organization for Nuclear Research (CERN) involves 23 countries, 15,000 researchers, billions of dollars a year, and the biggest machine in the worl Read more…

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

HPE Launches Storage Line Loaded with IBM’s Spectrum Scale File System

April 6, 2021

HPE today launched a new family of storage solutions bundled with IBM’s Spectrum Scale Erasure Code Edition parallel file system (description below) and featu Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Saudi Aramco Unveils Dammam 7, Its New Top Ten Supercomputer

January 21, 2021

By revenue, oil and gas giant Saudi Aramco is one of the largest companies in the world, and it has historically employed commensurate amounts of supercomputing Read more…

Quantum Computer Start-up IonQ Plans IPO via SPAC

March 8, 2021

IonQ, a Maryland-based quantum computing start-up working with ion trap technology, plans to go public via a Special Purpose Acquisition Company (SPAC) merger a Read more…

Leading Solution Providers

Contributors

Can Deep Learning Replace Numerical Weather Prediction?

March 3, 2021

Numerical weather prediction (NWP) is a mainstay of supercomputing. Some of the first applications of the first supercomputers dealt with climate modeling, and Read more…

Livermore’s El Capitan Supercomputer to Debut HPE ‘Rabbit’ Near Node Local Storage

February 18, 2021

A near node local storage innovation called Rabbit factored heavily into Lawrence Livermore National Laboratory’s decision to select Cray’s proposal for its CORAL-2 machine, the lab’s first exascale-class supercomputer, El Capitan. Details of this new storage technology were revealed... Read more…

New Deep Learning Algorithm Solves Rubik’s Cube

July 25, 2018

Solving (and attempting to solve) Rubik’s Cube has delighted millions of puzzle lovers since 1974 when the cube was invented by Hungarian sculptor and archite Read more…

African Supercomputing Center Inaugurates ‘Toubkal,’ Most Powerful Supercomputer on the Continent

February 25, 2021

Historically, Africa hasn’t exactly been synonymous with supercomputing. There are only a handful of supercomputers on the continent, with few ranking on the Read more…

The History of Supercomputing vs. COVID-19

March 9, 2021

The COVID-19 pandemic poses a greater challenge to the high-performance computing community than any before. HPCwire's coverage of the supercomputing response t Read more…

HPE Names Justin Hotard New HPC Chief as Pete Ungaro Departs

March 2, 2021

HPE CEO Antonio Neri announced today (March 2, 2021) the appointment of Justin Hotard as general manager of HPC, mission critical solutions and labs, effective Read more…

AMD Launches Epyc ‘Milan’ with 19 SKUs for HPC, Enterprise and Hyperscale

March 15, 2021

At a virtual launch event held today (Monday), AMD revealed its third-generation Epyc “Milan” CPU lineup: a set of 19 SKUs -- including the flagship 64-core, 280-watt 7763 part --  aimed at HPC, enterprise and cloud workloads. Notably, the third-gen Epyc Milan chips achieve 19 percent... Read more…

Microsoft, HPE Bringing AI, Edge, Cloud to Earth Orbit in Preparation for Mars Missions

February 12, 2021

The International Space Station will soon get a delivery of powerful AI, edge and cloud computing tools from HPE and Microsoft Azure to expand technology experi Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire