LLNL, ANL and GSK Provide Early Glimpse into Cerebras AI System Performance

By John Russell

October 13, 2020

AI chip and systems startup Cerebras was one of many AI companies showcased at the AI Hardware Summit which concluded last week. Rather than dwell on its technology, which is widely known, Cerebras invited collaborators from Lawrence Livermore National Laboratory, Argonne National Laboratory and GlaxoSmithKline (GSK) to talk about their early work on Cerebras machines and future plans.

It was interesting for a change to hear of concrete activities involving one of the new class of emerging AI machines, and in Cerebras’s case, the activities were varied from nascent to quite far along.

First, some background. If you’re an AI watcher, no doubt Cerebras is a familiar name. Flush with cash and early customers, the Los Altos-based company has been an early success story. Its ‘chip’ – the wafer scale engine (WSE) – is enormous. It consumes 46,000 square millimeters of silicon and has 400,000 so-called Sparse Linear Algebra Compute (SLAC) cores with roughly 1.2 trillion transistors.

The cores are designed for sparse linear algebra primitives that underpin all neural network computation, says the company. The WSE also 18 gigabytes of on-chip memory, nine petabytes of memory bandwidth and 100 petabytes per sec of fabric bandwidth. Its big reveal took place at Hot Chips in 2019 (see HPCwire coverage).

Cerebras founder and CEO Andrew Feldman’s presentation included a slide showing the WSE next to Nvidia A100 GPU. “You (can) see our wafer scale engine. It’s the size of a dinner plate.” This is case where size clearly matters and the early users seem to agree. The WSE’s sheer size allows porting entire models, big models at that, onto the chip, which in combination with the design greatly reduces latency and speeds throughput.

Feldman was quick to emphasize Cerebras is a systems company – not a chip company – and as he put it, one wouldn’t build a Ferrari engine and put it in a Volkswagen. Cerebras’ first system is the CS-1, a 15 RU tall system (~26 inches) and you can get three in a standard rack, said Feldman (system specs).

The CS-1 system

The point of the panel was less a description of the Cerebras technology and more an effort to show the CS-1 being put to early, practical use. So: how’s it going?

LLNL Tackles Physics Simulation with CS-1

Let’s start with LLNL’s Livermore Computing Center (LCC). LCC already has substantial computing capabilities including its number two supercomputer, Lassen, which is an IBM-built system featuring Power9 CPUs, Nvidia V100s and InfiniBand fabric. It’s a 23-petaflops system with 44 racks that was #10 on the Top500 when first stood up and was #14 on the most recent Top500.

“So why am I here talking about Cerebras,” said Bronis de Supinski, CTO for LCC. “We have this vision for performing cognitive simulation using what we call a heterogeneous system architecture. Most supercomputers have the same node over and over. That is what the Lassen’s base architecture was. By adding the CS-1, we now have a volume that is specifically tuned and intended for running machine learning models.”

“We have integrated the CS-1 connecting it directly to Lassen on the InfiniBand fabric. I’d like to note our integration team has told me that siting was one of the smoothest sitings that they’ve been involved with, particularly for the first time a company has cited something on our machinery floor,” said de Supinski.

The goal is to build machine learning into physics simulations. “There’s a variety of places we can do this, from coordinating input to individual simulations, to the many time steps, or even very tightly integrated into the simulation and computing values that are used within every time step or every few time steps of the simulation. That’s what we’re looking to do,” he said.

Fast communication between the systems is critical. “Can we have the interaction between these different volumes of the system occur quickly enough? As we’ve already heard there’s quite a bit of bandwidth in and out of this system (CS-1). We’ve begun looking at this, and we’re actually building into a simulation called CRETIN[i] the ability to do some of the physics in every time step. What we’ve found is even though we need to ship a fairly large number of values into the CS-1 in order to bring back a fair number of values, that the inference happens quickly enough with the CS-1 that we can actually make this work,” said de Supinski.

“We’re also looking at additional things like being able to update models on the fly as we generate more data from other parts of the simulation, and also potentially running the detailed physics simulation to be able to have a more accurate model. Again, the capability the CS-1 provides us with being able to run these large-scale models is such that we believe that this will work,” he said.

GSK – It’s All About the Data Flood

By now, it’s a mantra that much of modern drug discovery is about sieving the data flood provided by modern experimental instruments. “To give you an illustration of the [problem], in the first quarter of this year, we generated more data than GSK had generated [during its] this entire 300-year history before. We need these AI methods to pull this data together to process it,” said Kim Branson, SVP, global head of AI and ML. “What’s happening in drug discovery is that AI methods can be used everywhere.”

Like most of biopharma, GSK has long relied heavily upon modern IT but only recently took a deep plunge into AI.

“Our AI team was established in 2019, and is currently distributed between San Francisco and London, our two main hubs, and Boston, Heidelberg, Philadelphia, and Tel Aviv. We [have] about 100 team members which means there’s a lot of people wanting a lot of CPU cycles. The unique nature of the problems we have, and the data sources and the size of the data sources led us to explore the world of custom deep learning silicon,” said Branson.

The rapid advance of functional genomics and the ability to sequence individual single cells, measuring the RNA and DNA content on those cells, is generating data at a massive scale. Moreover, the range of informatics being deployed – from varied statistical analysis to first-principle-like simulation – can all benefit from AI.

“We actually make predictions, we design experiments, we generate data, just purely to build a model and feed that back again. That’s our that’s a core tenet. We routinely feature [large and diverse] datasets such as medical images or pathology images,” said Branson noting Cerebras’s high data movement bandwidth was important.

Ease of use was also an issue. “When GSK was looking for novel computational platforms to really allow us to exploit the data we’re gathering, we did a survey of the custom computing landscape and Cerebras stood out as a company that actually had silicon you could use but more importantly also had a technological stack. That was also a key. They know how to take things in PyTorch and TensorFlow and actually deploy them,” Branson said.

ANL Gets Substantial Speedups with of CS-1

Argonne is one of the longest-term Cerebras collaborators. More generally, ANL is very active in AI research for the Department of Energy and has an active AI testbed on which ANL hopes to test as many different AI chips as it can. That said, ANL is actively using the CSI-1 on an array of ongoing projects, including recent COVID-19 research.

Rick Stevens

“We have over 100 AI projects at the lab,” said Rick Stevens, associate laboratory director for computing, environment and life sciences. “We’ve picked a couple of them to be our primary targets for working on the CS-1.” He briefly summarized four:

  • LIGO. “This is a project that’s working on gravity wave detection [and] is building models that are suited for AI processing, the data coming off of things like LIGO to look for subtle signals of gravity waves. The project has been ongoing for about a year and is making good progress and achieving good speed ups on the CS-1,” said Stevens.
  • Cancer Moonshot. “Probably our flagship project is cancer drug response prediction. This is a joint project between Argonne, the National Cancer Institute and collaborator institutions. We’ve been working on a series of models that run on the CS-1 predicting tumor response to drugs and these models are achieving speed ups of many hundreds of times on the CS-1 compared to our GPU baselines. We’re quite happy with these. These are models that are running on large scale machines across the different architectures,” Stevens said.
  • COVID-19. “We’re also doing generative molecular design. This program is really aimed at generating new molecules that could be used as drugs. We’ve been applying this to our work in COVID-19. It’s also related to our cancer work in other areas where we’re trying to rapidly search through large spaces of drug molecules to find [candidate] molecules that meet certain criteria. Again, we’re getting great speedups there.”
  • Dark Matter. “Finally, we have an active program in dark matter search. This is part of a large software activity, getting ready for the Large Synoptic Survey Telescope (LSST). With this application we’re trying to look for the signatures of halos around galaxies, the mass distributions that are indicative of dark matter halos. Gravitational lensing events are a classic example and we’re doing this both in simulated data and in observation,” he said.

One of the more interesting efforts by ANL is an early program to examine how easy it is (or isn’t) to use the CS-1.

“Most of our models are in TensorFlow, but some of them are coming in PyTorch. We feel pretty confident about the ease-of-use. This fall, [ANL director of data science and learning division] Ian Foster and I are opening up our CS-1 to about 35 graduate students who are coming in with a little bit of experience, but are going to be kind of fresh users. We’re going to use that [experience] to really kind of understand how easy it is for users to get on this machine and to build new models and to train.

“Now it’s still early in terms of the state of the software. There’s still layer types and features that we need and that we are collaborating on but the CS-1 is not that hard to use. If your teams are comfortable developing in TensorFlow or PyTorch, they’ll be right at home using this machine.”

ANL has more plans to aggressively use its CS-1. “Going forward, we have a large-scale program in AI for science at the lab and a rather novel strategy that we’re pursuing that will involve the CS-1 integrating advanced AI hardware with our laboratory robotic system [for] near-autonomous discovery in biology, chemistry, and materials science. So that’s what we’re looking forward to in the future. And we’re just extremely pleased with our collaboration,” said Stevens.

[i] CRETIN is a 1D, 2D, and 3D non-local thermodynamic equilibrium (NLTE) atomic kinetics/radiation transport code which follows the time evolution of atomic populations and photon distributions as radiation interacts with a plasma. It can provide detailed spectra for comparing with experimental diagnostics. https://wci.llnl.gov/simulation/computer-codes

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Illinois Considers $20 Billion Quantum Manhattan Project Says Report

May 7, 2024

There are multiple reports that Illinois governor Jay Robert Pritzker is considering a $20 billion Quantum Manhattan-like project for the Chicago area. According to the reports, photonics quantum computer developer PsiQu Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of what it is like to orbit and enter a black hole. And yes, it c Read more…

2024 Winter Classic: Meet the Mentors Round-up

May 6, 2024

To make navigating easier, we have compiled a collection of all the mentor interviews and placed them in this single page round-up. Meet the HPE Mentors The latest installment of the 2024 Winter Classic Studio Update S Read more…

2024 Winter Classic: The Complete Team Round-up

May 6, 2024

To make navigating easier, we have compiled a collection of all the teams and placed them in this single page round-up. Meet Team Lobo This is the other team from University of New Mexico, since there are two, right? T Read more…

How Nvidia Could Use $700M Run.ai Acquisition for AI Consumption

May 6, 2024

Nvidia is touching $2 trillion in market cap purely on the brute force of its GPU sales, and there's room for the company to grow with software. The company hopes to fill a big software gap with an agreement to acquire R Read more…

2024 Winter Classic: Oak Ridge Score Reveal

May 5, 2024

It’s time to reveal the results from the Oak Ridge competition module, well, it’s actually well past time. My day job and travel schedule have put me way behind, but I am dedicated to getting all this great content o Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

How Nvidia Could Use $700M Run.ai Acquisition for AI Consumption

May 6, 2024

Nvidia is touching $2 trillion in market cap purely on the brute force of its GPU sales, and there's room for the company to grow with software. The company hop Read more…

Hyperion To Provide a Peek at Storage, File System Usage with Global Site Survey

May 3, 2024

Curious how the market for distributed file systems, interconnects, and high-end storage is playing out in 2024? Then you might be interested in the market anal Read more…

Qubit Watch: Intel Process, IBM’s Heron, APS March Meeting, PsiQuantum Platform, QED-C on Logistics, FS Comparison

May 1, 2024

Intel has long argued that leveraging its semiconductor manufacturing prowess and use of quantum dot qubits will help Intel emerge as a leader in the race to de Read more…

Stanford HAI AI Index Report: Science and Medicine

April 29, 2024

While AI tools are incredibly useful in a variety of industries, they truly shine when applied to solving problems in scientific and medical discovery. Research Read more…

IBM Delivers Qiskit 1.0 and Best Practices for Transitioning to It

April 29, 2024

After spending much of its December Quantum Summit discussing forthcoming quantum software development kit Qiskit 1.0 — the first full version — IBM quietly Read more…

Shutterstock 1748437547

Edge-to-Cloud: Exploring an HPC Expedition in Self-Driving Learning

April 25, 2024

The journey begins as Kate Keahey's wandering path unfolds, leading to improbable events. Keahey, Senior Scientist at Argonne National Laboratory and the Uni Read more…

Quantum Internet: Tsinghua Researchers’ New Memory Framework could be Game-Changer

April 25, 2024

Researchers from the Center for Quantum Information (CQI), Tsinghua University, Beijing, have reported successful development and testing of a new programmable Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Leading Solution Providers

Contributors

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire