Supercomputing’s Critical Role in the Fusion Ignition Breakthrough

By Oliver Peckham

December 21, 2022

On December 5th, the research team at the National Ignition Facility (NIF) at Lawrence Livermore National Laboratory (LLNL) achieved a historic win in energy science: for the first time ever, more energy was produced by an artificial fusion reaction than was consumed – 3.15 megajoules (MJ) produced versus 2.05 megajoules in laser energy to cause the reaction. High-performance computing was key to this breakthrough (called “ignition”), and HPCwire recently had the chance to speak with Brian Spears, deputy lead modeler for inertial confinement fusion (ICF) at the NIF, about HPC’s role in making this fusion ignition a reality.


The Basics

Fusion energy has long been the Holy Grail of energy science, promising enormous amounts of zero-carbon, low-radioactivity nuclear energy – but, sadly, the precise and extreme conditions required to generate fusion reactions have, until now, proved prohibitive to producing net energy gains in a fusion reactor.

At the NIF – which supports the National Nuclear Security Administration’s Stockpile Stewardship Program – researchers are working on the laser indirect drive approach to inertial confinement fusion. Spears explained the basics: first, laser beams are fired into a small gold can called a hohlraum (German for “empty space”). In the hohlraum, the lasers create a bath of x-rays, which cooks a target capsule made of ultrananocrystalline diamond. Inside that capsule: frozen deuterium and tritium. When the x-rays heat up the capsule enough, it explodes, shoving the nuclear fuel inside to extremely high temperatures and densities, which causes fusion to occur and produces substantial energy. The target chamber for the NIF is pictured in the header.

A render of lasers hitting the target. Image courtesy of LLNL.

The processes involved in performing these tests are mind-boggling. “We have something called cross-beam energy transfer that is literally what happens in Ghostbusters,” Spears said (referring to how, in the movie Ghostbusters, the titular Ghostbusters are warned of the dangers of crossing the streams of their proton packs). “We have two cones of lasers, some that come in closer to the top of the capsule and some further away, and when those two beams cross, laser power is actually transferred from one set of beams to the other, and we use that to control the asymmetry of the implosion so we don’t push one too hard on one side.”

The target is also the subject of exacting precision. “The outside of that capsule – the ablator – that explodes is polished to something like a hundred times smoother than a mirror, but it still has more defects than we’d like,” Spears said, explaining that these defects allow carbon to mix with the nuclear fuel, cooling the reaction. “This shot actually had a headwind of a capsule which was a little bit defective in our eyes. It’s still one of the most highly polished pieces of object that humans have ever made on the planet, but it wasn’t the best one we could make.”


The Methods

One of the issues that has hindered the development of fusion energy has been the difficulty and expense (and, as a result, the infrequency) of running each individual experiment. “Our laser can fire a couple of times a day,” Spears said. “These very large, demonstration-class experiments can happen once a week to once a month, so they’re pretty sparse in time.” At the NIF, Spears and his team apply HPC and AI to ensure that every time they fire the laser, they’re learning as much as they can from the results and using what they learn to refine subsequent experiments.

Spears said there were essentially two pieces to their predictive capability. First, their fundamental design capability – using hundreds of thousands of lines of radiation hydrodynamics code to run massive simulations of fusion reactions on leadership-class supercomputers. Second: cognitive simulation, or CogSim.

“We recognize that simulated prediction can sometimes be overly optimistic in the fusion world – which is why you’ve heard the line, maybe, that fusion is 30 years away and always will be,” Spears said – and this, he continued, is where CogSim helps out.

“What we call cognitive simulation is really the merging of simulation predictions with experimental reality that we’ve actually observed using AI,” he said. “What CogSim does is give us an assessment tool. Our standard, traditional HPC says, ‘Look, this is the way this implosion is going to work. If you do these things, it looks really nice.’ And then CogSim comes in and says, ‘Well, okay, but in the real world there are all these degradations that your design is going to have to contend with, and this is what it looks like when you actually field something at NIF, give me that new design and I’ll tell you whether it’s robust or not.’” (This isn’t the first time CogSim has hit our radar: ahead of the successful run, HPCwire awarded its 2022 Editors’ Choice Award for Best Use of HPC in Energy to LLNL and the NIF for their use of CogSim in fusion energy modeling.)

“We use almost all of our [HPC] resources across this process,” he said. The traditional design happens mostly on CPU-based systems – LLNL’s Intel-based Jade system, he added, is a workhorse for that segment of the research.

The Jade supercomputer at LLNL. Image courtesy of LLNL.

“When we transition into the cognitive simulation world, we use two supercomputers,” he continued. “One for capacity is the Trinity system at Los Alamos, so we use that quite heavily to do hundreds of thousands of rad-hydro predictions. When we actually do the machine learning piece, we sometimes put that on Sierra and use our largest system to do that. But we are pretty much on every machine that’s in our secure computing facility, both for capacity reasons and for the capabilities that the different machines bring.”

LLNL also has a Cerebras CS-2 system hooked up to its Lassen supercomputer, but Spears said that, while the CS-2 was used for research and development work and will be important for the next generation of research, it was not used principally in this work.

“We will use that kind of thinking developed [with the CS-2] for what we call Ice Cap, which is a project that we have built for when we take delivery of [the exascale supercomputer] El Capitan … in the next year,” Spears said. “So some of the techniques that are developed there are helping us understand how to move back and forth between high-precision workloads and low-precision machine learning workloads.”


The Shot

On September 19th, the team began putting more energy through the laser – something the computational work had indicated would help the experiment surmount the “ignition cliff” and reach the fabled “ignition plateau.” With that extra energy – the same 2.05 MJ put in during the December shot – the September shot was more successful, producing over 1 MJ of yield. However, it was hindered by an implosion that was “a little bit pancake-shaped,” so Spears and his colleagues asked the code: what would happen if they made it round by changing the way the power was distributed among the laser beams?

“The code said, ‘If you do that, you’re going to get two to five times more energy out than you did this time,’” Spears said. “And then we went through that CogSim loop of saying, ‘In the face of all the degradations, is that really true?’ And it looked like the answer was yes, and so we went pedal to the metal and did this.”

“We had gone through for a couple months with these traditional design tools and the answer looked pretty good. But I’ve been doing this for 18 years, and the answer has looked pretty good many times. With the CogSim tools, this was the first time that – by around Thanksgiving – we were combining the experiment with the simulation and we were getting things that looked qualitatively different from anything that we had seen before. And by the week before the shot, we had a very solid assessment that said more likely than not, this is going to exceed the amount of energy that goes into the target by laser. And that’s the first time we’d ever seen that number. Sometimes, in the previous shots before that, that number may have been 15% likely that that could happen … sometimes it was way, way, less than 15%. This time, it was over 50%.”

“We essentially, from the prediction side, called that shot, and said: hey, look, based on HPC, combined with experiments and a little AI magic thrown in, this looks really good,” he said. “I hit send on that email to our senior management team with a little bit of trepidation, because after 18 years of that not happening, statistically speaking, it’s not all that likely to happen.”

But it did.


What’s Next for the NIF

“There are insights that have come out of this,” Spears said. “We now understand very clearly the role that asymmetries play in our implosions. One of the principal things that we did was to use that ‘Ghostbuster effect’ – the cross-beam energy transfer – to make these implosions more spherical.” Spears added there would be four or five seminal papers emerging from the event, but cautioned that many of the other takeaways would be opaque to people who weren’t “deep, deep insiders.”

Of course, the path to scalable fusion – and even to radically improved experiments based on the successful shot – remains a long one. “The timescale is not short yet,” Spears said. “When we do an experiment, it takes us a few days in order to actually fully reduce the diagnostics and analyze them and understand what we saw. … Design ideas for that are probably in play within days to a week after we’ve seen the experiment. Then we go into a real production loop – but that loop is quite long, because to make our targets takes months and months and months, so if you look at an experiment and say, ‘Hey, look, I think we oughta thicken up the ablator just a little bit more’ – to build those capsules takes three months, and if you stick on the front of that a careful specification of that target of about a month, it’s hard for us to react much more quickly than six months.”

And, Spears said, don’t count on the successful shot being effortlessly reproduced. “We should point out, for the future – a little greater than 50% odds is still only a little better than a flip of a coin, right? This is not going to happen every time we pull the trigger on the laser, there are going to be things that happen and we’re going to learn from them just as we have for the last 18 years.”

Still, there are plenty of reasons for optimism in the wake of the successful shot.

“The next batch of shots will use targets that we’re pretty satisfied with so far, so we’re actually secretly even a little more optimistic for things in the future,” Spears said. “But it’s fusion, and stuff happens, so you never really know until you see the diagnostics the next morning.”

The researchers also expect the energy gains to scale dramatically with energy input. We expect it to be strongly nonlinear, and it will only get better as we build designs that accommodate the increase in energy,” Spears said. “For some perspective, between the last event and this one, we put in 8% more energy in the laser and we got 230% more energy out in fusion.”

The laser isn’t yet ready to be fired at that high intensity again, but Spears said that the next attempts to conduct large shots will likely be in January and February, then throughout the spring and summer. And, by next October, he expects the shots will be based largely on the lessons learned from the successful shot fired earlier this month.

“You’ll hear some splash again about ICF when we receive [El Capitan] and we essentially use more cycles than have ever been used in ICF to polish or refine one of these designs in a way that only El Cap can do it,” Spears added.

A render of El Capitan.

Spears also shared that the lab has been receiving an “overwhelming” amount of support in the wake of its win. “It’s hard not to be excited about getting up in the morning to go to work and do this,” he said.


The Future of Fusion Energy Research

Spears doesn’t think that the current workflow is how things will always be done, citing a self-driving laser project LLNL is working on with Colorado State University.

“There are laser systems that we’ve developed – both at Livermore and elsewhere – that can fire multiple times per second,” he said. Through a self-driving laser research loop, an experiment would be run and observed, and then a deep neural network surrogate would compare the experimental results to its predictions and intelligently choose the next experiment on the fly every few hundred milliseconds with no designer in the loop and an entire array of targets already available.

“This discovery loop that we’ve been talking about now becomes tightly closed,” Spears said. “These self-driving knowledge discovery loops – that’s the future. That’s the direction that we’re going. … This is making HPC a real-time commodity for experiments, and entirely blurring or erasing the boundary between HPC and experiment.”

“I expect that we’re going to see a petaflop-scale high-performance computer sitting in all of the experimental facilities, helping us run operations like this – but still calling home to enormous, El Capitan-sized computers for very high-precision stuff.”


Beyond Fusion Energy

The team’s methods are also useful for fields outside of fusion energy. For instance, Spears said, because the code they use is predictive enough for ICF, it also is good at understanding extreme ultraviolet lithography. “So it’s been a transformational capability in making semiconductors – which is kind of useful for doing HPC!” he said. “So in a nice virtuous cycle, we’ve made our computational capabilities so good that we can work with industry partners to predict better ways to make the semiconductors that we need to make better computers. It’s far away from ICF, but because it’s really extreme plasma conditions to do those things, we know how to do it and it has a nice benefit for us – and for your audience, too.”

CogSim is also finding uses in the life sciences, where researchers at LLNL applied it to Covid research last winter. “Our team used some of these CogSim [tools] to, over a Christmas sprint in two weeks, completely redesign an antibody for Covid-19,” Spears said. “By January of this year, just three weeks after that Christmas, those molecules were in trials with AstraZeneca.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Edge-to-Cloud: Exploring an HPC Expedition in Self-Driving Learning

April 25, 2024

The journey begins as Kate Keahey's wandering path unfolds, leading to improbable events. Keahey, Senior Scientist at Argonne National Laboratory and the University of Chicago, leads Chameleon. This innovative projec Read more…

Quantum Internet: Tsinghua Researchers’ New Memory Framework could be Game-Changer

April 25, 2024

Researchers from the Center for Quantum Information (CQI), Tsinghua University, Beijing, have reported successful development and testing of a new programmable quantum memory framework. “This work provides a promising Read more…

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Point. The system includes Intel's research chip called Loihi 2, Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Research senior analyst Steve Conway, who closely tracks HPC, AI, Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, and this day of contemplation is meant to provide all of us Read more…

Intel Announces Hala Point – World’s Largest Neuromorphic System for Sustainable AI

April 22, 2024

As we find ourselves on the brink of a technological revolution, the need for efficient and sustainable computing solutions has never been more critical.  A computer system that can mimic the way humans process and s Read more…

Shutterstock 1748437547

Edge-to-Cloud: Exploring an HPC Expedition in Self-Driving Learning

April 25, 2024

The journey begins as Kate Keahey's wandering path unfolds, leading to improbable events. Keahey, Senior Scientist at Argonne National Laboratory and the Uni Read more…

Quantum Internet: Tsinghua Researchers’ New Memory Framework could be Game-Changer

April 25, 2024

Researchers from the Center for Quantum Information (CQI), Tsinghua University, Beijing, have reported successful development and testing of a new programmable Read more…

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Poin Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Resear Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Leading Solution Providers

Contributors

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire