Exascale Computing to Help Accelerate Drive for Clean Fusion Energy

By Jon Bashor, Lawrence Berkeley National Laboratory Computing Sciences

October 2, 2017

Editor’s note: One of the U.S. Exascale Computing Project’s mandates is to explain how exascale computing power will enhance scientific discovery and society broadly. This article from ECP not only examines the need for exascale computing power to advance research on fusion reactor design but it also highlights the potential for collaboration with industry partners who will require this kind of power.

For decades, scientists have struggled to create a clean, unlimited energy source here on Earth by recreating the conditions that drive our sun. Called a fusion reactor, the mechanism would use powerful magnetic fields to confine and compress gases four times as hot as our sun. By using the magnetic fields to squeeze the gases, the atoms would fuse and release more energy than was used to power the reactor. But to date, that has only worked in theory.

Achieving fusion energy production would benefit society by providing a power source that is non-polluting, renewable and using fuels such as the hydrogen isotopes found in seawater and boron isotopes found in minerals.

Early fusion research projects in the 1950s and ‘60s relied on building expensive magnetic devices, testing them and then building new ones and repeating the cycle. In the mid-1970s, fusion scientists began using powerful computers to simulate how the hot gases, called plasmas, would be heated, squeezed and fused to produce energy. It’s an extremely complex and difficult problem, one that some fusion researchers have likened to holding gelatin together with rubber bands.

Using supercomputers to model and simulate plasma behavior, scientists have made great strides toward building a working reactor. The next generation of supercomputers on the horizon, known as exascale systems, will bring the promise of fusion energy closer.

The best-known fusion reactor design is called a tokamak, in which a donut-shaped chamber is used to contain the hot gases, inside. Because the reactors are so expensive, only small-scale ones have been built. ITER, an international effort to build the largest-ever tokamak-in the south of France. The project, conceived in 1985, is now scheduled to have its first plasma experiments in 2025 and begin fusion experiments in 2035. The estimated cost is 14 billion euros, with the European Union and six other nations footing the bill.

Historically, fusion research around the world has been funded by governments due to the high cost and long-range nature of the work.

But in the Orange County foothills of Southern California, a private company is also pursuing fusion energy, but taking a far different path than that of ITER and other tokamaks. Tri Alpha Energy’s cylindrical reactor design is completely different in its design philosophy, geometry, fuels and method of heating the plasma, all built with a different funding model. Chief Science Officer Toshiki Tajima says their approach makes them mavericks in the fusion community.

But the one thing both ITER and similar projects and Tri Alpha Energy have consistently relied on is using high-performance computers to simulate conditions inside the reactor as they seek to overcome the challenges inherent in designing, building and operating a machine that will replicate the processes of the sun here on Earth.

As each generation of supercomputers has come online, fusion scientists have been able to study plasma conditions in greater detail, helping them understand how the plasma will behave, how it may lose energy and disrupt the reactions, and what can be done to create and maintain fusion. With exascale supercomputers that are 50 times more powerful than today’s top systems looming on the horizon, Tri Alpha Energy sees great possibilities in accelerating the development of their reactor design. Tajima is one of 18 members of the industry advisory council for the U.S. Department of Energy’s (DOE) Exascale Computing Project (ECP).

“We’re very excited by the promise of exascale computing – we are currently fund-raising for our next-generation machine, but we can build a simulated reactor using a very powerful computer, and for this we would certainly need exascale,” Tajima said. “This would help us accurately predict if our idea would work, and if it works as predicted, our investors would be encouraged to support construction of the real thing.”

The Tri Alpha Energy fusion model builds on the experience and expertise of Tajima and his longtime mentor, the late Norman Rostoker, a professor of physics at the University of California, Irvine (UCI). Tajima first met Rostoker as a graduate student, leaving Japan to study at Irvine in 1973. In addition to his work with TAE, Tajima holds the Norman Rostoker Chair in Applied Physics at UCI. In 1998, Rostoker co-founded TAE, which Tajima joined in 2011.

In it for the long run

It was also in the mid-1970s, that the U.S. Atomic Energy Commission, the forerunner of DOE, created a computing center to support magnetic fusion energy research, first with a cast-off computer from classified defense programs, but then with a series of ever-more capable supercomputers. From the outset, Tajima was an active user, and still remembers he was User No. 1100 at the Magnetic Fusion Energy Computer Center. The Control Data Corp. and Cray supercomputers were a big leap ahead of the IBM 360 he had been using.

“The behavior of plasma could not easily be predicted with computation back then and it was very hard to make any progress,” Tajima said. “I was one of the very early birds to foul up the machines. When the Cray-1 arrived, it was marvelous and I fell in love with it.”

At the time, the tokamak was seen as the hot design and most people in the field gravitated in this direction, Tajima said, and he followed. But after learning about plasma-driven accelerators under Professor Rostoker, in 1976 he went to UCLA to work with Prof. John Dawson. “He and I shared a vision of new accelerators and we began using large-scale computation in 1975, an area in which I wanted to learn more from him,” Tajima said.

As a result, the two men wrote a paper entitled “Laser Electron Accelerator,” which appeared in Physical Review Letters in 1979. The seminal paper explained how firing an intense electromagnetic pulse (or beam of particles) into a plasma can create a wake in the plasma and that electrons, and perhaps ions, trapped in this wake can be accelerated to very high energies.

TAE’s philosophy, built on Rostoker’s ideas, is to combine both accelerator and fusion plasma research. In a tokamak, the deuterium-tritium fuel needs to be heated and confined at an energy level of 10,000 eV (electron volts) for fusion to occur. The TAE reactor, however, needs to be 30 times hotter. In a tokamak, the same magnetic fields that confine the plasma also heat it to 3 billion degrees C. In the TAE machine, the energy will be injected using a particle accelerator. “A 100,000 eV beam is nothing for an accelerator,” Tajima said, pointing to the 1G eV BELLA device at DOE’s Lawrence Berkeley National Laboratory. “Using a beam-driven plasma is relatively easy but it may be counterintuitive that you can get higher energy with more stability — the more energetic the wake is, the more stable it becomes.”

But this approach is not without risk. With the tokamak, the magnetic fields protect the plasma, much like the exoskeleton of a beetle protects the insect’s innards, Tajima said. But the accelerator beam creates a kind of spine, which creates the plasma by its weak magnetic fields, a condition known as Reverse Field Configuration. One of Rostoker’s concerns was that the plasma would be too vulnerable to other forces in the early stages of its formation. However, in the 40-centimeter diameter cylindrical reactor, the beam forms a ring like a bicycle tire, and like a bicycle, the stability increases the faster the wheels spin.

“The stronger the beam is, the more stable the plasma becomes,” Tajima said. “This was the riskiest problem for us to solve, but in early 2000 we showed the plasma could survive and this reassured our investors. We call this approach of tackling the hardest problem first ‘fail fast’.”

Another advantage of TAE’s approach is that the main fuel, Boron-11, does not produce neutrons as a by-product; instead it produces three alpha particles, which is the basis of the company’s name. A tokamak, using hydrogen-isotope fuels, generates neutrons, which can penetrate and damage materials, including the superconducting magnets that confine the tokamak plasma. To prevent this, the tokamak reactor requires one-meter-thick shielding. Without the need to contain neutrons, the TAE reactor does not need heavy shielding. This also helps reduce construction costs.

Computation Critical to Future Progress

With his 40 years of experience using HPC to advance fusion energy, Tajima offers a long-term perspective, from the past decades to exascale systems in the early 2020s. As a principal investigator on the Numerical Tokamak project in the early 1990s, he has helped build much of the HPC ecosystem for fusion research.

At the early stage of modeling fusion behavior, the codes focus on the global plasma at very fast time scales. These codes, known as MHD codes (magnetohydrodynamics), are not as computationally “expensive,” meaning they do not require as many computing resources, and at TAE were run on in-house clusters.

The next step is to model the more minute part of the plasma instability, known as kinetic instability, which requires more sophisticated codes that can simulate the plasma in greater detail over longer time scales. Achieving this requires more sophisticated systems. Around 2008-09, TAE stabilized this stage of the problem using its own computing system and by working with university collaborators who have access to federally funded supercomputing centers, such as those supported by DOE. “Our computing became more demanding during this time,” Tajima said.

The third step, which TAE is now tackling, is to make a plasma that can “live” longer, which is known as the transport issue in the fusion community. “This is a very, very difficult problem and consumes large amounts of computing resources as it encompasses a different element of the plasma,” Tajima said, “and the plasma becomes much more complex.”

The problem involves three distinct functions:

  • The core of the field reverse configuration, which is where the plasma is at the highest temperature
  • The “scrape-off layer,” which is the protective outer layer of ash on the core and which Tajima likens to an onion’s skin
  • The “ash cans,” or diverters, that are at each end of the reactor. They remove the ash, or impurities, from the scrape-off layer, which can make the plasma muddy and cause it to behave improperly.

“The problem is that the three elements behave very, very differently in both the plasma physics as well as in other properties,” Tajima said. “For example, the diverters are facing the metallic walls so you have to understand the interaction of the cold plate metals and the out-rushing impurities. And those dynamics are totally different than the core which is very high temperature and very high energy and spinning around like a bicycle tire, and the scrape-off layer.”

These factors are all coupled to each other using very complex geometries and in order to see if the TAE approach is feasible, researchers need to simulate the entirety of the reactor in order to understand and eventually control the reactions.

“We will run a three-layered simulation of our fusion reactor on the computer, with the huge particle code, the transport code and the neural net on the simulation – that’s our vision and we will certainly need an exascale machine to do this,” Tajima said. “This will allow us to predict if our concept works or not in advance of building machine so that our investors’ funds are not wasted.”

The overall code will have three components. At the basic level will be a representative simulation of particles in each part of the plasma. The second layer will be the more abstract transport code, which tracks heat moving in and out of the plasma. But even on exascale systems, the transport code will not be able to run fast enough to keep up with real-time changes in the plasma. Instabilities which affect the heat transport in the plasma come and go in milliseconds.

“So, we need a third layer that will be an artificial neural net, which will be able to react in microseconds, which is a bit similar to a driverless auto, and will ‘learn’ how to control the bicycle tire-shaped plasma, Tajima said. This application will be run on top of transport code and it will observe experimental data and react appropriately to keep the simulation running.

“Doing this will certainly require exascale computing,” Tajima said. “Without it we will take up to 30 years to finish – and our investors cannot wait that long. This project has been independent of the government funding, so that our investors’ fund provided an independent, totally different path toward fusion. This could amount to a means of national security to provide an alternative solution to a problem as large as fusion energy. Society will also benefit from a clean source of energy and our exascale-driven reactor march will be a very good thing for the nation and the world.”

Advanced Accelerators are Pivotal

Both particle accelerators and fusion energy are technologies important to the nation’s scientific leadership, with research funded over many decades by the Department of Energy and its predecessor agencies.

Not only are particle accelerators a vital part of the DOE-supported infrastructure of discovery science and university research, they also have private-sector applications and a broad range of benefits to industry, security, energy, the environment and medicine.

Since Toshiki Tajima and John Dawson published their paper “Laser Electron Accelerator” in 1979, the idea of building smaller accelerators, with the length measure in meters instead of kilometers, has gained traction. In these new accelerators, particles “surf” in the plasma wake of injected particles, reaching very high energy levels in very short distances.

According to Jean-Luc Vay, a researcher at DOE’s Lawrence Berkeley National Laboratory, taking full advantage of accelerators’ societal benefits, game-changing improvements in the size and cost of accelerators are needed. Plasma-based particle accelerators stand apart in their potential for these improvements, according to Vay, and turning this from a promising technology into a mainstream scientific tool depends critically on high-performance, high-fidelity modeling of complex processes that develop over a wide range of space and time scales.

To help achieve this goal, Vay is leading a project called Exascale Modeling of Advanced Particle Accelerators as part of DOE’s Exascale Computing Project. This project supports the practical economic design of smaller, less-expensive plasma-based accelerators.

As Tri Alpha Energy pursues its goal of using a particle accelerator (though this accelerator is not related to wakefield accelerators) to achieve fusion energy, the company is also planning to apply its experience and expertise in accelerator research for medical applications. Not only will this effort produce returns for the company’s investors, but it should also help advance TAE’s understanding of accelerators and using them to create a fusion reactor.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

At SC18: AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

Dell EMC’s HPC Chief on Strategy and Emerging Processor Diversity

November 16, 2018

Last January Thierry Pellegrino, a long-time Dell/Dell EMC veteran, became vice president of HPC. His tenure comes at a time when the very definition of HPC is blurring with AI writ large (data analytics, machine learnin Read more…

By John Russell

IBM’s AI-HPC Combine for ‘Intelligent Simulation’: Eliminating the Unnecessary 

November 16, 2018

A powerhouse concept in attaining new knowledge is the notion of the “emergent property,” the combination of formerly stovepiped scientific disciplines and exploratory methods to form cross-disciplinary intelligence Read more…

By Doug Black

HPE Extreme Performance Solutions

AI Can Be Scary. But Choosing the Wrong Partners Can Be Mortifying!

As you continue to dive deeper into AI, you will discover it is more than just deep learning. AI is an extremely complex set of machine learning, deep learning, reinforcement, and analytics algorithms with varying compute, storage, memory, and communications needs. Read more…

IBM Accelerated Insights

From Deep Blue to Summit – 30 Years of Supercomputing Innovation

This week, in honor of the 30th anniversary of the SC conference, we are highlighting some of the most significant IBM contributions to supercomputing over the past 30 years. Read more…

How the United States Invests in Supercomputing

November 14, 2018

The CORAL supercomputers Summit and Sierra are now the world's fastest computers and are already contributing to science with early applications. Ahead of SC18, Maciej Chojnowski with ICM at the University of Warsaw discussed the details of the CORAL project with Dr. Dimitri Kusnezov from the U.S. Department of Energy. Read more…

By Maciej Chojnowski

At SC18: AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

Dell EMC’s HPC Chief on Strategy and Emerging Processor Diversity

November 16, 2018

Last January Thierry Pellegrino, a long-time Dell/Dell EMC veteran, became vice president of HPC. His tenure comes at a time when the very definition of HPC is Read more…

By John Russell

IBM’s AI-HPC Combine for ‘Intelligent Simulation’: Eliminating the Unnecessary 

November 16, 2018

A powerhouse concept in attaining new knowledge is the notion of the “emergent property,” the combination of formerly stovepiped scientific disciplines and Read more…

By Doug Black

How the United States Invests in Supercomputing

November 14, 2018

The CORAL supercomputers Summit and Sierra are now the world's fastest computers and are already contributing to science with early applications. Ahead of SC18, Maciej Chojnowski with ICM at the University of Warsaw discussed the details of the CORAL project with Dr. Dimitri Kusnezov from the U.S. Department of Energy. Read more…

By Maciej Chojnowski

At SC18: Humanitarianism Amid Boom Times for HPC

November 14, 2018

At SC18 in Dallas, the feeling on the ground is one of forward-looking buoyancy. Like boom times that cycle through the Texas oil fields, the HPC industry is en Read more…

By Doug Black

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

New Panasas High Performance Storage Straddles Commercial-Traditional HPC

November 13, 2018

High performance storage vendor Panasas has launched a new version of its ActiveStor product line this morning featuring what the company said is the industry’s first plug-and-play, portable parallel file system that delivers up to 75 Gb/s per rack on industry standard hardware combined with “enterprise-grade reliability and manageability.” Read more…

By Doug Black

SC18 Student Cluster Competition – Revealing the Field

November 13, 2018

It’s November again and we’re almost ready for the kick-off of one of the greatest computer sports events in the world – the SC Student Cluster Competitio Read more…

By Dan Olds

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

TACC Wins Next NSF-funded Major Supercomputer

July 30, 2018

The Texas Advanced Computing Center (TACC) has won the next NSF-funded big supercomputer beating out rivals including the National Center for Supercomputing Ap Read more…

By John Russell

IBM at Hot Chips: What’s Next for Power

August 23, 2018

With processor, memory and networking technologies all racing to fill in for an ailing Moore’s law, the era of the heterogeneous datacenter is well underway, Read more…

By Tiffany Trader

Requiem for a Phi: Knights Landing Discontinued

July 25, 2018

On Monday, Intel made public its end of life strategy for the Knights Landing "KNL" Phi product set. The announcement makes official what has already been wide Read more…

By Tiffany Trader

House Passes $1.275B National Quantum Initiative

September 17, 2018

Last Thursday the U.S. House of Representatives passed the National Quantum Initiative Act (NQIA) intended to accelerate quantum computing research and developm Read more…

By John Russell

CERN Project Sees Orders-of-Magnitude Speedup with AI Approach

August 14, 2018

An award-winning effort at CERN has demonstrated potential to significantly change how the physics based modeling and simulation communities view machine learni Read more…

By Rob Farber

Summit Supercomputer is Already Making its Mark on Science

September 20, 2018

Summit, now the fastest supercomputer in the world, is quickly making its mark in science – five of the six finalists just announced for the prestigious 2018 Read more…

By John Russell

New Deep Learning Algorithm Solves Rubik’s Cube

July 25, 2018

Solving (and attempting to solve) Rubik’s Cube has delighted millions of puzzle lovers since 1974 when the cube was invented by Hungarian sculptor and archite Read more…

By John Russell

Leading Solution Providers

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

TACC’s ‘Frontera’ Supercomputer Expands Horizon for Extreme-Scale Science

August 29, 2018

The National Science Foundation and the Texas Advanced Computing Center announced today that a new system, called Frontera, will overtake Stampede 2 as the fast Read more…

By Tiffany Trader

At SC18: AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

HPE No. 1, IBM Surges, in ‘Bucking Bronco’ High Performance Server Market

September 27, 2018

Riding healthy U.S. and global economies, strong demand for AI-capable hardware and other tailwind trends, the high performance computing server market jumped 28 percent in the second quarter 2018 to $3.7 billion, up from $2.9 billion for the same period last year, according to industry analyst firm Hyperion Research. Read more…

By Doug Black

Intel Announces Cooper Lake, Advances AI Strategy

August 9, 2018

Intel's chief datacenter exec Navin Shenoy kicked off the company's Data-Centric Innovation Summit Wednesday, the day-long program devoted to Intel's datacenter Read more…

By Tiffany Trader

Germany Celebrates Launch of Two Fastest Supercomputers

September 26, 2018

The new high-performance computer SuperMUC-NG at the Leibniz Supercomputing Center (LRZ) in Garching is the fastest computer in Germany and one of the fastest i Read more…

By Tiffany Trader

Houston to Field Massive, ‘Geophysically Configured’ Cloud Supercomputer

October 11, 2018

Based on some news stories out today, one might get the impression that the next system to crack number one on the Top500 would be an industrial oil and gas mon Read more…

By Tiffany Trader

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This