Lighting a Fire Under Combustion Simulation

By Michael Feldman

June 1, 2011

Combustion simulation might seem like the ultimate in esoteric technologies, but auto companies, aircraft firms and fuel designers need increasingly sophisticated software to serve the needs of 21st century engine designs. And as government and industry demand better fuel efficiency and cleaner emissions, combustion software has become a hot topic.

HPCwire recently got the opportunity to take a look at Reaction Design, one of the premier makers of combustion simulation software, and talk with its CEO, Bernie Rosenthal.

A privately-funded company based in San Diego, Reaction Design has been around since 1995 and today employs about 30 people. Its claim to fame is providing state-of-the-art combustion simulation by bringing high-fidelity computational chemistry into the realm of computational fluid dynamics (CFD).

As with many types of computer aided engineering, the idea behind combustion simulation is to allow companies to replace millions of dollars in physical mockups and experiments with software models. According to Rosenthal, that requires sophisticated algorithms plus an intimate understanding of the different types of engines and fuels used by industry. The goal is to predict the thermodynamic behavior of combustion as well as the undesirable byproducts — carbon soot, nitrous oxide compounds (NOx), unburnt hydrocarbons, and carbon monoxide.

As far as the chemistry goes, Reaction Design developed its core competency early on. A couple of years after it was founded, the company acquired exclusive rights to a chemistry kinetics solver from Sandia National Laboratories, which the lab had developed to simulate rocket plumes and missile reactor designs. In conjunction with the software, Reaction Design picked up some of the key Sandia developers and brought them on-board. The solver was subsequently productized into CHEMKIN, the package that forms the basis of most of Reaction Design’s software offerings.

What sets Reaction Design apart is their ability to combine their computational chemistry codes with third-party CFD packages for combustion simulation. According to Rosenthal, they are the only vendor that combines the two components with a level of detail that he refers to as “real fuel modeling.”

Engine fuels are not simple formulations. Even refined gasoline is made up of thousands of different molecules that interact with each other during the combustion process. “Over the last 15 years that CFD has been available, most simulations have been approximating that fuel as one molecule,” notes Rosenthal. In general, he says, those simple simulations have worked. At least they did a good enough job to provide a 90 percent reduction in undesirable byproducts.

But emission standards now mandate combustion byproducts in the parts-per-million or even parts-per-billion range, which can be an expensive proposition for the end user. For example, a $65 thousand diesel engine could require a $15,000 after-treatment system just to deal with undesirable tailpipe emissions. In addition, the need for better fuel efficiency as well as the changing nature of the fuel itself — which can be anything from standard gasoline, to diesel, ethanol, liquified natural gas, biodiesel, propane, rapeseed oil, or some combination thereof — necessitates a more complex engine design.

To model all this in software requires a lot of number crunching. That’s especially true in the realm of turbine engines, which have very large geometries. Even with the benefit of parallelization on a medium-sized cluster with a dozen or two CPUs, a CFD simulation using a multi-million cell computational mesh requires multiple days of run-time execution — and that’s without any complex chemistry involved. “These guys were taking around a week to get an answer for a single cycle of the combustor,” says Rosenthal.

Reaction Design’s initial approach for the turbine engine community was to hook their existing chemistry solver onto CFD codes like FLUENT, STAR-CD and CFX, which are the ones most commonly used by manufacturers. Essentially they mapped the chemistry kinetics onto the CFD by splitting the combustion into a number of distinct regions, applying the chemistry to the CFD output, and then aggregating the results. The ensuing product, ENERGICO, is now used by a number of turbine firms, including the world’s largest gas turbine manufacturer for aircraft and power generation.

The problem was that this approach didn’t really offer true CFD-chemistry integration, and that is what automobile companies and other internal combustion engine manufacturers were demanding. This sector has traditionally looked to HPC to reduce compute times and increase capability by running their simulations on ever-larger and more powerful clusters. In general, simulation times for a combustion cycle were in the 8 to 12 hour range, meaning designers could initiate on an overnight run and analyze the results in the morning.

But newer engine designs, more complex injection and pressure schemes, and stricter emission requirements meant the simulations would have to do a lot more computation. Not only did the manufacturers want real fuel chemistry to be a part of this, they also wanted to keep their half-day simulation times.

Unfortunately, such chemistry is quite compute-intensive. According to Rosenthal, 80 to 90 percent of the run-time was going to be spent in the chemistry computation if they used their existing algorithms. So the Reaction Design developers took a second look at their software and were able to squeeze a 10-fold improvement in the algorithmic performance.

But even that wasn’t enough to keep the simulation run-times in the overnight realm. To accomplish that, they needed to parallelize their algorithms, which they did in typical MPI fashion. By doing so, users could scale the chemistry computation linearly just by adding more compute nodes, at least for moderate-sized clusters.

The second part to the solution was to merge the chemistry and CFD codes. There was just one problem: Reaction Design had no in-house CFD code, so they had to develop their own. The company now offers this as a standalone product called CHEMKIN CFD.

But the chemistry-integrated version, called FORTE, was the real breakthrough. It’s a complete HPC solution that supports advanced, 3D internal combustion engine design with real fuel chemistry hooked into a CFD solver. FORTE was officially announced in April, and a number of large auto firms in the US, Europe and Japan have already signed on, says Rosenthal.

FORTE may well scale up into hundreds of nodes, which would put simulation run-times into the 60-minute realm. But most manufacturers would probably use such large clusters to run multiple simulations using different parameters, rather than opt for shorter turnarounds on a single design run. The company’s next step is to see if they can scale their integrated CFD-chemistry approach to the larger geometries of the turbine engine, and offer a FORTE-like product for that industry.

Beyond that, Rosenthal is looking at GPU computing to further accelerate their codes. At this point, he’s wondering if he should invest development cycles in CUDA or OpenCL technology or wait for higher-level development tools to offer a more transparent way to tap into GPUs. Like most developers, he would rather the compiler and runtime do the heavy lifting in order to simplify any GPU-specific source code changes on his part. “But the real question to me is: what do my customers have?” says Rosenthal. “And they don’t have these… yet.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire