Exawind Project Demonstrates Blade-Resolved Simulation of NREL 5 MW Reference Wind Turbine

October 25, 2018

Oct. 25, 2018 — In 2017 wind generated 6.3% of the United States’s electricity, according to the US Energy Information Administration. If the nation can use its abundant wind resources to generate 30% of its electric power, the societal and economic impact will be profound. US energy security will be reinforced by the greater diversity in the energy supply. Cost-competitive electricity could be provided to key regions of the country, greenhouse-gas emissions reduced, and the quantity of water required for thermo-electric power generation lessened.

Geometry-resolved large-eddy simulation of the NREL 5 MW reference wind turbine, which shows velocity isosurfaces at 5.5 m/s after three rotor revolutions. The simulation was performed on the NERSC Cori system with Nalu-Wind, an open-source unstructured grid, low-Mach-number computational fluid dynamics code.

A key challenge for wide-scale deployment of wind power without subsidy is plant-level inefficiencies. Plant-level performance losses can be as high as 20–30% due to complex terrain, unique atmospheric flow phenomena, and the complex flow interactions that occur in large wind farms, which comprise multiple arrays with significant turbine-turbine wake interactions. Addressing the challenge in reducing plant-level losses requires more knowledge of their dynamics to inform optimization of existing plants (through, for example, new control strategies), optimized layout of new plants, and the creation of new wind turbine technology.

The U.S. Department of Energy’s (DOE) Exascale Computing Project (ECP), a joint collaboration of two DOE sponsoring organizations, the Office of Science (DOE-SC) and the National Nuclear Security Administration (NNSA), has funded a research effort named ExaWind under its Application Development focus area.

ECP’s ExaWind project aims to advance the fundamental comprehension of whole wind plant performance by examining wake formation, the impacts of complex terrain, and the effects of turbine-turbine wake interactions. When validated by targeted experiments, the predictive physics-based high-fidelity computational models at the center of the ExaWind project, and the new knowledge derived from their solutions, provide an effective path to optimizing wind plants.

Large-eddy simulation (LES) is a well-known mathematical computational fluid dynamics (CFD) approach to capturing the turbulent flow structures in engineering applications. The ExaWind team recently performed a LES of multiple revolutions of the National Renewable Energy Laboratory (NREL) 5 MW reference turbine with the open-source CFD code called Nalu-Wind. The NREL 5 MW reference turbine is a notional turbine fully defined in the open domain that has the key features of large modern wind turbines.

A body-fitted mesh—or computational geometric representation—of the blades, nacelle (the wind turbine’s encased generating components), and tower geometry were created. The modeling applied in this effort will be used to better understand weak and strong scaling performance of the ExaWind incompressible-flow-solver software stack and to advance the movement toward next-generation computing architectures. The software stack consists of software libraries for setting up and solving the system of equations, whose solution constitutes a numerical approximation of the physical-system continuum solution. This recent success of ExaWind establishes a new baseline capability for simulating modern turbines with body-resolved meshes and enables scientists and engineers to begin understanding the complex flow physics in multi-turbine wind farms that will take advantage of future ExaWind modeling and simulation capability.

Research Context and Objectives

Current methods for modeling wind energy cost and performance fall short due to insufficient model fidelity and inadequate treatment of fundamental phenomena such as atmospheric inflow structure, turbine wake development, and subsequent turbine-turbine wake flow interactions, which are persistently significant factors, especially in complex terrain. Most design and analysis tools simplify the flow physics with empirical representations that fail to capture first principles and/or are not computationally possible with existing simulation capabilities. Among the common deficiencies of the current methods are an inability to accurately model wake structure, a lack of understanding of the impact of different atmospheric turbulent conditions, and inadequate numerical schemes for transition from mesoscale (numerical weather prediction) to LES in CFD for flow within wind farms.

The impediment to predicting and minimizing energy losses and creating new technology options that will maximize performance has compelled ExaWind researchers to develop a predictive simulation capability that the team will use to simulate a wind plant composed of more than 100 multi-megawatt-scale wind turbines located within 100 square kilometers of complex terrain. These simulations require hundreds of billions of grid points to adequately resolve the flow physics and dynamic interactions.

The primary modeling and simulation environment of ExaWind is Nalu-Wind, which is based on the Nalu code developed at Sandia National Laboratories (SNL). Nalu-Wind is an unstructured-grid code that solves the acoustically incompressible Navier-Stokes equations; these equations are well suited for solving the low-Mach-number aerodynamics around the complex moving geometry of wind turbines and wind farms. Nalu-Wind is built on the Trilinos Sierra Toolkit (STK) library and can employ either the Hypre or Trilinos linear-system solver stacks. The ultimate objective of the ExaWind project is to create a predictive wind simulation capability that will run on an exascale-class computer by 2022.

To read more about this simulation, visit the full article here.


Source: Exascale Computing Project

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Research: A Survey of Numerical Methods Utilizing Mixed Precision Arithmetic

August 5, 2020

Within the past years, hardware vendors have started designing low precision special function units in response to the demand of the machine learning community and their demand for high compute power in low precision for Read more…

By Hartwig Anzt and Jack Dongarra

Implement Photonic Tensor Cores for Machine Learning?

August 5, 2020

Researchers from George Washington University have reported an approach for building photonic tensor cores that leverages phase change photonic memory to implement a neural network (NN). Their novel architecture, reporte Read more…

By John Russell

HPE Keeps Cray Brand Promise, Reveals HPE Cray Supercomputing Line

August 4, 2020

The HPC community, ever-affectionate toward Cray and its eponymous founder, can breathe a (virtual) sigh of relief. The Cray brand will live on, encompassing the pinnacle of HPE's HPC portfolio. After announcing its i Read more…

By Tiffany Trader

Machines, Connections, Data, and Especially People: OAC Acting Director Amy Friedlander Charts Office’s Blueprint for Innovation

August 3, 2020

The path to innovation in cyberinfrastructure (CI) will require continued focus on building HPC systems and secure connections between them, in addition to the increasingly important goals of data best practices and work Read more…

By Ken Chiacchia, Pittsburgh Supercomputing Center/XSEDE

Nvidia Said to Be Close on Arm Deal

August 3, 2020

GPU leader Nvidia Corp. is in talks to buy U.K. chip designer Arm from parent company Softbank, according to several reports over the weekend. If consummated, analysts said the acquisition would cement Nvidia’s stat Read more…

By George Leopold

AWS Solution Channel

AWS announces the release of AWS ParallelCluster 2.8.0

AWS ParallelCluster is a fully supported and maintained open source cluster management tool that makes it easy for scientists, researchers, and IT administrators to deploy and manage High Performance Computing (HPC) clusters in the AWS cloud. Read more…

Intel® HPC + AI Pavilion

Supercomputing the Pandemic: Scientific Community Tackles COVID-19 from Multiple Perspectives

Since their inception, supercomputers have taken on the biggest, most complex, and most data-intensive computing challenges—from confirming Einstein’s theories about gravitational waves to predicting the impacts of climate change. Read more…

Summer Reading: Here’s a Quantum Advantage You Can Bet On!

August 3, 2020

While quantum computing researchers today vigorously chase a demonstration of a quantum advantage – an application which when run on a quantum computer provides sufficient advantage to warrant switching from a classica Read more…

By John Russell

HPE Keeps Cray Brand Promise, Reveals HPE Cray Supercomputing Line

August 4, 2020

The HPC community, ever-affectionate toward Cray and its eponymous founder, can breathe a (virtual) sigh of relief. The Cray brand will live on, encompassing th Read more…

By Tiffany Trader

Machines, Connections, Data, and Especially People: OAC Acting Director Amy Friedlander Charts Office’s Blueprint for Innovation

August 3, 2020

The path to innovation in cyberinfrastructure (CI) will require continued focus on building HPC systems and secure connections between them, in addition to the Read more…

By Ken Chiacchia, Pittsburgh Supercomputing Center/XSEDE

Nvidia Said to Be Close on Arm Deal

August 3, 2020

GPU leader Nvidia Corp. is in talks to buy U.K. chip designer Arm from parent company Softbank, according to several reports over the weekend. If consummated Read more…

By George Leopold

Intel’s 7nm Slip Raises Questions About Ponte Vecchio GPU, Aurora Supercomputer

July 30, 2020

During its second-quarter earnings call, Intel announced a one-year delay of its 7nm process technology, which it says it will create an approximate six-month shift for its CPU product timing relative to prior expectations. The primary issue is a defect mode in the 7nm process that resulted in yield degradation... Read more…

By Tiffany Trader

PEARC20 Plenary Introduces Five Upcoming NSF-Funded HPC Systems

July 30, 2020

Five new HPC systems—three National Science Foundation-funded “Capacity” systems and two “Innovative Prototype/Testbed” systems—will be coming onlin Read more…

By Ken Chiacchia, Pittsburgh Supercomputing Center/XSEDE

Nvidia Dominates Latest MLPerf Training Benchmark Results

July 29, 2020

MLPerf.org released its third round of training benchmark (v0.7) results today and Nvidia again dominated, claiming 16 new records. Meanwhile, Google provided e Read more…

By John Russell

$39 Billion Worldwide HPC Market Faces 3.7% COVID-related Drop in 2020

July 29, 2020

Global HPC market revenue reached $39 billion in 2019, growing a healthy 8.2 percent over 2018, according to the latest analysis from Intersect360 Research. A 3 Read more…

By Tiffany Trader

Agenting Change: PEARC20 Keynote Encourages Cultural Change to Make Tech Better, More Diverse

July 29, 2020

The tech world will need to become more diverse if it is to thrive and survive, said Cherri Pancake, director of the Northwest Alliance for Computational Resear Read more…

By Ken Chiacchia, Pittsburgh Supercomputing Center/XSEDE

Supercomputer Modeling Tests How COVID-19 Spreads in Grocery Stores

April 8, 2020

In the COVID-19 era, many people are treating simple activities like getting gas or groceries with caution as they try to heed social distancing mandates and protect their own health. Still, significant uncertainty surrounds the relative risk of different activities, and conflicting information is prevalent. A team of Finnish researchers set out to address some of these uncertainties by... Read more…

By Oliver Peckham

Supercomputer-Powered Research Uncovers Signs of ‘Bradykinin Storm’ That May Explain COVID-19 Symptoms

July 28, 2020

Doctors and medical researchers have struggled to pinpoint – let alone explain – the deluge of symptoms induced by COVID-19 infections in patients, and what Read more…

By Oliver Peckham

Intel’s 7nm Slip Raises Questions About Ponte Vecchio GPU, Aurora Supercomputer

July 30, 2020

During its second-quarter earnings call, Intel announced a one-year delay of its 7nm process technology, which it says it will create an approximate six-month shift for its CPU product timing relative to prior expectations. The primary issue is a defect mode in the 7nm process that resulted in yield degradation... Read more…

By Tiffany Trader

Supercomputer Simulations Reveal the Fate of the Neanderthals

May 25, 2020

For hundreds of thousands of years, neanderthals roamed the planet, eventually (almost 50,000 years ago) giving way to homo sapiens, which quickly became the do Read more…

By Oliver Peckham

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

Neocortex Will Be First-of-Its-Kind 800,000-Core AI Supercomputer

June 9, 2020

Pittsburgh Supercomputing Center (PSC - a joint research organization of Carnegie Mellon University and the University of Pittsburgh) has won a $5 million award Read more…

By Tiffany Trader

Nvidia Said to Be Close on Arm Deal

August 3, 2020

GPU leader Nvidia Corp. is in talks to buy U.K. chip designer Arm from parent company Softbank, according to several reports over the weekend. If consummated Read more…

By George Leopold

Nvidia’s Ampere A100 GPU: Up to 2.5X the HPC, 20X the AI

May 14, 2020

Nvidia's first Ampere-based graphics card, the A100 GPU, packs a whopping 54 billion transistors on 826mm2 of silicon, making it the world's largest seven-nanom Read more…

By Tiffany Trader

Leading Solution Providers

Contributors

Honeywell’s Big Bet on Trapped Ion Quantum Computing

April 7, 2020

Honeywell doesn’t spring to mind when thinking of quantum computing pioneers, but a decade ago the high-tech conglomerate better known for its control systems waded deliberately into the then calmer quantum computing (QC) waters. Fast forward to March when Honeywell announced plans to introduce an ion trap-based quantum computer whose ‘performance’ would... Read more…

By John Russell

Australian Researchers Break All-Time Internet Speed Record

May 26, 2020

If you’ve been stuck at home for the last few months, you’ve probably become more attuned to the quality (or lack thereof) of your internet connection. Even Read more…

By Oliver Peckham

15 Slides on Programming Aurora and Exascale Systems

May 7, 2020

Sometime in 2021, Aurora, the first planned U.S. exascale system, is scheduled to be fired up at Argonne National Laboratory. Cray (now HPE) and Intel are the k Read more…

By John Russell

‘Billion Molecules Against COVID-19’ Challenge to Launch with Massive Supercomputing Support

April 22, 2020

Around the world, supercomputing centers have spun up and opened their doors for COVID-19 research in what may be the most unified supercomputing effort in hist Read more…

By Oliver Peckham

Joliot-Curie Supercomputer Used to Build First Full, High-Fidelity Aircraft Engine Simulation

July 14, 2020

When industrial designers plan the design of a new element of a vehicle’s propulsion or exterior, they typically use fluid dynamics to optimize airflow and in Read more…

By Oliver Peckham

$100B Plan Submitted for Massive Remake and Expansion of NSF

May 27, 2020

Legislation to reshape, expand - and rename - the National Science Foundation has been submitted in both the U.S. House and Senate. The proposal, which seems to Read more…

By John Russell

John Martinis Reportedly Leaves Google Quantum Effort

April 21, 2020

John Martinis, who led Google’s quantum computing effort since establishing its quantum hardware group in 2014, has left Google after being moved into an advi Read more…

By John Russell

Google Cloud Debuts 16-GPU Ampere A100 Instances

July 7, 2020

On the heels of the Nvidia’s Ampere A100 GPU launch in May, Google Cloud is announcing alpha availability of the A100 “Accelerator Optimized” VM A2 instance family on Google Compute Engine. The instances are powered by the HGX A100 16-GPU platform, which combines two HGX A100 8-GPU baseboards using... Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This