Berkeley Lab Researchers Lead Development of Workflow to Predict Ground Movement

October 9, 2017

Oct. 9, 2017 — With emerging exascale supercomputers, researchers will soon be able to accurately simulate the ground motions of regional earthquakes quickly and in unprecedented detail, as well as predict how these movements will impact energy infrastructure—from the electric grid to local power plants—and scientific research facilities.

Currently, an interdisciplinary team of researchers from the Department of Energy’s (DOE’s) Lawrence Berkeley (Berkeley Lab) and Lawrence Livermore (LLNL) national laboratories, as well as the University of California at Davis are building the first-ever end-to-end simulation code to precisely capture the geology and physics of regional earthquakes, and how the shaking impacts buildings. This work is part of the DOE’s Exascale Computing Project (ECP), which aims to maximize the benefits of exascale—future supercomputers that will be 50 times faster than our nation’s most powerful system today—for U.S. economic competitiveness, national security and scientific discovery.

Transforming hazard into risk: Researchers at Berkeley Lab, LLNL and UC Davis are utilizing ground motion estimates from a regional-scale geophysics model to drive infrastructure assessments. (Image Courtesy of David McCallen)

“Due to computing limitations, current geophysics simulations at the regional level typically resolve ground motions at 1-2 hertz (vibrations per second). Ultimately, we’d like to have motion estimates on the order of 5-10 hertz to accurately capture the dynamic response for a wide range of infrastructure,” says David McCallen, who leads an ECP-supported effort called High Performance, Multidisciplinary Simulations for Regional Scale Seismic Hazard and Risk Assessments. He’s also a guest scientist in Berkeley Lab’s Earth and Environmental Sciences Area.

One of the most important variables that affect earthquake damage to buildings is seismic wave frequency, or the rate at which an earthquake wave repeats each second. Buildings and structures respond differently to certain frequencies. Large structures like skyscrapers, bridges, and highway overpasses are sensitive to low frequency shaking, whereas smaller structures like homes are more likely to be damaged by high frequency shaking, which ranges from 2 to 10 hertz and above. McCallen notes that simulations of high frequency earthquakes are more computationally demanding and will require exascale computers.

In preparation for exascale, McCallen is working with Hans Johansen, a researcher in Berkeley Lab’s Computational Research Division (CRD), and others to update the existing SW4 code—which simulates seismic wave propagation—to take advantage of the latest supercomputers, like the National Energy Research Scientific Computing Center’s (NERSC’s) Cori system. This manycore system contains 68 processor cores per chip, nearly 10,000 nodes and new types of memory. NERSC is a DOE Office of Science national user facility operated by Berkeley Lab.  The SW4 code was developed by a team of researchers at LLNL, led by Anders Petersson, who is also involved in the exascale effort.

With recent updates to SW4, the collaboration successfully simulated a 6.5 magnitude earthquake on California’s Hayward fault at 3-hertz on NERSC’s Cori supercomputer in about 12 hours with 2,048 Knights Landing nodes. This first-of-a-kind simulation also captured the impact of this ground movement on buildings within a 100-square kilometer (km) radius of the rupture, as well as 30km underground. With future exascale systems, the researchers hope to run the same model at 5-10 hertz resolution in approximately five hours or less.

“Ultimately, we’d like to get to a much larger domain, higher frequency resolution and speed up our simulation time, ” says McCallen. “We know that the manner in which a fault ruptures is an important factor in determining how buildings react to the shaking, and because we don’t know how the Hayward fault will rupture or the precise geology of the Bay Area, we need to run many simulations to explore different scenarios. Speeding up our simulations on exascale systems will allow us to do that.”

This work was recently published in the Institute of Electrical and Electronics Engineers (IEEE) Computer Society’s Computers in Science and Engineering.

Predicting Earthquakes: Past, Present and Future 

Historically, researchers have taken an empirical approach to estimating ground motions and how the shaking stresses structures. So to predict how an earthquake would affect infrastructure in the San Francisco region, researchers might look at a past event that was about the same size—it might even have happened somewhere else—and use those observations to predict ground motion in San Francisco. Then they’d select some parameters from those simulations based on empirical analysis and surmise how various buildings may be affected.

“It is no surprise that there are certain instances where this method doesn’t work so well,” says McCallen. “Every single site is different—the geologic makeup may vary, faults may be oriented differently and so on. So our approach is to apply geophysical research to supercomputer simulations and accurately model the underlying physics of these processes.”

To achieve this, the tool under development by the project team employs a discretization technique that divides the Earth into billions of zones. Each zone is characterized with a set of geologic properties. Then, simulations calculate the surface motion for each zone. With an accurate understanding of surface motion in a given zone, researchers also get more precise estimates for how a building will be affected by shaking.

The team’s most recent simulations at NERSC divided a 100km x 100km x 30km region into 60 billion zones. By simulating 30km beneath the rupture site, the team can capture how surface-layer geology affects ground movements and buildings.

Eventually, the researchers would like to get their models tuned up to do hazard assessments. As Pacific Gas & Electric (PG&E) begins to implement a very dense array of accelerometers into their SmartMeters—a system of sensors that collects electric and natural gas use data from homes and businesses to help the customer understand and reduce their energy use—McCallen is working with the utility company about potentially using that data to get a more accurate understanding of how the ground is actually moving in different geologic regions.

“The San Francisco Bay is an extremely hazardous area from a seismic standpoint and the Hayward fault is probably one of the most potentially risky faults in the country,” says McCallen. “We chose to model this area because there is a lot of information about the geology here, so our models are reasonably well-constrained by real data. And, if we can accurately measure the risk and hazards in the Bay Area, it’ll have a big impact.”

He notes that the current seismic hazard assessment for Northern California identifies the Hayward Fault as the most likely to rupture with a magnitude 6.7 or greater event before 2044. Simulations of ground motions from large—magnitude 7.0 or more—earthquakes require domains on the order of 100-500 km and resolution on the order of about one to five meters, which translates into hundreds of billions of grid points. As the researchers aim to model even higher frequency motions between 5 to 10 hertz, they will need denser computational grids and finer time-steps, which will drive up computational demands. The only way to ultimately achieve these simulations is to exploit exascale computing, McCallen says.

In addition to leading an ECP project, McCallen is also a Berkeley lab research affiliate and Associate Vice President at the University of California Office of the President.

This work was done with support from the Exascale Computing Project, a collaborative effort between the DOE’s Office of Science and National Nuclear Security Agency. NERSC is a DOE Office of Science User Facility.

About NERSC and Berkeley Lab

The National Energy Research Scientific Computing Center (NERSC) is a U.S. Department of Energy Office of Science User Facility that serves as the primary high-performance computing center for scientific research sponsored by the Office of Science. Located at Lawrence Berkeley National Laboratory, the NERSC Center serves more than 6,000 scientists at national laboratories and universities researching a wide range of problems in combustion, climate modeling, fusion energy, materials science, physics, chemistry, computational biology, and other disciplines. Berkeley Lab is a DOE national laboratory located in Berkeley, California. It conducts unclassified scientific research and is managed by the University of California for the U.S. DOE Office of Science. »Learn more about computing sciences at Berkeley Lab.


Source: NERSC

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

D-Wave Breaks New Ground in Quantum Simulation

July 16, 2018

Last Friday D-Wave scientists and colleagues published work in Science which they say represents the first fulfillment of Richard Feynman’s 1982 notion that simulating physical systems could be done most effectively Read more…

By John Russell

RIKEN and CEA Mark One Year of Exascale-focused Collaboration

July 16, 2018

RIKEN in Japan and the French Alternative Energies and Atomic Energy Commission (CEA) formed a five-year cooperative research effort on January 11, 2017, to advance HPC and prepare for exascale computing (see HPCwire co Read more…

By Nishi Katsuya

AI Thought Leaders on Capitol Hill

July 14, 2018

On Thursday, July 12, the House Committee on Science, Space, and Technology heard from four academic and industry leaders – representatives from Berkeley Lab, Argonne Lab, GE Global Research and Carnegie Mellon University – on the opportunities springing from the intersection of machine learning and advanced-scale computing. Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

Introducing the First Integrated System Management Software for HPC Clusters from HPE

How do you manage your complex, growing cluster environments? Answer that big challenge with the new HPC cluster management solution: HPE Performance Cluster Manager. Read more…

IBM Accelerated Insights

Are Your Software Licenses Impeding Your Productivity?

In my previous article, Improving chip yield rates with cognitive manufacturing, I highlighted the costs associated with semiconductor manufacturing, and how cognitive methods can yield benefits in both design and manufacture.  Read more…

HPC Serves as a ‘Rosetta Stone’ for the Information Age

July 12, 2018

In an age defined and transformed by its data, several large-scale scientific instruments around the globe might be viewed as a ‘mother lode’ of precious data. With names seemingly created for a ‘techno-speak’ glossary, these interferometers, cyclotrons, sequencers, solenoids, satellite altimeters, and cryo-electron microscopes are churning out data in previously unthinkable and seemingly incomprehensible quantities -- billions, trillions and quadrillions of bits and bytes of electro-magnetic code. Read more…

By Warren Froelich

D-Wave Breaks New Ground in Quantum Simulation

July 16, 2018

Last Friday D-Wave scientists and colleagues published work in Science which they say represents the first fulfillment of Richard Feynman’s 1982 notion that Read more…

By John Russell

AI Thought Leaders on Capitol Hill

July 14, 2018

On Thursday, July 12, the House Committee on Science, Space, and Technology heard from four academic and industry leaders – representatives from Berkeley Lab, Argonne Lab, GE Global Research and Carnegie Mellon University – on the opportunities springing from the intersection of machine learning and advanced-scale computing. Read more…

By Tiffany Trader

HPC Serves as a ‘Rosetta Stone’ for the Information Age

July 12, 2018

In an age defined and transformed by its data, several large-scale scientific instruments around the globe might be viewed as a ‘mother lode’ of precious data. With names seemingly created for a ‘techno-speak’ glossary, these interferometers, cyclotrons, sequencers, solenoids, satellite altimeters, and cryo-electron microscopes are churning out data in previously unthinkable and seemingly incomprehensible quantities -- billions, trillions and quadrillions of bits and bytes of electro-magnetic code. Read more…

By Warren Froelich

Tsinghua Powers Through ISC18 Field

July 10, 2018

Tsinghua University topped all other competitors at the ISC18 Student Cluster Competition with an overall score of 88.43 out of 100. This gives Tsinghua their s Read more…

By Dan Olds

HPE, EPFL Launch Blue Brain 5 Supercomputer

July 10, 2018

HPE and the Ecole Polytechnique Federale de Lausannne (EPFL) Blue Brain Project yesterday introduced Blue Brain 5, a new supercomputer built by HPE, which displ Read more…

By John Russell

Pumping New Life into HPC Clusters, the Case for Liquid Cooling

July 10, 2018

High Performance Computing (HPC) faces some daunting challenges in the coming years as traditional, industry-standard systems push the boundaries of data center Read more…

By Scott Tease

Meet the ISC18 Cluster Teams: Up Close & Personal

July 6, 2018

It’s time to meet your ISC18 Student Cluster Competition teams. While I was able to film them live at the ISC show, the trick was finding time to edit the vid Read more…

By Dan Olds

PRACEdays18 Keynote Allan Williams (Australia/NCI): We’re Open for Business Down Under!

July 5, 2018

The University of Ljubljana in Slovenia hosted the third annual EHPCSW18 and fifth annual PRACEdays18 events which opened with a plenary session on May 29, 2018 Read more…

By Elizabeth Leake (STEM-Trek for HPCwire)

Leading Solution Providers

SC17 Booth Video Tours Playlist

Altair @ SC17

Altair

AMD @ SC17

AMD

ASRock Rack @ SC17

ASRock Rack

CEJN @ SC17

CEJN

DDN Storage @ SC17

DDN Storage

Huawei @ SC17

Huawei

IBM @ SC17

IBM

IBM Power Systems @ SC17

IBM Power Systems

Intel @ SC17

Intel

Lenovo @ SC17

Lenovo

Mellanox Technologies @ SC17

Mellanox Technologies

Microsoft @ SC17

Microsoft

Penguin Computing @ SC17

Penguin Computing

Pure Storage @ SC17

Pure Storage

Supericro @ SC17

Supericro

Tyan @ SC17

Tyan

Univa @ SC17

Univa

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This