Aurora and Upgraded Advanced Photon Source to Power Discovery at Argonne

December 8, 2022

Dec. 8, 2022 — Metals crack, neurons misfire, viruses mutate — all at scales of size and time we can barely fathom, let alone study.

The more data the upgraded APS feeds to Aurora, the greater Aurora’s overall quality and speed of analysis. The better the analysis Aurora provides to APS, the more rapidly researchers will be able to narrow the focus of their experiments. Credit: Argonne.

To understand how processes work or fail in the natural or mechanical world, to solve the riddles of science, requires the ability to probe more deeply and expose layers of detail never before observed.

To gain that understanding — that knowing — requires access to powerful scientific tools.

Tools of such magnitude are used throughout the global scientific community. Some of them call the U.S. Department of Energy’s (DOE) Argonne National Laboratory home, and for decades, researchers from around the world have chosen Argonne as a place to conduct their work. Two of the main reasons are the supercomputing power of the Argonne Leadership Computing Facility (ALCF) and the high-energy X-ray beams generated at its Advanced Photon Source (APS), both DOE Office of Science user facilities.

But it seems the more we know about a subject, the more complex the questions become. So much so that even the technical capabilities at hand are no longer effective or efficient enough to pursue them.

To meet such challenges, Argonne is creating a formidable arsenal of tools through transformative upgrades at both the APS and ALCF.

Already renowned for its modeling, simulation, data analysis and artificial intelligence (AI) capabilities, the ALCF will introduce Aurora, a machine at the forefront of supercomputing, that will adapt new strategies in AI to more powerful hardware.

And the APS, currently one of the brightest hard X-ray synchrotrons in the Western hemisphere, will radically increase its performance to probe the inner workings of both animate and inanimate objects at unprecedented scales.

Remarkable as each machine will be on its own, a plan to integrate the two will create something of a single scientific mega-instrument designed to rapidly deliver analyzed data for on-the-fly adjustments to experiments and faster time to solutions.

To succeed, developers have to overcome a dilemma of their own making: the crush of big data. And the upgrades currently underway at both facilities will dramatically increase the amount of data they already produce and analyze.

Researchers estimate that experiments at the APS will collect hundreds of times more raw data from experiments after the upgrade than they do currently. In fact, the amount of new information will exceed all the songs in the entire Spotify catalog, for example. This increase in data is a result of the extraordinary speeds at which the upgraded APS will be able to capture snapshots of materials and processes and the phenomenal resolution of those snapshots. Scientists will be able to see much smaller things much more clearly and much more quickly.

Researchers are betting on Aurora’s advanced computational methods to aggressively digest that information, learn from it and represent it in a more concise and targeted form, and much faster than any assemblage of scientists ever could.

This tight coupling of APS experiment instruments with ALCF supercomputers will require an immense networking effort. In the future, the two facilities will be connected by dozens of fiber optic cables capable of transmitting trillions of data bits per second, allowing for hypercharged input of data and output of analysis.

Such technical supermergers could transform the way science is conducted, not only at Argonne but around the world. They could reduce time-to-discovery from years and months to weeks and days. But access to the machines, particularly on-demand access, could prove a stumbling block if not administered efficiently.

The ALCF will support on-demand data analysis by providing immediate access to Aurora to investigate topics that could have critical national importance, such as COVID-19. Keeping up with access demands is a complex problem with a lot of moving parts, but resolution could have implications beyond just Aurora and the APS.

“This is a new paradigm for the traditional supercomputing center,” said Bill Allcock, director of operations for the ALCF. ​“By supporting on-demand work, we’ll be capable of supporting new classes of science that, traditionally, have not been well-supported on large, high performance computing systems. There’s more science out there that can use this type of collaboration, and our goal is to support it all.”

Dynamic Duo of Technology

The APS acts like a powerful microscope that allows researchers to study complex materials and systems. Currently, the APS delivers X-ray beams that are up to a billion times brighter than those used by your dentist to penetrate deep inside materials. They are so bright, in fact, that scientists can use them to ​see, in real time, what happens inside a working battery or detail the atomic structure of a single protein.

Aerial view of the APS. Credit: Argonne National Laboratory.

The APS was built in the 1990s, and upgrading it into a state-of-the-art facility is a gargantuan endeavor. The original electron storage ring, two-thirds of a mile around, will be removed and replaced by a new, completely updated ring. Fueled by this powerhouse engine, the upgraded APS will generate X-ray beams that are up to 500 times brighter than those produced by the current machine, allowing scientists to examine materials and processes in greater detail and with greater precision than ever before.

With the upgrade, researchers can zoom in on matter with unprecedented depth — down to less than a nanometer, or more than 50,000 times smaller than the width of a human hair. They can also capture processes at ultrafast speeds, like taking a rapid-fire snapshot of nature in motion in less than one-billion­­th of a second.

Meanwhile, across the Argonne campus, the ALCF is installing Aurora. A behemoth of a supercomputer, it will not only interpret the details of all of those images and return them to researchers with great efficiency, but also help steer experiments at Argonne.

One of the first exascale computers in the world, Aurora will be capable of performing two billion billion calculations per second, making it nearly 50 times faster than ALCF’s current supercomputer, Polaris.

Aurora comprises refrigerator-sized cabinets that house more than 10,000 nodes, the miniature workhorses of the machine. Each node contains a combination of processing units to help Aurora perform many functions simultaneously. And these processors, combined with powerful networking and high performance storage, allow all those nodes to work together for greater speed and efficiency.

One of Aurora’s distinguishing features will be its ability to seamlessly integrate the important scientific tools of data analysis, modeling and simulation, and AI. This trifecta produces a feedback mechanism that plays a significant role in redirecting the design of new simulations and experiments.

Introducing such immense computational power will help researchers construct infinitely more accurate models in a diversity of scientific domains and complex dynamic systems, such as climate, battery storage and cosmic evolution.

But even with the exascale computing power of Aurora, the data rates produced by scientific instruments are becoming so large that conventional approaches to analyzing data will still take too long.

Researchers are looking toward novel ways to reduce and process data in intervals close to real time. This will combine new methods in AI with traditional approaches such as modeling and simulation, image processing and statistical analysis.

One key AI technique is called machine learning, which creates models that get more accurate as they acquire more data — no problem there: The mountain of data from APS will be used to train AI models on Aurora.

With the ability to highlight important features in the data, these AI models can significantly speed up computations of complex systems by as much as 1,000 times and, eventually, be used to make predictions. They can also expedite some of the most intensive computational tasks, boosting scientists’ ability to unravel complex phenomena, like those that occur when a battery charges or when solar panels convert sunlight into electricity.

“This is critical if we want to steer or autonomously guide an experiment,” said Nicholas Schwarz, Scientific Software Engineering and Data Management group leader at the APS, who is responsible for the collaboration between APS and Aurora. ​“The experiments are so complex, the data so big, events happen so quickly, that humans can’t interpret them fast enough. This combination of Aurora, the upgraded APS and AI will present opportunities in scientific exploration at speeds and scales previously inaccessible to science.”

Transformative Science

Jonathan Almer’s team at the APS is among the many groups that will benefit from this near-real-time approach provided by the collaboration. The team supports industry and military collaborators studying various alloys for use in aerospace and nuclear applications. These scientists want to see how their materials react to external stimuli, such as elevated temperatures and mechanical loads.

Carefully simulating those conditions at the beamline, high-energy X-ray beams explore the structure of the material and create a large cache of 2D images, often producing mountains of data.

“Researchers want us to reconstruct this extremely fast so they can have an image of their sample to help them decide which load or temperature to go to next,” explained Almer, group leader in Argonne’s X-ray Science division.

To handle the data explosion, the X-ray detectors that capture these images will stream this data to Aurora, which delivers highly detailed 3D reconstructions of the sample. Then it will filter results back to APS experimenters.

This rapid turnaround will allow researchers to better home in on pivotal pieces of information, quickly narrow the focus of their experiments, and accelerate discovery.

The combined power of Aurora and APS will also allow scientists to map brain structure and function at an unprecedented scale, probing large brain volumes in a single experiment. Like deciphering a map, scientists are already plotting the axons and neurons, or highways and arteries, of a brain’s communication pathways.

Work has focused primarily on the smaller mouse brain, but ultimately researchers would like to zoom in on the back streets and alleys of the larger, more complicated human brain.

But building a complete road map of the entire human brain will produce a deluge of data, much of it packed into the innumerable X-ray images produced at the upgraded APS. Ciphering through the data to ferret out the billions of connections will require the integration of Aurora and AI algorithms to make sense of it all.

“Let’s say we did a nano X-ray of an entire primate,” said Narayanan ​“Bobby” Kasthuri, a neuroscience researcher at Argonne and the University of Chicago. ​“Essentially, we would have so many axons, there’s not enough time or humans to trace every one of them manually. So, we work with our collaborators at the ALCF to come up with AI algorithms that will automatically trace these axons throughout the brain.”

By studying the structure and connectivity between cells within the brain, scientists aim to understand the relationship between neural structure and mental function.

“All told, this confluence of technologies, the massive increase in computing from Aurora and AI and the brighter, higher energy, more coherent X-rays of the APS, will enable a transformation in the way we do science,” Schwarz noted. ​“On their own, we could not successfully conduct science on these grand scales. But when we combine them, a whole new world of science is open to us.”

Aurora is currently being installed at Argonne and will start up in 2023. The upgraded APS is scheduled to come online in 2024.


Source: John Spizzirri And Leah Hesla, Argonne National Laboratory

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

2024 Winter Classic: Texas Two Step

April 18, 2024

Texas Tech University. Their middle name is ‘tech’, so it’s no surprise that they’ve been fielding not one, but two teams in the last three Winter Classic cluster competitions. Their teams, dubbed Matador and Red Read more…

2024 Winter Classic: The Return of Team Fayetteville

April 18, 2024

Hailing from Fayetteville, NC, Fayetteville State University stayed under the radar in their first Winter Classic competition in 2022. Solid students for sure, but not a lot of HPC experience. All good. They didn’t Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use of Rigetti’s Novera 9-qubit QPU. The approach by a quantum Read more…

2024 Winter Classic: Meet Team Morehouse

April 17, 2024

Morehouse College? The university is well-known for their long list of illustrious graduates, the rigor of their academics, and the quality of the instruction. They were one of the first schools to sign up for the Winter Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire