What’s New in HPC Research: E3SM Diags, Sunway Supercomputer, AxoNN, Snellius Supercomputer & More

By Mariana Iriarte

April 28, 2022

In this regular feature, HPCwire highlights newly published research in the high-performance computing community and related domains. From parallel programming to exascale to quantum computing, the details are here.


Figure 1. A schematic overview of E3SM Diags structure and workflow.

The E3SM diagnostics package: a python-based diagnostics package for earth system models evaluation

E3SM Diags is an open-source Python software package that was released in 2017 and developed to support the Department of Energy Energy Exascale Earth System 5 Model project. A multi-institutional team of researchers modeled E3SM Diags after the atmospheric model working group diagnostics package from the National Center for Atmospheric Research. In this pre-print open access paper, researchers detail version 2.6’s new features including “more process-oriented and phenomenon-based evaluation 10 diagnostics have been implemented, such as analysis of the Quasi-biennial Oscillation, El Niño – Southern Oscillation, streamflow, diurnal cycle of precipitation, tropical cyclones and ozone.” Researchers designed the tool to be flexible and added “new observational datasets and new diagnostic 15 algorithms.” In this latest version, the software package was “extended significantly beyond the initial goal to be a Python equivalent of the NCL AMWG package.” 

Authors: Chengzhu Zhang, Jean-Christophe Golaz, Ryan Forsyth, Tom Vo, Shaocheng Xie, Zeshawn Shaheen, Gerald L. Potter, Xylar S. Asay-Davis, Charles S. Zender, Wuyin Lin, Chih-Chieh Chen, Chris R. Terai, Salil Mahajan, Tian Zhou, Karthik Balaguru, Qi Tang, Cheng Tao, Yuying Zhang, Todd Emmenegger, and Paul Ullrich

21296exponentially complex quantum many-body simulation via scalable deep learning method

Chinese researchers pose their justification for their bid to win the Gordon Bell Prize in this paper from a multi-institutional team of researchers. The researchers report that “a deep learning-based simulation protocol can [solve the quantum many-body problem] with state-of-the-art precision in the Hilbert space as large as 21296 for spin system and 3144 for fermion system, using a HPC-AI hybrid framework on the new Sunway supercomputer.” Using up to 40 million heterogeneous SW26010pro cores, the applications achieved 94 percent weak scaling efficiency and 72 percent strong scaling efficiency,” according to the research team.

Authors: Xiao Liang, Mingfan Li, Qian Xiao, Hong An, Lixin He, Xuncheng Zhao, Junshi Chen, Chao Yang, Fei Wang, Hong Qian, Li Shen, Dongning Jia, Yongjian Gu, Xin Liu, and Zhiqiang Wei

Fig. 1. Inter-layer parallelism on four GPUs. The green and yellow boxes represent the forward and backward passes of a microbatch respectively. The numbers inside each box represent the microbatch number. We assume that the backward pass takes twice as much time as the forward pass

AxoNN: an asynchronous, message-driven parallel framework for extreme-scale deep learning

Researchers from the department of computer science at the University of Maryland in College Park, Maryland, introduce a parallel deep learning framework named AxoNN, which “exploits asynchrony and message-driven execution to schedule neural network operations on each GPU, thereby reducing GPU idle time and maximizing hardware efficiency.” With the implementation of AxoNN, memory consumption is reduced by four times“by using the CPU memory as a scratch space for offloading data periodically during training.” The reduction in memory consumption enabled the researchers to “increase the number of parameters per GPU by four times, thus reducing the amount of communication and increasing performance by over 13%.” Researchers demonstrated that “when tested against large transformer models with 12–100 billion parameters on 48–384 NVIDIA Tesla V100 GPUs, AxoNN achieves a per-GPU throughput of 49.4– 54.78% of theoretical peak and reduces the training time by 22- 37 days (15–25% speedup) as compared to the state-of-the-art.”

Authors: Siddharth Singh and Abhinav Bhatele 

Evaluation of finite difference based asynchronous partial differential equations solver for reacting flows

In this paper, an international team of researchers from Texas A&M University in Texas, California Institute of Technology and Sandia National Laboratories in California, and the Indian Institute of Science in India present “an effective methodology of implementing temporal discretization using a multi-stage Runge-Kutta method with asynchrony-tolerant (AT) schemes.” The researchers argue that the combination of using the multi-stage Runge-Kutta method with AT schemes provides minimal overheads to scale next-generation exascale machines with extreme parallelism. “Together these schemes are used to perform asynchronous simulations of canonical reacting flow problems, demonstrated in one-dimension including auto-ignition of a premixture, premixed flame propagation and non-premixed autoignition.” In addition, the paper also dives into the loss of accuracy of weighted essentially non-oscillatory schemes when used in conjunction with relaxed synchronization. “To overcome this loss of accuracy, high-order AT-WENO schemes are derived and tested on linear and non-linear equations.” Lastly “AT-WENO schemes are demonstrated in the propagation of a detonation wave with delays at PE boundaries.”

Authors: Komal Kumari, Emmet Cleary, Swapnil Desai, Diego A. Donzis, Jacqueline H. Chen, and Konduri Aditya

Figure 1: Cyclic distribution in several dimensions, indicated by colors.

Minimizing communication in the multidimensional FFT

Mathematicians from the Mathematical Institute at the Utrecht University in the Netherlands “present a parallel algorithm for the fast Fourier transform (FFT) in higher dimensions.” According to Thomas Koopman and Rob H. Bisseling, “this algorithm generalizes the cyclic-to-cyclic one-dimensional parallel algorithm to a cyclic-to-cyclic multidimensional parallel algorithm while retaining the property of needing only a single all-to-all communication step.” Using the Dutch National Supercomputer Snellius, the researchers “show that FFTU is competitive with the state-of-the-art and that it allows to use of a larger number of processors, while keeping communication limited to a single all-to-all operation. For arrays of size 10243 and 645, FFTU achieves a speedup of a factor 149 and 176, respectively, on 4,096 processors.”

Authors: Thomas Koopman and Rob H. Bisseling

Deep neural networks for solving extremely large linear systems 

Hong Kong mathematicians from the University of Hong Kong in Pokfulam, Hong Kong study “deep neural networks for solving extremely large linear systems arising from physically relevant problems.” According to the authors, the biggest advantage of using the method is the amount of storage saved. The paper includes”‘examples arising from partial differential equations, queueing problems and probabilistic Boolean networks…to demonstrate that solutions of linear systems with sizes ranging from septillion (1024) to nonillion (1030) can be learned quite accurately.”

Authors: Yiqi Gu and Michale K. Ng

mpiQulacs: a distributed quantum computer simulator for A64FX-based cluster systems

Researchers from the ICT Systems Laboratory at Fujitsu LTD designed mpiQulacs, which is a distributed state vector simulator that is “optimized for large-scale simulation on A64FX based cluster systems.” In this paper, the researchers “evaluate weak and strong scaling of mpiQulacs with up to 36 qubits on a new 64-node A64FX-based cluster system named Todoroki.” Researchers compare mpiQulacs with other distributed state vector simulators demonstrating that mpiQulacs was able to perform really well with “large-scale simulation on tens of nodes while sustaining a nearly ideal scalability.” In addition, they define quantum B/F ratio, which “indicates the execution efficiency of state vector simulators running on cluster systems.”  Using quantum B/F ratio, the researchers also demonstrated “that mpiQulacs running on Todoroki fits the requirements of distributed state vector simulation rather than the existing simulators running on general purpose CPU-based or GPU-based cluster systems.”

Authors: Satoshi Imamura, Masafumi Yamazaki, Takumi Honda, Akihiko Kasagi, Akihiro Tabuchi, Hiroshi Nakao, Naoto Fukumoto, and Kohta Nakashima


Do you know about research that should be included in next month’s list? If so, send us an email at [email protected]. We look forward to hearing from you.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Like Nvidia, Google’s Moat Draws Interest from DOJ

October 14, 2024

A "moat" is a common term associated with Nvidia and its proprietary products that lock customers into their hardware and software. Another moat breakdown should have them concerned. The U.S. Department of Justice is Read more…

Recipe for Scaling: ARQUIN Framework for Simulating a Distributed Quantum Computing System

October 14, 2024

One of the most difficult problems with quantum computing relates to increasing the size of the quantum computer. Researchers globally are seeking to solve this “challenge of scale.” To bring quantum scaling closer Read more…

Nvidia Is Increasingly the Secret Sauce in AI Deployments, But You Still Need Experience

October 14, 2024

I’ve been through a number of briefings from different vendors from IBM to HP, and there is one constant: they are all leaning heavily on Nvidia for their AI services strategy. That may be a best practice, but Nvidia d Read more…

Zapata Computing, Early Quantum-AI Software Specialist, Ceases Operations

October 14, 2024

Zapata Computing, which was founded in 2017 as a Harvard spinout specializing in quantum software and later pivoted to an AI focus, is ceasing operations, according to an SEC filing last week. Zapata had gone public one Read more…

AMD Announces Flurry of New Chips

October 10, 2024

AMD today announced several new chips including its newest Instinct GPU — the MI325X — as it chases Nvidia. Other new devices announced at the company event in San Francisco included the 5th Gen AMD EPYC processors, Read more…

NSF Grants $107,600 to English Professors to Research Aurora Supercomputer

October 9, 2024

The National Science Foundation has granted $107,600 to English professors at US universities to unearth the mysteries of the Aurora supercomputer. The two-year grant recipients will write up what the Aurora supercompute Read more…

Nvidia Is Increasingly the Secret Sauce in AI Deployments, But You Still Need Experience

October 14, 2024

I’ve been through a number of briefings from different vendors from IBM to HP, and there is one constant: they are all leaning heavily on Nvidia for their AI Read more…

NSF Grants $107,600 to English Professors to Research Aurora Supercomputer

October 9, 2024

The National Science Foundation has granted $107,600 to English professors at US universities to unearth the mysteries of the Aurora supercomputer. The two-year Read more…

VAST Looks Inward, Outward for An AI Edge

October 9, 2024

There’s no single best way to respond to the explosion of data and AI. Sometimes you need to bring everything into your own unified platform. Other times, you Read more…

Google Reports Progress on Quantum Devices beyond Supercomputer Capability

October 9, 2024

A Google-led team of researchers has presented more evidence that it’s possible to run productive circuits on today’s near-term intermediate scale quantum d Read more…

At 50, Foxconn Celebrates Graduation from Connectors to AI Supercomputing

October 8, 2024

Foxconn is celebrating its 50th birthday this year. It started by making connectors, then moved to systems, and now, a supercomputer. The company announced it w Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks. These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvi Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago this week emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whateve Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed t Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

Granite Rapids HPC Benchmarks: I’m Thinking Intel Is Back (Updated)

September 25, 2024

Waiting is the hardest part. In the fall of 2023, HPCwire wrote about the new diverging Xeon processor strategy from Intel. Instead of a on-size-fits all approa Read more…

Ansys Fluent® Adds AMD Instinct™ MI200 and MI300 Acceleration to Power CFD Simulations

September 23, 2024

Ansys Fluent® is well-known in the commercial computational fluid dynamics (CFD) space and is praised for its versatility as a general-purpose solver. Its impr Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Leading Solution Providers

Contributors

IBM Develops New Quantum Benchmarking Tool — Benchpress

September 26, 2024

Benchmarking is an important topic in quantum computing. There’s consensus it’s needed but opinions vary widely on how to go about it. Last week, IBM introd Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qu Read more…

US Implements Controls on Quantum Computing and other Technologies

September 27, 2024

Yesterday the Commerce Department announced export controls on quantum computing technologies as well as new controls for advanced semiconductors and additive Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire