AN INTERVIEW WITH SUPERCOMPUTING LEGEND JACK DONGARRA

May 24, 2002

by Martin Meuer, Prometeus GmbH

Jack Dongarra, director of the Center for Information Technology Research (CITR) and of the Innovative Computing Laboratory (ICL) at the University of Tennessee, is a legend in the supercomputing field since twenty years: LINPACK, MPI and the TOP500 are closely related to him.

Jack Dongarra will give the keynote presentation on Friday, June 21, 2002 at the 17th International Supercomputer Conference in Heidelberg, Germany. The title of his talk is: “High Performance Computing, Computational Grid, and Numerical Libraries”.

This interview for HPCwire was conducted by Martin Meuer, Prometeus GmbH, Germany.

HPCwire: Jack, you belong to the most renowned benchmark and HPC experts in the world. How long have you been personally involved in benchmarking and in the field of HPC?

DONGARRA: The original LINPACK Benchmark is, in some sense, an accident. So I guess that makes me an “Accidental Benchmarker”. It was originally designed to assist users of the LINPACK numerical software package by providing information on execution times required to solve a system of linear equations. The first “LINPACK Benchmark” report appeared as an appendix in the LINPACK Users’ Guide in 1979. The appendix of the Users’ Guide collected performance data for one commonly used path in the LINPACK software package. Results were provided for a matrix problem of size 100, on a collection of widely used computers (23 computers in all). This was done so users could estimate the time required to solve their matrix problem by extrapolation.

Over the years additional performance data was added, more as a hobby than anything else, and today the collection includes around 1500 different computer systems. In addition to the number of computers increasing, the scope of the benchmark has also expanded. The benchmark report describes the performance for solving a general dense matrix problem Ax=b at three levels of problem size and optimization opportunity: 100 by 100 problem (inner loop optimization), 1000 by 1000 problem (three loop optimization – the whole program), and a scalable parallel problem.

HPCwire: The Linpack benchmark has been used up until now to determine the TOP500-lists, of which you are a co-publisher. What are the strengths and weaknesses of Linpack for the evaluation of supercomputers?

DONGARRA: In order to fully exploit the increasing computational power of highly parallel computers, the application software must be scalable, that is, able to take advantage of larger machine configurations to solve larger problems with the same efficiency. The LINPACK benchmark addressed scalability by an introduction of a new category in 1991. This new category is referred to as a Highly-Parallel LINPACK (HPL) NxN benchmark. It requires solution of systems of linear equations by some method. The problem size is allowed to vary, and the best floating-point execution rate should be reported. In computing the execution rate, the number of operations should be 2n3/3+2n2 independent of the actual method used. If Gaussian elimination is chosen, partial pivoting must be used. The accuracy of solution is measured and reported. The following quantities of the benchmark are reported in the TOP500 list:

∑ Rmax is the performance in GF/s for the largest problem run on a computer, ∑ Nmax is the size of the largest problem run on a computer, ∑ N1/2 is the size where half the Rmax execution rate is achieved, ∑ Rpeak is the theoretical peak performance in GF/s for the computer.

As such this benchmark reports the performance of this one application with its floating point operations and message passing. So there is one application and one number that represents the performance for the computer system. The weakness is that this is just one application and one number.

HPCwire: How would you rate the “IDC balanced rating HPC benchmark”, which is considering other parameters like bandwidth/latency of the memory and the scalability of the system? Is it a more realistic approach than Linpack and therefore capable of replacing Linpack for the evaluation of supercomputers?

DONGARRA: IDC Benchmark looks on the surface to be a good starting point for a follow on benchmark. Unfortunately because of the way the rating procedure is implemented there are problems. Others have gone into detail on the problems.

As Aad van der Steen said in a recent article in the Primeur Weekly of February 13, 2002 when asked: How informative is the IDC Balanced Rating HPC Benchmark? He said “It is not. It looks like all the time and effort invested in the HPC Forum has yielded a sub-standard product that should be radically improved or withdrawn.” Presently it only adds to the confusion that already exists on the HPC Benchmark scene.

HPCwire: Are the TOP500 authors searching for an alternative to Linpack as yardstick for such evaluations? Or will we be forced to remain satisfied with Linpack for the expected PF/s systems within the next 10 years?

DONGARRA: Yes, the organizers of the Top500 are actively looking to expand the scope of the benchmark reporting. It is important to include more performance characteristic and signatures for a given system. There are a number of alternatives to look at in expanding the effort such as, STREAMS benchmark, the EuroBen-DM benchmark, the NAS Parallel Benchmarks, and PARKBENCH. Each of these describes more about the scalability of systems and provides more insight than the IDC benchmark will.

HPCwire: In the TOP500-list, which will be published in June during the ISC2002 in Heidelberg, the Japanese Earth Simulator (ES) will replace the American ASCI White as the new #1, with a best Linpack Performance of 35.61 TF/s. How large was the dense system of linear equations relating to this, which ES solved for the new Linpack world record, and how long did it take for ES to solve it?

DONGARRA: The system of linear equations was of size 1,041,216; (8.7 TB of memory). This is the largest dense system of linear equations I have seen solved on a computer. The benchmark took 5.8 hours to run. The results of the computation were checked and were accurate to floating point arithmetic specifications. The algorithm used to solve the system was a standard LU decomposition with partial pivoting and the software environment, for the most part, was FORTRAN using MPI with special coding for the computational kernels.

HPCwire: You yourself have referred to ES in the New York Times as Computenik, alluding to the Russian Sputnik in 1957, which certainly motivated and accelerated the US space program decidedly. Was the rapid completion of ES really such a great surprise in the USA? As far as I know, the Japanese ES- Project, including the time schedule, was known worldwide from the beginning. Or was the scheduled completion of ES with an efficiency of 87% considered too optimistic?

DONGARRA: I think the time schedule was known, but the fact that it was completed on time and is up in running is impressive. The result achieved with 87% on the benchmark is also very impressive. The impressive features of the performance are that it is 5 times the performance of the ASCI White Pacific computer for the benchmark. This is the largest difference we have seen in the past 10 years.

Here are some additional impressive statistics of the Earth Simulator:

The performance of ES is approximately a fourth of the performance of all the TOP500 computers from the November 2001 list, greater than the performance of all the DOE computers together and greater than the sum of the Top 20 Computers in the US.

HPCwire: Do you think that ES will accelerate the ASCI-Program?

DONGARRA: People in the US are taking notice of the Japanese accomplishments in the Earth Simulator Computer and it is the hope of many of us that the US government will continue to invest in the future of high performance computing.

HPCwire: What does the ES mean for “vector parallel multiprocessing architectures”; will they gain terrain? Will not only NEC, but also Cray Inc. with the SV2, profit from ES?

DONGARRA: I’m not sure about the Cray SV2, but the NEC SX architecture is very impressive as a highly parallel system as composed in the Earth Simulator. I would guess there are a number of customers around the world who are very interested in acquiring such system composed of the SX nodes and the Earth Simulator’s switch fabric.

HPCwire: Currently grid computing is of major interest. For example, at the ISC2002 in Heidelberg in a tutorial, a whole day will be dedicated to this topic, and in the conference an entire session, in which you will supply the Friday keynote address of ISC2002 relating to the topic “High Performance Computing, Computational Grids, and Numerical Libraries”.

What are your current projects concerning grid computing?

DONGARRA: At the University of Tennessee we are working on three Grid related projects, NetSolve, GrADS, and Harness.

Since 1995 we have been working on an approach to grid computing called NetSolve. NetSolve provides easy access to computational resources that are distributed with respect to both geography and ownership. Using NetSolve, a user can access both hardware and software computational resources distributed across a network.

In addition we are working with a group of colleagues on a NSF funded program execution framework being developed by the Grid Application Development Software (GrADS) Project. The goal of this framework is to provide good resource allocation for Grid applications and to support adaptive reallocation if performance degrades because of changes in the availability of Grid resources.

HARNESS (Heterogeneous Adaptable Reconfigurable Networked SystemS) is an experimental metacomputing framework built around the services of a highly customizable and reconfigurable distributed virtual machine (DVM). A DVM is a tightly coupled computation and resource grid that provides a flexible environment to manage and coordinate parallel application execution. This is a collaboration of researchers at the University of Tennessee, Emory University, and Oak Ridge National Laboratory and funded by the Department of Energy.

HPCwire: How would you predict grid computing to affect the landscape of HPC in the future?

DONGARRA: The Grid will make it possible to implement dramatically new classes of applications. These applications ranging from new systems for scientific inquiry, through computing support for crisis management, to support for personal lifestyle management are characterized by three dominant themes: computing resources are no longer localized, but distributed and hence heterogeneous and dynamic; computation is increasingly sophisticated and multidisciplinary; and computation is integrated into our daily lives, and hence subject to stricter time constraints than at present.

A good resource on activity in the area is “The Grid: Blueprint for a New Computing Infrastructure”, edited by Globus originators Ian Foster and Carl Kesselman (Foster and Kesselman, 1998).

HPCwire: There are many non-commercial projects in the area of grid computing, for example SETI@home. When can we expect the first commercial applications?

DONGARRA: There are examples today. Companies like United Devices, Entropia, Avaki, Parabon Computation are trying to establish a viable commercial operation using the concepts and ideas that came from the SETI@home experience.

Web site: http://www.supercomp.de

============================================================

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Edge-to-Cloud: Exploring an HPC Expedition in Self-Driving Learning

April 25, 2024

The journey begins as Kate Keahey's wandering path unfolds, leading to improbable events. Keahey, Senior Scientist at Argonne National Laboratory and the University of Chicago, leads Chameleon. This innovative projec Read more…

Quantum Internet: Tsinghua Researchers’ New Memory Framework could be Game-Changer

April 25, 2024

Researchers from the Center for Quantum Information (CQI), Tsinghua University, Beijing, have reported successful development and testing of a new programmable quantum memory framework. “This work provides a promising Read more…

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Point. The system includes Intel's research chip called Loihi 2, Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Research senior analyst Steve Conway, who closely tracks HPC, AI, Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, and this day of contemplation is meant to provide all of us Read more…

Intel Announces Hala Point – World’s Largest Neuromorphic System for Sustainable AI

April 22, 2024

As we find ourselves on the brink of a technological revolution, the need for efficient and sustainable computing solutions has never been more critical.  A computer system that can mimic the way humans process and s Read more…

Shutterstock 1748437547

Edge-to-Cloud: Exploring an HPC Expedition in Self-Driving Learning

April 25, 2024

The journey begins as Kate Keahey's wandering path unfolds, leading to improbable events. Keahey, Senior Scientist at Argonne National Laboratory and the Uni Read more…

Quantum Internet: Tsinghua Researchers’ New Memory Framework could be Game-Changer

April 25, 2024

Researchers from the Center for Quantum Information (CQI), Tsinghua University, Beijing, have reported successful development and testing of a new programmable Read more…

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Poin Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Resear Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Leading Solution Providers

Contributors

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire