Sometimes Accomplishment Is Starting Something New Rather than Finishing Something Old

By Thomas Sterling and Chirag Dekate

June 2, 2010

So perhaps it was of this last year of the first decade of the first century of the new millennium in the field of high performance computing. Not to minimize the continued progression of petaflops computing as we enter Year 3 AP (After Petaflops).  With the addition of new machines both deployed and planned, petaflops-scale applications, as acknowledged by the Gordon Bell Prize, steady increase in the number of cores per socket, and the uncomfortable marriage of GPUs in heterogeneous structures — the last year has been marked by continued and demonstrable advances. As petaflops computing has become truly international in scope and application, this emerging system class is no longer an ethereal fringe, but rather has gained firm traction at such power houses (yes, meant in more than one way) as Oak Ridge National Laboratory, where they now serve humanity as the heavy lifters in computational methods addressing the challenges of the modern world.

But one potentially important accomplishment in the last twelve months is not something that has been completed; instead, it is something that has been just initiated. Even as we gain a footing in the era of petaflops computing, we have set in motion the exploration of the undiscovered domain of exaflops computing. This year has seen the launching of multiple programs to develop the concepts, architectures, software stack, programming models, and new families of parallel algorithms necessary to enable the practical realization of exaflops capability prior to the end of this decade. These have involved unprecedented cooperation and coordination within government agencies and laboratories, industry, academia, and internationally. At the dawn of the petaflops era, the emerging focus on the performance regime three orders of magnitude beyond is unlike anything before it and in stark contrast to the grass-roots workshops towards petaflops back in the relaxed days of the run-up to teraflops in the 1990s.

There are good reasons for this. The challenges facing the continued delivered sustained performance across a broad range of application domains are dramatic and reflect a corner turning on the trends that have driven us forward, ultimately due to Moore’s Law and the semiconductor revolution. These, somewhat over simplistically, can be summarized as: concurrency, power, reliability, and productivity.

In the past, the double-whammy of increases in clock rate and increases in processor core complexity delivered two decades of sustained exponential growth in processor core performance which when integrated in clusters of SMP nodes has given us the iconic images of straight lines on semi-log graphs with respect to the passage of time. Now the S-curve is bending for a second time, and not in a good way. Power has hit the threshold of pain, and the architecture tricks have been largely exhausted. Increased resources have been dedicated to confronting the egregious impact of the memory wall and the latencies and blocking incurred. Ever decreasing efficiencies (single digit not uncommon) by several normalization factors (e.g., FLOPS, utilization, per transistor, per joule, per hectare) have exposed the soft underbelly of an ultimately unsustainable golden age: exponentials cannot go on forever.

Indeed, the authors have projected that “we will never achieve sustained zettaflops computing” using the hardware paradigm of Boolean logic gates and binary data storage. Due to the speed of light, Boltzmann’s Constant, and atomic granularity it is predicted that the wall, which is more like a very steep hill will occur at about 32 exaflops. But we are not there yet; indeed, there are a good four orders of magnitude to go. And that will be hard.

Three major activities can be cited that have just been created during the last year to engage the talents of the international community including experts in: hardware, software, algorithms, and domain science. These have resulted from at least two years of preliminary workshops and studies sponsored by diverse entities and internal industry planning as well. These are: IESP, DOE X-Stack, and DARPA UHPC. There are many smaller activities as well.

The International Exascale Software Project (IESP) has brought together the interests, talents, and resources of the international community to cooperate and coordinate long-term development of the necessary software infrastructure required to enable effective exaflops-scale performance before the end of this decade. Learning from past experiences where software always appeared to lag behind the hardware, this world-straddling endeavor is driven by the recognition that to succeed, the software needs to be there when the hardware is. More importantly, the hardware designs must be informed by the needs of the software so that there is minimum mismatch and the concomitant ensuing generations of unsatisfactory patches. But there is an even more critical imperative: the realization that without the right software, exaflops may not be achievable at all (except in special cases) and that no one nation can go it alone; the HPC community is just too small for multiple conflicting paths of a top to bottom software refactoring. In the last year, four multi-day meetings in France, Japan, and the UK among representatives of all of the major HPC nations have provided an emerging roadmap to inform future planning of the joint development of the full supporting software infrastructure for Exascale systems’ operation and programming.

The US DOE has also begun a new program of research with the release of its recent RFP to develop the components of the “X-Stack,” the software required to enable a new generation of science and technology applications with the advent of future exaflops capable systems. These elements include operating systems, runtime systems, programming models and tools, and methods for reliability and mass storage and I/O. The winners, not yet announced, will represent a new wave of research in the US combining partners in the national laboratories, industry, and academia driven by the requirements of major mission-critical applications. This and other related DOE programs were developed in part from an extensive series of community workshops through the preceding year on application domains, hardware and software systems, and mathematical algorithms. This research will join other programs around the world in the first concerted effort to turn the corner and set a new trajectory for future HPC system software architecture, design, and implementation.

Perhaps most dramatic and at the same time risky undertaking is the new DARPA Ubiquitous High Performance Computing (UHPC) research program. UHPC is intended to attack the above challenges through nothing less than revolutionizing HPC system design. Through a lengthy program development process that involved three separate studies in technology, software, and resiliency engaging the talents of experts throughout the US, UHPC evolved an energetic research charter to reinvent computing prior to the end of this decade. The program was not explicitly targeted to exascale but rather to the mid-range of one or some unspecified number of interconnected and interoperable racks, each capable of approximately 1 petaflops sustained performance with a power budget of less than 60 kilowatts.

At the foundation of this program is the call for a new model of parallel computation to replace the venerable and highly successful message-passing model that has dominated for the last two decades. A major emphasis is on power reduction with an average energy of 25 Pico-Joules per floating point operation. A thousand such racks if sufficiently efficient would deliver 1 exaflops for 20 megawatts.

Emphasis is placed on the co-design of both hardware and software components in response to challenge problems that will span the applications domains from some of the largest STEM problems to heavy real time I/O streaming to knowledge management graph problems. Scaling down is as important as scaling up to UHPC, with single modules capable of multiple teraflops (and in mobile modules this is an important operating point).

The program may run eight or nine years and result in one or more prototypes of fully-operational systems. The first half of the program, Phases 1 and 2 spanning four years, will begin this summer with the winning teams to be announced in a month’s time. Atypical of such programs is the expectation of strong cooperation among competing teams and the delivery of much of the techniques and technology to the research community throughout the four phases of the program.

This year has indeed been a very productive year, both for its accomplishments in the deployment and application of petaflops-scale systems and for its forward-looking inauguration of the exaflops era.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pressing needs and hurdles to widespread AI adoption. The sudde Read more…

Quantinuum Reports 99.9% 2-Qubit Gate Fidelity, Caps Eventful 2 Months

April 16, 2024

March and April have been good months for Quantinuum, which today released a blog announcing the ion trap quantum computer specialist has achieved a 99.9% (three nines) two-qubit gate fidelity on its H1 system. The lates Read more…

Mystery Solved: Intel’s Former HPC Chief Now Running Software Engineering Group 

April 15, 2024

Last year, Jeff McVeigh, Intel's readily available leader of the high-performance computing group, suddenly went silent, with no interviews granted or appearances at press conferences.  It led to questions -- what's Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Institute for Human-Centered AI (HAI) put out a yearly report to t Read more…

Crossing the Quantum Threshold: The Path to 10,000 Qubits

April 15, 2024

Editor’s Note: Why do qubit count and quality matter? What’s the difference between physical qubits and logical qubits? Quantum computer vendors toss these terms and numbers around as indicators of the strengths of t Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips are available off the shelf, a concern raised at many recent Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Computational Chemistry Needs To Be Sustainable, Too

April 8, 2024

A diverse group of computational chemists is encouraging the research community to embrace a sustainable software ecosystem. That's the message behind a recent Read more…

Hyperion Research: Eleven HPC Predictions for 2024

April 4, 2024

HPCwire is happy to announce a new series with Hyperion Research  - a fact-based market research firm focusing on the HPC market. In addition to providing mark Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

Leading Solution Providers

Contributors

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

Intel’s Xeon General Manager Talks about Server Chips 

January 2, 2024

Intel is talking data-center growth and is done digging graves for its dead enterprise products, including GPUs, storage, and networking products, which fell to Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire