2009-2019: A Look Back on a Decade of Supercomputing

By Andrew Jones

December 15, 2009

As we turn the decade into the 2020s, we take a nostalgic look back at the last ten years of supercomputing. It’s amazing to think how much has changed in that time. Many of our older readers will recall how things were before the official Planetary Supercomputing Facilities at Shanghai, Oak Ridge and Saclay were established. Strange as it may seem now, each country — in fact, each university or company — had its own supercomputer!

Hindsight is easier, of course, but it is interesting to review how this major change in supercomputing came to happen over the last few years.

At the start of the decade, each major university, research centre or company using simulation & modelling had its own HPC resources — they owned it or leased it, operated it, housed it, etc. In addition, some countries (US, UK, Germany, etc.) operated their own national resources for open research. The national facilities were larger than individual institutions could afford, and access to these was usually by a mechanism known as “peer review” — the prospective user would write a short case describing how their science would benefit from using the facility and a group of fellow scientists would judge if the science was worthy. (Note: they rated the science, almost never the quality of the computing implementation!) Very often these national supercomputers were reserved for capability computations, similar to today’s Strategic Simulation category at Shanghai.

The highest profile facilities were those in major research centres (e.g., universities, US DOE labs, etc.) but many commercial organisations had very large facilities too, although these weren’t as well publicised since companies had begun to recognise their use of HPC as a strategic competitive asset. The world’s fastest supercomputers were ranked twice yearly on the TOP500 list. One of the key uses of the TOP500 was for tracking the increasing performance of supercomputing power, usually through a plot showing performance on a vertical logarithmic axis against years on a horizontal axis, and especially two trends on this plot: the reasonably linear growth (on the log scale) of the performance of the fastest machine at any one time; and the smooth linearly (log scale) increasing sum of performance of the 500 systems on the list. The first spark towards the Planetary Supercomputing Facilities came when someone asked “what if we could actually use the compute power of that sum line at once?”

Another factor was the increasing cost of the facilities provision — from computer acquisition (capital) to power (both capital for infrastructure and recurrent for operations) to site management (recurrent and capital, project management, etc.).

Based on this, a number of collaborations started to occur. In Europe, over 20 countries joined together for the two-year PRACE initiative to explore how a pan-European supercomputer service could work in practice. Much was learned from that project and the influences can be seen in the three Planetary Supercomputing Facilities. In the US, ORNL, originally a DOE open science national supercomputing centre, started to host other national facilities (initially for NSF, NOAA and DoD). In fact, ORNL was probably the first planetary supercomputing facility in practice, even though, as we know, Shanghai was the first official Planetary Supercomputing Facility.

People started to realise that operating these large supercomputers was not the interesting part of HPC, and was in fact a very specialist job. As more and more aggregation between national operating sites occurred, and as the scale limited the potential sites (due to power constraints, etc.), it became apparent that there would only be a few sites worldwide capable of fulfilling the growth predicted by the original TOP500 trends.

Then of course came what I call “the public realisation”. Politicians, the public, and Boards finally got it. Supercomputing made a difference. It wasn’t just big rooms of computers costing lots of tax dollars. It was a tool to underpin science, and often to propel it forward. It was a tool for accelerating any properly-formulated computational task, many even with impact on daily life. Better weather predictions. Better design and safety testing of household products. Consumer video/image processing (I remember trying to do early video processing on my own PC!). Speech processing — think how that has revolutionized mobile communications since the early days of typing email messages on BlackBerrys and the like.

And then the critical step — businesses and researchers finally understood that their competitive asset was the capabilities of their modelling software and user expertise — not the hardware itself. Successful businesses rushed to establish a lead over their competitors by investing in their modelling capability — especially robustness (getting trustable predictions/analysis), scalability (being able to process much larger datasets than before) and performance (driving down time to solutions).

As this “software arms race” was put into practice (led by the commercial users) — slowly at first but then with a surge of investment in robust scalable high performance software — money spent on hardware ceased to be the competitive difference. Coupled with the massive increase in demand for HPC resources following the public realisation, and the challenges of managing large facilities, this led to the announcement of the first Planetary Supercomputer Facility in Shanghai. Whilst there was initially preferential access for Chinese domestic users, anyone in the world could use the facility — from consumers to researchers to businesses. After years of trying to exploit commodity components, HPC itself became a commodity service. And this was true HPC, supporting tightly-coupled large simulations, not the earlier attempts at something daftly called “cloud computing,” which only really supported large numbers of very small jobs. The facility shocked the world with its scale — being larger not only than the then top machine on the TOP500, but also larger than the sum of the 500 systems.

The business case for individual ownership of HPC facilities worldwide suddenly became dramatically tougher to justify, with Shanghai providing all classes of computer resources at scale, including the various specialist processing types. Everyone got better HPC, whether capacity or capability, and cheaper HPC than they could ever provide locally. The consumer demand drove innovations in ease-of-use and accounting that previously were only ambitions of seemingly-perpetual academic research.

The international agreements from research funding agencies on behalf of their user communities and from consumer HPC brokers soon followed, confirming the official Planetary Supercomputing Facility status. Within a year, the US had followed suit, securing global agreement for Oak Ridge as the second official Planetary Supercomputing Facility, and of course deployed even more powerful resources than Shanghai.

Soon, the main security concerns had been solved. Network bandwidth that plagued earlier global collaborations went away, as data rarely needed to leave the facilities (or if so, only to transfer between Oak Ridge and Shanghai, which now had massive dedicated bandwidth). Anything that might be done with the data could be done at Oak Ridge or Shanghai — the data never needed to go anywhere else.

With the opening last year of the third and final Planetary Supercomputing Facility at Saclay, the world’s HPC is now ready to sprint into the next decade. We have now left the housing and daily care of the hardware to the specialists. The volume of public and private demand has set the scene for strong HPC provision into the future. We have the three official global providers to ensure consumer choice, with its competitive benefits, but few enough providers to underpin their business cases for the most capable possible HPC infrastructure.

With the pervasiveness of HPC in consumer, business and research arenas, and the long overdue acceptance of the truth that the software capabilities and performance at scale was the competitive asset, “can program HPC at scale” is now more than ever a valuable item for your CV.

For all this astounding progress, I wonder how quaint today’s world will seem when we look back from 2030. After all, just imagine someone reading this in 2009!

2009 Author’s Note: This is not intended to be a prediction nor vision for the next decade, merely some seasonal fun looking at some unlikely extremes of how our community might develop. After all, we’ve had reports saying “it’s the software” for years — so are the chances of us finally doing anything about it more or less likely than the Planetary Supercomputing Facilities?

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Penguin Computing Brings Cascade Lake-AP to OCP Form Factor

July 7, 2020

Penguin Computing, a subsidiary of SMART Global Holdings, Inc., is announcing a new Tundra server, Tundra AP, that is the first to implement the Intel Xeon Scalable 9200 series processors (codenamed Cascade Lake-AP) in t Read more…

By Tiffany Trader

Google Cloud Debuts 16-GPU Ampere A100 Instances

July 7, 2020

On the heels of the Nvidia's Ampere A100 GPU launch in May, Google Cloud is announcing alpha availability of the A100 "Accelerator Optimized" VM A2 instance family on Google Compute Engine. The instances are powered by t Read more…

By Tiffany Trader

Q&A: HLRS’s Bastian Koller Tackles HPC and Industry in Germany and Europe

July 6, 2020

HPCwire: Let's start with HLRS and work our way up to the European scale. HLRS has stood out in the HPC world for its support of both scientific and industrial research. Can you discuss key developments in recent years? Read more…

By Steve Conway, Hyperion

The Barcelona Supercomputing Center Offers a Virtual Tour of Its MareNostrum Supercomputer

July 6, 2020

With the COVID-19 pandemic continuing to threaten the world and disrupt normal operations, facility tours remain a little difficult to operate, with many supercomputing centers having shuttered facility tours for visitor Read more…

By Oliver Peckham

What’s New in Computing vs. COVID-19: Fugaku, Congress, De Novo Design & More

July 2, 2020

Supercomputing, big data and artificial intelligence are crucial tools in the fight against the coronavirus pandemic. Around the world, researchers, corporations and governments are urgently devoting their computing reso Read more…

By Oliver Peckham

AWS Solution Channel

Maxar Builds HPC on AWS to Deliver Forecasts 58% Faster Than Weather Supercomputer

When weather threatens drilling rigs, refineries, and other energy facilities, oil and gas companies want to move fast to protect personnel and equipment. And for firms that trade commodity shares in oil, precious metals, crops, and livestock, the weather can significantly impact their buy-sell decisions. Read more…

Intel® HPC + AI Pavilion

Supercomputing the Pandemic: Scientific Community Tackles COVID-19 from Multiple Perspectives

Since their inception, supercomputers have taken on the biggest, most complex, and most data-intensive computing challenges—from confirming Einstein’s theories about gravitational waves to predicting the impacts of climate change. Read more…

OpenPOWER Reboot – New Director, New Silicon Partners, Leveraging Linux Foundation Connections

July 2, 2020

Earlier this week the OpenPOWER Foundation announced the contribution of IBM’s A21 Power processor core design to the open source community. Roughly this time last year, IBM announced open sourcing its Power instructio Read more…

By John Russell

Google Cloud Debuts 16-GPU Ampere A100 Instances

July 7, 2020

On the heels of the Nvidia's Ampere A100 GPU launch in May, Google Cloud is announcing alpha availability of the A100 "Accelerator Optimized" VM A2 instance fam Read more…

By Tiffany Trader

Q&A: HLRS’s Bastian Koller Tackles HPC and Industry in Germany and Europe

July 6, 2020

HPCwire: Let's start with HLRS and work our way up to the European scale. HLRS has stood out in the HPC world for its support of both scientific and industrial Read more…

By Steve Conway, Hyperion

OpenPOWER Reboot – New Director, New Silicon Partners, Leveraging Linux Foundation Connections

July 2, 2020

Earlier this week the OpenPOWER Foundation announced the contribution of IBM’s A21 Power processor core design to the open source community. Roughly this time Read more…

By John Russell

Hyperion Forecast – Headwinds in 2020 Won’t Stifle Cloud HPC Adoption or Arm’s Rise

June 30, 2020

The semiannual taking of HPC’s pulse by Hyperion Research – late fall at SC and early summer at ISC – is a much-watched indicator of things come. This yea Read more…

By John Russell

Racism and HPC: a Special Podcast

June 29, 2020

Promoting greater diversity in HPC is a much-discussed goal and ostensibly a long-sought goal in HPC. Yet it seems clear HPC is far from achieving this goal. Re Read more…

Top500 Trends: Movement on Top, but Record Low Turnover

June 25, 2020

The 55th installment of the Top500 list saw strong activity in the leadership segment with four new systems in the top ten and a crowning achievement from the f Read more…

By Tiffany Trader

ISC 2020 Keynote: Hope for the Future, Praise for Fugaku and HPC’s Pandemic Response

June 24, 2020

In stark contrast to past years Thomas Sterling’s ISC20 keynote today struck a more somber note with the COVID-19 pandemic as the central character in Sterling’s annual review of worldwide trends in HPC. Better known for his engaging manner and occasional willingness to poke prickly egos, Sterling instead strode through the numbing statistics associated... Read more…

By John Russell

ISC 2020’s Student Cluster Competition Winners Announced

June 24, 2020

Normally, the Student Cluster Competition involves teams of students building real computing clusters on the show floors of major supercomputer conferences and Read more…

By Oliver Peckham

Supercomputer Modeling Tests How COVID-19 Spreads in Grocery Stores

April 8, 2020

In the COVID-19 era, many people are treating simple activities like getting gas or groceries with caution as they try to heed social distancing mandates and protect their own health. Still, significant uncertainty surrounds the relative risk of different activities, and conflicting information is prevalent. A team of Finnish researchers set out to address some of these uncertainties by... Read more…

By Oliver Peckham

[email protected] Turns Its Massive Crowdsourced Computer Network Against COVID-19

March 16, 2020

For gamers, fighting against a global crisis is usually pure fantasy – but now, it’s looking more like a reality. As supercomputers around the world spin up Read more…

By Oliver Peckham

[email protected] Rallies a Legion of Computers Against the Coronavirus

March 24, 2020

Last week, we highlighted [email protected], a massive, crowdsourced computer network that has turned its resources against the coronavirus pandemic sweeping the globe – but [email protected] isn’t the only game in town. The internet is buzzing with crowdsourced computing... Read more…

By Oliver Peckham

Global Supercomputing Is Mobilizing Against COVID-19

March 12, 2020

Tech has been taking some heavy losses from the coronavirus pandemic. Global supply chains have been disrupted, virtually every major tech conference taking place over the next few months has been canceled... Read more…

By Oliver Peckham

Supercomputer Simulations Reveal the Fate of the Neanderthals

May 25, 2020

For hundreds of thousands of years, neanderthals roamed the planet, eventually (almost 50,000 years ago) giving way to homo sapiens, which quickly became the do Read more…

By Oliver Peckham

DoE Expands on Role of COVID-19 Supercomputing Consortium

March 25, 2020

After announcing the launch of the COVID-19 High Performance Computing Consortium on Sunday, the Department of Energy yesterday provided more details on its sco Read more…

By John Russell

Steve Scott Lays Out HPE-Cray Blended Product Roadmap

March 11, 2020

Last week, the day before the El Capitan processor disclosures were made at HPE's new headquarters in San Jose, Steve Scott (CTO for HPC & AI at HPE, and former Cray CTO) was on-hand at the Rice Oil & Gas HPC conference in Houston. He was there to discuss the HPE-Cray transition and blended roadmap, as well as his favorite topic, Cray's eighth-gen networking technology, Slingshot. Read more…

By Tiffany Trader

Honeywell’s Big Bet on Trapped Ion Quantum Computing

April 7, 2020

Honeywell doesn’t spring to mind when thinking of quantum computing pioneers, but a decade ago the high-tech conglomerate better known for its control systems waded deliberately into the then calmer quantum computing (QC) waters. Fast forward to March when Honeywell announced plans to introduce an ion trap-based quantum computer whose ‘performance’ would... Read more…

By John Russell

Leading Solution Providers

Contributors

Neocortex Will Be First-of-Its-Kind 800,000-Core AI Supercomputer

June 9, 2020

Pittsburgh Supercomputing Center (PSC - a joint research organization of Carnegie Mellon University and the University of Pittsburgh) has won a $5 million award Read more…

By Tiffany Trader

‘Billion Molecules Against COVID-19’ Challenge to Launch with Massive Supercomputing Support

April 22, 2020

Around the world, supercomputing centers have spun up and opened their doors for COVID-19 research in what may be the most unified supercomputing effort in hist Read more…

By Oliver Peckham

Nvidia’s Ampere A100 GPU: Up to 2.5X the HPC, 20X the AI

May 14, 2020

Nvidia's first Ampere-based graphics card, the A100 GPU, packs a whopping 54 billion transistors on 826mm2 of silicon, making it the world's largest seven-nanom Read more…

By Tiffany Trader

Australian Researchers Break All-Time Internet Speed Record

May 26, 2020

If you’ve been stuck at home for the last few months, you’ve probably become more attuned to the quality (or lack thereof) of your internet connection. Even Read more…

By Oliver Peckham

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

15 Slides on Programming Aurora and Exascale Systems

May 7, 2020

Sometime in 2021, Aurora, the first planned U.S. exascale system, is scheduled to be fired up at Argonne National Laboratory. Cray (now HPE) and Intel are the k Read more…

By John Russell

Summit Supercomputer is Already Making its Mark on Science

September 20, 2018

Summit, now the fastest supercomputer in the world, is quickly making its mark in science – five of the six finalists just announced for the prestigious 2018 Read more…

By John Russell

TACC Supercomputers Run Simulations Illuminating COVID-19, DNA Replication

March 19, 2020

As supercomputers around the world spin up to combat the coronavirus, the Texas Advanced Computing Center (TACC) is announcing results that may help to illumina Read more…

By Staff report

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This