Since 1987 - Covering the Fastest Computers in the World and the People Who Run Them

January 29, 2013

Experts Discuss the Future of Supercomputers

Tiffany Trader

Science Live hosted an online chat last Thursday, entitled The Future of Supercomputers, featuring special guests and prominent HPC experts Jack Dongarra from the University of Tennessee and Horst Simon from Lawrence Berkeley National Laboratory. The discussion focused on the next big thing in supercomputing, the coming class of exascale systems and all that entails, namely, developing useful machines that are hundreds of times faster than the current best-of-class.

Science Magazine‘s Robert Service begins by asking a basic, yet crucial question: Why? Why do we need this level of computing power? Jack Dongarra responds that the benefits reach into virtually every segment of technology and science, from energy research to life science, manufacturing and even entertainment. He argues that the more powerful a nation’s computer capability, the better it can compete on the global playing field. What’s more, there are “considerable flow-down benefits” to the entire IT field, from smaller computer systems all the way to handheld consumer devices.

Concerning the main challenges to fielding such systems, Horst Simon echoed a common opinion: power consumption.

“Extrapolation from today’s technology to the exascale would lead to systems with 100 MW or more power requirements,” noted Simon, using the current TOP500 chart-topper Titan as an example. The ORNL machine requires 8 MW to output 20 petaflops, so a similar exaflops machine would require 400 MW. He estimates the cost to operate such a system at $400 million per year.

The HPCers both emphasized that breaking the exascale barrier will require a concerted effort on multiple fronts, as well as major changes to hardware, software, algorithms and application – it’s an ecosystem.

In Simon’s words: “There is no single revolutionary strategy that will get us to Exaflops. It will have to be several breakthroughs that need to be achieved in the same timeframe. That makes it so hard.”

Simon and Dongarra also stressed the major investment that will be required, especially if the US is to maintain its technological lead.

Simon observed that China has the “budget and willingness” to beat the US to exascale, but in his opinion, they tend to be emulators rather than true innovators. Sometimes this strategy pays off, however, as when China leveraged US GPU expertise to achieve supercomputing dominance in 2010 with Tianhe-1A.

Just the cost of building an exascale-class system in the 2020 timeframe is $200 million, according to Dongarra’s best estimate, and that figure is only for the actual machine – it does not take into account the difficult research that comes beforehand.

At about the half-way mark, we were reminded that this was very much an open forum when an audience member chimed in to ask what “HPC” stood for. What a perfect opportunity to consider the role of community outreach in encouraging public support for HPC.

The remainder of this one-hour talk delves into a multitude of important topics as they relate to exascale computing, including the relevance of the Linpack benchmark, programmability challenges, the stagnation of government funding, environmental implications, and the role of international collaboration. The entire transcript is available online and is well worth the 10 or 15 minutes it takes to read.

Thank you to Science Magazine and writer/moderator Robert Service for hosting this important conversation.

Share This