HPC Movers and Shakers: Thomas Sterling

By Nicole Hemsoth

October 29, 2010

Ah, New Orleans. The Big Easy. What a great selection for this year’s Supercomputing Conference. It was, in fact, in New Orleans where I first met Thomas Sterling. Thomas, or “Tron” to his friends, graciously invited me to join a small group for dinner one evening during an IEEE conference last summer. We dined at one of the city’s sumptuous restaurants in the Quarter, of course. It was over dinner that warm summer evening that I discovered that this highly respected man, the “father of Beowulf clusters,” has a sharp wit and irrepressible sense of humor that I found absolutely irresistible. There was no doubt in my mind about who my next column would feature as we head to the home of Louisiana State University for SC10.

HPCwire: Thomas, I love your nickname, “Tron.” How did you come by that handle?

Thomas: I had a feeling you were going to bring this up. My nickname Tron goes way back to my Navy days where I was a “tweet”, or aviation electronics technician, repairing F4 Phantom aircraft avionics. When something went really screwy with the electronics, the guys in the shop would blame the “tron god”. Now, I may not have been the most motivated squid in the “nav”, but when there was a really hard problem I was usually the one to tackle it. That’s where “tron” came from; some association with this notional tron god. It’s kind of a compliment, I guess. Later, when serving as an editor on my school paper I used “Tron” as a pen name, and then as my login at MIT, as well as my email address for the inchoate ARPAnet. It’s stayed with me ever since. As Herman Melville might have written, “Call me Tron”.

HPCwire: Speaking of names, you are highly regarded as “the father of Beowulf clusters” and for your research on petaflops computing architecture. I remember that Woody Allen once said you should never take a course where they make you read “Beowulf.” (Laughs) Seriously, though, why did you pick that name for your approach to commodity clusters?

Thomas: (Chuckling) The name was, I should be embarrassed to admit, a complete accident. I was sitting in my office at the Goddard Space Flight Center, and the Program Manager called saying they were sending in the paperwork for my new project and they needed a name for it. I had been putting her off for close to a month. She said she would not hang up until she had a name, or there would not be a project.

I respond well to threats. I needed inspiration and looked around my office in desperation when I noticed at the top of a stack of books my mother’s old copy of Beowulf. I remember saying to the PM, and this is a quote: “Oh hell, just call it ‘Beowulf’; nobody will ever hear of it anyway!” And seriously, that’s how it happened. Someone in the press started calling our Linux clusters “Beowulf-class systems,” so they, whoever they were, get credit for giving the actual clusters that name.

I think there is a lesson here; I’m just not sure what it is.

HPCwire: What’s your stance on shared versus distributed memory? Will we continue to build petascale computers with global shared memory, similar to the legacy products from Sun or SGI, or do you think the programming will be different?

Thomas: This is an important topic and reflects the diversity of experiences that drive perspective, and hence conflict. A major problem is our terminology; our words do not provide us with an effective lexicon to consider all possibilities. For example: “distributed memory.” Does this mean physically separate with intervening distance, blocks of memory not sharing a unified name space, both, or something else? When memory access times are dominated not by the DRAM cycle time but the latency of communication, it is distributed memory. I expect this to be the case for the largest machines of the future. I also expect that hardware support for global address space and unified name spaces are required for efficiency, programmability, and scalability. That sounds like “shared memory.” But now there is that last issue: is it cache coherent? And the answer there is: No, not in the usual sense. But the full explanation to this is too long for a brief interview!

HPCwire: Speaking of which, there are still people out there who long for the days when supercomputers were specialized custom designs and not built out of consumer-grade electronics like Beowulf clusters. What do you say to those people? Will we ever return to using technology specifically crafted for HPC?

Thomas: I am among those who feel that design driven by HPC requirements is essential to advance the field towards Exascale. However, those same changes will be useful for general-purpose and commercial computing as well.

Many ideas first realized in “specialized” HPC designs have migrated into the common general-purpose microprocessor of today. My expectation is that we will continue to use general-purpose devices, but they will change in accordance with the needs of scalability, efficiency, and parallel programmability.

HPCwire: I’ve heard you refer to new “execution models” too. What does that mean? What’s wrong with the current execution models?

Thomas: An execution model is a set of governing principles guiding the co-design and operation of the many interoperable layers of a computing system. It permits the use of the notion of the “decision chain” that recognizes a set of contributing influences. Understanding the decision chain contributes to determining why the operation was performed where and when it was.

Throughout the extraordinary evolution of supercomputing, spanning twelve orders of magnitude in a single lifetime, advancing technologies have required adjustments to the way we organize structures and methods of operation. For instance, a change of balance in bandwidth versus capacity. At least five times we have experienced a revolution in supercomputing. A “6th phase change” is due, and is best represented as a new model of computation.

HPC is in the midst of such a phase change because the technologies are already seen to require different ways of organizing systems, such as multi-core and GPU accelerators. CSP and MPI will not fulfill the needs of all applications on all system classes.

HPCwire: So, you agree with the people who think that MPI is at the end of its rope, and that we need something new, soon. Are you working on alternative programming models?

Thomas: MPI is not near the end of its rope, but we do need something new soon. MPI even in its current form will serve many applications on many systems for many years.

HPCwire: One of your Caltech friends put me up to asking you this one; what do you miss about Caltech?

Thomas: No question, Caltech is a special place, and even after five years I get homesick for it occasionally. There is a mindset there that any fundamental question in science or engineering can be investigated by bright minds to reveal some, if not all, of its secrets. It is the right to strive, to exceed, to understand, that I miss most. Oh, and the strawberry lemonades at the Athenaeum, of course.

HPCwire: So, are there things you can do at LSU that are easier than if you were still at Caltech?

Thomas: Joining the faculty at LSU was a risk, both for LSU and for me! A large state school is a very different environment than boutique intellectual environments such as Caltech, and that concerned me. I had not served as a tenured Professor before and that was a risk for LSU as well. They wanted someone who would rapidly expand their research program in the area of HPC systems with high national exposure. LSU, with support of the state of Louisiana, established the Center for Computation and Technology that complemented the capabilities of the academic departments by providing an advanced environment for interdisciplinary research to foster the goals of both LSU and Louisiana. This has turned out to be a great fit and both LSU and I have benefitted from this new relationship.

HPCwire: What can you share with us about the research you are doing at LSU with “ParalleX”?

Thomas: The research we have undertaken at LSU is risky and driven by the premise that the field of HPC is in that 6th phase change, as discussed above. The ParalleX execution model is a new synthesis of a collection of abstract constructs, relationships, and functional mechanisms to address starvation, latency, overhead, and contention in systems comprising more than a billion simultaneous executing entities with worst-case latencies on the order of a hundred thousand cycles. Recently, the LSU group has teamed with Guang Gao at the University of Delaware, an expert in many related fields, to expand and improve the product of this research under the auspices of the DARPA UHPC Program, the Sandia-led X-Caliber team, and the Intel-led Runnemede team.

HPCwire: This is my ‘live your dream question’: If you could wave a wand and change something about the way HPC is today, what would that thing be?

Thomas: (pause) It basically comes down to having cores designed to operate efficiently in the context of a billion other like cores on a single computational problem, returning us to the notion of a single computer rather than merely a large loose collection of cores using software-managed I/O.

Maybe this does sound like something out of Hogwarts.

HPCwire: I have experienced firsthand your irrepressible sense of humor. What’s the funniest thing you’ve seen in this business?

Thomas: (Laughs) I always get into trouble when I think something is funny but the humor is often lost on others. One time when giving a talk at Los Alamos, I brought up the early energy-efficient computer, “Green Destiny” which, unfortunately, did not perform well, at least initially. I commented: “I don’t get it. Why don’t they just unplug the thing? They’d save even more power and get almost the same performance!” I was never invited back.

At a panel at the Supercomputing conference one year, a member of the audience asked about multithreading and I (without thinking of the consequences) commented that “we can thank Intel for associating the word ‘hyper’ with the number ‘2’.”

I didn’t get invited to a single whisper suite session that year.

HPCwire: Can you share with me two or three interesting things about you that relatively few (or none) of your colleagues or friends know?

Thomas: Attempting to find a human side of me may be a futile endeavor. It may not exist, and I am doubtful that your readers will find anything but my contributions of any interest. However, here goes. One: I love sailing; I used to have a J-105 sloop named “No Compromise.” Two: I am fascinated with the history of Bronze Age cultures and how emerging technologies drive them to ever more complex structures of civilization. Three: I used to enjoy long-distance river kayaking in the Arctic; and I still kayak today, though not to the extent that I used to.

HPCwire: Lastly, what do you consider to be your greatest personal achievement?

Thomas: Not to be trite, but I don’t believe I’ve done it yet. I expect to contribute to the new class of systems capable of Exaflops performance and effective dynamic graph processing for symbolic computing. I believe that such contributions will take the form of the new model of computation such as ParalleX, as a guiding abstraction and its manifestation as a new system software structure and new core architecture for symbiotic operation among billions of executing elements. If I can contribute in some small way to this “6th Phase of HPC,” I will consider that my best personal achievement.

About the Author

An avid HPC watcher and established technology marketing professional; Caroline resides in the California Bay Area and recently joined the HPCwire team as a contributing editor. You can reach her at caroline.connor@longstonegroup.com.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

NSF Awards $10M to Extend Chameleon Cloud Testbed Project

September 19, 2017

The National Science Foundation has awarded a second phase, $10 million grant to the Chameleon cloud computing testbed project led by University of Chicago with partners at the Texas Advanced Computing Center (TACC), Ren Read more…

By John Russell

NERSC Simulations Shed Light on Fusion Reaction Turbulence

September 19, 2017

Understanding fusion reactions in detail – particularly plasma turbulence – is critical to the effort to bring fusion power to reality. Recent work including roughly 70 million hours of compute time at the National E Read more…

Kathy Yelick Charts the Promise and Progress of Exascale Science

September 15, 2017

On Friday, Sept. 8, Kathy Yelick of Lawrence Berkeley National Laboratory and the University of California, Berkeley, delivered the keynote address on “Breakthrough Science at the Exascale” at the ACM Europe Conferen Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

HPE Prepares Customers for Success with the HPC Software Portfolio

High performance computing (HPC) software is key to harnessing the full power of HPC environments. Development and management tools enable IT departments to streamline installation and maintenance of their systems as well as create, optimize, and run their HPC applications. Read more…

U of Illinois, NCSA Launch First US Nanomanufacturing Node

September 14, 2017

The University of Illinois at Urbana-Champaign together with the National Center for Supercomputing Applications (NCSA) have launched the United States's first computational node aimed at the development of nanomanufactu Read more…

By Tiffany Trader

Kathy Yelick Charts the Promise and Progress of Exascale Science

September 15, 2017

On Friday, Sept. 8, Kathy Yelick of Lawrence Berkeley National Laboratory and the University of California, Berkeley, delivered the keynote address on “Breakt Read more…

By Tiffany Trader

DARPA Pledges Another $300 Million for Post-Moore’s Readiness

September 14, 2017

The Defense Advanced Research Projects Agency (DARPA) launched a giant funding effort to ensure the United States can sustain the pace of electronic innovation vital to both a flourishing economy and a secure military. Under the banner of the Electronics Resurgence Initiative (ERI), some $500-$800 million will be invested in post-Moore’s Law technologies. Read more…

By Tiffany Trader

IBM Breaks Ground for Complex Quantum Chemistry

September 14, 2017

IBM has reported the use of a novel algorithm to simulate BeH2 (beryllium-hydride) on a quantum computer. This is the largest molecule so far simulated on a quantum computer. The technique, which used six qubits of a seven-qubit system, is an important step forward and may suggest an approach to simulating ever larger molecules. Read more…

By John Russell

Cubes, Culture, and a New Challenge: Trish Damkroger Talks about Life at Intel—and Why HPC Matters More Than Ever

September 13, 2017

Trish Damkroger wasn’t looking to change jobs when she attended SC15 in Austin, Texas. Capping a 15-year career within Department of Energy (DOE) laboratories, she was acting Associate Director for Computation at Lawrence Livermore National Laboratory (LLNL). Her mission was to equip the lab’s scientists and research partners with resources that would advance their cutting-edge work... Read more…

By Jan Rowell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

MIT-IBM Watson AI Lab Targets Algorithms, AI Physics

September 7, 2017

Investment continues to flow into artificial intelligence research, especially in key areas such as AI algorithms that promise to move the technology from speci Read more…

By George Leopold

Need Data Science CyberInfrastructure? Check with RENCI’s xDCI Concierge

September 6, 2017

For about a year the Renaissance Computing Institute (RENCI) has been assembling best practices and open source components around data-driven scientific researc Read more…

By John Russell

IBM Advances Web-based Quantum Programming

September 5, 2017

IBM Research is pairing its Jupyter-based Data Science Experience notebook environment with its cloud-based quantum computer, IBM Q, in hopes of encouraging a new class of entrepreneurial user to solve intractable problems that even exceed the capabilities of the best AI systems. Read more…

By Alex Woodie

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Leading Solution Providers

IBM Clears Path to 5nm with Silicon Nanosheets

June 5, 2017

Two years since announcing the industry’s first 7nm node test chip, IBM and its research alliance partners GlobalFoundries and Samsung have developed a proces Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

GlobalFoundries: 7nm Chips Coming in 2018, EUV in 2019

June 13, 2017

GlobalFoundries has formally announced that its 7nm technology is ready for customer engagement with product tape outs expected for the first half of 2018. The Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This