Should I Buy GPGPUs or Blue Gene?

By Christopher Lazou

November 4, 2010

The new Tianhe-1A Chinese system with a Linpack performance of 2.5 petaflops, placing it in the number one spot of the new TOP500 list to be presented at SC10 in New Orleans this month, has put “the cat amongst the pigeons” — or should I say the “River in the Sky” — as far as HPC politics in the USA are concerned. But away from the headlines there might be a more tempered reality.

I received a paper from the Department of Computer Science at the University of Warwick, a shorter version of which recently won Best Paper at the Daresbury GPU workshop. An extended version is to be presented at the PMBS workshop at SC10 on Monday, November 15. This paper, “Performance Analysis of a Hybrid MPI/CUDA Implementation of the NAS-LU Benchmark,” (PDF) describes some interesting work being done at Warwick and with access to machines at Lawrence Livermore National Laboratory (LLNL). Essentially their study asks the question: As an organization, should I commit to a platform based on general-purpose GPUs (GPGPUs) or an IBM Blue Gene?

In procuring a new supercomputer, one takes many factors into consideration. Performance, availability and software; potential of the system for future scientific delivery; and viability of the company marketing it, are but a few. This is why the odds are often stacked in favor of established companies to deliver the next successful product. The Dahrendorf dictum that “history proceeds by changing the subject,” however, provides the necessary optimism for aspiring new vendors of radical architectures. And there is a lot of fast-moving history happening in HPC.

As the reader knows there are lots of technical issues tied up in evaluating computer systems and making an informed decision: CPU speed, memory size and bandwidth, communication latency, scalability, capability, electrical power consumption, ease of supporting legacy code, etc. Indeed one needs to take on board the integral of all resources that contribute to the total cost of ownership (TCO). I think this study from the University of Warwick potentially captures the essence of the interesting crossroads at which current HPC finds itself, as ORNL, LLNL and others are now demonstrating.

Using benchmarking and performance modeling, the Warwick team was able to address some of the underlying technical issues, speculating as to the likely performance and power footprint of possible large-scale solutions based on GPGPU and Blue Gene platforms.

Before I offer a perspective of their findings let me clarify what the Warwick study focuses on. After discussing the potential problems facing the HPC industry in its aspiration to deliver exascale systems by 2015-18, they then compared the performance of pipelined wavefront computations (a class of parallel application), running across multiple GPU nodes against an InfiniBand-based cluster of AMD processors and an IBM Blue Gene/P. They augment these runtimes with projections from a recently-developed analytical model of NAS-LU, a computational fluid dynamics benchmark that employs the wavefront algorithm. This study says nothing about other mainstream supercomputers from IBM, Cray, HP, SGI, NEC, Fujitsu, and so on, or other classes of computations, but one can clearly see where their work is heading.

As the reader is aware an interesting race is emerging in supercomputing. In 2011/12 Lawrence Livermore National Laboratory will deploy their 20 petaflops Blue Gene/Q Sequoia system based on future IBM Blue Gene technology. At the same time, Nebulae and Tianhe-1A at the Chinese National Supercomputing Centers and, at a future date, Jaguar at the Oak Ridge Leadership Computing Facility (OLCF), are employing NVIDIA GPUs to attain multi-petaflops systems.

Of course large computing facilities such as LLNL and OLCF buy both, but for those organizations with more modest budgets, a choice must be made?

What makes these architectures different?

The Blue Gene, currently in its fourth technology iteration, owes its design to a previous debate in the late 1990s on how to achieve petaflops for a specific application, namely protein folding. At that time, general-purpose computers could not deliver the needed performance within reasonable power and footprint constraints. To overcome these constraints IBM aptly adopted a reduced instruction set design. To paraphrase Einstein: “A computer (theory) should be as simple as possible, but not simpler.”

The Blue Gene approach to building large supercomputers is to take a large number of relatively-simple processing cores and to connect these via a low latency, highly-scalable interconnect. This has the advantage of creating a high aggregate memory bandwidth (as each core is connected directly to its own memory) whilst maintaining low power consumption because of the low clock frequency and simple design of the processor. The simple nature of the cores makes porting of existing MPI-based codes easier as few modifications are needed, assuming the code presents good scalability. In order to maintain efficient power usage and use of physical space, the Blue Gene/P has a maximum limit of 1GB of memory per execution core.

The Blue Gene architecture is highly rated. The project was awarded the National Medal of Technology and Innovation by U.S. President Barack Obama in late 2009. Its main architect, Alan Gara, is to be awarded the prestigious Seymour Cray medal by IEEE at this year’s Supercomputing Conference in New Orleans.

In contrast, GPU-based machines are being produced from high-end designs based on consumer-grade video and graphics cards — an example of history proceeding by changing the subject. Because of the significant economies, this has the potential to offer high performance at lower cost. The approach utilizes parallelism in the form of a large number of lightweight threads which provide good performance provided each thread executes the same instructions. If the control flow diverges, the penalties can be very costly. In a sense these are a modern equivalent of vector processors but with the ability to simultaneously execute considerably larger numbers of instructions. Currently, most GPU clusters are small scale and are connected by InfiniBand, which requires messages to be copied from the GPU to the main host memory and then from the memory to the remote node.

This “double-penalty” creates a high cost in exchanging data between cards, unlike the Blue Gene system where the low latency interconnect makes message passing relatively inexpensive. The high compute power per GPU concentrates the equivalent processing power into fewer numbers of nodes helping to reduce, but not eliminate, the scalability requirements of the application. However, where communication is needed it is expensive, creating significant problems for applications which need to scale to thousands of GPU devices. Current GPU designs have either 3 GB or 6 GB of memory which, when divided between the execution threads, yields a very small amount of memory per thread — considerably less than conventional clusters based on general-purpose processors or a Blue Gene/P system.

The GPGPU-Blue Gene debate is not simply one of hardware. Application developers are also preparing for change. For many years HPC experts have warned that performance gains to applications from higher clock speeds and more memory per core, such as that seen in the blistering Intel Westmere, are not guaranteed in future architectures. The Blue Gene/P typically has 1 GB of memory per core, which for many application developers is like squeezing an elephant into a mini. An investment is needed to modify the application code to meet this memory constraint. GPU solutions require an even tighter squeeze (6 GB shared memory per 448-core device), not to mention the contortion needed to engineer core code kernels for the GPUs (whilst avoiding canceling out any benefits because of data transfers, etc).

Given that HPC code development and maintenance is the bread and butter of supercomputing programs, and occupies the largest proportion of the overall cost, it is not unreasonable to ask in which direction we should be steering application effort.

What can be learned from current Blue Gene and GPU-based systems?

There are significant installations of both Blue Gene and GPU-based systems. In the June TOP500 list, Lawrence Livermore’s Dawn system, based on Blue Gene/P, clocked in at 415 teraflops and Nebulae, based on GPUs, clocked in at 1.271 petaflops. So what lessons if any can be drawn from these systems?
 
The study from the University of Warwick addresses this question: “Given what we can benchmark on current GPUs and Blue Genes, can we model how an application will behave on such systems at petascale?” The authors of this study, Pennycook, Hammond, Mudalige and Jarvis consider not only what this means in terms of raw performance, or time to solution, but also what this costs in terms of power budget.

Pennycook and his colleagues ask how many Blue Gene cores are needed to get equivalent performance to that achievable from a GPU-based solution. Their work uses extensive benchmarking of HPC-capable GPUs, including the NVIDIA C2050 built on the ‘Fermi’ architecture, alongside Nehalem-class CPUs and the Dawn Blue Gene/P system at LLNL. Performance models are built, for each class of system, which allow them to investigate the performance of applications at scale. Such performance modeling techniques are also used in benchmarking and procurement.

Their work provides some eye-catching results:

1. Taking the NAS-LU parallel benchmark code as an example, the equivalent Class E time-to-solution requires a Blue Gene/P to have 8,192 cores compared to 256 Tesla C2050 cores, 32 times more processing elements than a GPU-based system. This large difference may tempt you, but before running to your nearest GPU outlet to place an order, reflect on this: the processing elements of the Blue Gene solution require around 33 kW, whereas the smaller GPU system requires a maximum of 60 kW.

2. The theoretical peak of the GPU solution is nearly five times that of the Blue Gene. Is this another reason to visit the GPU store? If you are interested in your position in the TOP500 List, then yes, go GPU, but if you are interested in higher sustained performance as a percentage of peak then proceed with caution. The GPU solution clearly outguns the Blue Gene on peak, but achieves an equivalent time-to-solution in the NAS benchmark test.

3. Peak versus actual performance is hotly debated, and this study stokes the fire. The performance results of China’s Nebulae system are revealing, and supportive of this argument. The machine has a theoretical peak of nearly 3 petaflops, but Linpack can currently only deliver 1.271 petaflops of that peak. In contrast the Dawn Blue Gene/P at LLNL has a theoretical peak of 0.5 petaflops and delivers a Linpack performance of 0.415 petaflops. This begs the question: what hope is there for applications, and should an organization be investing in peak or in achievable?

4. GPU single-node performance is second to none. Pennycook et al acknowledge that the single node performance of a GPU is a real win. The same NAS-LU example ran approximately 7 times quicker on the GPU than it did on state-of-the-art CPU-only solutions from Intel and AMD.

But Pennycook is quick to point out that “these headline figures often fail to consider interconnect overheads; we still need to connect these GPUs somehow.”

An interesting observation in their results is that the Blue Gene scales well. So much so that at around 16,000 Blue Gene/P cores, the equivalent time to solution would only need four times fewer GPU processing elements. What this demonstrates is that the GPU-to-Blue Gene ratio is high for smaller systems, but it decreases as the systems get larger. This is significant in terms of power; 16,000 Blue Gene cores require around 66 kW, 4,000 Tesla C2050s require a maximum of 974 kW.

So where does this lead?

The authors of this study state: “The performance of these architectures raises interesting questions about the future direction of HPC architectures – in one case we might expect smaller clusters of SIMT or GPU-based solutions which will favor kernels of highly vectorized code or, alternatively, we might expect highly parallel solutions typified by the Blue Gene/P, where ‘many-core’ will mean massively parallel quantities of independently operating cores.”

The Pennycook study is application specific, “but at the end of the day this is what these supercomputers are designed to support,” he says. Their work is also being extended to applications from Rolls-Royce, AWE and others.

Re-engineering applications for both types of platforms requires significant investment: Blue Genes are memory constrained, have low clock rates and clearly excel at scale, which our current algorithms in many cases do not. GPUs on the other hand require the careful porting of core kernels, which will undoubtedly result in performance gains, but nevertheless needs clustering through effective interconnects, else any gains will be lost.

So what is it going to be, GPU or Blue Gene? It all depends on the size of the system. On first inspection, the GPUs show promising power efficiency, but this is just half the story. Utilizing the available peak of a GPU is a difficult challenge. The Blue Gene, however, is closer to traditional designs, so realizing performance on these platforms presents fewer programming challenges, as long as the algorithms themselves scale.

In my view, this study by the University of Warwick is an invaluable contribution to the debate about emerging architectures and algorithms, in which the HPC industry needs to engage in its pursuit of exascale systems.
 
Enough for now. Just go along to the PMBS 10 workshop on Monday, November 15, in New Orleans and join the debate.

Note: The International Performance Modeling, Benchmarking and Simulation of High Performance Computing Systems workshop (PMBS 10) is part of the SC10 Technical Program. The workshop will take place on Monday, November 15, in rooms 278/279 of the Ernest N. Morial Convention Centre in New Orleans, La.

—–

Copyright ©: Christopher Lazou

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment firm International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

Weekly Twitter Roundup (Jan. 19, 2017)

January 19, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

France’s CEA and Japan’s RIKEN to Partner on ARM and Exascale

January 19, 2017

France’s CEA and Japan’s RIKEN institute announced a multi-faceted five-year collaboration to advance HPC generally and prepare for exascale computing. Among the particulars are efforts to: build out the ARM ecosystem; work on code development and code sharing on the existing and future platforms; share expertise in specific application areas (material and seismic sciences for example); improve techniques for using numerical simulation with big data; and expand HPC workforce training. It seems to be a very full agenda. Read more…

By Nishi Katsuya and John Russell

ARM Waving: Attention, Deployments, and Development

January 18, 2017

It’s been a heady two weeks for the ARM HPC advocacy camp. At this week’s Mont-Blanc Project meeting held at the Barcelona Supercomputer Center, Cray announced plans to build an ARM-based supercomputer in the U.K. while Mont-Blanc selected Cavium’s ThunderX2 ARM chip for its third phase of development. Last week, France’s CEA and Japan’s Riken announced a deep collaboration aimed largely at fostering the ARM ecosystem. This activity follows a busy 2016 when SoftBank acquired ARM, OpenHPC announced ARM support, ARM released its SVE spec, Fujistu chose ARM for the post K machine, and ARM acquired HPC tool provider Allinea in December. Read more…

By John Russell

HPE Extreme Performance Solutions

Remote Visualization: An Integral Technology for Upstream Oil & Gas

As the exploration and production (E&P) of natural resources evolves into an even more complex and vital task, visualization technology has become integral for the upstream oil and gas industry. Read more…

Women Coders from Russia, Italy, and Poland Top Study

January 17, 2017

According to a study posted on HackerRank today the best women coders as judged by performance on HackerRank challenges come from Russia, Italy, and Poland. Read more…

By John Russell

Spurred by Global Ambitions, Inspur in Joint HPC Deal with DDN

January 17, 2017

Inspur, the fast-growth cloud computing and server vendor from China that has several systems on the current Top500 list, and DDN, a leader in high-end storage, have announced a joint sales and marketing agreement to produce solutions based on DDN storage platforms integrated with servers, networking, software and services from Inspur. Read more…

By Doug Black

Weekly Twitter Roundup (Jan. 12, 2017)

January 12, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

NSF Seeks Input on Cyberinfrastructure Advances Needed

January 12, 2017

In cased you missed it, the National Science Foundation posted a “Dear Colleague Letter” (DCL) late last week seeking input on needs for the next generation of cyberinfrastructure to support science and engineering. Read more…

By John Russell

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment firm International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

France’s CEA and Japan’s RIKEN to Partner on ARM and Exascale

January 19, 2017

France’s CEA and Japan’s RIKEN institute announced a multi-faceted five-year collaboration to advance HPC generally and prepare for exascale computing. Among the particulars are efforts to: build out the ARM ecosystem; work on code development and code sharing on the existing and future platforms; share expertise in specific application areas (material and seismic sciences for example); improve techniques for using numerical simulation with big data; and expand HPC workforce training. It seems to be a very full agenda. Read more…

By Nishi Katsuya and John Russell

ARM Waving: Attention, Deployments, and Development

January 18, 2017

It’s been a heady two weeks for the ARM HPC advocacy camp. At this week’s Mont-Blanc Project meeting held at the Barcelona Supercomputer Center, Cray announced plans to build an ARM-based supercomputer in the U.K. while Mont-Blanc selected Cavium’s ThunderX2 ARM chip for its third phase of development. Last week, France’s CEA and Japan’s Riken announced a deep collaboration aimed largely at fostering the ARM ecosystem. This activity follows a busy 2016 when SoftBank acquired ARM, OpenHPC announced ARM support, ARM released its SVE spec, Fujistu chose ARM for the post K machine, and ARM acquired HPC tool provider Allinea in December. Read more…

By John Russell

Spurred by Global Ambitions, Inspur in Joint HPC Deal with DDN

January 17, 2017

Inspur, the fast-growth cloud computing and server vendor from China that has several systems on the current Top500 list, and DDN, a leader in high-end storage, have announced a joint sales and marketing agreement to produce solutions based on DDN storage platforms integrated with servers, networking, software and services from Inspur. Read more…

By Doug Black

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

UberCloud Cites Progress in HPC Cloud Computing

January 10, 2017

200 HPC cloud experiments, 80 case studies, and a ton of hands-on experience gained, that’s the harvest of four years of UberCloud HPC Experiments. Read more…

By Wolfgang Gentzsch and Burak Yenier

A Conversation with Women in HPC Director Toni Collis

January 6, 2017

In this SC16 video interview, HPCwire Managing Editor Tiffany Trader sits down with Toni Collis, the director and founder of the Women in HPC (WHPC) network, to discuss the strides made since the organization’s debut in 2014. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

AWS Beats Azure to K80 General Availability

September 30, 2016

Amazon Web Services has seeded its cloud with Nvidia Tesla K80 GPUs to meet the growing demand for accelerated computing across an increasingly-diverse range of workloads. The P2 instance family is a welcome addition for compute- and data-focused users who were growing frustrated with the performance limitations of Amazon's G2 instances, which are backed by three-year-old Nvidia GRID K520 graphics cards. Read more…

By Tiffany Trader

US, China Vie for Supercomputing Supremacy

November 14, 2016

The 48th edition of the TOP500 list is fresh off the presses and while there is no new number one system, as previously teased by China, there are a number of notable entrants from the US and around the world and significant trends to report on. Read more…

By Tiffany Trader

Vectors: How the Old Became New Again in Supercomputing

September 26, 2016

Vector instructions, once a powerful performance innovation of supercomputing in the 1970s and 1980s became an obsolete technology in the 1990s. But like the mythical phoenix bird, vector instructions have arisen from the ashes. Here is the history of a technology that went from new to old then back to new. Read more…

By Lynd Stringer

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

Container App ‘Singularity’ Eases Scientific Computing

October 20, 2016

HPC container platform Singularity is just six months out from its 1.0 release but already is making inroads across the HPC research landscape. It's in use at Lawrence Berkeley National Laboratory (LBNL), where Singularity founder Gregory Kurtzer has worked in the High Performance Computing Services (HPCS) group for 16 years. Read more…

By Tiffany Trader

Dell EMC Engineers Strategy to Democratize HPC

September 29, 2016

The freshly minted Dell EMC division of Dell Technologies is on a mission to take HPC mainstream with a strategy that hinges on engineered solutions, beginning with a focus on three industry verticals: manufacturing, research and life sciences. "Unlike traditional HPC where everybody bought parts, assembled parts and ran the workloads and did iterative engineering, we want folks to focus on time to innovation and let us worry about the infrastructure," said Jim Ganthier, senior vice president, validated solutions organization at Dell EMC Converged Platforms Solution Division. Read more…

By Tiffany Trader

Lighting up Aurora: Behind the Scenes at the Creation of the DOE’s Upcoming 200 Petaflops Supercomputer

December 1, 2016

In April 2015, U.S. Department of Energy Undersecretary Franklin Orr announced that Intel would be the prime contractor for Aurora: Read more…

By Jan Rowell

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. Read more…

By John Russell

Leading Solution Providers

D-Wave SC16 Update: What’s Bo Ewald Saying These Days

November 18, 2016

Tucked in a back section of the SC16 exhibit hall, quantum computing pioneer D-Wave has been talking up its new 2000-qubit processor announced in September. Forget for a moment the criticism sometimes aimed at D-Wave. This small Canadian company has sold several machines including, for example, ones to Lockheed and NASA, and has worked with Google on mapping machine learning problems to quantum computing. In July Los Alamos National Laboratory took possession of a 1000-quibit D-Wave 2X system that LANL ordered a year ago around the time of SC15. Read more…

By John Russell

CPU Benchmarking: Haswell Versus POWER8

June 2, 2015

With OpenPOWER activity ramping up and IBM’s prominent role in the upcoming DOE machines Summit and Sierra, it’s a good time to look at how the IBM POWER CPU stacks up against the x86 Xeon Haswell CPU from Intel. Read more…

By Tiffany Trader

Nvidia Sees Bright Future for AI Supercomputing

November 23, 2016

Graphics chipmaker Nvidia made a strong showing at SC16 in Salt Lake City last week. Read more…

By Tiffany Trader

Beyond von Neumann, Neuromorphic Computing Steadily Advances

March 21, 2016

Neuromorphic computing – brain inspired computing – has long been a tantalizing goal. The human brain does with around 20 watts what supercomputers do with megawatts. And power consumption isn’t the only difference. Fundamentally, brains ‘think differently’ than the von Neumann architecture-based computers. While neuromorphic computing progress has been intriguing, it has still not proven very practical. Read more…

By John Russell

The Exascale Computing Project Awards $39.8M to 22 Projects

September 7, 2016

The Department of Energy’s Exascale Computing Project (ECP) hit an important milestone today with the announcement of its first round of funding, moving the nation closer to its goal of reaching capable exascale computing by 2023. Read more…

By Tiffany Trader

Dell Knights Landing Machine Sets New STAC Records

November 2, 2016

The Securities Technology Analysis Center, commonly known as STAC, has released a new report characterizing the performance of the Knight Landing-based Dell PowerEdge C6320p server on the STAC-A2 benchmarking suite, widely used by the financial services industry to test and evaluate computing platforms. The Dell machine has set new records for both the baseline Greeks benchmark and the large Greeks benchmark. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

What Knights Landing Is Not

June 18, 2016

As we get ready to launch the newest member of the Intel Xeon Phi family, code named Knights Landing, it is natural that there be some questions and potentially some confusion. Read more…

By James Reinders, Intel

  • arrow
  • Click Here for More Headlines
  • arrow
Share This