Visit additional Tabor Communication Publications
November 19, 2010
If there was a dominating theme at the Supercomputing Conference this year, it had to be GPU computing. From the influx of GPU-accelerated systems on the TOP500, including the number one system in the world to the inclusion of GPGPUs into nearly every discussion of exascale machines to the visibility of GPUs across the exhibition hall, the technology seemed to be ubiquitous at SC10.
Arguably, the biggest vendor announcement at the show was the launch of SGI's Prism XL machine, and although that system is designed as a general-purpose platform for various kinds of HPC accelerators, it's almost a given that the vast majority will be shipped with GPUs.
Today, every major and minor HPC system vendor now offers GPU-equipped servers, with plans by many to expand their portfolio over the next year. And that can only mean the customer demand for such technology is now palpable. In fact, if you aspire to be an HPC OEM or software provider and don't have a GPU strategy, the next few years are going to be mighty lonely.
But not everyone at SC10 was hopping on the GPU bandwagon. (And I'm not just talking about the Convey folks.) There is a definite divide in the HPC application community about the value of graphics processors for science codes. I spoke with a number of developers who had played with GPUs and found they couldn't realize that magical 10X performance bump they felt they needed to commit their applications to a new platform. Although there are plenty of technical computing applications that have been ported to CUDA, many -- the majority, in fact -- have not.
CAPS enterprise, makers of GPU-friendly compiler tools, offers a support service for porting codes to GPUs and found that 10X speedups should be considered quite good for an HPC application. According the them, getting to 100X or beyond would be attainable only by those algorithms that are not memory-bound, that is, those dominated by computation rather than memory access. Most of the customer applications they've worked with have been able to achieve between 2X and 10X performance increases when ported to GPUs, and sometimes that's not enough for to justify a platform change. In some cases, reworking of the CPU component, alone, achieved a significant speedup. Only about half of the CAPS customers that were considering ports have made the jump to GPGPUs.
In talking with people here at SC10 and at NVIDIA's GPU Technology Conference in September, my impression is that the bigger, older codes are more resistance to being ported to GPUs than smaller and newer ones. And it makes perfect sense. In many cases, those older codes are no longer attached to their original developers, which makes transforming the algorithms into a GPU-friendly design (or any design) that much harder. Also, legacy codes tend to have accumulated kludges and tweaks that make such redesigns extremely painful. This feeds into the human aspect of software engineering, where the if-it-aint-broke-don't-fix-it crowd often dominates the software maintenance mentality.
This might help to explain the slow response of the US and Europe to adopt GPU-equipped supercomputers, at least at the level of the large national labs and universities. After all, this is where many of those legacy HPC codes are developed and maintained. That said, I suspect there are actually more GPU-accelerated clusters in the US and Europe than anywhere else; it's the petascale systems that have not been forthcoming. At this point, the West is at least a year behind China and Japan in the GPGPU supercomputer arms race.
GPU computing skeptics can also point to evidence that there are better architectures for supercomputing already out there, or soon to be launched. For example, despite the enviable performance per watt of the graphics processor, the number one system on the just-announced Green500 list is a Blue Gene/Q prototype system. Of course, that's cheating a bit, given that production Blue Gene/Q systems don't yet exist. But the prototype Q did manage to beat the state-of-the-art TSUBAME 2.0 GPU supercomputer rather handily -- 1684 megaflops/watt to 984 megaflops/watt. I suspect the "green" matchup will be much closer in 2011, when NVIDIA's next-generation "Kepler" hardware and Blue Gene/Q are both in the field.
Also, the top system on the new Graph 500 list was the IBM Blue Gene/P system at Argonne National Lab. The Graph 500 attempts to measure the suitability of platforms for data analytics-type workloads, which is not the strong suit of the graphics processor, at least in its current incarnation. Graph problems require an architecture that can do a lot of random data accesses across memory at a very high rate. Few conventional computing architectures -- CPU, GPU or otherwise -- are any good at this.
Committed GPU computing dissenters are likely pinning their hopes on Intel's Many Integrated Core (MIC) architecture, which is designed to address the same problem space as GPGPUs, but does so with a conventional x86 architecture. For the risk-averse, there is certainly an allure to recompiling your legacy source code with a future Intel compiler that will automagically spit out MIC code. But waiting until 2012 to see if that chip and compiler deliver as advertised could be the riskiest bet of all. Of course, we'll have to wait until SC12 to see how this story turns out.
Posted by Michael Feldman - November 19, 2010 @ 3:41 PM, Pacific Standard Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.