Researchers Squeeze GPU Performance from 11 Big Science Apps

By Michael Feldman

July 18, 2012

The GPGPU faithful received another round of encouraging news this week. In a report  published this week, researchers documented that GPU-equipped supercomputers enabled application speedups between 1.4x and 6.1x across a range of well-known science codes. While those results aren’t the order of magnitude performance increases that were being bandied about in the early days of GPU computing, the researchers were encouraged that the technology is producing consistently good results with some of the most popular HPC science applications in the world.

The work was presented in March at the Accelerating Computational Science Symposium, an event devoted to understanding the use of hybrid supercomputers for scientific research. The ensuing report published by the Oak Ridge Leadership Computing Facility, detailed the performance GPU acceleration across the science application spectrum — biology, chemical physics, combustion, nuclear fission and fusion, material science, seismology, molecular dynamics, and climatology.

The 11 simulation codes tested –  S3D, Denovo, LAMMPS, WL-LSMS, CAM-SE, NAMD, Chroma, QMCPACK, SPECFEM-3D, GTC, and CP2K — are used by tens of thousands of researchers worldwide. NAMD alone has over 50 thousand users.

It should be noted that all of the principle participants at the symposium, including Oak Ridge National Laboratory (ORNL), the National Center for Supercomputing Applications (NCSA) and the Swiss National Supercomputing Center (CSCS), not to mention symposium sponsors Cray and NVIDIA, have a stake in proving the viability of GPU-accelerated supercomputing. The three supercomputing centers recently made substantial investments in GPU-based HPC, ORNL with its upcoming 20-plus-petaflop Titan system, NCSA with the 10-petaflop Blue Waters supercomputer, and CSCS with its currently installed 176-node Todi machine.

Titan, Blue Waters and Todi are all Cray supercomputers with varying amounts of AMD Opteron and NVIDIA Tesla horsepower, although none with greater than a 1:1 GPU-to-CPU ratio. That assumes a certain balance in the application between the sequential pieces of the code that would best be run on the CPU and the parallel components that would be candidates for the GPU. But applications can have very different needs in this regard, so that hardware ratio may not always be optimal. Vendors such as HP, Dell, Appro and others offer systems with much higher ratios of GPU to CPUs.

To level the playing field as much as possible, the performance runs for the science apps were made on CSCS’s Monte Rosa, a Cray XE6 machine equipped with two AMD “Interlagos” (Opteron 6200) CPUs per node, and TitanDev, a XK6 Titan-based testbed that consists of hybrid nodes, each of which contain one NVIDIA Fermi GPU and one Interlagos CPU . So in essence, the applications were tested on the same two systems, one of which replaced the second CPU with a GPU in each node. Here are the results:

Application

Performance

XK6 vs XE6

Software Framework

S3D

Turbulent combustion

1.4 OpenACC

NAMD

Molecular dynamics

1.4 CUDA

CP2K

Chemical physics

1.5  CUDA

CAM-SE

Community atmosphere model

1.5 PGI CUDA Fortran

WL-LSMS

Statistical mechanics of magnetic materials

1.6  CUDA

GTC/GTC-GPU

Plasma physics for fusion energy

 1.6  CUDA

 SPECFEM-3D

Seismology

 2.5  CUDA

 QMCPACK

Electronic structure of materials

 3.0  CUDA

 LAMMPS

Molecular dynamics

 3.2  CUDA

 Denovo

3D neutron transport for nuclear reactors

 3.3  CUDA

 Chroma

Lattice quantum chromodynamics

 6.1  CUDA

According to this, the Fermi GPU-equipped XK6 was able to extract between 140 and 610 percent of the application performance compared to the CPU-only XE6. As CSCS director Thomas Schulthess observed at the symposium, that takes into account the fact the Interlagos Opteron is a new x86 processor, while Fermi is a two-year-old design. The implication is that the upcoming Kepler K20 GPU, which is supposed to be available later this year (and which will be deployed in Titan and Blue Waters), should widen the CPU-GPU performance gap even more.

“It’s going to be interesting to see in the next few years if there’s going to be a small avalanche, or is a big avalanche coming that’s really going to revolutionize computational science.” said Schulthess.

Even though the researchers provided an apples-to-apples comparison from a hardware perspective, the application software implementation for the two architectures is, by definition, rather different. Although the report did not delve too deeply into the software frameworks, most of these GPU codes incorporated CUDA or CUDA-based libraries. Only two of the applications, CAM-SE and S3D, used a higher level programming approach: PGI’s CUDA Fortran compiler for CAM-SE and OpenACC directives (compiler unknown) for the S3D implementation. Neither of these did particularly well, relative to the performance increases for the other applications, but there are not enough examples here to make any generalizations.

The other thing to keep in mind is that is no guarantee that the code implementations for either the CPU-only or hybrid versions are optimal at extracting the maximum performance from the silicon. A Fermi-class Tesla M2090 module delivers 665 gigaflops of peak performance, which is about 5 or 6 times that of a high-end Opteron 6200. The only code that appeared to fully exploit the performance advantage of the GPU was Chroma, the code for high energy and nuclear physics. Since applications vary significantly in their potential to utilize a highly threaded architecture like a GPU, this should come as no surprise.

Another aspect that needs to be taken into account is power usage. Although the performance comparison between the two processors is a useful one, if codes can scale equally well on a CPU as a GPU, performance per watt becomes a more valid criteria. Since these GPU accelerators consume about twice the power of a high-end x86 under full load, that means each hybrid node uses 50 percent more power than the corresponding CPU-only one when those systems are running at peak.

That suggests that the GPU-accelerated version of these codes should probably run at least 1.5 times as fast in this configuration to keep performance per watt in line. (Note that half of these codes are clustered around that break-even point.) To be fair, that’s not precisely true, since when the graphics engine is not being fully utilized it won’t be drawing anything near its maximum wattage; in general the GPU is much more efficient at throughput computing than its CPU brethren. But the fact remains that the power-performance behavior of the codes needs to be factored in when you’re considering the advantages of GPU acceleration.

Another missing piece of this comparison is how well these same applications would run on NVIDIA’s HPC competition, namely Intel’s Xeon Phi (aka MIC) coprocessor and its very different software ecosystem. Of course, there is no Xeon Phi yet, so that comparison can’t yet be made. But by this time next year, teraflop-capable MIC and Kepler chips should be in crunching away at applications on production machines. At that point, the case for accelerated science codes could be even more compelling.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Long Flights to Cluster Fights: Meet the Asian Student Cluster Teams

November 22, 2017

Five teams from Asia traveled thousands of miles to compete at the SC17 Student Cluster Competition in Denver. Our cameras were there to meet ‘em, greet ‘em, and grill ‘em about their clusters and how they’re doi Read more…

By Dan Olds

Japan Unveils Quantum Neural Network

November 22, 2017

The U.S. and China are leading the race toward productive quantum computing, but it's early enough that ultimate leadership is still something of an open question. The latest geo-region to throw its hat in the quantum co Read more…

By Tiffany Trader

Perspective: What Really Happened at SC17?

November 22, 2017

SC is over. Now comes the myriad of follow-ups. Inboxes are filled with templated emails from vendors and other exhibitors hoping to win a place in the post-SC thinking of booth visitors. Attendees of tutorials, workshop Read more…

By Andrew Jones

HPE Extreme Performance Solutions

HPE Wins “Best HPC Server” for the Apollo 6000 Gen10 System

Hewlett Packard Enterprise (HPE) was nominated for 14 HPCwire Readers’ and Editors’ Choice Awards—including “Best High Performance Computing (HPC) Server Product or Technology” and “Top Supercomputing Achievement.” The HPE Apollo 6000 Gen10 was named “Best HPC Server” of 2017. Read more…

Turnaround Complete, HPE’s Whitman Departs

November 22, 2017

Having turned around the aircraft carrier the Silicon Valley icon had become, Meg Whitman is leaving the helm of a restructured Hewlett Packard. Her successor, technologist Antonio Neri will now guide what Whitman assert Read more…

By George Leopold

Long Flights to Cluster Fights: Meet the Asian Student Cluster Teams

November 22, 2017

Five teams from Asia traveled thousands of miles to compete at the SC17 Student Cluster Competition in Denver. Our cameras were there to meet ‘em, greet ‘em Read more…

By Dan Olds

Perspective: What Really Happened at SC17?

November 22, 2017

SC is over. Now comes the myriad of follow-ups. Inboxes are filled with templated emails from vendors and other exhibitors hoping to win a place in the post-SC Read more…

By Andrew Jones

SC Bids Farewell to Denver, Heads to Dallas for 30th Anniversary

November 17, 2017

After a jam-packed four-day expo and intensive six-day technical program, SC17 has wrapped up another successful event that brought together nearly 13,000 visit Read more…

By Tiffany Trader

SC17 Keynote – HPC Powers SKA Efforts to Peer Deep into the Cosmos

November 17, 2017

This week’s SC17 keynote – Life, the Universe and Computing: The Story of the SKA Telescope – was a powerful pitch for the potential of Big Science projects that also showcased the foundational role of high performance computing in modern science. It was also visually stunning. Read more…

By John Russell

How Cities Use HPC at the Edge to Get Smarter

November 17, 2017

Cities are sensoring up, collecting vast troves of data that they’re running through predictive models and using the insights to solve problems that, in some Read more…

By Doug Black

Student Cluster LINPACK Record Shattered! More LINs Packed Than Ever before!

November 16, 2017

Nanyang Technological University, the pride of Singapore, utterly destroyed the Student Cluster Competition LINPACK record by posting a score of 51.77 TFlop/s a Read more…

By Dan Olds

Hyperion Market Update: ‘Decent’ Growth Led by HPE; AI Transparency a Risk Issue

November 15, 2017

The HPC market update from Hyperion Research (formerly IDC) at the annual SC conference is a business and social “must,” and this year’s presentation at S Read more…

By Doug Black

Nvidia Focuses Its Cloud Containers on HPC Applications

November 14, 2017

Having migrated its top-of-the-line datacenter GPU to the largest cloud vendors, Nvidia is touting its Volta architecture for a range of scientific computing ta Read more…

By George Leopold

US Coalesces Plans for First Exascale Supercomputer: Aurora in 2021

September 27, 2017

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, in Arlington, Va., yesterday (Sept. 26), it was revealed that the "Aurora" supercompute Read more…

By Tiffany Trader

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

AMD Showcases Growing Portfolio of EPYC and Radeon-based Systems at SC17

November 13, 2017

AMD’s charge back into HPC and the datacenter is on full display at SC17. Having launched the EPYC processor line in June along with its MI25 GPU the focus he Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

GlobalFoundries Puts Wind in AMD’s Sails with 12nm FinFET

September 24, 2017

From its annual tech conference last week (Sept. 20), where GlobalFoundries welcomed more than 600 semiconductor professionals (reaching the Santa Clara venue Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Leading Solution Providers

SC17 Booth Video Tours

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

Intel Launches Software Tools to Ease FPGA Programming

September 5, 2017

Field Programmable Gate Arrays (FPGAs) have a reputation for being difficult to program, requiring expertise in specialty languages, like Verilog or VHDL. Easin Read more…

By Tiffany Trader

HPC Chips – A Veritable Smorgasbord?

October 10, 2017

For the first time since AMD's ill-fated launch of Bulldozer the answer to the question, 'Which CPU will be in my next HPC system?' doesn't have to be 'Whichever variety of Intel Xeon E5 they are selling when we procure'. Read more…

By Dairsie Latimer

Flipping the Flops and Reading the Top500 Tea Leaves

November 13, 2017

The 50th edition of the Top500 list, the biannual publication of the world’s fastest supercomputers based on public Linpack benchmarking results, was released Read more…

By Tiffany Trader

IBM Advances Web-based Quantum Programming

September 5, 2017

IBM Research is pairing its Jupyter-based Data Science Experience notebook environment with its cloud-based quantum computer, IBM Q, in hopes of encouraging a new class of entrepreneurial user to solve intractable problems that even exceed the capabilities of the best AI systems. Read more…

By Alex Woodie

Share This