What Power8 and OpenPOWER Might Mean for HPC

By Timothy Prickett Morgan

April 23, 2014

IBM is making a big play in hybrid computing, seeking to marry its POWER8 processors with various kinds of accelerators and high-speed networking and opening up its chip and system software through the OpenPOWER Foundation. At the Open Innovation Summit in San Francisco today, IBM and its foundation partners talked about how their collaboration would deliver machines better tuned for hyperscale Web applications and data analytics, but the same technologies will no doubt also be deployed in traditional supercomputing environments.

One could argue that what IBM is trying to accomplish with the OpenPOWER Foundation in an open way with many partners up and down the hardware and software stacks is based on ideas it put to the test many years ago in the “Roadrunner” petascale-class x86-Cell hybrid system at Los Alamos National Laboratory. This was the first large-scale accelerated system, and IBM seeks to make what was exotic with Roadrunner six years ago easier to do today and something more akin to normal in the years ahead.

IBM OpenPOWER Rosamilia Balog
At the Open Innovation Summit in San Francisco co-hosted by IBM and the OpenPOWER Foundation, Senior Vice President of IBM’s Systems and Technology Group Tom Rosamilia (left) and IBM General Manager of Power Systems Doug Balog (right) revealed the company’s POWER8 Systems. (Photo credit: Monica Davey/Feature Photo Service for IBM)

The POWER8 chip that was previewed today is about more than just increasing the core count by 50 percent and boosting the performance per socket on the order of 2X to 2.5X for various workloads compared to the POWER7 processors. IBM wants for all manner of accelerators to link very tightly with POWER8 processors and for scale-out clusters based on these chips to have the right kind of acceleration – be it a GPU, an FPGA, a DSP, or some other component – necessary for their particular jobs.

“The way I like to think about it is that with the POWER8 processor, we have created a superlane highway, and now we need cars driving on it,” explained Doug Balog, general manager of IBM’s Power Systems division, at the summit when announcing the new POWER8-based systems from IBM. “Very fast, very efficient vehicles. It is about that open interface that allows for that accelerator to attach to the POWER system.”

One of the interfaces that will be available to link accelerators to the POWER8 processors is IBM’s own Coherent Accelerator Processor Interface, or CAPI for short. This is an overlay on top of the PCI-Express 3.0 controllers on the POWER8 chip that will allow for high-speed linking between the CPU and accelerators and, more importantly, present a shared virtual memory space to applications across the memory attached to the CPU and any memory attached to the accelerator. Moving data back and forth between the CPU memory and the accelerator memory is a big hindrance to performance on accelerated machines, so making all memory addressable to all compute components is important.

Separately from IBM’s CAPI effort, but aligning nicely with it, NVIDIA has come up with its own NVLink interconnect, which will be used to hook its Tesla GPUs to POWER8 (and perhaps other) processors as well as to each other. Sumit Gupta, general manager of Tesla GPU Accelerated Computing Business unit at NVIDIA, said that NVLink would be incorporated into a future POWER processor design from IBM, and reminded everyone that the interconnect was part of the “Pascal” generation of GPUs from NVIDIA due in 2016 or so. NVIDIA is also going to license the technology behind NVLink to members of the OpenPOWER Foundation, Gupta said, and added that the two companies were in the meantime working on accelerating applications that combine POWER and Tesla compute. In the fourth quarter of this year, Gupta said, NVIDIA will deliver full support of its CUDA development environment for CPU-GPU hybrids on POWER processors, and in fact, IBM will also start shipping POWER8 systems that include Tesla GPU coprocessors.

“At the end of the day, the system is only as good as the software that takes advantage of it,” Gupta explained, “and that is why the future and the long-term focus is going to be around software.”

IBM is working with FPGA makers Xilinx and Altera to show the benefits of a hybrid setup running over the CAPI interface, so this is not just about GPU acceleration. Next week at the Impact2014 event, IBM and Xilinx will show a Memcached key value store application being accelerated by FPGAs and showing a factor of 35X better performance and an order of magnitude lower latency. A Monte Carlo simulation running on POWER machines accelerated by Altera FPGAs will show a factor of 200X speedup. Network adapter and switch maker Mellanox Technologies is also working with IBM to show how using Remote Direct Memory Access (RDMA) with a different key value store application boosted throughput and cut latencies by a factor of 10X.

There are a number of other benefits that come with POWER8 chips aside from high-speed links between processors and accelerators. The chips have a native little endian memory storage and accessing method, which is what x86 processors have and which stands in contrast to big endian ordering in memory. The important thing is this: with both x86 and POWER8 supporting little endian memory, applications that are coded from x86 systems to POWER8 systems are “a recompilation, a test, and a go,” as Balog put it. This is particularly important for C, C++, and Fortran and obviously has no bearing on interpreted languages like Java.

While the OpenPOWER Foundation members have not explicitly said they are seeking to accelerate traditional HPC applications, it is clear that these are core technologies that have been used – often first – in HPC systems. And there is no reason to believe that if IBM and its friends, including search engine giant Google, come up with more efficient ways of running plain vanilla POWER systems as well as hybrid machines that mix and match POWER chips and accelerators that these scale out systems will not end up in government and academic HPC centers. It will, as always, come down to having machines that suit a particular workload and at the right price. Suffice it to say that if IBM and its OpenPOWER Foundation partners can make a server that is low-cost enough to appeal to hyperscale datacenter operators, this will no doubt pique the interest of HPC centers.

In the meantime, IBM has divulged some details of the five new machines that it will sell based on the POWER8 processors. These are called the POWER Systems S-Class, and the “S” stands for scale out. Given the overwhelming adoption of Linux by the HPC community, the two Linux-only machines in the new line are probably the most appropriate, but one box that has slightly faster clock speeds might be interesting to HPC customers depending on what IBM charges for them.

The POWER S822L is a two-socket machine that fits into a 2U chassis, and it is probably the one that HPC shops will look at first. The system has room for two processor cards, which plug into the system board, and each one of them has a maximum of 512 GB of main memory. Customers can choose a POWER8 processor with ten of its twelve cores activated and running at 3.42 GHz or one with all twelve cores activated but running at only 3.02 GHz. This machine does not support cheap SATA disks, but does have room for a dozen 2.5-inch SAS drives or SSDs. It also has a storage cage with room for six 1.8-inch SSDs and has nine PCI-Express 3.0 slots. This machine will be available on June 10, and it supports Red Hat Enterprise Linux 6.5, SUSE Linux Enterprise Server 11 SP3, and Canonical Ubuntu Server 14.04 LTS. It can also run the POWERKVM hypervisor, a variant of the KVM hypervisor that IBM has created in conjunction with Red Hat and Canonical.

The POWER S812L comes with the same 2U chassis but only has one processor in it (the same options as the two-socket machine above) and only has six PCI-Express 3.0 peripheral slots. This machine will not ship until August, and will probably be less appealing to HPC shops because of its lower compute density compared to the POWER S822L.

The third machine that might see some HPC play is the POWER S822, which can run IBM’s AIX 6.1 or 7.1 operating systems or RHEL 6.5 or SLES 11 SP3. (It has not been certified to run Ubuntu Server, however.) This server can have one or two processor cards, and customers can choose between two different processors: a six-core POWER8 humming along at 3.89 GHz and a ten-core variant spinning at 3.42 GHz. Each socket in this machine supports up to 512 GB of memory, and it has six PCI slots with one processor card and nine with two processor cards.

IBM said that a base POWER Systems S-Class server would cost $7,973, but did not say what configuration or what specific system that cost was tied to. Big Blue is expected to put out pricing information on the new S-Class systems next week.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Advancing Modular Supercomputing with DEEP and DEEP-ER Architectures

February 24, 2017

Knowing that the jump to exascale will require novel architectural approaches capable of delivering dramatic efficiency and performance gains, researchers around the world are hard at work on next-generation HPC systems. Read more…

By Sean Thielen

Weekly Twitter Roundup (Feb. 23, 2017)

February 23, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

HPE Server Shows Low Latency on STAC-N1 Test

February 22, 2017

The performance of trade and match servers can be a critical differentiator for financial trading houses. Read more…

By John Russell

HPC Financial Update (Feb. 2017)

February 22, 2017

In this recurring feature, we’ll provide you with financial highlights from companies in the HPC industry. Check back in regularly for an updated list with the most pertinent fiscal information. Read more…

By Thomas Ayres

HPE Extreme Performance Solutions

O&G Companies Create Value with High Performance Remote Visualization

Today’s oil and gas (O&G) companies are striving to process datasets that have become not only tremendously large, but extremely complex. And the larger that data becomes, the harder it is to move and analyze it – particularly with a workforce that could be distributed between drilling sites, offshore rigs, and remote offices. Read more…

Rethinking HPC Platforms for ‘Second Gen’ Applications

February 22, 2017

Just what constitutes HPC and how best to support it is a keen topic currently. Read more…

By John Russell

HPC Technique Propels Deep Learning at Scale

February 21, 2017

Researchers from Baidu’s Silicon Valley AI Lab (SVAIL) have adapted a well-known HPC communication technique to boost the speed and scale of their neural network training and now they are sharing their implementation with the larger deep learning community. Read more…

By Tiffany Trader

IDC: Will the Real Exascale Race Please Stand Up?

February 21, 2017

So the exascale race is on. And lots of organizations are in the pack. Government announcements from the US, China, India, Japan, and the EU indicate that they are working hard to make it happen – some sooner, some later. Read more…

By Bob Sorensen, IDC

ExxonMobil, NCSA, Cray Scale Reservoir Simulation to 700,000+ Processors

February 17, 2017

In a scaling breakthrough for oil and gas discovery, ExxonMobil geoscientists report they have harnessed the power of 717,000 processors – the equivalent of 22,000 32-processor computers – to run complex oil and gas reservoir simulation models. Read more…

By Doug Black

Advancing Modular Supercomputing with DEEP and DEEP-ER Architectures

February 24, 2017

Knowing that the jump to exascale will require novel architectural approaches capable of delivering dramatic efficiency and performance gains, researchers around the world are hard at work on next-generation HPC systems. Read more…

By Sean Thielen

HPC Technique Propels Deep Learning at Scale

February 21, 2017

Researchers from Baidu’s Silicon Valley AI Lab (SVAIL) have adapted a well-known HPC communication technique to boost the speed and scale of their neural network training and now they are sharing their implementation with the larger deep learning community. Read more…

By Tiffany Trader

IDC: Will the Real Exascale Race Please Stand Up?

February 21, 2017

So the exascale race is on. And lots of organizations are in the pack. Government announcements from the US, China, India, Japan, and the EU indicate that they are working hard to make it happen – some sooner, some later. Read more…

By Bob Sorensen, IDC

TSUBAME3.0 Points to Future HPE Pascal-NVLink-OPA Server

February 17, 2017

Since our initial coverage of the TSUBAME3.0 supercomputer yesterday, more details have come to light on this innovative project. Of particular interest is a new board design for NVLink-equipped Pascal P100 GPUs that will create another entrant to the space currently occupied by Nvidia's DGX-1 system, IBM's "Minsky" platform and the Supermicro SuperServer (1028GQ-TXR). Read more…

By Tiffany Trader

Tokyo Tech’s TSUBAME3.0 Will Be First HPE-SGI Super

February 16, 2017

In a press event Friday afternoon local time in Japan, Tokyo Institute of Technology (Tokyo Tech) announced its plans for the TSUBAME3.0 supercomputer, which will be Japan’s “fastest AI supercomputer,” Read more…

By Tiffany Trader

Drug Developers Use Google Cloud HPC in the Fight Against ALS

February 16, 2017

Within the haystack of a lethal disease such as ALS (amyotrophic lateral sclerosis / Lou Gehrig’s Disease) there exists, somewhere, the needle that will pierce this therapy-resistant affliction. Read more…

By Doug Black

Azure Edges AWS in Linpack Benchmark Study

February 15, 2017

The “when will clouds be ready for HPC” question has ebbed and flowed for years. Read more…

By John Russell

Is Liquid Cooling Ready to Go Mainstream?

February 13, 2017

Lost in the frenzy of SC16 was a substantial rise in the number of vendors showing server oriented liquid cooling technologies. Three decades ago liquid cooling was pretty much the exclusive realm of the Cray-2 and IBM mainframe class products. That’s changing. We are now seeing an emergence of x86 class server products with exotic plumbing technology ranging from Direct-to-Chip to servers and storage completely immersed in a dielectric fluid. Read more…

By Steve Campbell

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

US, China Vie for Supercomputing Supremacy

November 14, 2016

The 48th edition of the TOP500 list is fresh off the presses and while there is no new number one system, as previously teased by China, there are a number of notable entrants from the US and around the world and significant trends to report on. Read more…

By Tiffany Trader

Lighting up Aurora: Behind the Scenes at the Creation of the DOE’s Upcoming 200 Petaflops Supercomputer

December 1, 2016

In April 2015, U.S. Department of Energy Undersecretary Franklin Orr announced that Intel would be the prime contractor for Aurora: Read more…

By Jan Rowell

D-Wave SC16 Update: What’s Bo Ewald Saying These Days

November 18, 2016

Tucked in a back section of the SC16 exhibit hall, quantum computing pioneer D-Wave has been talking up its new 2000-qubit processor announced in September. Forget for a moment the criticism sometimes aimed at D-Wave. This small Canadian company has sold several machines including, for example, ones to Lockheed and NASA, and has worked with Google on mapping machine learning problems to quantum computing. In July Los Alamos National Laboratory took possession of a 1000-quibit D-Wave 2X system that LANL ordered a year ago around the time of SC15. Read more…

By John Russell

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. Read more…

By John Russell

IBM Wants to be “Red Hat” of Deep Learning

January 26, 2017

IBM today announced the addition of TensorFlow and Chainer deep learning frameworks to its PowerAI suite of deep learning tools, which already includes popular offerings such as Caffe, Theano, and Torch. Read more…

By John Russell

HPC Startup Advances Auto-Parallelization’s Promise

January 23, 2017

The shift from single core to multicore hardware has made finding parallelism in codes more important than ever, but that hasn’t made the task of parallel programming any easier. Read more…

By Tiffany Trader

Tokyo Tech’s TSUBAME3.0 Will Be First HPE-SGI Super

February 16, 2017

In a press event Friday afternoon local time in Japan, Tokyo Institute of Technology (Tokyo Tech) announced its plans for the TSUBAME3.0 supercomputer, which will be Japan’s “fastest AI supercomputer,” Read more…

By Tiffany Trader

Leading Solution Providers

CPU Benchmarking: Haswell Versus POWER8

June 2, 2015

With OpenPOWER activity ramping up and IBM’s prominent role in the upcoming DOE machines Summit and Sierra, it’s a good time to look at how the IBM POWER CPU stacks up against the x86 Xeon Haswell CPU from Intel. Read more…

By Tiffany Trader

Nvidia Sees Bright Future for AI Supercomputing

November 23, 2016

Graphics chipmaker Nvidia made a strong showing at SC16 in Salt Lake City last week. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

TSUBAME3.0 Points to Future HPE Pascal-NVLink-OPA Server

February 17, 2017

Since our initial coverage of the TSUBAME3.0 supercomputer yesterday, more details have come to light on this innovative project. Of particular interest is a new board design for NVLink-equipped Pascal P100 GPUs that will create another entrant to the space currently occupied by Nvidia's DGX-1 system, IBM's "Minsky" platform and the Supermicro SuperServer (1028GQ-TXR). Read more…

By Tiffany Trader

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment firm International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

Dell Knights Landing Machine Sets New STAC Records

November 2, 2016

The Securities Technology Analysis Center, commonly known as STAC, has released a new report characterizing the performance of the Knight Landing-based Dell PowerEdge C6320p server on the STAC-A2 benchmarking suite, widely used by the financial services industry to test and evaluate computing platforms. The Dell machine has set new records for both the baseline Greeks benchmark and the large Greeks benchmark. Read more…

By Tiffany Trader

Is Liquid Cooling Ready to Go Mainstream?

February 13, 2017

Lost in the frenzy of SC16 was a substantial rise in the number of vendors showing server oriented liquid cooling technologies. Three decades ago liquid cooling was pretty much the exclusive realm of the Cray-2 and IBM mainframe class products. That’s changing. We are now seeing an emergence of x86 class server products with exotic plumbing technology ranging from Direct-to-Chip to servers and storage completely immersed in a dielectric fluid. Read more…

By Steve Campbell

What Knights Landing Is Not

June 18, 2016

As we get ready to launch the newest member of the Intel Xeon Phi family, code named Knights Landing, it is natural that there be some questions and potentially some confusion. Read more…

By James Reinders, Intel

  • arrow
  • Click Here for More Headlines
  • arrow
Share This