Fast Forward: Five HPC Predictions for 2018

By John Russell

December 21, 2017

What’s on your list of high (and low) lights for 2017? Volta 100’s arrival on the heels of the P100? Appearance, albeit late in the year, of IBM’s Power9? Exascale Computing Project leadership shuffling? AMD’s return from the dead in the datacenter? Scandal at PEZY? Aurora’s stumble? Trump? There’s lots to choose from.

Whether you’re thinking ‘good riddance’ or ‘stay a little longer’ about 2017 – it feels like a year where there’s not a lot in between. It’s probably best focus on the future. Here are a few 2018 predictions mostly accenting the positive; indeed there is quite a bit to be positive about amid ever present dark clouds. Along the way there’s a few observations about 2017, and links to HPCwire coverage of note.

1. Big Blue Gets Its Mojo Back

Let’s be candid. Since dumping its x86 business, IBM has endured a bumpy ride. Building a new ecosystem has a definite “what were we thinking” level of difficulty. That doesn’t mean it can’t be done. OpenPOWER and IBM have done much that’s right, but getting to payoff is a costly, painful struggle. Power8 systems, despite the much-praised Power instruction set and scattered public support from systems builders and hyperscalers, mostly fizzled; some would argue it was swamped by timing and anticipation of Power9. Pricing may also have played a role.

Three events now seem poised to reenergize IBM and OpenPOWER.

  • First is arrival of the Power9 processor in December. It’s being promoted as a from-the-ground-up AI optimized chip able to leverage all kinds of accelerator (FPGA, GPU, etc.) high-speed interconnect (NVLink, OpenCAPI, etc.), and high memory bandwidth technology. It’s available in IBM’s AC922 server based on the same architecture as the Department of Energy CORAL supercomputers, Summit and Sierra. The Power9 wait is over.
  • Second, the Aurora project being led by Intel has been delayed. True, it is now scheduled to be the first U.S. exascale machine, deployed in 2021 at Argonne National Laboratory, but it clearly missed its mark as one of the scheduled pre-exascale machines. There’s also an open question as to which processor will be used for Aurora. And 2021 still seems quite distant. Overall Aurora’s trouble is IBM’s serendipity.
  • Third, expectations are high the IBM Summit machine will be stood up and tested in time to top the next Top500 list in June 2018. It’s expected to hit 150-200 petaflops peak performance. That would be a huge boost for IBM and its advanced Power architecture from a public awareness perspective. China has dominated the recent lists (ten consecutive ‘wins’) and the top performing U.S. machine, Titan, fell from for fourth to fifth in November. BTW, Titan is powered by AMD Opteron processors.
IBM Power9 AC922 rendering

IBM will attack the market in force with its ‘Summit-based’ servers. It will also likely get stronger buy-in from the OpenPOWER community, most of whom must still support Intel systems. Power9 system price points are also expected to be more attractive. Finally, with U.S. national competitiveness juices bubbling – amplified by Trump’s ‘America First’ mantra – the current U.S. administration is likely to talk up the IBM Top500 Summit achievement.

Bottom line: Big Blue will start making early hay in the server market (HPC and otherwise) after what must seem like a very long growing season. Time to start the harvest. Also, let’s not forget IBM is a giant in computing generally with a growing cloud business, a rapidly advancing quantum computing research and commercial program, a neuromorphic chip and research effort, and extensive portfolio of storage, software, mainframe, and services offerings. (For me, the jury is still out on Watson.) Big Blue is getting its mojo back.

Links to relevant articles:

IBM Begins Power9 Rollout with Backing from DOE, Google

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

Flipping the Flops and Reading the Top500 Tea Leaves 


2. AMD’s Datacenter Revival Looks Good – Don’t Blow It!

AMD has had many lives and it’s crossed swords with Intel over x86 matters (technology and markets) with regularity. Sometimes enough is enough. The company largely abandoned the datacenter a few years ago for a number of reasons. David versus Goliath doesn’t always end well for David. This year AMD has plunged back in and its bet is a big one that encompasses solid technology, price performance, and as of SC17, considerable commitment from some important systems makers to support the EPYC processor line.

“It’s not enough to come back with one product, you’ve got to come back with a product cadence that moves as the market moves. So not only are we coming back with EPYC, we’re also [discussing follow-on products] so when customers move with us today on EPYC they know they have a safe home and a migration path with Rome,” said Scott Aylor, AMD corporate VP and GM of enterprise solutions business, at the time of the launch.

Lower cost is clearly part of the strategy and AMD has been touting cost-performance comparisons. A portion of the EPYC line has been designed for single socket servers which are nearly extinct in the datacenter these days. AMD argues that around 50 percent of the market buys two-socket solutions because there was no alternative; now, says AMD, there is. In fact, Microsoft Azure recently announced an instance based on a single socket EPYC solution.

To meet a broad range of applications, AMD is tiering products in 32, 24, and 16-core ranges. The top end is aimed at scale out and HPC workloads. Indeed, AMD showcased ‘Project 47 supercomputer’ at SIGGRAPH in the summer which is based on the EPYC 7601 and AMD Radeon Instinct MI25 GPUs. A full 20-server rack of P47 systems achieves 30.05 gigaflops per watt in single-precision performance, but is less impressive on double precision arithmetic.

Bottom line: AMD is back, at least for now, charging into the datacenter.

Links to relevant articles:

AMD Showcases Growing Portfolio of EPYC and Radeon-based Systems at SC17

AMD Charges Back into the Datacenter and HPC Workflows with EPYC Processor

AMD Stuffs a Petaflops of Machine Intelligence into 20-Node Rack

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration


3. The Quantum Computing Haze will Thicken Not Thin

Ok, I admit it. Quantum computing pretty much baffles me. Despite the mountain of rhetoric surrounding quantum computing, I suspect am not alone. Universal Quantum Computers. Quantum Adiabatic Annealing Computers. An expanding zoo of qubit types. Qudits. Quantum simulation on classical computers. Quantum Supremacy. Google, Microsoft, IBM, D-wave, a handful of academic and national lab quantum computing programs. Something called the Chicago Quantum Exchange under David Awschalom, associated with UChicago, Argonne, Fermilab, and located in the Institute for Molecular Engineering.

Feynman would chuckle. The saying is ‘where there’s smoke there’s fire’ and while that’s true enough, the plentiful smoke around quantum computing today is awfully hard to see through. Obviously there is something important going on but how important or when it will be important (let alone mainstream) is very unclear.

An IBM cryostat wired for a prototype 50 qubit system. (PRNewsfoto/IBM)

Philip Ball’s recent Nature piece on quantum supremacy (Race for quantum supremacy hits theoretical quagmire, Nature, 11/14/17) is both informative and entertaining. Quantum supremacy is the stage at which the capabilities of a quantum computer exceed those of any available classical computer. Of course the latter keep advancing.

Ball wrote, “Computer scientists and engineers are rather more phlegmatic about the notion of quantum supremacy than excited commentators who foresee an impending quantum takeover of information technology. They see it not as an abrupt boundary but as a symbolic gesture: a conceptual tool on which to peg a discussion of the differences between the two methods of computation. And, perhaps, a neat advertising slogan.”

Actually there’s a fair amount of good literature on quantum computing. Just a few of the current challenges include size (how many qubits) of current machines, needed error correction, nascent software, decoherency, exotic machines – think supercooled superconductors as an example – optimum qubit types…You get the idea.

Bottom line: The haze surrounding quantum computing’s future won’t lift for a few more years. Maybe a few specific quantum communication applications will emerge sooner.

Links to relevant articles:

Microsoft Wants to Speed Quantum Development

House Subcommittee Tackles US Competitiveness in Quantum Computing

Intel Delivers 17-Qubit Quantum Chip to European Research Partner

IBM Breaks Ground for Complex Quantum Chemistry

Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

IBM Launches Commercial Quantum Network with Samsung, ORNL


4. AI will Continue Sucking the Air Out of the Room

Maybe this is a good thing. AI writ large is blanketing the computer landscape. Its language is everywhere and dominating the marketing landscape. Every vendor, it seems, has an AI box or service or chip(s). More interesting is what’s happening in developing and using ‘AI’ technology. The CANcer Distributed Learning Environment (CANDLE) project – tasked with developing deep learning tools for the war on cancer and putting them to use – is a good example; it has released the early version of its infrastructure on GitHub. This includes algorithms, frameworks, and all manner of relevant tools.

CANDLE has already developed a model able to predict tumor response to drug pairs for a particular cancer type at 93 percent. The data sets are huge and machine learning is the only way to chew through them to build models. It’s working on a model to handle triplet drug combos. “There will be drugs I predict in clinical trials based on the results that we achieve this year,” Rick Stevens, one of the PIs on CANDLE and a senior researcher at Argonne National Laboratory, told HPCwire at SC17.

There’s a wealth of new (and some rather old data analytics) technology to support AI. New frameworks. Advancing accelerator technology. The rise of mixed precision machines – Japan’s plans for a 130-petaflops (half-precision) supercomputer by early 2018 called ABCI for AI Bridging Cloud Infrastructure is a good high-end example. There’s too much to cover here beyond saying AI is a game changer on its own for many applications and will also prove to be incredibly powerful in speeding up traditional floating-point intensive HPC applications such as molecular modeling.

In a memo to employees this week, Intel CEO Brian Krzanich wrote, “It’s almost impossible to perfectly predict the future, but if there’s one thing about the future I am 100 percent sure of, it is the role of data. Anything that produces data, anything that requires a lot of computing.” AI computing will be an important part of nearly all computing going forward.

Bottom line: Brace for more AI.

Links to relevant articles:

Japan Plans Super-Efficient AI Supercomputer

AI Speeds Astrophysics Image Analysis by 10,000x

Cray Brings AI and HPC Together on Flagship Supers

Nvidia CEO Predicts AI ‘Cambrian Explosion’

Intel Unveils Deep Learning Framework, BigDL


5. The HPC Identity Crisis will Continue in Force (Does it Matter?)

Ok, a better phrasing is what constitutes HPC today and do we even know how many HPC workers there are? We talk about this inside HPCwire all the time. The blending (broadening) of HPC with big data/AI computing is one element. Simple redefinition by fiat is another. Various constituents offer differing perspectives.

When someone says HPC it means something really specific to traditional HPC folks; it’s tightly coupled, we’ve got some sort of low latency interconnect, parallel file systems, designed to run high performance, highly scalable custom applications. But today, this has changed. HPC has come to mean pretty much any form of scientific computing and as a result, its breadth has grown in terms of what kind of applications we need to support.” – Gregory Kurtzer, Singularity (HPC container software).

Hyperion Research pegs the number of HPC sites in the U.S. at 759 (academic, government, commercial) and suggests there could be around 120,000 HPCers in the U.S. and perhaps a quarter of a million worldwide.

Making sense of the collision between traditional HPC and big data (and finding ways to harmonize the two) has been a hot topic at least since 2015 when it was identified as an objective in National Strategic Computing Initiative. There’s even been a series of five international workshops (in US, Japan, the EU and China) on Big Data and Extreme-scale Computing (BDEC) and Jack Dongarra and colleagues working on the project have just issued a reportPathways to Convergence: Towards a Shaping Strategy for a Future Software and Data Ecosystem for Scientific Inquiry. HPCwire will dig into the report’s findings at a subsequent time.

The point here is that change is overwhelming how HPC is looked at and what it is considered to be. HPC census and market sizing is an ongoing challenge. One astute industry observer noted:

“The idea of framing out the real HPC TAM (total available market) is an interesting one.  If I live in a big DoE facility and run code on the Titan HPC, I know I am an HPC guy. But if I am a car part designer that subs to GM, who uses Autodesk for visualization for the design of a driver’s side mirror, I may not think of myself as such (I sure as hell will not attend  SC17) .

“That and the fact that I saw so many vendors at SC that have products that address some of the less technically aggressive aspects of HPC (i.e. tape storage) that really aren’t HPC specific but that can be relevant to HPC users. So it’s hard to say what the TAM is because reaching out to customers who may be HPC, but don’t move in the HPC world per se is complicated at best.

“Even worse, figuring out how to count marketing dollars that reaches some indeterminant percentage of a loosely defined HPC market is fraught with intrigue.”

Bottom line: The HPC Who-am-I? pathos will continue in 2018 but preoccupation with delivering AI will mute some of the debate.


6. Lesser but Still Interesting 2018/2017 Glimpses

Doug Kothe, ECP director

The container craze will continue because it solves a real problem. ECP, now led by Doug Kothe, will shift into its next gear as the first U.S pre-exascale machines are stood up. Forget the doubters – the Nvidia juggernaut will keep rolling, though perhaps there won’t be another V100-like blockbuster introduced in 2018. Intel’s impressive Skylake chip line arrived and is in systems everywhere. Vendors’ infatuation with selling so-called easier-to-deploy HPC solutions into the enterprise – think vertical solutions – will fade; they’ve tried selling these but mostly without success for many reasons.

ARM will continue its march into new markets. This topic doesn’t rise to greater prominence here because we still need to see more systems on line, whether at the very high end such as the post K computer, or sales of ARM server systems such as HPE’s recently introduced Apollo 70, the company’s first ARM-based HPC server. The earlier ARM-based Moonshot offering fared poorly.

Unexpected scandal marred the end of the year with the arrest of PEZY founder, president and CEO Motoaki Saito and another PEZY employee, Daisuke Suzuki, on suspicion of defrauding a government institution of 431 million yen (~$3.8 million) and is unsettling. HPC seems reasonably free of such misbehavior. Maybe that’s my misperception.

It was sad to see what amounts to the end of the line for SPARC with Oracle’s discontinuance of development efforts and related layoffs.

On a positive note: There’s a new book from Thomas Sterling, professor of electrical engineering and director of the Center for Research in Extreme Scale Technologies, Indiana University – High Performance Computing: Modern Systems and Practices, co-written with colleagues Matthew Anderson and Maciej Brodowicz. It’s available now (link to publisher: https://www.elsevier.com/books/high-performance-computing/sterling/978-0-12-420158-3?start_rank=1&sortby=sortByDateDesc&imprintname=Morgan%20Kaufmann).

As always, there was a fair amount of personnel shuffling this year. Diane Bryant left Intel and joined Google. AI pioneer Andrew Ng left his post at Baidu. Intel lured GPU designer Raja Koduri from AMD; he was SVP and chief of architecture for Radeon Technology Group. Meg Whitman is stepping down as chairman of HPE – the company she helped bring into existence by overseeing the spit up of HP a year ago – and will be succeeded by Antonio Neri.

Obviously there is so much more to talk about. The HPC world is a vibrant, fascinating place, and tremendous force in science and society today.

Happy holidays and a hopeful new year to all. On to 2018!

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

InfiniBand Still Tops in Supercomputing

July 19, 2018

In the competitive global HPC landscape, system and processor vendors, nations and end user sites certainly get a lot of attention--deservedly so--but more than ever, the network plays a crucial role. While fast, perform Read more…

By Tiffany Trader

HPC for Life: Genomics, Brain Research, and Beyond

July 19, 2018

During the past few decades, the life sciences have witnessed one landmark discovery after another with the aid of HPC, paving the way toward a new era of personalized treatments based on an individual’s genetic makeup Read more…

By Warren Froelich

WCRP’s New Strategic Plan for Climate Research Highlights the Importance of HPC

July 19, 2018

As climate modeling increasingly leverages exascale computing and researchers warn of an impending computing gap in climate research, the World Climate Research Programme (WCRP) is developing its new Strategic Plan – and high-performance computing is slated to play a critical role. Read more…

By Oliver Peckham

HPE Extreme Performance Solutions

Introducing the First Integrated System Management Software for HPC Clusters from HPE

How do you manage your complex, growing cluster environments? Answer that big challenge with the new HPC cluster management solution: HPE Performance Cluster Manager. Read more…

IBM Accelerated Insights

Are Your Software Licenses Impeding Your Productivity?

In my previous article, Improving chip yield rates with cognitive manufacturing, I highlighted the costs associated with semiconductor manufacturing, and how cognitive methods can yield benefits in both design and manufacture.  Read more…

U.S. Exascale Computing Project Releases Software Technology Progress Report

July 19, 2018

As is often noted the race to exascale computing isn’t just about hardware. This week the U.S. Exascale Computing Project (ECP) released its latest Software Technology (ST) Capability Assessment Report detailing progress so far. Read more…

By John Russell

InfiniBand Still Tops in Supercomputing

July 19, 2018

In the competitive global HPC landscape, system and processor vendors, nations and end user sites certainly get a lot of attention--deservedly so--but more than Read more…

By Tiffany Trader

HPC for Life: Genomics, Brain Research, and Beyond

July 19, 2018

During the past few decades, the life sciences have witnessed one landmark discovery after another with the aid of HPC, paving the way toward a new era of perso Read more…

By Warren Froelich

D-Wave Breaks New Ground in Quantum Simulation

July 16, 2018

Last Friday D-Wave scientists and colleagues published work in Science which they say represents the first fulfillment of Richard Feynman’s 1982 notion that Read more…

By John Russell

AI Thought Leaders on Capitol Hill

July 14, 2018

On Thursday, July 12, the House Committee on Science, Space, and Technology heard from four academic and industry leaders – representatives from Berkeley Lab, Argonne Lab, GE Global Research and Carnegie Mellon University – on the opportunities springing from the intersection of machine learning and advanced-scale computing. Read more…

By Tiffany Trader

HPC Serves as a ‘Rosetta Stone’ for the Information Age

July 12, 2018

In an age defined and transformed by its data, several large-scale scientific instruments around the globe might be viewed as a ‘mother lode’ of precious data. With names seemingly created for a ‘techno-speak’ glossary, these interferometers, cyclotrons, sequencers, solenoids, satellite altimeters, and cryo-electron microscopes are churning out data in previously unthinkable and seemingly incomprehensible quantities -- billions, trillions and quadrillions of bits and bytes of electro-magnetic code. Read more…

By Warren Froelich

Tsinghua Powers Through ISC18 Field

July 10, 2018

Tsinghua University topped all other competitors at the ISC18 Student Cluster Competition with an overall score of 88.43 out of 100. This gives Tsinghua their s Read more…

By Dan Olds

HPE, EPFL Launch Blue Brain 5 Supercomputer

July 10, 2018

HPE and the Ecole Polytechnique Federale de Lausannne (EPFL) Blue Brain Project yesterday introduced Blue Brain 5, a new supercomputer built by HPE, which displ Read more…

By John Russell

Pumping New Life into HPC Clusters, the Case for Liquid Cooling

July 10, 2018

High Performance Computing (HPC) faces some daunting challenges in the coming years as traditional, industry-standard systems push the boundaries of data center Read more…

By Scott Tease

Leading Solution Providers

SC17 Booth Video Tours Playlist

Altair @ SC17

Altair

AMD @ SC17

AMD

ASRock Rack @ SC17

ASRock Rack

CEJN @ SC17

CEJN

DDN Storage @ SC17

DDN Storage

Huawei @ SC17

Huawei

IBM @ SC17

IBM

IBM Power Systems @ SC17

IBM Power Systems

Intel @ SC17

Intel

Lenovo @ SC17

Lenovo

Mellanox Technologies @ SC17

Mellanox Technologies

Microsoft @ SC17

Microsoft

Penguin Computing @ SC17

Penguin Computing

Pure Storage @ SC17

Pure Storage

Supericro @ SC17

Supericro

Tyan @ SC17

Tyan

Univa @ SC17

Univa

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This