SHPCP Talk: What Is Disaggregated Composable Infrastructure and Should You Care?

By John Russell

December 14, 2020

What’s your take on disaggregated composable infrastructure? Do you know what it is? Speaking at the Society of HPC Professional’s annual technology (virtual) meeting last week, Earl Dodd, HPC and HPDA architect at technology services giant World Wide Technology, delivered a strong pitch for DCI as the wave of the future for mainstream HPC.

DCI is another version of software defined everything that takes aim at the datacenter. Dodd argued DCI can slash latency and boost TCO.

“We’ve done a really good job in HPC of trying to scale out. OK, we used to scale up, [but] we don’t scale up anymore. We used to have the big SMPs and the CC-NUMA machines and then we got away from that. Now, I see a lot of 4-processor, 8-processor, 16-processor systems. I see these big systems growing again. So how do you seamlessly [and] dynamically put this whole scale-up capability together with that scale-out capability? I believe only forms of disaggregated composability will allow us to drive that,” said Dodd.

Earl Dodd, World Wide Technology

Leaving aside supercomputers and the hyperscalers – vast creatures in their own rights – Dodd took aim at commercial and more mainstream HPC infrastructure. The culprit stymieing modern HPC performance and adoption, he maintained, is latency which in turn is depressing utilization rates which in turn reduces TCO.

Dodd cited Nvidia CEO Jensen Huang’s recent mantra – “[Jensen says] in the datacenter world, and that includes HPC, all things computing must be disaggregated, composable and accelerated. I totally agree with that. I’m mainly talking about the capability market and specifically, the capability market in the cooperation that’s going on with HPC, big data, AI, ML, and DL [where] we’re seeing very, very low utilization rates.”

That change is afoot in HPC is something no one can deny. Dodd and his colleague Zach Splaingard have written a brief paper (Primer Series: Rack-Scale Composable Infrastructure) and argue in it:

“As more applications introduce support for accelerators (GPUs and FPGAs), which can reduce time to result from weeks to literally minutes, users are clamoring for more of these expensive resources. Yet, industry data shows they are only utilized 15 to 20 percent of the time, stranded behind the traditional data center’s rigid architecture.

“Legacy data center infrastructures were not designed for today’s workflow requirements. The scalable modern data center needs a solution that can integrate compute, storage and other communication I/O into a single-system cluster fabric, scaling resources up and out across the cluster as needed. This solution should free resources from their silos to be shared with other network users who draw from these resource pools through a disaggregated composable infrastructure (DCI), an emerging category of infrastructure designed to maximize IT resource usage and improve business agility.”

Broadly, this is not a new idea. It’s virtualization by another name. New software and getting rid of traditional network bottlenecks are among the key enablers, argued Dodd. The PCIe bus (or something like CXL or Gen-Z), contended Dodd, is an excellent candidate for replacing InfiniBand/Ethernet at many points in today’s datacenter.

“In the old days, I’ve got my nodes – thin nodes, fat nodes, high nodes, half-nodes, whatever you want to call them. I’ve got a computing element on computing system on the left side (slide above), and I got thingies in it. That’s a very technical term, by the way thingies. So, compute, memory, GPUs, FPGAs, smart NICs, etc. The idea is those can be disaggregated, and put into other systems, into other enclosures, other boxes, and treated as dynamically available resource pools,” said Dodd in his SHPCP talk.

“If you look at this chart, I’ve got a server on one side and server on the other side. We’re using something [like this] on the right at the labs now. We’ve been actively testing GigaIO‘s FabreX environment. In the old days, you had to go all the way down to that NIC layer, and you had to get across some form of a network and get back up and talk to something. And that’s even when you had to pool multiple nodes together. Okay, so we create fatter nodes, 4U boxes, and 7U boxes and 9U enclosures, and then put a midplane in there and put these things together. That whole thing is eliminated when I can talk PCIe-to-PCIe,” he said.

Currently, said Dodd, there are many misconceptions around DCI vendor lock-in, performance penalties, and whether DCI will sort of happen on its own.

“A lot of people are very much worried about vendor lock-in. You’ll hear a lot about, “Oh, I have a composable infrastructure,” but then they’ll say, and “I have a midplane, or I have a backplane and thou shalt only have stuff to fit into my mid plane, or backplane.” That creates vendor lock-in. And there is this concept of a tax, as in performance penalty associated with it by going through another midplane, another set of chipsets, another translation that goes on in there and that just adds latency. Latency overall is the big killer. It’s not the bandwidth, it’s the latency, latency, latency.

“[Another misbelief] is that composability is going to happen on its own, it’s going to have the Darwin effect; if we’ve gone from a converged infrastructure to hyperconverged infrastructure, [then] composability will naturally come out of that. It’ll happen on its own. Wrong, that is not going to happen. It has to be actually thought [through], because there is a lot of software needed to help make software-defined datacenter [actually become] software defined. Another misconception is you have to be close: you know, the memory must stick right next to the processor. The GPU must be really close and in the same box as the processor, the FPGA, or the persistent memory. Or you need the smart NIC sitting right next to the box. That is completely wrong. Within reason today I can disaggregate things up to about 30 meters, and still maintain the same performance characteristics. Now, that is a game changer.”

One important issue, said Dodd, was the tendency of the HPC community to think of itself as defined by workloads; today that is anachronistic, he suggested. What matters more, he argued, is workflow because it directly influences latency. Disaggregating resources in the manner described he says will drive down latency and drive up HPC utilization.

There was certainly a product pitch element to Dodd’s presentation and not too much detail. That said, it was fascinating and he was quick to emphasize that the GigaIO technology he was using as an example was based on open standards including Redfish and PCIe. Others will certainly develop their own version of DCI technology. Indeed Liqid’s composable architecture creates a flexible shared resource pool over the PCIe bus.

“This allows you to take almost any of these nodes and make some of the resources available to other node types in the system via the PCIe switch,” said George Moncrief, chief technologist at ERDC’s DoD Supercomputing Resource Center (DSRC), in an interview with HPCwire.

It will be interesting to watch how DCI writ large develops.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Graphcore Introduces Larger-Than-Ever IPU-Based Pods

October 22, 2021

Graphcore and its “Intelligent Processing Units” (IPUs) emerged from stealth in 2016 and launched its second-generation IPU in 2020. While the company has also launched its IPUs in a variety of form factors over the Read more…

Quantum Chemistry Project to Be Among the First on EuroHPC’s LUMI System

October 22, 2021

Finland’s CSC has just installed the first module of LUMI, a 550-peak petaflops system supported by the European Union’s EuroHPC Joint Undertaking. While LUMI -- pictured in the header -- isn’t slated to complete i Read more…

Killer Instinct: AMD’s Multi-Chip MI200 GPU Readies for a Major Global Debut

October 21, 2021

AMD’s next-generation supercomputer GPU is on its way – and by all appearances, it’s about to make a name for itself. The AMD Radeon Instinct MI200 GPU (a successor to the MI100) will, over the next year, begin to power three massive systems on three continents: the United States’ exascale Frontier system; the European Union’s pre-exascale LUMI system; and Australia’s petascale Setonix system. Read more…

D-Wave Embraces Gate-Based Quantum Computing; Charts Path Forward

October 21, 2021

Earlier this month D-Wave Systems, the quantum computing pioneer that has long championed quantum annealing-based quantum computing (and sometimes taken heat for that approach), announced it was expanding into gate-based Read more…

LLNL Prepares the Water and Power Infrastructure for El Capitan

October 21, 2021

When it’s (ostensibly) ready in early 2023, El Capitan is expected to deliver in excess of two exaflops of peak computing power – around four times the power of Fugaku, the current top-ranked supercomputer in the wor Read more…

AWS Solution Channel

Royalty-free stock illustration ID: 537899029

Running GROMACS on GPU instances

Comparing the performance of real applications across different Amazon Elastic Compute Cloud (Amazon EC2) instance types is the best way we’ve found for finding optimal configurations for HPC applications here at AWS. Read more…

Faster Optical Switch that Operates at ‘Room Temp’ Developed by IBM, Skolkovo Researchers

October 19, 2021

Optical switching technology holds great promise for many applications but hot operating temperatures have been a key obstacle slowing progress. Now, a new optical switching device that can operate at room temperatures a Read more…

Killer Instinct: AMD’s Multi-Chip MI200 GPU Readies for a Major Global Debut

October 21, 2021

AMD’s next-generation supercomputer GPU is on its way – and by all appearances, it’s about to make a name for itself. The AMD Radeon Instinct MI200 GPU (a successor to the MI100) will, over the next year, begin to power three massive systems on three continents: the United States’ exascale Frontier system; the European Union’s pre-exascale LUMI system; and Australia’s petascale Setonix system. Read more…

D-Wave Embraces Gate-Based Quantum Computing; Charts Path Forward

October 21, 2021

Earlier this month D-Wave Systems, the quantum computing pioneer that has long championed quantum annealing-based quantum computing (and sometimes taken heat fo Read more…

LLNL Prepares the Water and Power Infrastructure for El Capitan

October 21, 2021

When it’s (ostensibly) ready in early 2023, El Capitan is expected to deliver in excess of two exaflops of peak computing power – around four times the powe Read more…

Intel Reorgs HPC Group, Creates Two ‘Super Compute’ Groups

October 15, 2021

Following on changes made in June that moved Intel’s HPC unit out of the Data Platform Group and into the newly created Accelerated Computing Systems and Graphics (AXG) business unit, led by Raja Koduri, Intel is making further updates to the HPC group and announcing... Read more…

Quantum Workforce – NSTC Report Highlights Need for International Talent

October 13, 2021

Attracting and training the needed quantum workforce to fuel the ongoing quantum information sciences (QIS) revolution is a hot topic these days. Last week, the U.S. National Science and Technology Council issued a report – The Role of International Talent in Quantum Information Science... Read more…

Eni Returns to HPE for ‘HPC4’ Refresh via GreenLake

October 13, 2021

Italian energy company Eni is upgrading its HPC4 system with new gear from HPE that will be installed in Eni’s Green Data Center in Ferrera Erbognone (a provi Read more…

The Blueprint for the National Strategic Computing Reserve

October 12, 2021

Over the last year, the HPC community has been buzzing with the possibility of a National Strategic Computing Reserve (NSCR). An in-utero brainchild of the COVID-19 High-Performance Computing Consortium, an NSCR would serve as a Merchant Marine for urgent computing... Read more…

UCLA Researchers Report Largest Chiplet Design and Early Prototyping

October 12, 2021

What’s the best path forward for large-scale chip/system integration? Good question. Cerebras has set a high bar with its wafer scale engine 2 (WSE-2); it has 2.6 trillion transistors, including 850,000 cores, and was fabricated using TSMC’s 7nm process on a roughly 8” x 8” silicon footprint. Read more…

Enter Dojo: Tesla Reveals Design for Modular Supercomputer & D1 Chip

August 20, 2021

Two months ago, Tesla revealed a massive GPU cluster that it said was “roughly the number five supercomputer in the world,” and which was just a precursor to Tesla’s real supercomputing moonshot: the long-rumored, little-detailed Dojo system. Read more…

Esperanto, Silicon in Hand, Champions the Efficiency of Its 1,092-Core RISC-V Chip

August 27, 2021

Esperanto Technologies made waves last December when it announced ET-SoC-1, a new RISC-V-based chip aimed at machine learning that packed nearly 1,100 cores onto a package small enough to fit six times over on a single PCIe card. Now, Esperanto is back, silicon in-hand and taking aim... Read more…

US Closes in on Exascale: Frontier Installation Is Underway

September 29, 2021

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, held by Zoom this week (Sept. 29-30), it was revealed that the Frontier supercomputer is currently being installed at Oak Ridge National Laboratory in Oak Ridge, Tenn. The staff at the Oak Ridge Leadership... Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer... Read more…

Intel Completes LLVM Adoption; Will End Updates to Classic C/C++ Compilers in Future

August 10, 2021

Intel reported in a blog this week that its adoption of the open source LLVM architecture for Intel’s C/C++ compiler is complete. The transition is part of In Read more…

Intel Reorgs HPC Group, Creates Two ‘Super Compute’ Groups

October 15, 2021

Following on changes made in June that moved Intel’s HPC unit out of the Data Platform Group and into the newly created Accelerated Computing Systems and Graphics (AXG) business unit, led by Raja Koduri, Intel is making further updates to the HPC group and announcing... Read more…

Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was Read more…

AMD-Xilinx Deal Gains UK, EU Approvals — China’s Decision Still Pending

July 1, 2021

AMD’s planned acquisition of FPGA maker Xilinx is now in the hands of Chinese regulators after needed antitrust approvals for the $35 billion deal were receiv Read more…

Leading Solution Providers

Contributors

HPE Wins $2B GreenLake HPC-as-a-Service Deal with NSA

September 1, 2021

In the heated, oft-contentious, government IT space, HPE has won a massive $2 billion contract to provide HPC and AI services to the United States’ National Security Agency (NSA). Following on the heels of the now-canceled $10 billion JEDI contract (reissued as JWCC) and a $10 billion... Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

The Latest MLPerf Inference Results: Nvidia GPUs Hold Sway but Here Come CPUs and Intel

September 22, 2021

The latest round of MLPerf inference benchmark (v 1.1) results was released today and Nvidia again dominated, sweeping the top spots in the closed (apples-to-ap Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Intel Unveils New Node Names; Sapphire Rapids Is Now an ‘Intel 7’ CPU

July 27, 2021

What's a preeminent chip company to do when its process node technology lags the competition by (roughly) one generation, but outmoded naming conventions make i Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Quantum Computer Market Headed to $830M in 2024

September 13, 2021

What is one to make of the quantum computing market? Energized (lots of funding) but still chaotic and advancing in unpredictable ways (e.g. competing qubit tec Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire