Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

By John Russell

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was on display at Hot Chips this week. Arm, Nvidia, and Intel all gave talks on the emerging class of devices and while much of the technical detail about their latest offerings wasn’t new, it was notable that DPUs and IPUs warranted a session at all. Also interesting is that Arm technology is being used in all of the new offerings – Octeon10 (Marvell/Arm), the BlueField line (Nvidia), Mount Evans (Intel’s new chip).

Once discounted as little more than SmartNICs on steroids, DPU/IPUs are gaining advocates. There’s a growing consensus that DPUs and IPUs –  or whatever we end up calling them – tackle a growing problem and offer significant advantages. Broadly, the idea is to unload many house-keeping chores, such as networking control, storage management, and security now being run on host CPUs. These tasks have steadily consumed more and more of CPU resources within datacenters and the cloud.

“Research from Google and Facebook has shown [infrastructure workloads] consume from 22 percent to 80 percent of CPU cycles across a variety of micro service workloads,” said Brad Burres of Intel in his talk. “You can see from that data how offloading infrastructure applications provides meaningful benefit to the cloud operators in several key areas.”

Intel is the most recent big player to jump onto the DPU/IPU wagon. Burres, an Intel fellow, leads IPU architecture work in Intel’s newly created networking and edge group. Mount Evans, Intel’s first ASIC IPU, was just announced a week ago at Intel’s Architecture Day. Notably, Mount Evans shed x86 cores in favor Arm Neoverse N1 cores, not least for power consumption advantages.

Planting a flag for Intel, Burres said, “IPU stands for infrastructure processing unit and is a term Intel recently introduced. We used to call these SmartNICs. But as more and more infrastructure applications beyond networking move into the platform, especially with their associated control planes, we wanted to use a more accurate informative name. [It] is sometimes called the DPU, but that name confuses a lot of our customers, because everything does data processing. More importantly, an IPU represents a revolution in datacenter architecture.”

Let the competitive jockeying (and device naming) begin.

One of the nice things about Hot Chips is its tech talk bias and the slides often provide a useful glimpse into approaches and capabilities. To varying degrees, that was the case here. Here’s quick snapshot of the three presentations:

  • Arm Neoverse N2 and Octeon. Andrea Pellegrini, distinguished engineer at Arm, reviewed the Arm Neoverse platform roadmap and dug into Neoverse N2 advances relative to N1. Much of that material presented was familiar (see HPCwire coverage). Among N2 advances are SVE2 support, CMN-700 (coherent mesh network), IPC lift (40 percent), and significant branch prediction improvements. Directly relevant to the session, was his brief description of Marvell’s forthcoming Octeon10 DPU family, introduced in June and which will leverage Neoverse N2 architecture. It will be fabbed using TSMC’s 5nm process.
  • Nvidia’s BlueField Lineup. Principal architect Idan Burstein presented Nvidia’s vision although many of technical details were widely familiar (see HPCwire coverage). Nvidia – via its Mellanox acquisition – has pursued the evolution of smart networking into full-function DPUs the longest. BlueField-2 is now shipping, BlueField-3 is expected next year, and BlueField-4 plans are underway. Nvidia’s vision for the DPU is grand and encompasses not only commonly-cited chores (networking, storage, security) but also expansion to include other application pieces. BlueField-2 is a 7-billion-transistor SOC and BlueField-3 will have 22 billion transistors including Arm cores and GPU cores.
  • Intel’s Mount Evans Climb. Burres broadly walked through its functionality without digging deeply into implementation. Mount Evans is Intel’s first ASIC IPU. It has had FPGA-based IPU SOCs and, broadly speaking, it has leveraged assets from its acquisition of BareFoot Networks in 2019. Its early target market seems to be cloud providers where there is a huge x86 installed base, and Burres said Mount Evans was designed in collaboration with a major cloud provider though he didn’t say who.

Presented here are a few points from each of the presentations along with a few slides that provide a glimpse into design choices and implementations.

Arm’s Ambition for Neoverse Includes DPUs

Arm, of course, is a relative newcomer to large datacenter and Pellegrini noted N2, the second generation of Neoverse, is intended to serve a wider infrastructure market. “Partners can use this platform to build optimized, local systems with constraining power envelope systems, like the ones needed for 5G deployments,” he said. “On the other hand, they can build high-core count, high frequency, high-memory bandwidth systems for the datacenter.” N2’s efficiency profile allows users to pack many cores per socket, making it a great design for power efficient, specialized designs, he said.

Pellegrini singled Marvell’s Octeon 10 DPU as an example of a special purpose design. “This DPU relies on Neoverse N2 platform for its general-purpose computer system and they have up to 36 N2 cores. Marvell augmented the design with several specialized IPs to enable high-speed packet processing and network connectivity. When we look at instantiation of this design and its capabilities, we see how a partner like Marvell can extract the most out of the Neoverse N2 platform to take advantage of the bleeding edge technology, such as DDR five [and] MPC agent files to deliver groundbreaking networking speeds. We’re talking about up to 400 Gigabit Ethernet here.”

He noted that Neoverse N2 introduces memory partition and monitoring (MPAM), which is a technology that can “help users monitor and partition shared system resources, such that they can ensure more reliable consistent performance even in heavily contended multi-tenant systems.” He added that Neoverse N2 significantly advances support for utilization, “introducing better hardware capabilities for handling nested virtual machines, reducing the overhead and cost of commonly used virtual machine operations.”

Does Nvidia’s BlueField Have Head Start?

Echoing others, Burstein said soaring network bandwidth demands (think AI requirements) and the fact that 25-to-50 percent of CPU cycles are consumed for “infrastructure” needs is forcing the move to DPUs.

“There is a need to isolate the datacenter workloads and to accelerate them in order to support these higher bandwidth demands. The naive approach to solving this problem is positioning those infrastructure-processing elements that were running on the on the application processor to be running on processors embedded in the networking device (think SmartNICs). This has no gain in performance over efficiency. It solves the isolation problem, but it is not scalable to higher bandwidth requirements, and it will require significant system modifications as CPUs are consuming more power,” said Burstein.

The BlueField-3 specs are impressive: 22 billion transistors, the first 400 gigabits-per-second networking chip, 16 Arm CPUs to run the entire virtualization software stack, for instance, running VMware ESX. “BlueField-3 takes security to a whole new level, fully offloading [and] accelerating IPsec and TLS cryptography, secret key management and regular expression processing. We’re on a pace to introduce a new BlueField generation every 18 months. BlueField-3 will do 400 gigabits per second, and be 10x the processing capability of Bluefield-2, and BlueField-4 will do 800 gigabits per second, and add Nvidia’s AI computing technologies to get another 10x boost,” reported Nvidia at BlueField-3’s introduction in April.

Burstein said, “BlueField is a complicated system. It’s about 400 gigabits of Ethernet and InfiniBand with pipelined with crypto and security acceleration. It has 36 lanes of PCI. It supports two times 370 packet per second, and it supports two times 40 million packets per second, all at the scale of millions of flows.”

Broadly, the company has said BlueField is critical to its cloud-native supercomputing architecture strategy. The company has reported that Dell Technologies, Inspur, Lenovo and Supermicro are integrating BlueField DPUs into their systems and that several cloud providers, including Baidu, JD.com and UCloud, for example, are or plan to integrate BlueField DPU. The proof will be in deployments and subsequent traction.

DOCA (Data Center Infrastructure-on-a-Chip Architecture) is Nvidia’s programming framework for BlueField: “DOCA software consists of an SDK and a runtime environment. The DOCA SDK provides industry-standard open APIs and frameworks, including Data Plane Development Kit (DPDK) and P4 for networking and security and the Storage Performance Development Kit (SPDK) for storage. The frameworks simplify application offload with integrated NVIDIA acceleration packages. The DOCA-based services are exposed in the compute nodes as industry-standard I/O interfaces, enabling infrastructure virtualization and isolation.”

Climbing Mount Evans – Intel’s IPU Aims for the Clouds

Burres broadly talked about Mount Evans’ technology and zeroed in on the case for adoption by cloud providers.

“First, the IPU allows for a separation of functions between the service provider and the tenants. This provides for greater security and isolation for all parties. It also enables important use cases like bare metal hosting to run on the same exact hardware platforms, using the same services as virtual machines. It lets tenants have full control over their CPU. They can do things like run their own hypervisor. And in that case, the cloud operators still fully control of the infrastructure functions such as networking, storage and security because those live out in the IPU,” said Burres.

“Second, the IPU provides [an] infrastructure-optimized execution environment. This includes a significant investment in hardware accelerators in the IPU, which enable the IPU to process the infrastructure tasks very efficiently. That allows better tuning of software and cores for these types of workloads. Overall, this optimizes the performance and the cloud operator can now rent out 100 percent CPU to his guest, which also maximizes revenues. Lastly, the IPU can help enable new service models for storage by abstracting the storage initiated from the tenant,” he contended.

Burres argued Mount Evans programmable packet processing capability was perhaps its most impressive feature. It supports “use cases like vSwitch offload, firewalls, telemetry functions, while supporting up to 200 million packets per second performance on real world implementations. This is enhanced with a fully featured transmit traffic shaper.” Mount Evans provides inline IPsec to secure every packet being sent across the network and supports up to 16 million secure connections.

“On the right-hand side (block diagram below), our compute complex is built on the Arm Neoverse architecture using the N1 cores, with up to 16 cores running up to 3 GHz. This is backed by a large 32-megabytes system level cache, and a three dual-mode LP DDR4 controller that can write a theoretical 102 gigabytes per second of memory bandwidth. Together these give us the bandwidth and horsepower to take on larger production workloads. The compute complex is tightly coupled with the network subsystem, allowing the NSS accelerators to use a system level cache as their own last level cache. The mesh providing high bandwidth low latency connections between the two sides,” he said.

Burres also emphasized the new Intel IPU was designed from-the-ground up; he said he was responding to comments he’d seen that Intel was simply gluing existing IP together. Burres said little about the choice of Arm over IA processor cores other than it was done as part of Intel’s routine evaluation. Likely, Arm’s reduced power requirement was a factor.

Whether IPUs and DPU win a place in the growing library of processor acronyms is likely to emerge over the next couple of years as the devices actually come to market and their functionality (performance and cost) gets tested.

Analysts have generally been friendly to the idea. Karl Freund, principal at Cambrian AI Research told HPCwire, “This market for Smart NICS+ is just beginning to take shape. The early adopters are the hyperscalers typically using an ASIC (AWS) or FPGA (Microsoft).  Nvidia however foresees needing something much more powerful with a GPU along with Arm cores on the IPU. In three years, that could be a game changer for very large composable infrastructure.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

The Case for an Edge-Driven Future for Supercomputing

September 24, 2021

“Exascale only becomes valuable when it’s creating and using data that we care about,” said Pete Beckman, co-director of the Northwestern-Argonne Institute of Science and Engineering (NAISE), at the most recent HPC Read more…

Three Universities Team for NSF-Funded ‘ACES’ Reconfigurable Supercomputer Prototype

September 23, 2021

As Moore’s law slows, HPC developers are increasingly looking for speed gains in specialized code and specialized hardware – but this specialization, in turn, can make testing and deploying code trickier than ever. Now, researchers from Texas A&M University, the University of Illinois at Urbana... Read more…

Qubit Stream: Monte Carlo Advance, Infosys Joins the Fray, D-Wave Meeting Plans, and More

September 23, 2021

It seems the stream of quantum computing reports never ceases. This week – IonQ and Goldman Sachs tackle Monte Carlo on quantum hardware, Cambridge Quantum pushes chemistry calculations forward, D-Wave prepares for its Read more…

Asetek Announces It Is Exiting HPC to Protect Future Profitability

September 22, 2021

Liquid cooling specialist Asetek, well-known in HPC circles for its direct-to-chip cooling technology that is inside some of the fastest supercomputers in the world, announced today that it is exiting the HPC space amid multiple supply chain issues related to the pandemic. Although pandemic supply chain... Read more…

TACC Supercomputer Delves Into Protein Interactions

September 22, 2021

Adenosine triphosphate (ATP) is a compound used to funnel energy from mitochondria to other parts of the cell, enabling energy-driven functions like muscle contractions. For ATP to flow, though, the interaction between the hexokinase-II (HKII) enzyme and the proteins found in a specific channel on the mitochondria’s outer membrane. Now, simulations conducted on supercomputers at the Texas Advanced Computing Center (TACC) have simulated... Read more…

AWS Solution Channel

Introducing AWS ParallelCluster 3

Running HPC workloads, like computational fluid dynamics (CFD), molecular dynamics, or weather forecasting typically involves a lot of moving parts. You need a hundreds or thousands of compute cores, a job scheduler for keeping them fed, a shared file system that’s tuned for throughput or IOPS (or both), loads of libraries, a fast network, and a head node to make sense of all this. Read more…

The Latest MLPerf Inference Results: Nvidia GPUs Hold Sway but Here Come CPUs and Intel

September 22, 2021

The latest round of MLPerf inference benchmark (v 1.1) results was released today and Nvidia again dominated, sweeping the top spots in the closed (apples-to-apples) datacenter and edge categories. Perhaps more interesti Read more…

The Case for an Edge-Driven Future for Supercomputing

September 24, 2021

“Exascale only becomes valuable when it’s creating and using data that we care about,” said Pete Beckman, co-director of the Northwestern-Argonne Institut Read more…

Three Universities Team for NSF-Funded ‘ACES’ Reconfigurable Supercomputer Prototype

September 23, 2021

As Moore’s law slows, HPC developers are increasingly looking for speed gains in specialized code and specialized hardware – but this specialization, in turn, can make testing and deploying code trickier than ever. Now, researchers from Texas A&M University, the University of Illinois at Urbana... Read more…

Qubit Stream: Monte Carlo Advance, Infosys Joins the Fray, D-Wave Meeting Plans, and More

September 23, 2021

It seems the stream of quantum computing reports never ceases. This week – IonQ and Goldman Sachs tackle Monte Carlo on quantum hardware, Cambridge Quantum pu Read more…

Asetek Announces It Is Exiting HPC to Protect Future Profitability

September 22, 2021

Liquid cooling specialist Asetek, well-known in HPC circles for its direct-to-chip cooling technology that is inside some of the fastest supercomputers in the world, announced today that it is exiting the HPC space amid multiple supply chain issues related to the pandemic. Although pandemic supply chain... Read more…

TACC Supercomputer Delves Into Protein Interactions

September 22, 2021

Adenosine triphosphate (ATP) is a compound used to funnel energy from mitochondria to other parts of the cell, enabling energy-driven functions like muscle contractions. For ATP to flow, though, the interaction between the hexokinase-II (HKII) enzyme and the proteins found in a specific channel on the mitochondria’s outer membrane. Now, simulations conducted on supercomputers at the Texas Advanced Computing Center (TACC) have simulated... Read more…

The Latest MLPerf Inference Results: Nvidia GPUs Hold Sway but Here Come CPUs and Intel

September 22, 2021

The latest round of MLPerf inference benchmark (v 1.1) results was released today and Nvidia again dominated, sweeping the top spots in the closed (apples-to-ap Read more…

Why HPC Storage Matters More Now Than Ever: Analyst Q&A

September 17, 2021

With soaring data volumes and insatiable computing driving nearly every facet of economic, social and scientific progress, data storage is seizing the spotlight. Hyperion Research analyst and noted storage expert Mark Nossokoff looks at key storage trends in the context of the evolving HPC (and AI) landscape... Read more…

GigaIO Gets $14.7M in Series B Funding to Expand Its Composable Fabric Technology to Customers

September 16, 2021

Just before the COVID-19 pandemic began in March 2020, GigaIO introduced its Universal Composable Fabric technology, which allows enterprises to bring together Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer called Dojo to process truly vast amounts of video data. It’s a beast! … A truly useful exaflop at de facto FP32.” Read more…

Enter Dojo: Tesla Reveals Design for Modular Supercomputer & D1 Chip

August 20, 2021

Two months ago, Tesla revealed a massive GPU cluster that it said was “roughly the number five supercomputer in the world,” and which was just a precursor to Tesla’s real supercomputing moonshot: the long-rumored, little-detailed Dojo system. “We’ve been scaling our neural network training compute dramatically over the last few years,” said Milan Kovac, Tesla’s director of autopilot engineering. Read more…

Esperanto, Silicon in Hand, Champions the Efficiency of Its 1,092-Core RISC-V Chip

August 27, 2021

Esperanto Technologies made waves last December when it announced ET-SoC-1, a new RISC-V-based chip aimed at machine learning that packed nearly 1,100 cores onto a package small enough to fit six times over on a single PCIe card. Now, Esperanto is back, silicon in-hand and taking aim... Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months after Red Hat deprecated its support for the widely popular, free CentOS server operating system. The Rocky Linux development effort... Read more…

Intel Completes LLVM Adoption; Will End Updates to Classic C/C++ Compilers in Future

August 10, 2021

Intel reported in a blog this week that its adoption of the open source LLVM architecture for Intel’s C/C++ compiler is complete. The transition is part of In Read more…

Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was Read more…

AMD-Xilinx Deal Gains UK, EU Approvals — China’s Decision Still Pending

July 1, 2021

AMD’s planned acquisition of FPGA maker Xilinx is now in the hands of Chinese regulators after needed antitrust approvals for the $35 billion deal were receiv Read more…

Google Launches TPU v4 AI Chips

May 20, 2021

Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I Read more…

Leading Solution Providers

Contributors

HPE Wins $2B GreenLake HPC-as-a-Service Deal with NSA

September 1, 2021

In the heated, oft-contentious, government IT space, HPE has won a massive $2 billion contract to provide HPC and AI services to the United States’ National Security Agency (NSA). Following on the heels of the now-canceled $10 billion JEDI contract (reissued as JWCC) and a $10 billion... Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Intel Unveils New Node Names; Sapphire Rapids Is Now an ‘Intel 7’ CPU

July 27, 2021

What's a preeminent chip company to do when its process node technology lags the competition by (roughly) one generation, but outmoded naming conventions make it seem like it's two nodes behind? For Intel, the response was to change how it refers to its nodes with the aim of better reflecting its positioning within the leadership semiconductor manufacturing space. Intel revealed its new node nomenclature, and... Read more…

Intel Announces Sapphire Rapids with HBM, Reveals Ponte Vecchio Form Factors

June 28, 2021

From the ISC 2021 Digital event, Intel announced it will offer Sapphire Rapids with integrated HBM, detailed new Xe-HPC GPU form factors, and introduced commercial support for DAOS (distributed application object storage). Intel also announced a new Ethernet solution, aimed at smaller-scale HPC. With integrated High Bandwidth Memory (HBM), the forthcoming Intel Xeon Scalable processors... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire