Intel’s Rental Service on Chips Could Face Buyer Backlash

By Agam Shah

January 11, 2023

Intel is bringing subscription and rental services to semiconductors as it explores new business models, but it remains to be seen if buyers warm up to the idea of paying extra to unlock features on a chip.

Intel is bringing an “on-demand” feature to its new Xeon CPUs codenamed Sapphire Rapids, which the company launched on Tuesday after long delays.

The Intel On Demand program involves paying a fee to activate specific features on chips. The on-demand service is like renting a movie – you pay a fee to unlock the content.

Intel in a press document explained on-demand as a feature to “expand and/or upgrade accelerators and hardware-enhanced features available on most 4th Gen Xeon processor SKUs.”

Intel’s fourth-gen Xeon Sapphire Rapids CPU

Some on-chip features that customers could rent include accelerators integrated into the CPU that provide extra juice to applications in artificial intelligence, analytics, networking and storage.

The on-demand features are available in 36 of the 52 new processors with Sapphire Rapids CPUs announced by the chipmaker.

The chips with on-demand are for mainline servers, and big iron in public and private cloud server infrastructures. The chips with on-demand features have up to 56 cores and 105MB of cache, and are priced between $624 and $10,710.

The on-demand features are not available on the Xeon Max series high-performance computing chips, which are designed for supercomputers. The Xeon Max chips have tightly integrated components that include memory and accelerators.

The point of on-demand was to provide chips with certain features turned off, but can be activated when customers know the software development cycle is complete and ready to take advantage of the accelerators, said Ronak Singhal, a senior fellow at Intel.

Software development typically has a long tail and requires additional tuning to work with specific accelerators.

Customers may not know if applications would use those accelerators on day zero, Singhal said, adding that “the on-demand capability allows you to… not pay for something you’re not going to use.”

The cost of silicon is an important part of on-demand offerings, said Lisa Spelman, corporate vice president and general manager for Xeon at Intel, during a press briefing.

“It not only provides that flexibility for when you utilize that portion of the silicon, it’s not consuming energy as well,” Spelman said.

Schema depicting how On Demand is activated in the field (Source: Intel)

The on-demand feature is also an opportunity for Intel to explore and build up a new business model, Spelman said. The company has tried the on-demand business model with some customers with the previous two generations of Xeon chips.

“We’ve reached a more production level, so this is kind of the start of it. You’ll see it continue as we evolve into our next generations as well. It is in-step and in-line with the way that just about every other product is consumed. Even a car has become an on-demand type of offering,” Spelman said.

The additional benefit of the new business model is that it also lowers the cost of silicon ahead of customers deploying new servers, Intel officials said.

The pricing for on-demand services remains unclear, and Intel did not respond to requests for comment on how customers will be expected to pay for such services. But there were hints on how the services could be bundled through cloud offerings.

Microsoft announced it was bringing the fourth-gen Intel Xeon Scalable processors to its Azure cloud service with confidential computing features based on TDX (Trust Domain Extensions). TDX are on-chip instructions that create a trusted execution environment to securely store and move data and run applications. The instructions create a secure boundary in virtual machines hosting guest OS and applications, which can’t be accessed by hypervisors.

Microsoft didn’t provide pricing of the Sapphire Rapids instances with TDX. But it could be priced in line with Azure’s confidential computing instances such as DCasv5 and ECasv5 based on AMD Epyc chips, which have the SEV-SNP secure enclave feature. The DCasv5 starts at $69 per month with a barebone configuration, and goes up to $3,314 per month with 96 Epyc CPU cores and 384GB of memory.

Intel’s main rival AMD isn’t charging customers rental or subscription fees for features on its chips, and is taking more of a general-purpose computing approach with Epyc chips. Intel is focusing more silicon space on acceleration for specific workloads, and charging for it, which could negatively impact the business. But only time will tell if Intel’s gamble on renting out that real estate on silicon was a good move, analysts said.

“Maybe this is not a winning strategy, which is making users pay for silicon,” said Dylan Patel, who is founder of SemiAnalysis, a semiconductor research and consulting firm.

The next-generation Epyc chips from AMD, on the other hand, could have chiplets with Xilinx and XDNA AI accelerators, and customers could have more choice in customizing chips. AMD could retain its general-purpose computing message, while offering customers the choice of an AI chiplet or a plain CPU, while saying “we’re not going to make you pay for silicon that’s not used,” Patel said.

For Intel, which has its own factories, there are a lot more things to consider.

“At the end of the day, in your manufacturing costs, you put those circuits on the die, and you’re manufacturing them. In that case, it is a trade-off of less cores. Intel, with Sapphire Rapids, and obviously Granite Rapids, they will do the same with more accelerators. But with Sierra Forest, they will do the opposite, which is a bunch of lower performance cores, a ton of them. It is an interesting divergence,” Patel said.

The disparate approaches to server chips by AMD and Intel are stark, but it’s hard to tell which chipmaker holds the advantage. It depends on what applications emerge, and what kind of chips they require. Intel’s on-demand approach to accelerators may work out to the company’s advantage.

“These workloads don’t all exist today. But as they start to get developed, it will be a big deal,” Patel said.

This is not the first time Intel is charging customers to unlock features on CPUs. In 2010, Intel charged $50 to unlock features on the Pentium G6951 chip, which caused consumer backlash.

“In theory, it’s a good idea – you only pay for what you need. But we will have to wait and see how the market adapts to it,” said Jim McGregor, principal analyst at Tirias Research.

It may be difficult for Intel to justify on-demand features for customers buying silicon directly from the chipmaker or through its server partners for on-prem use. Switching features on and off in on-premises servers may present challenges as Intel may not have access to servers in such installations. (A license can also be activated at the time of purchase, according to Intel.)

“It’s easier for hyperscalers and IT service providers to charge for it. That makes more sense. If you go and get AWS or Azure, you pay for what you need, or you pay by the minute, or you pay for the resources. It is kind of hard to do that at the silicon level,” McGregor said.

The cost of renting silicon space on-demand will be tacked on to the base price of the Sapphire Rapids chips, but it is yet to be seen on how it will fit into Intel’s business model.

“It is going to be awkward to have to go back and decide ‘I need this accelerator.’ It will take several years for the IT community, especially the enterprise community, to become accustomed to that,” McGregor said.

Intel does not see a market for on-demand in the high-performance computing market as features like high-bandwidth memory and specific accelerators are core parts of the system, not an upgrade option, McGregor said.

The accelerators that could be rented through on-demand include the In-Memory Analytics Accelerator (IAA), which is an engine that boosts database and big data performance by taking on analytics processing offloaded by CPUs. The accelerator focuses on analytics primitives, compression and decompression, and boosts performance by increasing the query throughput and decreasing the memory footprint, which allows for more effective bandwidth.

The IAA was able to boost the performance of RocksDB, which was created by Meta and stores values differently to relational databases, by up to three times, with 2.2x better performance per watt versus the 3rd Gen Xeon chips, Intel said during a press briefing.

Intel said the purchasing model also applies to its Dynamic Load Balancer (DLB), Data Streaming Accelerator (DSA), Quick Assist Technology (QAT) and Software Guard Extensions (SGX). Provider partners include H3C, HPE, Inspur, Lenovo, Supermicro, PhoenixNAP, and Variscale.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Point. The system includes Intel's research chip called Loihi 2, Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Research senior analyst Steve Conway, who closely tracks HPC, AI, Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, and this day of contemplation is meant to provide all of us Read more…

Intel Announces Hala Point – World’s Largest Neuromorphic System for Sustainable AI

April 22, 2024

As we find ourselves on the brink of a technological revolution, the need for efficient and sustainable computing solutions has never been more critical.  A computer system that can mimic the way humans process and s Read more…

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

Intel’s Silicon Brain System a Blueprint for Future AI Computing Architectures

April 24, 2024

Intel is releasing a whole arsenal of AI chips and systems hoping something will stick in the market. Its latest entry is a neuromorphic system called Hala Poin Read more…

Anders Dam Jensen on HPC Sovereignty, Sustainability, and JU Progress

April 23, 2024

The recent 2024 EuroHPC Summit meeting took place in Antwerp, with attendance substantially up since 2023 to 750 participants. HPCwire asked Intersect360 Resear Read more…

AI Saves the Planet this Earth Day

April 22, 2024

Earth Day was originally conceived as a day of reflection. Our planet’s life-sustaining properties are unlike any other celestial body that we’ve observed, Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

Intel’s Xeon General Manager Talks about Server Chips 

January 2, 2024

Intel is talking data-center growth and is done digging graves for its dead enterprise products, including GPUs, storage, and networking products, which fell to Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire