Intel Announces Falcon Shores CPU-GPU Combo Architecture for 2024

By Oliver Peckham

February 18, 2022

Intel held its 2022 investor meeting yesterday, covering everything from the imminent Sapphire Rapids CPUs to the hotly anticipated (and delayed) Ponte Vecchio GPUs. But somewhat buried in its summary of the meeting was a new namedrop: “Falcon Shores,” described as “a new architecture that will bring x86 and Xe GPU together into a single socket.”

The reveal was brief, delivered by Raja Koduri (senior vice president and general manager of the Accelerated Computing Systems and Graphics [AXG] Group at Intel) over the course of just a couple minutes toward the end of a virtual breakout session.

Falcon Shores

“We are working on a brand new architecture codenamed Falcon Shores,” Koduri said. “Falcon Shores will bring x86 and Xe GPU acceleration together into a Xeon socket, taking advantage of next-generation packaging, memory, and I/O technologies, giving huge performance and efficiency improvements for systems computing large datasets and training gigantic AI models.”

Image courtesy of Intel.

“We expect Falcon Shores to deliver more than 5× performance per watt, more than [a] 5× compute density increase and more than 5× memory capacity and bandwidth improvement [relative to current platforms], all in a single socket with a vastly simplified GPU programming model,” he continued. “Falcon Shores is built on top of an impressive array of technologies … including the Angstrom-era process technology, next-generation packaging, new extreme bandwidth shared memory being developed by Intel, and industry-leading I/O. We are super excited about this architecture as it brings acceleration to a much broader range of workloads than the current discrete solutions.”

According to the roadmap, Falcon Shores is slated for 2024.

Coup de Grace?

One might compare the concept of Intel’s Falcon Shores to Nvidia’s forthcoming Grace CPU—its first Arm-based processor, and one designed (like Falcon Shores) to tightly couple with its in-house GPUs. In Nvidia’s case, Grace will use next-generation Nvidia NVLink and future-generation Arm Neoverse cores to couple with GPUs at a one-to-one CPU-GPU ratio. Grace was announced almost a year ago (April 2021) and Nvidia is slating it for 2023, a year earlier than Falcon Shores.

Sapphire Rapids

These are all, however, a ways off—and Koduri also took the time to highlight items much closer on the horizon than Falcon Shores. Sapphire Rapids, for instance—Intel’s next-gen Xeon—will start shipping next month to select customers.

“Sapphire raises the bar and sets a new standard in the industry for workload-optimized performance,” said Sandra Rivera, executive vice president and general manager of the Datacenter and AI Group at Intel, during the event. “Sapphire Rapids will also lead the industry in important memory and interconnect standards. For example, PCIe [and] DDR5, as well as the new high-speed cache coherent interconnect CXL, a standard that Intel led on in the industry.”

Sapphire Rapids will also come in a high-bandwidth memory (HBM) flavor, which Intel stressed throughout the event. “Our strategy is to build on this foundation [with Xeon] and extend this to even higher compute and memory bandwidth,” Koduri said. “First, we are bringing high-bandwidth memory—or HBM—integrated into the package with the Xeon CPU, offering GPU-like memory bandwidths to CPU workloads.” Sapphire Rappids with HBM is expected to ship, Koduri said, in the second half of 2022.

Image courtesy of Intel.

Ponte Vecchio

Slightly further out than Sapphire Rapids is Ponte Vecchio, Intel’s discrete Xe GPU that is slated to serve as the heart of the United States’ Aurora exascale supercomputer. Aurora—initially scheduled for 2021—is now (ostensibly) scheduled for late 2022, in lockstep with Intel’s process node challenges and the resulting delays for Ponte Vecchio. Originally intended to use Intel’s 7nm process (now known as “Intel 4”), the main compute tile of Ponte Vecchio is now being manufactured on TSMC’s N5 process.

“We are on track to deliver this GPU for [the] Aurora supercomputer program later this year,” Koduri assured listeners. “We are making steady progress on this product, and we are excited to show some early leadership performance results on this GPU.” To that end, he presented performance comparisons against an unidentified competitor product on a financial services workload benchmark, saying pre-production Ponte Vecchio units showed a “significant performance improvement over the best solution in the market today.”

Image courtesy of Intel.

oneAPI

Gelling Intel’s near-term strategy is its oneAPI programming model.

“Combining Ponte Vecchio with Xeon HBM is great from a hardware perspective,” Koduri said, “but equally important is bringing a way to seamlessly and transparently take advantage of that hardware technology with the existing base of Xeon HPC and AI software. That is where our OneAPI open ecosystem comes into play. We architected OneAPI to leverage the Xeon software ecosystem seamlessly, allowing software developers to work across a range of CPUs and accelerators with a single codebase. By making OneAPI an open ecosystem, we move the barriers of [the] closed, proprietary programming environment that current GPU accelerators use in HPC and AI.”

Koduri added that Intel expects “tremendous momentum in developer adoption of our open approach this year as Xeon HBM and [Ponte Vecchio] get more accessible to everyone.”

… and beyond

“We have follow-on products for Xeon HBM and PVC in 2023 coming as well,” he said. Among those: Emerald Rapids, Intel’s next-generation Xeon processor on the Intel 7 process node.

2024, then, brings us back to Falcon Shores. 2024 will also bring Granite Rapids, another next-generation Xeon product—a P-core processor previously slated for the Intel 4 process, but now upgraded to the Intel 3 process. Granite Rapids will launch alongside Sierra Forest, an E-core Xeon processor also on the Intel 3 process.

Intel’s updated Xeon roadmap shows dual-track Performance/Efficiency schema. Image courtesy of Intel.

 

Intel’s process node roadmap and 2022 milestones. Presented by Pat Gelsinger at Intel’s investor meeting on Feb. 17, 2022.
Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

2024 Winter Classic: Texas Two Step

April 18, 2024

Texas Tech University. Their middle name is ‘tech’, so it’s no surprise that they’ve been fielding not one, but two teams in the last three Winter Classic cluster competitions. Their teams, dubbed Matador and Red Read more…

2024 Winter Classic: The Return of Team Fayetteville

April 18, 2024

Hailing from Fayetteville, NC, Fayetteville State University stayed under the radar in their first Winter Classic competition in 2022. Solid students for sure, but not a lot of HPC experience. All good. They didn’t Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use of Rigetti’s Novera 9-qubit QPU. The approach by a quantum Read more…

2024 Winter Classic: Meet Team Morehouse

April 17, 2024

Morehouse College? The university is well-known for their long list of illustrious graduates, the rigor of their academics, and the quality of the instruction. They were one of the first schools to sign up for the Winter Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire