Q&A with Intel’s Jeff McVeigh, an HPCwire Person to Watch in 2022

By HPCwire Editorial Team

June 17, 2022

HPCwire presents our interview with Jeff McVeigh, vice president and general manager, Super Compute Group, Intel Corporation, and an HPCwire 2022 Person to Watch. McVeigh shares Intel’s plans for the year ahead, his perspective on HPC-AI synergies and the differentation in Intel’s AI-directed products. Also covered are quantum computing, HPC trends, and more. This interview was conducted by email at the beginning of April.

Jeff, congratulations on your selection as a 2022 HPCWire Person to Watch and on your promotion! Summarize the major milestones achieved last year for your division and briefly outline your HPC/AI agenda for 2022.

Thank you, it’s truly an honor! Last year, when Pat Gelsinger returned to Intel as our CEO, we established ambitious goals regarding the future of computing using the superpowers of ubiquitous computing, pervasive connectivity, artificial intelligence, and cloud-to-edge infrastructure. To achieve our goals, we brought together our business and product portfolio covering our high-end Xeon processors that power over 85% of the today’s HPC deployments with our upcoming high-end datacenter GPU products. We combine these hardware products with a comprehensive HPC software stack from drivers to tools to applications that provide our customers with complete solutions to help solve the world’s most challenging problems, faster.

Over the past twelve months, we launched the 3rd-generation Intel Xeon Scalable process (previously named Ice Lake), which delivered up to 30% better performance gen-to-gen on a vast set of HPC workloads; we demonstrated world-record performance on the forthcoming Next Gen Intel Xeon Scalable processors, codenamed Sapphire Rapids with integrated High Bandwidth Memory that offers applications up to 4x more raw memory bandwidth, and up to 2.8x the workload performance compared to current platforms; and our flagship Intel datacenter GPU, codenamed Ponte Vecchio that will power the Aurora supercomputer, started sampling to customers including Argonne National Labs. We also significantly ramped the oneAPI ecosystem with two new Centers of Excellence, which provides developers with a fully open programming model across diverse architectures and even vendors.

As you can see, we have a super year ahead as we get ready to ship compelling products and solutions targeted for HPC & AI.

How do you see the relationship between HPC and AI, both broadly and more specifically at Intel? Tie this to specific Intel products or goals.

At the most basic level, AI and HPC workloads need all the computing density and memory bandwidth that we can muster. While different workloads will have different needs and configurations (maximum core count, or frequency, or bandwidth), we must always provide balanced systems that match the workload characteristics. As AI models become more sophisticated, we’re seeing those get infused into traditional HPC codes. At first, these are used to augment computational models to speed iterations and overtime they’re expected to fully merge. That’s why we are increasingly integrating AI into our HPC products and traditional HPC requirements into our AI products.

For example, Xeon processors are the only x86 CPUs with built-in AI acceleration optimized to analyze the massive data sets in HPC workloads. With Next Gen Intel Xeon Scalable processors, codenamed Sapphire Rapids will deliver nearly 2x performance gain for end-to-end AI workloads. Additionally, our flagship datacenter GPU – Ponte Vecchio – has industry-leading, full rate double-precision support in addition to a massive L2 cache, high-bandwidth memory, dense vector/matrix compute, and glue less scale-up/out. This hardware portfolio is supported by the foundational oneAPI programming model to enable productive performance for HPC and AI applications.

Beyond the products in market today and those that will be launched later this year, we’re also working on a brand-new architecture codenamed Falcon Shores. Falcon Shores will bring x86 and Xe GPU acceleration together into the Xeon socket, taking advantage of next-generation packaging, memory, and I/O technologies for huge performance gains. We will also see efficiency improvements in systems computing large data sets and training gigantic AI models. We expect Falcon Shores to deliver more than 5x performance-per-watt, compute density, and memory capacity and bandwidth – all in a single socket with a vastly simplified GPU programming model. Products like Falcon Shores will accelerate the convergence of HPC & AI.

Intel has several different devices that offer “AI capabilities.” How do you differentiate these different product lines, including the Xe GPU?

Artificial intelligence is the fastest-growing compute workload from datacenter to mobile to the intelligent edge. Powering AI’s proliferation requires a holistic approach that accounts for vastly different use cases, workloads, power requirements and more. That’s why our silicon portfolio is comprised of a diverse mix of architectures from CPUs, GPUs, FPGAs and ASICs, which address specific customer requirements. Some of these products target power-constrained environments (client and IOT devices), others target cloud computing environments with diverse needs across AI and non-AI services, and others target dedicated AI training use cases. One size doesn’t fit all from a hardware perspective, but we need one, unified and open programming model across them – the oneAPI industry initiative provides this foundation.

It has been proposed that quantum accelerators will be integrated into either an HPC architecture or workflow. How do you see HPC and quantum coming together?

Quantum computing has the potential to solve problems that conventional computing – even the world’s most powerful supercomputers – can’t address in reasonable compute times or feasible level of energy consumption. However, before we integrate quantum computing into scaled HPC deployments, we need to solve some of the fundamental technical hurdles.

One of these hurdles is the underlying physics of qubits. We’ve uniquely pioneered research in silicon spin qubit technology, an alternative approach to other qubit technologies being pursued in the quantum field. Spin qubits resemble the electron transistors Intel has spent the last 50 years manufacturing at very large scale.

Additional hurdles include the achievement of high-fidelity multi qubit operations, efficient interconnect or wiring between qubits, fast and low-power qubit control electronics, plus high-performance, energy-efficient error correction to reduce the impact of qubit fragility.

While we’re still in the early stages of quantum computing research, we continue to make progress and produce positive outcomes, which will bring the world closer to realizing the true potential of quantum computing. Partnerships, like the one we have with QuTech/Delft University and Argonne National Laboratories will help address these technically complex challenges.

Computational milestones are inspiring and exciting. What is your vision for the exascale era?

While reaching a specific compute milestone can be treated as just another number, I feel that the exascale era will create a mind shift that will fundamentally change the way algorithms are designed to achieve a multiplicative affect. Instead of just speeding up an existing simulation by an order-of-magnitude, what if that simulation could be 10-times more accurate to develop novel materials that help to reverse climate change, or 100X more variations could be run to quickly discover cures to chronic diseases.

Evolutions in architectures or algorithms bring about incremental change; revolutions occur when both advance simultaneously. This is the door that opens with the exascale era.

Intel’s top execs are emphasizing a new era of openness. Can you comment on this and offer a few examples?

An open approach has always been part of our DNA. Throughout Intel’s history we’ve taken the approach of driving open platforms and industry-shaping standards with the APIs that enabled them for the good of the computing ecosystem. Open standards drive ubiquitous computing, which drives demand for more computing. A small sampling of these includes USB, Wi-Fi, and UCIe. These enable the entire ecosystem to be successful – end-users, developers, partners, and enterprises – instead the few when constrained to a proprietary solution.

An open ecosystem is the north star that guides our roadmaps and how we bring products to market. Our mission for the Super Compute Group is to drive these standards and partner with the ecosystem to create the foundational hardware platforms the industry can rally behind and build upon in the service of helping our customers solve the world’s most challenging problems, faster.

One of our most recent examples of this is how we incubated the oneAPI ecosystem and ensured its broad adoption across architectures and vendors. This provides developers with choice through collaborations with the broad ecosystem and working together with our industry partners to solve systemic challenges of security and distribution. We will also collaborate with academia and developers to innovate on future challenges like neuromorphic and quantum computing – all in the open.

Where do you see HPC headed? What trends – and in particular emerging trends – do you find most notable? Any areas you are concerned about, or identify as in need of more attention/investment?

Delivering supercomputing for all requires scale but it is equally important to reduce energy consumption. The increasingly open and distributed nature of compute, growth in AI/ML applications and heterogeneous computing will help HPC scale. The industry will have to make progress in lowering the power and cooling costs of these systems to make them more efficient and accessible to the broader community. I’m excited about the possibilities technologies like high bandwidth memory will bring to HPC and AI workloads. Intel has connected the dots between HPC and AI by building products that are optimized for these workloads.

To advance HPC from Exa-scale to Zetta-scale we need significant, revolutionary gains in architecture, power efficiency, thermal management, process and packaging technology, memory and IO capacity and bandwidth – all supporting rapid evolution of the software applications that run on top.

More generally, what excites you about working in high-performance computing?

Today, we are at the threshold of a new generation of HPC, where the technology’s scalability, ubiquity and accessibility can transform all our lives. I am very pleased that we’ve democratized HPC and put it in the hands of scientists who now have easy access to supercomputers to advance cutting edge research.

Nowhere has this been more obvious than with the battle against Covid-19. From the start of the pandemic, the scientific and research communities tapped these advanced supercomputers – both within research labs and in cloud HPC-as-a-service environments – to monitor, study, treat and eventually develop the drugs used to treat the SARS-CoV-2 novel coronavirus. The speed at which this was done was breathtaking and would have been impossible had it not been for the broad availability of HPC technologies.

I am excited and humbled to be leading Intel’s Super Compute Group responsible for building the technology foundations to advance HPC and make it accessible. We are energized with the new vision and strategy that Pat laid out and are committed to executing at a torrid pace to deliver our new products to customers this year.

What led you to pursue a career in the computing field and what are your suggestions for engaging the next generation of IT professionals?

I’ve always been fascinated with technology, even from my early childhood when I would disassemble our family telephone or TV to see how it worked – often to the frustration of my parents! My PhD thesis was on efficient compression of multi-view video signals – basically, how to use computational power and machine learning to predict different viewpoints. Now this was in the mid-1990s, so both the compute resources and AI algorithms were far from today’s capabilities, but that experience gave me the foundations on the interplay of architecture and algorithms.

My initial work at Intel carried this forward during the early days of video conferencing and streaming (well before we all used Zoom for work or YouTube even existed). I then worked on Intel’s integrated graphics products and then our developer tools. These experiences taught me about the importance of end-to-end solutions to solve customer problems and ensuring both performance and productivity – one without the other is always an uphill battle.

For the next generation, I recommend seeking out opportunities that match your passion as it’s always best to work on projects that you’re curious about and get you motivated to put in the effort every day. I also stress the importance of communication to convey the “so what” behind data or trends or issues that need to be fixed. If you can’t communicate clearly in all directions – peers, employees, or leaders – you’ll be fighting those battles alone.

Outside of the professional sphere, what can you tell us about yourself – family stories, unique hobbies, favorite places, etc.? Is there anything about you your colleagues might be surprised to learn?

I love the outdoors – skiing, running, hiking, mountaineering – anything that connects me with nature. To feed this thirst, I recently bought a sprinter van that I’m outfitting for the sole purpose of chasing powder for skiing. I’m also a big football fan. I enjoy these hobbies with my awesome wife, two terrific daughters and our sometimes-misbehaving dog Tillamook (who got his name because of his love of eating butter off the kitchen counter).

McVeigh is one of 12 HPCwire People to Watch for 2022. You can read the interviews with the other honorees at this link.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire