Understanding the GPU —The Catalyst of the Current AI Revolution

By Manasi Rashinkar

July 30, 2024

GPUs are often likened to the “gold” of artificial intelligence, vital to today’s generative AI age. This article aims to explain why AI is unachievable without GPUs. Let’s start with a simple processor task — displaying an image on the screen (shown below).

As straightforward as it seems, this task involves several steps: geometry transformation, rasterization, fragment processing, framebuffer operations, and output merging. These outline the GPU pipeline’s process for rendering 3D graphics.

(Source: https://img-new.cgtrader.com/items/708323/72453d4a2d/teapot-3d-model-c4d.jpg)

In a GPU pipeline, an image is converted into a polygon mesh representation, as seen below:

A teapot image is transformed into a polygon mesh that can rendered individually (in parallel)

A single teapot image is transformed into a mesh structure made of hundreds of triangles, each processed separately in the same manner.

What does a GPU offer that a CPU cannot in handling this “simple” task? High-end server CPUs can have up to 128 cores so that a CPU can simultaneously process 128 triangles in the teapot. The user sees a partially rendered teapot that slowly completes as the CPU cores finish and picks new triangles to render. Imagine playing Grand Theft Auto (GTA) and seeing the scene rendered in parts—it would ruin the experience, making even the old snake game seem more fun.

How does a GPU provide the complete GTA gaming experience? The answer is “parallelism” because of its tens of thousands of cores. A GPU can render all the triangles of the teapot simultaneously due to its many threads working on each triangle in parallel. Essentially, CPUs handle serial computing, whereas GPUs are built for parallel processing.

This captivating video showcases GPU computing’s power.

Initially created to boost 3D graphics rendering, GPUs have become more versatile and programmable over time. They add capabilities for better visual effects and realistic scenes through advanced lighting and shadowing, revolutionizing gaming. But it didn’t stop there. Developers saw GPUs’ untapped potential. Returning to our teapot example, GPUs perform vector-based mathematical calculations and matrix multiplications to render the image. Rendering a simple teapot requires about 192 bytes, while a complex GTA scene with 100 objects needs around 10KB.

GPUs’ built-in parallelism and high throughput lead to accelerated computing, pushing researchers to use GPUs for tasks like protein folding simulations and physics calculations. These early achievements showed that GPUs could speed up computation-heavy tasks beyond graphics rendering, such as matrix and vector operations used in neural networks. Although neural networks were achievable without GPUs, their capabilities were constrained by the available computational power. The advent of GPUs provided the necessary resources to train deep and complex neural networks effectively, driving rapid advancements and widespread adoption of deep learning techniques.

To allow GPUs to handle a wide range of tasks effectively, Nvidia has developed different types of GPU cores specialized for various functions:

  • CUDA Cores: These are for general-purpose parallel processing, including rendering graphics, scientific computations, and basic machine learning tasks.
  • Tensor Cores: Designed for deep learning and AI, they speed up tensor operations like matrix multiplications, which are essential for training and inference in neural networks.
  • RT Cores: Focused on real-time ray tracing, these provide realistic lighting, shadows, and reflections in graphics.

Does this mean GPUs can replace CPUs? Absolutely not! The CPU is like the brain of the computer, excelling in swiftly managing individual tasks with its fewer but more powerful cores. The CPU is oriented towards latency, reflecting system response time, while the GPU is about throughput, indicating system capacity. The GPU’s journey from merely a graphics accelerator to playing a pivotal role in supercomputers marks a tale of rapid technological progress and expanding applications. Machine learning used to be slow and inaccurate, but the integration of GPUs revolutionized large neural networks, driving advancements in fields like autonomous driving and image/object recognition. High-performance computing, now a leading enterprise technology, has largely been propelled by GPUs.


Manasi Rashinkar holds a Master of Science in Electrical Engineering from Santa Clara University and is currently the senior ASIC Engineer—timing Lead at Nvidia. This article is Manasi’s own work and does not represent Nvidia.

 

 

 

 

 

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

At 50, Foxconn Celebrates Graduation from Connectors to AI Supercomputing

October 8, 2024

Foxconn is celebrating its 50th birthday this year. It started by making connectors, then moved to systems, and now, a supercomputer. The company announced it would build the supercomputer with Nvidia's Blackwell GPUs an Read more…

ZLUDA Takes Third Wack as a CUDA Emulator

October 7, 2024

The ZLUDA CUDA emulator is back in its third invocation. At one point, the project was quietly funded by AMD and demonstrated the ability to run unmodified CUDA applications with near-native performance on AMD GPUs. Cons Read more…

Quantum Companies D-Wave and Rigetti Again Face Stock Delisting

October 4, 2024

Both D-Wave (NYSE: QBTS) and Rigetti (Nasdaq: RGTI) are again facing stock delisting. This is a third time for D-Wave, which issued a press release today following notification by the SEC. Rigetti was notified of delisti Read more…

Alps Scientific Symposium Highlights AI’s Role in Tackling Science’s Biggest Challenges

October 4, 2024

ETH Zürich recently celebrated the launch of the AI-optimized “Alps” supercomputer with a scientific symposium focused on the future possibilities of scientific AI thanks to increased compute power and a flexible ar Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks. These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvidia GPUs). Recently, MLCommons introduced the results of its Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago this week emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whatever physical processor they want, without making code changes, Read more…

At 50, Foxconn Celebrates Graduation from Connectors to AI Supercomputing

October 8, 2024

Foxconn is celebrating its 50th birthday this year. It started by making connectors, then moved to systems, and now, a supercomputer. The company announced it w Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks. These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvi Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago this week emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whateve Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed t Read more…

How GenAI Will Impact Jobs In the Real World

September 30, 2024

There’s been a lot of fear, uncertainty, and doubt (FUD) about the potential for generative AI to take people’s jobs. The capability of large language model Read more…

IBM and NASA Launch Open-Source AI Model for Advanced Climate and Weather Research

September 25, 2024

IBM and NASA have developed a new AI foundation model for a wide range of climate and weather applications, with contributions from the Department of Energy’s Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Building the Quantum Economy — Chicago Style

September 24, 2024

Will there be regional winner in the global quantum economy sweepstakes? With visions of Silicon Valley’s iconic success in electronics and Boston/Cambridge� Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Granite Rapids HPC Benchmarks: I’m Thinking Intel Is Back (Updated)

September 25, 2024

Waiting is the hardest part. In the fall of 2023, HPCwire wrote about the new diverging Xeon processor strategy from Intel. Instead of a on-size-fits all approa Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Ansys Fluent® Adds AMD Instinct™ MI200 and MI300 Acceleration to Power CFD Simulations

September 23, 2024

Ansys Fluent® is well-known in the commercial computational fluid dynamics (CFD) space and is praised for its versatility as a general-purpose solver. Its impr Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Leading Solution Providers

Contributors

IBM Develops New Quantum Benchmarking Tool — Benchpress

September 26, 2024

Benchmarking is an important topic in quantum computing. There’s consensus it’s needed but opinions vary widely on how to go about it. Last week, IBM introd Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qu Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

US Implements Controls on Quantum Computing and other Technologies

September 27, 2024

Yesterday the Commerce Department announced export controls on quantum computing technologies as well as new controls for advanced semiconductors and additive Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire