Is the GenAI Bubble Finally Popping?

By Alex Woodie

August 21, 2024

Doubt is creeping into discussion over generative AI, as industry analysts begin to publicly question whether the huge investments in GenAI will ever pay off. The lack of a “killer app” besides coding co-pilots and chatbots is the most pressing concern, critics in a Goldman Sachs Research letter say, while data availability, chip shortages, and power concerns also provide headwinds. However, many remain bullish on the long-term prospects of GenAI for business and society.

The amount of sheer, unadulterated hype layered onto GenAI over the past year and a half certainly caught the attention of seasoned tech journalists, particularly those who lived through the dot-com boom and ensuing bust at the turn of the century, not to mention the subsequent rise of cloud computing and smartphones with the introduction of Amazon Web Services and the Apple iPhone in 2006 and 2007, respectively.

The big data boom of the early 2010s was the next tech obsession, culminating with the coronation of Hadoop as The New New Thing, to paraphrase Michael Lewis’ illuminating 1999 book into Silicon Valley’s fixation on continuous technological reinvention. After the collapse of Hadoop–slowly at first, and then all of a sudden in 2019–the big data marketing machine subtly shifted gears and AI was the hot new thing. Several other new (new) things made valiant runs for attention and VC dollars along the way–Blockchain will change the world! 5G will turbocharge edge computing! Self-driving cars are almost here! Smart dust is new oil!–but nothing really seemed to really gain traction, and the big data world made incremental gains with traditional machine learning while wondering what these newfangled neural networks would ever be good for.

GenAI is the newest new thing

That is, until OpenAI dropped a new large language model (LLM) called ChatGPT onto the world in late 2022. Since then, the hype level for neural network-powered AI, and transformer network-based GenAI in particular, has been eerily reminiscent of these previous Big Moments In Tech. It’s worth pointing out that some of these big moments turned out to be actual inflection points, such as mobile and cloud, some had us asking ourselves “What were we thinking (blockchain, 5G), while it took years for the full lessons from other technological breakthroughs to become apparent (the dot-com boom, even Hadoop-style computing).

So the big question for us now is: Which of those categories will we be putting GenAI into in five years? One of the voices suggesting AI may go the way of 5G and blockchain is none other than Goldman Sachs. In a much-read report from the June edition of the Goldman Sachs Research Newsletter titled “Gen AI: too much spend, too little benefit?” Editor Allison Nathan ponders whether AI will pan out.

“The promise of generative AI technology to transform companies, industries, and societies continues to be touted, leading tech giants, other companies, and utilities to spend an estimated ~$1tn on capex in coming years, including significant investments in data centers, chips, other AI infrastructure, and the power grid,” she writes. “But this spending has little to show for it so far beyond reports of efficiency gains among developers.”

Nathan interviewed MIT Professor Daron Acemoglu, who said that only a quarter of tasks that AI is supposed to automate will actually be automated in a cost-effective manner. Overall, Acemoglu estimates that only 5% of all tasks will be automated within 10 years, raising the overall productivity of the United States by less than 1% over that time.

“Generative AI has the potential to fundamentally change the process of scientific discovery, research and development, innovation, new product and material testing, etc. as well as create new products and platforms,” Acemoglu told Nathan. “But given the focus and architecture of generative AI technology today, these truly transformative changes won’t happen quickly and few–if any–will likely occur within the next 10 years.”

Accelerating GenAI progress by ramping up production of its two core ingredients–data and GPUs–probably won’t work, as data quality is a big piece of the equation, Acemoglu said.

GenAI seems to attract irrational exuberance (Roman-Samborskyi/Shutterstock)

“Including twice as much data from Reddit into the next version of GPT may improve its ability to predict the next word when engaging in an informal conversation,” he said, “but it won’t necessarily improve a customer service representative’s ability to help a customer troubleshoot problems with their video service.”

A shortage in chips suitable for training GenAI models is another factor in Goldman’s pessimistic (some would say realistic) take on GenAI. That has benefited Nvidia enormously, which saw revenue grow by more than 260%, to $26 billion, for the quarter ended April 28. That helped pump its market cap over the $3-trillion market, joining Microsoft and Apple as the most valuable companies in the world.

“Today, Nvidia is the only company currently capable of producing the GPUs that power AI,” Jim Covello, Goldman’s head of global equity research, wrote in the newsletter. “Some people believe that competitors to Nvidia from within the semiconductor industry or from the hyperscalers–Google, Amazon, and Microsoft–themselves will emerge, which is possible. But that’s a big leap from where we are today given that chip companies have tried and failed to dethrone Nvidia from its dominant GPU position for the last 10 years.”

The huge costs involved in training and using GenAI act as headwinds against any productivity or efficiency gains that the GenAI may ultimately deliver, Covello said.

“Currently, AI has shown the most promise in making existing processes–like coding–more efficient, although estimates of even these efficiency improvements have declined, and the cost of utilizing the technology to solve tasks is much higher than existing methods,” he wrote.

Nvidia’s fortunes have skyrocketed thanks to GPU demand from GenAI

Covello was semiconductor analyst when smartphones were first introduced, and learned a few lessons about what it takes to actually realize monetary gains from technological innovation. For instance, the smartphone makers promised to integrate global positioning systems (GPS) into the phones, he said, and they had a roadmap that proved prescient.

“No comparable roadmap exists today” for AI, he said. “AI bulls seem to just trust that use cases will proliferate as the technology evolves. But eighteen months after the introduction of generative AI to the world, not one truly transformative–let alone cost-effective–application has been found.”

Finally, the amount of power required to train LLMs and other GenAI models has to be factored into the equation. It’s been estimated that AI currently consumes about 0.5% of the world’s energy, and that amount is expected to increase in the future.

“Utilities are fielding hundreds of requests for huge amounts of power as everyone chases the AI wave, but only a fraction of that demand will ultimately be realized,” says Brian Janous, the Co-founder of Cloverleaf Infrastructure and formerly the VP of energy at Microsoft.

The total capacity of power projects waiting to connect to the grid grew nearly 30% last year, with wait times currently ranging from 40-70 months, Janous said. With so many projects waiting for power, data centers looking for more power to fuel AI training will become “easy targets.”

The US needs to expand its grid to handle expected increase for power demand, but that isn’t likely to be done cheaply or efficiently, he said. “The US has unfortunately lost the ability to build large infrastructure projects–this is a task better suited for 1930s America, not 2030s America,” Janous said. “So, that leaves me a bit pessimistic.”

The enormous electricity demands of AI, and the US’s inabilty to build new power sources, also pose headwinds to AI success (BESTWEB/Shutterstock)

But not everyone is pessimistic about AI’s future. One GenAI optimist is Joseph Briggs, Goldman’s senior global economist. In his article countering  Acemoglu, Briggs estimates that GenAI ultimately will automate 25% of all work tasks and raise US productivity by 9% and GDP growth by 6.1% cumulatively over the next decade. What’s more, GenAI will not only automate some existing tasks currently done by humans, but will spur the creation of new tasks, he said.

“…[T]he full automation of AI exposed tasks that are likely to occur over a longer horizon could generate significant cost savings to the tune of several thousands of dollars per worker per year,” he wrote. “The cost of new technologies also tends to fall rapidly over time. Given that cost-saving applications of generative AI will likely follow a similar pattern, and that the marginal cost of deployment will likely be very small once applications are developed, we expect AI adoption and automation rates to ultimately far exceed Acemoglu’s 4.6% estimate.”

Kash Rangan is another GenAI believer. In an interview with the Goldman editor Nathan, the senior equity research analyst said he’s amazed at the pace of GenAI innovation and impressed at the infrastructure buildout of the cloud bigs. He acknowledged that GenAI hasn’t discovered its killer app yet, in the way that ERP dominated the 1990s, search and e-commerce dominated the 2000s, and cloud applications dominated the 2010s.

“But this shouldn’t come as a surprise given that every computing cycle follows a progression known as IPA—infrastructure first, platforms next, and applications last,” Rangan said. “The AI cycle is still very much in the infrastructure buildout phase, so finding the killer application will take more time, but I believe we’ll get there.”

His colleague, Eric Sheridan, joined him in a bullish stance.

“So, the technology is still very much a work in progress. But it’s impossible to sit through demonstrations of generative AI’s capabilities at company events or developer conferences and not come away excited about its long-term potential,” he said.

“So, while I would never say I’m not concerned about the possibility of no payback, I’m not particularly worried about it today, though I could become more concerned if scaled consumer applications don’t emerge over the next 6-18 [months],” Sheridan said.

The promise of GenAI remains high, if unfulfilled at the end of the day. The big question right now is whether GenAI’s returns will go up before the clock runs out. The clock is ticking.

Related Items:

Gartner Warns 30% of GenAI Initiatives Will Be Abandoned by 2025

GenAI Hype Bubble Refuses to Pop

When GenAI Hype Exceeds GenAI Reality

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Argonne’s HPC/AI User Forum Wrap Up

September 11, 2024

As fans of this publication will already know, AI is everywhere. We hear about it in the news, at work, and in our daily lives. It’s such a revolutionary technology that even established events focusing on HPC specific Read more…

Quantum Software Specialist Q-CTRL Inks Deals with IBM, Rigetti, Oxford, and Diraq

September 10, 2024

Q-CTRL, the Australia-based start-up focusing on quantum infrastructure software, today announced that its performance-management software, Fire Opal, will be natively integrated into four of the world's most advanced qu Read more…

Computing-Driven Medicine: Sleeping Better with HPC

September 10, 2024

As a senior undergraduate student at Fisk University in Nashville, Tenn., Ifrah Khurram's calculus professor, Dr. Sanjukta Hota, encouraged her to apply for the Sustainable Research Pathways Program (SRP). SRP was create Read more…

LLNL Engineers Harness Machine Learning to Unlock New Possibilities in Lattice Structures

September 9, 2024

Lattice structures, characterized by their complex patterns and hierarchical designs, offer immense potential across various industries, including automotive, aerospace, and biomedical engineering. With their outstand Read more…

NSF-Funded Data Fabric Takes Flight

September 5, 2024

The data fabric has emerged as an enterprise data management pattern for companies that struggle to provide large teams of users with access to well-managed, integrated, and secured data. Now scientists working at univer Read more…

xAI Colossus: The Elon Project

September 5, 2024

Elon Musk's xAI cluster, named Colossus (possibly after the 1970 movie about a massive computer that does not end well), has been brought online. Musk recently posted the following on X/Twitter: "This weekend, the @xA Read more…

Shutterstock 793611091

Argonne’s HPC/AI User Forum Wrap Up

September 11, 2024

As fans of this publication will already know, AI is everywhere. We hear about it in the news, at work, and in our daily lives. It’s such a revolutionary tech Read more…

Quantum Software Specialist Q-CTRL Inks Deals with IBM, Rigetti, Oxford, and Diraq

September 10, 2024

Q-CTRL, the Australia-based start-up focusing on quantum infrastructure software, today announced that its performance-management software, Fire Opal, will be n Read more…

NSF-Funded Data Fabric Takes Flight

September 5, 2024

The data fabric has emerged as an enterprise data management pattern for companies that struggle to provide large teams of users with access to well-managed, in Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Shutterstock 1897494979

What’s New with Chapel? Nine Questions for the Development Team

September 4, 2024

HPC news headlines often highlight the latest hardware speeds and feeds. While advances on the hardware front are important, improving the ability to write soft Read more…

Critics Slam Government on Compute Speeds in Regulations

September 3, 2024

Critics are accusing the U.S. and state governments of overreaching by including limits on compute speeds in regulations and laws, which they claim will limit i Read more…

Shutterstock 1622080153

AWS Perfects Cloud Service for Supercomputing Customers

August 29, 2024

Amazon's AWS believes it has finally created a cloud service that will break through with HPC and supercomputing customers. The cloud provider a Read more…

HPC Debrief: James Walker CEO of NANO Nuclear Energy on Powering Datacenters

August 27, 2024

Welcome to The HPC Debrief where we interview industry leaders that are shaping the future of HPC. As the growth of AI continues, finding power for data centers Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Leading Solution Providers

Contributors

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Intel’s Next-gen Falcon Shores Coming Out in Late 2025 

April 30, 2024

It's a long wait for customers hanging on for Intel's next-generation GPU, Falcon Shores, which will be released in late 2025.  "Then we have a rich, a very Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

Department of Justice Begins Antitrust Probe into Nvidia

August 9, 2024

After months of skyrocketing stock prices and unhinged optimism, Nvidia has run into a few snags – a  design flaw in one of its new chips and an antitrust pr Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

MLPerf Training 4.0 – Nvidia Still King; Power and LLM Fine Tuning Added

June 12, 2024

There are really two stories packaged in the most recent MLPerf  Training 4.0 results, released today. The first, of course, is the results. Nvidia (currently Read more…

xAI Colossus: The Elon Project

September 5, 2024

Elon Musk's xAI cluster, named Colossus (possibly after the 1970 movie about a massive computer that does not end well), has been brought online. Musk recently Read more…

Spelunking the HPC and AI GPU Software Stacks

June 21, 2024

As AI continues to reach into every domain of life, the question remains as to what kind of software these tools will run on. The choice in software stacks – Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire