Top Three Pitfalls to Avoid When Processing Data with LLMs

By Erik Landerholm

June 26, 2024

It’s a truism of data analytics: when it comes to data, more is generally better. But the explosion of AI-powered large language models (LLMs) like ChatGPT and Google Gemini (formerly Bard) challenges this conventional wisdom.

As organizations in every industry rush to enrich their own private data sets with LLMs, the quest for more and better data is unfolding at a scale never seen before, stretching the limits of present-day infrastructure in new and disruptive ways. Yet the sheer scale of the data sets ingested by LLMs raises an important question: Is more data really better if you don’t have the infrastructure to handle it?

Training LLMs on internal data poses many challenges for data and development teams. This entails the need for considerable compute budgets, access to powerful GPUs (graphics processing units), complex distributed compute techniques, and teams with deep machine learning (ML) expertise.

Outside of a few hyperscalers and tech giants, most organizations today simply don’t have that infrastructure readily available. That means they are forced to build it themselves, at great cost and effort. If the required GPUs are available at all, cobbling them together with other tools to create a data stack is prohibitively expensive. And it’s not how data scientists want to spend their time.

Three Pitfalls to Avoid

In the quest to pull together or bolster their infrastructure so that it can meet these new demands, what’s an organization to do? When setting out to train and tune LLMs against their data, what guideposts can they look for to make sure their efforts are on track and that they’re not jeopardizing the success of their projects? The best way to identify potential risks is to ask the following three questions:

1. Focusing too much on building the stack vs. analyzing the data

Time spent assembling a data stack is time taken away from the stack’s reason for being: analyzing your data. If you find yourself doing too much of it, look for a platform that automates the foundational elements of building your stack so your data scientists can focus on analyzing and extracting value from the data. You want to be able to pick the components, then have the stack generated for you so you can get to insights quickly.

2. Finding GPUs needed to process the data

Remember when all the talk was about managing cloud costs through multi-cloud solutions, cloud portability, and so on? Today, there’s an analogous conversation on the issue of GPU availability and right-sizing. What is the right GPU for your LLM, who provides it and at what hourly cost to analyze your data, and where do you want to run your stack? Making the right decisions requires balancing multiple factors, such as your computational needs, budget constraints, and future requirements. Look for a platform that is architected in a way that gives you the choice and flexibility to use the GPUs that fit your project and to run your stack wherever you choose, be it on different cloud providers or on your own hardware.

3. Running AI workloads against your data cost-effectively

Finally, given the high costs involved, no one wants to pay for idle resources. Look for a platform that offers ephemeral environments, which allow you to spin up and spin down your instances so you only pay when you’re using the system, not when it’s idle and waiting.

Déjà-vu All Over Again?

In many ways, data scientists seeking to extract insights from their data using LLMs face a similar dilemma to the one software developers faced in the early days of DevOps. Developers who just wanted to build great software had to take on the running of operations and their own infrastructure. That “shift left” eventually led to bottlenecks and other inefficiencies for dev teams, which ultimately hindered many organizations from reaping the benefits of DevOps.

(PopTika/Shutterstock)

This issue was somewhat solved by DevOps teams (and now increasingly platform engineering teams) tasked with building platforms that developers could code on top of. The idea was to recast developers as DevOps’ or PE teams’ customers, and in doing so free them up to write great code without having to worry about infrastructure.

The lesson for organizations caught up in the rush to gain new insights from their data by incorporating the latest LLMs is this: Don’t saddle your data scientists with infrastructure worries.

Let Data Scientists Be Data Scientists

In the brave new world opened up by LLMs and the next-gen GPUs that can handle data-intensive AI workloads, let your data scientists be data scientists. Let them use these astounding innovations to test hypotheses and gain insights that can help you train and optimize your data models and drive value that can help differentiate your organization in the market and lead to the creation of new products.

To navigate this golden age of opportunity effectively, choose a platform that helps you focus on your differentiators while automating the foundational elements of building your AI stack. Look for a solution that gives you choice and flexibility in GPU usage and where you run your stack. Lastly, find an option that offers ephemeral environments that allow you to optimize costs by paying only for the resources you use. Embracing these key principles will empower you to solve the infrastructure dilemma posed by today’s Gen AI gold rush—and position your organization for success.


 

About the author:  Erik Landerholm is a seasoned software engineering leader with over 20 years of experience in the tech industry. As the co-founder of Release.com and a Y Combinator alum from the summer of 2009, Erik has a rich history of entrepreneurial success. His previous roles include co-founder of CarWoo! and IMSafer, as well as Senior Vice President and Chief Architect at TrueCar.

Related Items:

Why A Bad LLM Is Worse Than No LLM At All

LLMs Are the Dinosaur-Killing Meteor for Old BI, ThoughtSpot CEO Says

GenAI Doesn’t Need Bigger LLMs. It Needs Better Data

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Quantum Software Specialist Q-CTRL Inks Deals with IBM, Rigetti, Oxford, and Diraq

September 10, 2024

Q-CTRL, the Australia-based start-up focusing on quantum infrastructure software, today announced that its performance-management software, Fire Opal, will be natively integrated into four of the world's most advanced qu Read more…

Computing-Driven Medicine: Sleeping Better with HPC

September 10, 2024

As a senior undergraduate student at Fisk University in Nashville, Tenn., Ifrah Khurram's calculus professor, Dr. Sanjukta Hota, encouraged her to apply for the Sustainable Research Pathways Program (SRP). SRP was create Read more…

LLNL Engineers Harness Machine Learning to Unlock New Possibilities in Lattice Structures

September 9, 2024

Lattice structures, characterized by their complex patterns and hierarchical designs, offer immense potential across various industries, including automotive, aerospace, and biomedical engineering. With their outstand Read more…

NSF-Funded Data Fabric Takes Flight

September 5, 2024

The data fabric has emerged as an enterprise data management pattern for companies that struggle to provide large teams of users with access to well-managed, integrated, and secured data. Now scientists working at univer Read more…

xAI Colossus: The Elon Project

September 5, 2024

Elon Musk's xAI cluster, named Colossus (possibly after the 1970 movie about a massive computer that does not end well), has been brought online. Musk recently posted the following on X/Twitter: "This weekend, the @xA Read more…

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with performance benchmarks. In the first paper, Understanding Data Mov Read more…

Quantum Software Specialist Q-CTRL Inks Deals with IBM, Rigetti, Oxford, and Diraq

September 10, 2024

Q-CTRL, the Australia-based start-up focusing on quantum infrastructure software, today announced that its performance-management software, Fire Opal, will be n Read more…

NSF-Funded Data Fabric Takes Flight

September 5, 2024

The data fabric has emerged as an enterprise data management pattern for companies that struggle to provide large teams of users with access to well-managed, in Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Shutterstock 1897494979

What’s New with Chapel? Nine Questions for the Development Team

September 4, 2024

HPC news headlines often highlight the latest hardware speeds and feeds. While advances on the hardware front are important, improving the ability to write soft Read more…

Critics Slam Government on Compute Speeds in Regulations

September 3, 2024

Critics are accusing the U.S. and state governments of overreaching by including limits on compute speeds in regulations and laws, which they claim will limit i Read more…

Shutterstock 1622080153

AWS Perfects Cloud Service for Supercomputing Customers

August 29, 2024

Amazon's AWS believes it has finally created a cloud service that will break through with HPC and supercomputing customers. The cloud provider a Read more…

HPC Debrief: James Walker CEO of NANO Nuclear Energy on Powering Datacenters

August 27, 2024

Welcome to The HPC Debrief where we interview industry leaders that are shaping the future of HPC. As the growth of AI continues, finding power for data centers Read more…

CEO Q&A: Acceleration is Quantinuum’s New Mantra for Success

August 27, 2024

At the Quantum World Congress (QWC) in mid-September, trapped ion quantum computing pioneer Quantinuum will unveil more about its expanding roadmap. Its current Read more…

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Leading Solution Providers

Contributors

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Intel’s Next-gen Falcon Shores Coming Out in Late 2025 

April 30, 2024

It's a long wait for customers hanging on for Intel's next-generation GPU, Falcon Shores, which will be released in late 2025.  "Then we have a rich, a very Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

Department of Justice Begins Antitrust Probe into Nvidia

August 9, 2024

After months of skyrocketing stock prices and unhinged optimism, Nvidia has run into a few snags – a  design flaw in one of its new chips and an antitrust pr Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

MLPerf Training 4.0 – Nvidia Still King; Power and LLM Fine Tuning Added

June 12, 2024

There are really two stories packaged in the most recent MLPerf  Training 4.0 results, released today. The first, of course, is the results. Nvidia (currently Read more…

Spelunking the HPC and AI GPU Software Stacks

June 21, 2024

As AI continues to reach into every domain of life, the question remains as to what kind of software these tools will run on. The choice in software stacks – Read more…

Quantum Watchers – Terrific Interview with Caltech’s John Preskill by CERN

July 17, 2024

In case you missed it, there's a fascinating interview with John Preskill, the prominent Caltech physicist and pioneering quantum computing researcher that was Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire