What the New HPC Means to Market Intelligence

By Addison Snell

March 23, 2007

Meet the new HPC, where Productivity is our middle name. High Productivity Computing is not a whim; it is a new set of market dynamics that is more relevant to today's global economy. For a growing number of users and vendors, HPC refers not to cores, cycles, or flops but to discovery, efficiency, or time to market.

DARPA notably used the new P for the High Productivity Computing Systems (HPCS) program, stating that “value will be determined by assessing many additional factors beyond just theoretical peak flops.” There is also a proliferation of vendors in the market now selling pieces of the HPC solution. Tabor Communications is not pushing the industry in a new direction; we are updating the definitions to head in the same direction the industry is already going.

It's not as if none of us cared about productivity before. Of course we did. Productivity has always been the main intended benefit of HPC, and we never suffered from abbreviation-induced heartburn. So why does P no longer stand for performance?

Performance (defined roughly as what the server could accomplish with a Utopian application load) and productivity (defined roughly as the advantage gained by the user) were at one time tightly correlated, such that an increase in raw performance would boost the speed of generating results. Thanks to a variety of immutable technical and market forces — clustering, data explosion, and multi-core come to mind — the correlation between raw performance and productivity has broken down.

Over time the industry has moved from uni-processor to SMP to distributed-memory clusters. Each step toward higher absolute performance has been taken at the cost of additional architectural intricacy. The newest challenge is the transition to dual-core, quad-core, and future higher-level multi-core chips. The introduction of a new level of complexity at the socket level drives another wedge in the gap between theoretical and realized performance. Most applications will not take full advantage of additional cores without changes to the codes themselves or to the productivity tools surrounding them.

We have reached a point where there are too may bottlenecks. On systems with the same processor counts and types, productivity can vary widely depending on the workload manager, file system, interconnect, or scheduler, just to name a few.

The challenge is not unique to HPC. It applies equally well to business computing and to PCs. This is both a comfort and a curse. It increases the odds that someone will develop compelling solutions for multi-core parallelism, but it decreases the odds that these solutions will be designed and built with HPC in mind.

All is not lost. There are technologies in development — schedulers, workload managers, accelerators, tools — that have promise in easing the multi-core productivity-performance gap. We need to give these technologies a fair chance in the market. We need constructs for analyzing the efficacy of these tools. As an industry, we need to measure and reward productivity, not performance.

Tabor Research has an inclusive view that defines High Productivity Computing as follows:

High Productivity Computing (HPC) is the use of servers, clusters, and supercomputers — plus associated operational components such as software tools, networks, storage subsystems, and services — for scientific, engineering, or analytical tasks that are particularly intensive in computation, memory usage, or data management. HPC is used by scientists and engineers both in research and in production across industry, government, and academia. Within industry, HPC can frequently be distinguished from general business computing in that companies generally will use HPC applications to gain advantage in their core endeavors — e.g., finding oil, designing automobile parts, or protecting clients' investments — as opposed to non-core endeavors such as payroll management or resource planning.

This updated view of the market requires a change in market research, because it adds users, vendors, and applications to the industry. Forecasting and analysis are the cornerstones of the business, and now have more things to count. Tabor Research is also incorporating demand-side research to gain a view toward how users spend their money — including hardware, software, facilities, and staff — to achieve real productivity with today's systems.

From a vendor perspective, this definition requires us to push research beyond big server vendors' main product lines to include their partners and solutions for all forms of workflow optimization. Second-tier vendors play a major role here and become an important part of the market census.

From a user perspective, there are categories of HPC applications that the old definition failed to cover. HPC technologies have become more accessible and are now being adopted more rapidly in commercial markets. Online gaming, for example, has been using HPC system configurations to meet the real-time I/O requirements of hosting massive multi-user domains. Non-traditional application areas across the spectrum of system sizes and configurations are becoming the face of the new HPC. (For more on our broadened application scope, register for free to read “Five Important Predictions for HPC in 2007” at www.taborresearch.com.)

This approach to gathering market intelligence requires more work, both from the analysts and from the community, but it is ultimately more rewarding for all of us. For the cost of answering a survey now and then, you not only gain a better understanding of where you stand in the community, but you also make your opinions heard and thus steer the course of HPC development.

Tabor Research isn't inventing the new HPC. It already exists. Our goal is to count it and forecast it. And if the industry gains a better understanding of the ways in which Productivity, not merely Performance, is at the heart of HPC, we'll all be better off.

—–

Addison Snell is the VP/GM of Tabor Research, providing actionable market intelligence for High Productivity Computing. You can register for free to download exclusive premium content and record your opinions on the market at the new Tabor Research website, debuting today at www.taborresearch.com.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire