Aggregating Clusters Through Virtualization; Virtual SMP Benefits

By Shai Fultheim

December 16, 2008

Aggregation is a virtualization technique that makes multiple physical systems appear to function as a single logical system. The building blocks for this approach are the same building blocks used in scale-out (clustering): industry-standard servers and high-speed interconnects. Aggregation fundamentally replaces the functionality of custom and proprietary chipsets with software, leveraging commodity interconnects such as InfiniBand. It utilizes only a tiny fraction of the system’s CPUs and RAM to provide chipset-level services without sacrificing system performance.

By running a single logical system, customers reduce the management costs associated with managing multiple cluster nodes, and take advantage of large contiguous memory and unified I/O architecture. To understand how aggregation works, we will first explain the architecture of a traditional SMP system, and then dive into the details of the aggregation approach.

Traditional Multi-Processor Systems

Traditional multi-processor systems run a single operating system (OS). The OS interacts with the system using a well-defined hardware interface, which provides the OS with predefined services to use and control the hardware. These interfaces may include hardware detection and probing, memory ordering semantics, I/O space access and interrupt delivery mechanisms.

Intel’s Multiprocessor Specification allows a single copy of an operating system to run on a single CPU system as well as on a multi-CPU system with up to 255 CPUs. It details a well-defined interface that allows the OS to know exactly how to probe the hardware to determine what kind of system is running underneath it, and then behaves accordingly. This interface also handles the coordination of the underlying system with the OS. For a traditional multi-processor system, such interface is implemented in a silicon chipset.

In addition to the hardware interface, the multi-processor system consists of CPUs, memory and I/O subsystems. These components are all connected together with a proprietary backplane, often implemented by a chipset. Examples of such chipsets and backplanes are Intel’s FSB (Front Side Bus), AMD’s HT (Hyper-Transport), SUN’s CrossBar, SGI’s NUMALINK and IBM’s XA.

The chipset and the proprietary backplane (system interconnect) are the elements where multi-processor systems differ the most from each other and where the major cost of a high-end multi-processor systems is derived. The system interconnect is expensive because the more processors that are added to a system, the more complex it becomes to connect them all together in a manner that ensures both coherency and performance. Traditional multi-processor systems require the creation of a custom chipset to implement the system interconnect to allow processor, memory and I/O communication.

Software Approach: Aggregation

Aggregation reduces the costs associated with custom chipsets and backplanes. It requires multiple high-volume, industry-standard x86 systems (processor speed and amount of memory across boards do not have to be the same). In addition, a high-speed, low-latency interconnect serves as backplane. InfiniBand infrastructure excels in this regard. The aggregation software is loaded on each system board below the operating system layer.

One System

Once loaded into the memory of each of the system boards, the virtualization software aggregates the compute, memory and I/O capabilities of each system and presents a unified virtual system to both the operating system and the applications running above the OS. The aggregation software uses a software-interception engine in the form of a Virtual Machine Monitor (VMM) to provide a uniform execution environment. The software also creates the required BIOS and Advanced Configuration and Power Interface (ACPI) environment to provide the OS (and the software stack above the OS) a coherent image of a single system.

Coherent Memory

The software maintains cache coherency between the individual boards using multiple advanced coherency algorithms. These complex algorithms operate concurrently on a per-block basis, based on real-time memory activity access patterns. Board local-memory is leveraged together with caching algorithms to minimize the effect of interconnect latencies.

Shared I/O

The virtualization software aggregates I/O resources across all boards into a unified PCI hierarchy and presents them as a common pool of I/O resources to the OS and the application. The OS is able to utilize all the system storage and networking controllers toward providing high-I/O system capabilities.

The Advantages of Aggregation

Ease of Use

Aggregation simplifies scale-deployments by having a single system to manage compared to the complexities involved with managing a cluster. A single system removes the need for cluster file systems, cluster interconnect provisioning, application provisioning and installation and updating multiple operating systems and applications. The use of one operating system instead of one per node results in a significant savings in time and money during installation, as well as in ongoing management costs.

Simplified I/O Architecture

I/O requirements for a scale-out model can be very complex and costly, involving networked storage with accompanying costs related to additional HBAs, and FC switch infrastructure. Aggregation technology consolidates each individual server’s network and storage interfaces. I/O resource consolidation reduces the number of drivers, HBAs, NICs, cables, and switch ports and all the associated maintenance overhead. The user needs fewer I/O devices to purchase, manage and service with increased availability, resiliency and runtime scalability of I/O resources.

Large Memory System

For workloads that require a large contiguous memory, customers have traditionally used the scale-up approach. Aggregation provides a cost-effective alternative to buying expensive and large proprietary shared-memory systems for such workloads. It enables an application requiring large amounts of memory to leverage the memory of multiple systems, and reduce the need to use a hard-drive for swap or scratch space. Application runtime can be dramatically reduced by running simulations with in-core solvers or by using memory instead of swap for large-memory footprint models.

Aggregation thus provides a cost-effective virtual x86 platform with a large shared memory that minimizes the physical infrastructure requirements and can run both distributed applications, as well as applications requiring a large memory footprint at optimal performance on the same physical infrastructure.

Compute-Intensive, Large Core-Count Requirements

For workloads that require a high core count coupled with shared memory, customers have traditionally used proprietary shared-memory systems. Aggregation provides a very cost effective x86 alternative to these expensive and proprietary RISC systems.
Aggregation technology combines memory-bandwidth across boards, as opposed to traditional SMP or NUMA architecture where memory bandwidth decreases as the machine scales. This enables solutions based on aggregation technology to show close-to-linear memory bandwidth scaling, thereby delivering excellent performance for threaded applications.

Improved Utilization

Even in large cluster deployments in data centers, it makes sense to deploy aggregation, since fewer larger nodes mean less cluster complexity and better utilization of the infrastructure due to reduced fragmentation of the resources. An example can be found in the financial services industry, where organizations need to run hundreds or thousands of simulations at once. A common deployment model will involve hundreds of servers, where each will execute a few simulations. In this example, each cluster node is running a single application at 80 percent utilization. By using aggregation to create fewer larger nodes, every four aggregated systems can run another copy of the application, leveraging the underutilized resources and driving an additional 25 percent utilization.

ScaleMP’s vSMP Foundation is an example of the aggregation concept. The company has taken an approach that makes a tradeoff between memory latency and memory bandwidth.

Memory Bandwidth vs. Memory Latency

vSMP Foundation uses caching technologies to provide parallel access to system memory. With vSMP Foundation, data migration and replication are employed to maximize system memory bandwidth. The additional system memory bandwidth is used to mask the backplane latencies. ScaleMP uses standard interconnect technology versus the custom built back planes used in traditional multi-processors. While the backplane latency of the InfiniBand interconnect is higher than traditional multi-processor backplanes, the additional memory bandwidth offsets this higher latency. One of the keys to appreciating vSMP Foundation’s ability to mask backplane latency and provide superior performance is the understanding of the fundamentals behind efficiency in memory management. At its elemental level, efficiency can be defined as:

Efficiency = 1 – (Access x Latency)

Where:

Access represents the number of times a processor has to reach out to memory that is not within the processor cache (i.e., on main memory, requiring access via the backplane).

Latency represents the amount of processor wait time such memory requires each time it is accessed.

The efficiency of a system can be improved by reducing the number of times the processor accesses the backplane or reducing the latency of each access to the backplane. Typically, the access is defined by the nature of the application, and latency is based on the technology of the backplane.

Historically, the industry has improved performance by focusing significant R&D on reducing latency in each new generation of products (backplanes, memory-speed, etc.). The assumption was that the access patterns were driven by the applications and, hence, largely out of the control of the system vendors.

ScaleMP’s basic approach was to replace proprietary chipsets and backplanes with InfiniBand interconnects and reduce the number of times a processor has to access the backplane for memory operations on another physical board.

Non-Uniformed Memory Architecture (NUMA) is a common multi-processor architecture which inherently support system scaling by adding additional system nodes. NUMA drawbacks are result of non-homogenous memory access latency, which require operating system and application awareness. ScaleMP utilizes a combination of NUMA and Cache Only Memory Architecture (COMA) in conjunction with a massive cache (typically 5-10 percent of the system’s RAM), to trade off the backplane latency with the use of redundant RAM for caching. The backplane latency is mitigated using software-driven adaptive caching techniques and achieving better systems economics by leveraging commodity memory costs versus proprietary backplanes and chipsets.

ScaleMP utilizes multiple memory coherency algorithms that are selected based on several aspects of the application behaviors, such as historical memory access pattern, on-the-fly code analysis and I/O behavioral analysis. In essence, in spite of having higher backplane latency versus traditional multi-processor systems, vSMP Foundation techniques for memory access reduction are designed to offset the disadvantage of higher-latency, commodity industry-standard interconnects. With the progress that InfiniBand is making in improving latency from SDR to DDR to QDR, the latency gap between an InfiniBand interconnect and a proprietary interconnect is shrinking.

By leveraging industry-standard processors and systems coupled with industry-standard interconnects, aggregation creates a new paradigm for high performance computing and represents a step toward delivering better performance with lower cost and less complexity. Aggregation overcomes the fundamental limitation of clusters for applications that require large shared memory, and it addresses the barriers many technical computing customers have when it comes to adopting clusters, namely a lack of appropriate IT skills to install and manage them. The drawbacks associated with traditional SMP systems, such as high cost and vendor lock-in, are likewise avoided.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

2024 Winter Classic: Texas Two Step

April 18, 2024

Texas Tech University. Their middle name is ‘tech’, so it’s no surprise that they’ve been fielding not one, but two teams in the last three Winter Classic cluster competitions. Their teams, dubbed Matador and Red Read more…

2024 Winter Classic: The Return of Team Fayetteville

April 18, 2024

Hailing from Fayetteville, NC, Fayetteville State University stayed under the radar in their first Winter Classic competition in 2022. Solid students for sure, but not a lot of HPC experience. All good. They didn’t Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use of Rigetti’s Novera 9-qubit QPU. The approach by a quantum Read more…

2024 Winter Classic: Meet Team Morehouse

April 17, 2024

Morehouse College? The university is well-known for their long list of illustrious graduates, the rigor of their academics, and the quality of the instruction. They were one of the first schools to sign up for the Winter Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire