The Last Mile of Virtualization

By Wolfgang Gentzsch

May 3, 2012

A review of eXludus’ new micro-virtualization technology for multicore environments

In a previous article (We Need More than Multicore), I discussed the evolution of multicore processors, and the dramatic effect this processor shift can have on compute cluster performance. Clearly, leveraging a lot of cores will require that many concurrent tasks – as opposed to a single massively parallel task – run safely and predictably within a system. These concurrent tasks will range from serial to multi-threaded to parallel tasks, and all will need to share the same system resources in a productive and reliable manner.

The question becomes how to do this in operating systems (OS) environments that were not designed with multicore architectures in mind. For example, Linux, which has become the pervasive operating OS for servers, is based on time slicing, which is somewhat analogous to suboptimal round-robin server farm dispatch. But it has limitations when running many concurrent tasks that access shared processors and memory. As the number of competing tasks increases, the likelihood of interference between tasks rises exponentially.

The operating system has limited tools that identify escalating resource access problems or proactive defenses to resolve such problems. With these inherent limitations, users often realize reduced system performance and/or reliability. Nor can the OS manage work prioritization between tasks very well, so the established workload management priorities are largely lost when a mix of jobs from different queues are dispatched to a compute node. In summary, the major issues with today’s multicore systems include:

  • For most applications, users are only able to leverage the capabilities of ALL cores (and thus experience high overall utilization) by running many iterations concurrently. The challenge then becomes a balancing act. Even slightly too much work, in terms of physical memory oversubscription (as little as 5 percent), leads to performance degradation and reliability problems. Too little work gets done and valuable resources sit idle.
  • It is manually impossible to continuously balance work against resources, as the use of these resources fluctuates during an application’s execution. And user memory hints, to the extent they are accurate, have to express the high-water usage even though an application may use much less than the high-water mark for a significant percentage of the time.
  • With multi-application/multi-tenant systems, it is difficult or even impossible to meet varying service level agreements (SLAs). Users and applications may not get the level of resources expected, committed, or paid for, and performance levels may vary widely from one iteration to the next.
  • With many concurrently running applications, all work becomes largely equal in the kernel, so high-value tasks can be slowed down by low-value tasks. Under standard Linux it is difficult or even impossible to set varying priority levels for the various executing applications.
  • Full server virtualization is too heavy-handed for running high performance applications. While legacy virtualization may allow you to segment a system in an attempt to improve system utilization, the added utilization rate may be offset in large part by the hypervisor overhead. Legacy server virtualization is useful for multi-OS requirements, but if the organization just needs to run millions of jobs under the same (Linux) OS then overhead, administration, and costs are too high.  

Multicore optimization specialist eXludus Technologies believes it has the answer. The company recently announced the industry’s first micro-virtualization solution. This software creates lightweight micro-containers that encapsulate one or more applications, and are based upon defined application or project policies. These containers have embedded resource allocation intelligence that applies predictive queuing algorithms in order to optimize allocation of micro-resources, such as cores and memory. And it does so in real time.

With negligible system and administrative overhead, the eXludus solution expands the use case for virtualization, making it suitable for performance-sensitive environments (most notably, HPC) that have previously avoided virtualization because of overhead concerns.

The software promises to extract up to 70 percent more throughput from the same resources, while acting as a safety net to avoid resource over-subscription that is detrimental to system performance. Since the micro-containers run within an OS, the eXludus software can be deployed separately from or together with existing server and storage virtualization solutions.

By applying virtualization underneath the OS, a number of benefits are exposed. For example, although it’s easy to load a system, the challenge is achieving maximum utilization while avoiding resource oversubscription, which leads to performance reduction and system instability. Using a lightweight framework, micro-virtualization automates the process of optimizing resources, safely allowing system utilization to be increased. In more detail, micro-virtualization is designed to:

  • Achieve more application processing power per system, aids server consolidation, resulting in fewer systems needed for given workload, and reduces data center power demands (system power, cooling, and space).
  • Maintain kernel level task priorities so that resources are steered to high-value work.
  • Ensure SLA can be met though simple and easy to define policies.

The support of SLAs is particularly important. Multi-tenancy invariably grows with core counts, either via cloud-like unrelated users or within the enterprise, where various departments or projects end up sharing systems to a greater degree. These multi-tenants pay in some manner, either directly or via budget contribution, so they demand to get what they pay for. Therefore, consistent and predictable results are important, i.e., a user can’t complete processing in X time on one iteration and 2X on the next iteration.

Micro-virtualization provides mechanisms that ensure that specific applications, users, and projects receive the CPU and memory resources that have been paid for or committed to. Specifically, administrators can declare CPU and memory percentages that guarantee resource levels.

Within the kernel, multi-tenant/multi-user/multi-project work falls subject to equal OS time-slice behavior. That’s true even though the work, which ranges from high-cost applications to open-source, is not all equal.  Users can easily run into situations where high-value work is slowed by low-value work. Micro-virtualization provides tools that allow for discrete task prioritization that can predictably steer resource allocations.

Within an enterprise there may be many iterations of an application running, but each iteration may have unique value. Consider a chip-design environment where multiple next-gen processors are under development. The soon-to-be-released processor has more enterprise value than a design to be completed in three years. Micro-virtualization easily accommodates resource steering to the high value work, even to the extent that work can be flagged as having exclusive access to system resources — think of a chip tape-out that needs last minute fixes, or rendering where a movie is about to be released and maximum performance is required.

The eXludus solution also promises to open up HPC to virtualization. Legacy server virtualization has not been very successful in these situations as the heavyweight hypervisor approach has high overhead costs. Micro-virtualization is truly lightweight, in the range of 1 to 2 percent overhead, whereas full server virtualization may entail an 18 to 20 penalty in these scenarios. And where full virtualization complexity is administratively cumbersome, micro-virtualization is simple and literally can be deployed within hours.  

A current limitation of eXludus micro-virtualization is that all work must be run within a single OS environment, in this case Linux. And the solution does not yield an easily predictable or consistent performance gain. Throughput increases are a function of the workload, which users may find difficult to comprehend.

In aggregate though, micro-virtualization as a means to effectively segment multicore systems and extract maximum efficiencies appears to be an idea whose time has come. Service efficiency is improved and resources are more accurately steered to the highest value tasks in support of business objectives.

—–

Wolfgang Gentzsch is an independent HPC consultant for cluster, grid, and cloud computing, technologies and a member of the Board of Directors for eXludus Technologies.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Google Cloud Makes Good on Promise to Add Nvidia P100 GPUs

September 21, 2017

Google has taken down the notice on its cloud platform website that says Nvidia Tesla P100s are “coming soon.” That's because the search giant has announced the beta launch of the high-end P100 Nvidia Tesla GPUs on t Read more…

By George Leopold

Cray Wins $48M Supercomputer Contract from KISTI

September 21, 2017

It was a good day for Cray which won a $48 million contract from the Korea Institute of Science and Technology Information (KISTI) for a 128-rack CS500 cluster supercomputer. The new system, equipped with Intel Xeon Scal Read more…

By John Russell

Adolfy Hoisie to Lead Brookhaven’s Computing for National Security Effort

September 21, 2017

Brookhaven National Laboratory announced today that Adolfy Hoisie will chair its newly formed Computing for National Security department, which is part of Brookhaven’s new Computational Science Initiative (CSI). Read more…

By John Russell

HPE Extreme Performance Solutions

HPE Prepares Customers for Success with the HPC Software Portfolio

High performance computing (HPC) software is key to harnessing the full power of HPC environments. Development and management tools enable IT departments to streamline installation and maintenance of their systems as well as create, optimize, and run their HPC applications. Read more…

PNNL’s Center for Advanced Tech Evaluation Seeks Wider HPC Community Ties

September 21, 2017

Two years ago the Department of Energy established the Center for Advanced Technology Evaluation (CENATE) at Pacific Northwest National Laboratory (PNNL). CENATE’s ambitious mission was to be a proving ground for near- Read more…

By John Russell

Stanford University and UberCloud Achieve Breakthrough in Living Heart Simulations

September 21, 2017

Cardiac arrhythmia can be an undesirable and potentially lethal side effect of drugs. During this condition, the electrical activity of the heart turns chaotic, Read more…

By Wolfgang Gentzsch, UberCloud, and Francisco Sahli, Stanford University

PNNL’s Center for Advanced Tech Evaluation Seeks Wider HPC Community Ties

September 21, 2017

Two years ago the Department of Energy established the Center for Advanced Technology Evaluation (CENATE) at Pacific Northwest National Laboratory (PNNL). CENAT Read more…

By John Russell

Exascale Computing Project Names Doug Kothe as Director

September 20, 2017

The Department of Energy’s Exascale Computing Project (ECP) has named Doug Kothe as its new director effective October 1. He replaces Paul Messina, who is stepping down after two years to return to Argonne National Laboratory. Kothe is a 32-year veteran of DOE’s National Laboratory System. Read more…

Takeaways from the Milwaukee HPC User Forum

September 19, 2017

Milwaukee’s elegant Pfister Hotel hosted approximately 100 attendees for the 66th HPC User Forum (September 5-7, 2017). In the original home city of Pabst Blu Read more…

By Merle Giles

Kathy Yelick Charts the Promise and Progress of Exascale Science

September 15, 2017

On Friday, Sept. 8, Kathy Yelick of Lawrence Berkeley National Laboratory and the University of California, Berkeley, delivered the keynote address on “Breakthrough Science at the Exascale” at the ACM Europe Conference in Barcelona. In conjunction with her presentation, Yelick agreed to a short Q&A discussion with HPCwire. Read more…

By Tiffany Trader

DARPA Pledges Another $300 Million for Post-Moore’s Readiness

September 14, 2017

The Defense Advanced Research Projects Agency (DARPA) launched a giant funding effort to ensure the United States can sustain the pace of electronic innovation vital to both a flourishing economy and a secure military. Under the banner of the Electronics Resurgence Initiative (ERI), some $500-$800 million will be invested in post-Moore’s Law technologies. Read more…

By Tiffany Trader

IBM Breaks Ground for Complex Quantum Chemistry

September 14, 2017

IBM has reported the use of a novel algorithm to simulate BeH2 (beryllium-hydride) on a quantum computer. This is the largest molecule so far simulated on a quantum computer. The technique, which used six qubits of a seven-qubit system, is an important step forward and may suggest an approach to simulating ever larger molecules. Read more…

By John Russell

Cubes, Culture, and a New Challenge: Trish Damkroger Talks about Life at Intel—and Why HPC Matters More Than Ever

September 13, 2017

Trish Damkroger wasn’t looking to change jobs when she attended SC15 in Austin, Texas. Capping a 15-year career within Department of Energy (DOE) laboratories, she was acting Associate Director for Computation at Lawrence Livermore National Laboratory (LLNL). Her mission was to equip the lab’s scientists and research partners with resources that would advance their cutting-edge work... Read more…

By Jan Rowell

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Leading Solution Providers

IBM Clears Path to 5nm with Silicon Nanosheets

June 5, 2017

Two years since announcing the industry’s first 7nm node test chip, IBM and its research alliance partners GlobalFoundries and Samsung have developed a proces Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

GlobalFoundries: 7nm Chips Coming in 2018, EUV in 2019

June 13, 2017

GlobalFoundries has formally announced that its 7nm technology is ready for customer engagement with product tape outs expected for the first half of 2018. The Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This