Kubernetes and HPC Applications in Hybrid Cloud Environments – Part II

By Daniel Gruber,Burak Yenier and Wolfgang Gentzsch, UberCloud

March 19, 2020

With the rise of cloud services, CIOs are recognizing that applications, middleware, and infrastructure running in various compute environments need a common management and operating model. Maintaining different application and middleware stacks on-premises and in cloud environments, by possibly using different specialized infrastructure and application management solutions for each cloud provider, adds lots of friction in dynamically allocating, using, and managing those resources.

Lack of a common management and operating model in hybrid cloud environments can cause:

  • Inhomogeneous, fragmented environments create additional complexity for managers, operators, and security.
  • Speed of innovation slows down due to hybrid environments without common management.
  • Cloud resources are hard to change or shutdown when dependent on a cloud provider’s specific services.
  • Workloads can’t be easily migrated back to on-premises environments when bound to specific cloud environment setups, and vice versa.

Kubernetes has become the de-facto standard container orchestrator as pointed out in a previous article. All major companies provide and build solutions on top of a standardized API which is available everywhere. CIOs are now looking into the applicability of Kubernetes for HPC in hybrid-cloud as it offers a common management and operating model for every environment.

Kubernetes: A Common Management and Operating Model for Hybrid Cloud

Kubernetes facilitates the use and administration of countless containers running on fleets of servers. It is the new standard platform for hybrid environments supported by many IT vendors and cloud providers. CIOs can now allocate a fully configured and supported container orchestrator as base for all of their application workloads.

Kubernetes, unlike proprietary infrastructure solutions, provides portability, ease of administration, high availability, integrability, and monitoring capabilities. When managing resources on Kubernetes CIOs are no longer bound to a specific infrastructure. They can offer their users the same set of functionalities, be it on-premises or in any cloud, using the same application stack. Users are not even aware that their applications are running on Kubernetes, nor on which infrastructure they are running: in their own data centers or at a specific cloud provider, like Google, Microsoft, or Amazon.

Reducing complexity in hybrid cloud environments by using a standardized software stack like Kubernetes comes with many advantages: improvements made for one platform can be made automatically available on other platforms; deployment and operational aspects can be simplified; and security audits are easier and rigorously to execute.

Kubernetes and HPC

Kubernetes is the de facto platform for AI and ML already. However, when it comes to traditional HPC, some challenges remain. There is still a set of features built into HPC workload managers not yet available in Kubernetes. We discussed the major differences already previously in our HPCwire Part I article. Major gaps of Kubernetes for HPC currently are: native support for distributed memory jobs, namely MPI applications, and a missing job queueing system compatible with existing HPC applications.

Kubernetes has built-in high availability on many layers. However, for HPC jobs, it is not enough to restart a single container that failed because the whole distributed job itself might have failed already. In this case, automatic rescheduling of the entire distributed memory job is required. This is something Kubernetes doesn’t handle.

Beside these challenges, Kubernetes comes with many benefits for HPC: for example, the environment for the engineer and for the containerized HPC application is always the same, be it on-premises or running in a cloud-based environment; and the capability to quickly change from one infrastructure to another allows the HPC team to align with their company’s cloud roadmap. The freedom to move workloads between infrastructures based on a common API – the Kubernetes API – is what becomes valuable.

Containerized HPC Applications on Kubernetes

Over the past five years, dozens of HPC applications have been containerized, be it commercial, like ANSYS, COMSOL, STAR-CCM+, or open source packages like OpenFOAM and GROMACS, along with HPC cluster schedulers like Univa Grid Engine and Slurm. Thanks to container technology, a constant stream of updates and improvements is provided which can be promptly and seamlessly updated by customers. Additionally, the container images allow users to go back at any time to a previous application version so that they always can reproduce their previous results.

Example HPC Application Cluster Architecture running on Managed Kubernetes.

In the meantime, many container environments have been implemented by using infrastructure and configuration management tools like Terraform and Puppet or by building cloud specific HPC integrations into existing portals. But with the advent of Kubernetes, container environments became easier to maintain and are much more dynamic. Rolling out a cluster, rescaling the worker nodes, using a constant set of preemptible instances, and high availability are driven by controllers which continuously drive the cluster to the desired state. Thus, major HPC gaps of Kubernetes have been closed. This way, today, distributed memory/MPI jobs can be supported in any Kubernetes environment, which provides a built-in HPC workload manager integration running inside HPC containers. That allows traditional HPC applications to run without any changes. Also, GPU and non-GPU enabled applications based on Ansys and COMSOL have been launched successfully, through a high-performance, GPU enabled desktop running inside a pod. Once logged in to the desktop the engineer can start submitting batch jobs or single MPI applications which are distributed across a set of pods allocated on multiple nodes.

Conclusions

Kubernetes not only supports microservice based enterprise applications, but also self-service engineering HPC applications. In summary, as this research has shown, the key advantages of using Kubernetes as a foundation for running containerized engineering applications are:

  • Unified application stack available on virtually any infrastructure
  • True hybrid cloud usage scenarios for engineering workload. For the engineers it is transparent where the application runs, be it on-premises or in the cloud
  • which leads to providing the best performance for running engineering applications by allocating always the newest and fastest machines available in the cloud
  • Building and resizing a self-contained HPC application and compute cluster as self-service for the engineer which is only limited by cloud quotas and budget per time period
  • Robust management stack, supported by many Cloud providers
  • Optimizing costs by only paying for what is used. No idle resources which need to be allocated before they are going to be used.
  • High security through self-contained dedicated compute clusters
  • Minimal operational overhead by self-provisioning and disposable components for which updates are simple destroy and re-create commands
  • Kubernetes based workload is easier to integrate in widely adopted continuous integration and deployment solutions (like Tekton, Concourse, or future versions of Jenkins)

In this research, container-based HPC application environments have been implemented on top of Kubernetes (e.g. on Google GCP and Amazon AWS) and also used as self-service test environments which can be deployed from scratch by HPC application specialists, not operators. It has also been used in CI/CD pipelines to automatically build test environments which run tests against existing container solutions and shut down the infrastructure afterwards. In customer environments, the IT group benefits from an easier to maintain system using a supported, managed Kubernetes which can ramp up, resized and deleted computing resources within minutes.

About the Authors

Daniel Gruber, Burak Yenier, and Wolfgang Gentzsch are with UberCloud, a company that started in 2013 with developing HPC container technology and containerized engineering applications, to facilitate access and use of engineering HPC workload in a shared on-premise or on-demand cloud environment. In this article and the part-one article published on HPCwire last September, they describe their experiences during the last 12 months using UberCloud HPC containers on Kubernetes.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Faster Optical Switch that Operates at ‘Room Temp’ Developed by IBM, Skolkovo Researchers

October 19, 2021

Optical switching technology holds great promise for many applications but hot operating temperatures have been a key obstacle slowing progress. Now, a new optical switching device that can operate at room temperatures a Read more…

Energy Exascale Earth System Model Version 2 Promises Twice the Speed

October 18, 2021

The Energy Exascale Earth System Model (E3SM) is an ongoing Department of Energy (DOE) earth system modeling, simulation and prediction project aiming to “assert and maintain an international scientific leadership posi Read more…

Intel Reorgs HPC Group, Creates Two ‘Super Compute’ Groups

October 15, 2021

Following on changes made in June that moved Intel’s HPC unit out of the Data Platform Group and into the newly created Accelerated Computing Systems and Graphics (AXG) business unit, led by Raja Koduri, Intel is making further updates to the HPC group and announcing... Read more…

Royalty-free stock illustration ID: 1938746143

MosaicML, Led by Naveen Rao, Comes Out of Stealth Aiming to Ease Model Training

October 15, 2021

With more and more enterprises turning to AI for a myriad of tasks, companies quickly find out that training AI models is expensive, difficult and time-consuming. Finding a new approach to deal with those cascading challenges is the aim of a new startup, MosaicML, that just came out of stealth... Read more…

NSF Awards $11M to SDSC, MIT and Univ. of Oregon to Secure the Internet

October 14, 2021

From a security standpoint, the internet is a problem. The infrastructure developed decades ago has cracked, leaked and been patched up innumerable times, leaving vulnerabilities that are difficult to address due to cost Read more…

AWS Solution Channel

Cost optimizing Ansys LS-Dyna on AWS

Organizations migrate their high performance computing (HPC) workloads from on-premises infrastructure to Amazon Web Services (AWS) for advantages such as high availability, elastic capacity, latest processors, storage, and networking technologies; Read more…

SC21 Announces Science and Beyond Plenary: the Intersection of Ethics and HPC

October 13, 2021

The Intersection of Ethics and HPC will be the guiding topic of SC21's Science & Beyond plenary, inspired by the event tagline of the same name. The evening event will be moderated by Daniel Reed with panelists Crist Read more…

Intel Reorgs HPC Group, Creates Two ‘Super Compute’ Groups

October 15, 2021

Following on changes made in June that moved Intel’s HPC unit out of the Data Platform Group and into the newly created Accelerated Computing Systems and Graphics (AXG) business unit, led by Raja Koduri, Intel is making further updates to the HPC group and announcing... Read more…

Royalty-free stock illustration ID: 1938746143

MosaicML, Led by Naveen Rao, Comes Out of Stealth Aiming to Ease Model Training

October 15, 2021

With more and more enterprises turning to AI for a myriad of tasks, companies quickly find out that training AI models is expensive, difficult and time-consuming. Finding a new approach to deal with those cascading challenges is the aim of a new startup, MosaicML, that just came out of stealth... Read more…

Quantum Workforce – NSTC Report Highlights Need for International Talent

October 13, 2021

Attracting and training the needed quantum workforce to fuel the ongoing quantum information sciences (QIS) revolution is a hot topic these days. Last week, the U.S. National Science and Technology Council issued a report – The Role of International Talent in Quantum Information Science... Read more…

Eni Returns to HPE for ‘HPC4’ Refresh via GreenLake

October 13, 2021

Italian energy company Eni is upgrading its HPC4 system with new gear from HPE that will be installed in Eni’s Green Data Center in Ferrera Erbognone (a provi Read more…

The Blueprint for the National Strategic Computing Reserve

October 12, 2021

Over the last year, the HPC community has been buzzing with the possibility of a National Strategic Computing Reserve (NSCR). An in-utero brainchild of the COVID-19 High-Performance Computing Consortium, an NSCR would serve as a Merchant Marine for urgent computing... Read more…

UCLA Researchers Report Largest Chiplet Design and Early Prototyping

October 12, 2021

What’s the best path forward for large-scale chip/system integration? Good question. Cerebras has set a high bar with its wafer scale engine 2 (WSE-2); it has 2.6 trillion transistors, including 850,000 cores, and was fabricated using TSMC’s 7nm process on a roughly 8” x 8” silicon footprint. Read more…

What’s Next for EuroHPC: an Interview with EuroHPC Exec. Dir. Anders Dam Jensen

October 7, 2021

One year after taking the post as executive director of the EuroHPC JU, Anders Dam Jensen reviews the project's accomplishments and details what's ahead as EuroHPC's operating period has now been extended out to the year 2027. Read more…

University of Bath Unveils Janus, an Azure-Based Cloud HPC Environment

October 6, 2021

The University of Bath is upgrading its HPC infrastructure, which it says “supports a growing and wide range of research activities across the University.” Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer... Read more…

Enter Dojo: Tesla Reveals Design for Modular Supercomputer & D1 Chip

August 20, 2021

Two months ago, Tesla revealed a massive GPU cluster that it said was “roughly the number five supercomputer in the world,” and which was just a precursor to Tesla’s real supercomputing moonshot: the long-rumored, little-detailed Dojo system. Read more…

Esperanto, Silicon in Hand, Champions the Efficiency of Its 1,092-Core RISC-V Chip

August 27, 2021

Esperanto Technologies made waves last December when it announced ET-SoC-1, a new RISC-V-based chip aimed at machine learning that packed nearly 1,100 cores onto a package small enough to fit six times over on a single PCIe card. Now, Esperanto is back, silicon in-hand and taking aim... Read more…

US Closes in on Exascale: Frontier Installation Is Underway

September 29, 2021

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, held by Zoom this week (Sept. 29-30), it was revealed that the Frontier supercomputer is currently being installed at Oak Ridge National Laboratory in Oak Ridge, Tenn. The staff at the Oak Ridge Leadership... Read more…

Intel Completes LLVM Adoption; Will End Updates to Classic C/C++ Compilers in Future

August 10, 2021

Intel reported in a blog this week that its adoption of the open source LLVM architecture for Intel’s C/C++ compiler is complete. The transition is part of In Read more…

Intel Reorgs HPC Group, Creates Two ‘Super Compute’ Groups

October 15, 2021

Following on changes made in June that moved Intel’s HPC unit out of the Data Platform Group and into the newly created Accelerated Computing Systems and Graphics (AXG) business unit, led by Raja Koduri, Intel is making further updates to the HPC group and announcing... Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months... Read more…

Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was Read more…

Leading Solution Providers

Contributors

AMD-Xilinx Deal Gains UK, EU Approvals — China’s Decision Still Pending

July 1, 2021

AMD’s planned acquisition of FPGA maker Xilinx is now in the hands of Chinese regulators after needed antitrust approvals for the $35 billion deal were receiv Read more…

HPE Wins $2B GreenLake HPC-as-a-Service Deal with NSA

September 1, 2021

In the heated, oft-contentious, government IT space, HPE has won a massive $2 billion contract to provide HPC and AI services to the United States’ National Security Agency (NSA). Following on the heels of the now-canceled $10 billion JEDI contract (reissued as JWCC) and a $10 billion... Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

The Latest MLPerf Inference Results: Nvidia GPUs Hold Sway but Here Come CPUs and Intel

September 22, 2021

The latest round of MLPerf inference benchmark (v 1.1) results was released today and Nvidia again dominated, sweeping the top spots in the closed (apples-to-ap Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Intel Unveils New Node Names; Sapphire Rapids Is Now an ‘Intel 7’ CPU

July 27, 2021

What's a preeminent chip company to do when its process node technology lags the competition by (roughly) one generation, but outmoded naming conventions make i Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire