The UberCloud Experiment Continues…

By Wolfgang Gentzsch and Burak Yenier

June 16, 2014

Improved product quality, faster time to market and ultimately increased ROI have long been associated with the adoption of HPC tools. The benefits that engineers and scientists can expect from using technical computing in their research, design and development processes can be huge. In spite of this, there is still unmet need with relatively few scientists and manufacturers using servers or clusters when designing and developing their products on computers. The majority of virtual prototyping and large-scale data modeling is still performed on desktop or laptop computers. It’s not surprising then that many of these users face problems stemming from the lack of performance of their machines. Obtaining additional computing power means either purchasing a server outright or using a cloud-based offering.

Many system vendors have developed a complete set of products, solutions and services for “high performance computing” (HPC) and buying an HPC server for an SME is no longer out of reach. Owning an HPC server, however, is not necessarily the best idea in terms of cost-efficiency or maintenance, because the Total Cost of Ownership (TCO) of such a machine is pretty high, and maintaining such a system requires additional manpower and expertise.

The other option today is to use an HPC Cloud solution that allows engineers and scientists to keep using their workstation for daily design and development work, and to “burst” larger, more complex jobs into the Cloud when needed. Thus, users have access to quasi infinite computing resources that offer higher-quality results. A Cloud solution helps reduce capital expenditure, offers greater business agility by dynamically scaling resources up and down as needed, and is only paid for when used.

The UberCloud Experiment Accelerating HPC Cloud Adoption

Since July 2012 the UberCloud Experiment has attracted more than 1500 organizations from 72 countries. We were able to build 152 teams in CFD, FEM, Biology, Finance, and other domains, and to publish many case studies reporting about different applications, experience, and lessons learned. UberCloud TechTalk has been founded providing educational lectures for our community. And the UberCloud Exhibit offers a Cloud services catalogue where community members can exhibit their Cloud related services or select the services which they want to use for their team experiment or for their daily work. Sponsored by Intel and HPCwire in 2013, the first Compendium with 25 HPC Cloud case studies received the HPCwire Readers Choice Award for the best HPC Cloud implementation. Now, in June 2014, the second Compendium of UberCloud case studies has been published, sponsored by Intel and HPCwire, which can be downloaded for free from here.

The UberCloud Experiment provides a platform for scientists and engineers to explore, learn and understand the end-to-end process of accessing and using HPC Cloud resources, and to identify and resolve the roadblocks. End-users, software providers, resource providers, and computing experts collaborate in teams to jointly solve the end-users’ application in the Cloud. Let’s start by defining what roles each stakeholder plays to make service-based high performance technical computing in the cloud come together:

– A typical example is a small or medium size manufacturer in the process of designing, prototyping and developing its next-generation product. These users are prime candidates for HPC-as-a-Service when in-house computation on workstations has become too lengthy and acquiring additional computing power in the form of a server is too cumbersome.

Application software provider
– These are software owners of all stripes, including ISVs, public domain software organizations and individual developers. The UberCloud Experiment prefers rock-solid software, which has the potential to be used on a wider scale.

Resource provider
– This pertains to anyone who owns HPC resources networked to the outside world. A classic HPC center would fall into this category as well as a standard datacenter used to handle batch jobs, or a cluster-owning commercial entity that is willing to offer up cycles to run non-competitive workloads during periods of low CPU-utilization.

Computing experts
– This group includes individuals and companies with HPC expertise, e.g. in areas like cluster management and software porting and optimizing. It also encompasses PhD-level domain specialists with in-depth application knowledge. These experts work as team leaders, with end-users, computer centers and software providers to help glue the pieces together.

For example, suppose the end-user is in need of additional compute resources to increase the quality of a product design or to speed up a product design cycle – say for simulating more sophisticated geometries or physics or for running many more simulations for a higher quality result. That suggests a specific software stack, domain expertise and even hardware configuration. The general idea is to look at the end-user’s tasks and software and select the appropriate resources and expertise that match certain requirements.

Cloud Case Studies from the UberCloud Experiment

As a glimpse into the wealth of practical use cases, we chose four HPC Cloud projects out of the 152 UberCloud Experiments: in life sciences teams 61 and 89, and in CAE teams 118 and 142.

Team 61: Molecular dynamics of the mutant PI3Kα protein

The end-user of this team was Zoe Cournia from the Biomedical Research Foundation of the Academy of Athens. Resource provider was the GRNET-Okeanos IaaS Cloud service, represented by Vangelis Floros and Stefanos Gerangelos from the Greek Research and Technology Network S.A. And HPC expert was Dominique Dehareng from the Center for Protein Engineering at the University of Liège. Open source software Gromacs 4.6.1 was used for the simulations.

Cancer is a leading cause of death worldwide, accounting for 7.6 million deaths in 2008 according to the World Health Organization. This number is expected to increase to 21 million by 2030. One of the signaling pathways which, when deregulated, becomes centrally involved in several types of cancers, like colon, mammary, and endometrial tumorigenesis, is served by phosphoinositide-3-kinase alpha (PI3Kα). The importance of PI3Kα in cancer is highlighted by the discovery that PIK3CA, the gene encoding the catalytic p110α subunit of PI3Kα, is frequently mutated in human malignancies. The goal of this project was to gain insights into the oncogenic mechanism of two commonly expressed PI3Kα mutants by studying their conformational changes with Molecular Dynamics (MD) simulations, in comparison with the PI3Kα wild-type (normal, non-cancerous) protein. The utilization of cloud computing in performing MD simulations of mutant PI3Kα with Gromacs was examined in this case study.


Figure 1: The protein PI3Ka is depicted in ribbons and is placed in a water box, shown as red dots.

The cloud computing service that was provided was a free and very efficient service. Cloud computing proved to be an extremely easy process as compared to building and maintaining your own cluster. The team was provided with a VM with 8 cores and the possibility of building a cluster of up to 64 cores connected via Ethernet network. In the 8 cores, the MD simulation of the mutant PI3Kα protein scaled linearly and did run faster when compared to in-house simulations. Team communications through BaseCamp were very efficient. Collaborative work over the Internet, using on-line resources like cloud computing hardware and open source software such as Gromacs, is an efficient alternative to in-person meetings and in-house calculation servers.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 89:
Performance Analysis of GROMACS Molecular Dynamics for Simulating Enzyme Substrate in the Cloud

The end-users of this team were Pravin Kumar R, Thirupathi Jatti, Soma Ghosh, Satish Nandagond, and Naveen Kulkarni from Polyclone Bioservices. Patrice Calegari and Marc Levrier were from resource provider Bull ServiWare providing eXtreme factory cloud services. Jaap Flohil from Foldyne was the team expert and Dennis Nagy from BeyondCAE the UberCloud mentor of the team.

The team focused on evaluating the performance of double precision MPI-enabled GROMACS 4.6.3 on 25 Bullx 510 blades (each 16-core Intel SNB, total 400 cores) for peta-scaling molecular dynamics simulations. The activities were organized into three tasks: install and optimize GROMACS performance on the Bull extreme factory cluster; install accessory tools to analyze simulation data; and test different multi-scale molecular systems involving enzyme substrate complexes on the cluster. The starting point of eEF (enzyme engineering framework), Polyclones’ framework for enzyme engineering, is to conduct molecular dynamics (MD) studies and calculate different parameters using the MD trajectories.


Figure 2: Protein enclosed in a box of water and ions. We can study proteins in atomic detail, down to the movements of individual water molecules (red/white balls and stick model), ions (purple), and the protein itself shown in the surface model (blue and green represents macromolecular dimer form of the protein).

Using 400 cores for MD studies was extremely helpful in estimating the time and resource for the partial completion of an enzyme engineering project using eEF. Bandwidth and latency were excellent. The remote visualization system was used to analyze huge (100 GB) simulation data. The web interface from the eXtreme factory cloud helped in shooting the jobs, allocating the cores for different jobs simultaneously, and also organizing the output results. The XF team installed and built a web interface for Gromacs, which can be used seamlessly to run and analyze molecular dynamics.


Figure 3: Performance trend as the number of cores increases for a system with ~120K particles.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 118:
Coupling In-house FE Code with ANSYS Fluent CFD

The end user was Hubert Dengg from Rolls-Royce Deutschland, software providers were Wim Slagter and René Kapa from ANSYS, resource providers and team experts were Thomas Gropp and Alexander Heine from CPU 24/7, and Marius Swoboda from Rolls-Royce Deutschland acted as HPC/CAE expert.

In the present test case, a jet engine high pressure compressor assembly was the subject of a transient aerothermal analysis using FEA/CFD coupling technique. Coupling is achieved through an iterative loop with smooth exchange of information between the FEA and CFD simulations at each time step, ensuring consistency of temperature and heat flux on the coupled interfaces between the metal and the fluid domains. The aim of the HPC Experiment was to link ANSYS Fluent with an in-house FEA code. This was done by extracting heat flux profiles from the Fluent CFD model and applying them to the FE model. The FE model provides metal temperatures in the solid domain.

This conjugate heat transfer process is very consuming in terms of computing power, especially when 3D CFD models with more than 10 million cells are required. As a consequence, we thought that using cloud resources would have a beneficial effect regarding computing time.

The computation was performed on the 32 cores of two nodes with dual Intel Xeon processors. The calculation was done in cycles in which the FE code and Fluent CFD ran alternating, exchanging their results.


Figure 4: Contours of total temperature for a jet engine component.

Outsourcing of the computational workload to an external cluster allowed the end user to distribute computing power in an efficient way – especially when the in-house computing resources were already at their limit.

Bigger models usually give more detailed insights into the physical behavior of the system.

In addition, the end user benefited from the HPC provider’s knowledge of how to setup a cluster, run applications in parallel based on MPI, create a host file, handle licenses, and prepare everything needed for turn-key access to the cluster.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 142:
Virtual Testing of Severe Service Control Valve

End User was Mark A. Lobo from Lobo Engineering.
Autodesk provided Simulation CFD 360 (SimCFD) and the supporting cloud infrastructure. And the HPC/CAE application experts were Jon den Hartog and Heath Houghton from Autodesk.


Figure 5: a control valve model with idealized flow path was used to minimize effects of a complex body cavity and trim design.

Flow control valve specifications include performance ratings in order for a valve to be properly applied in fluid management systems. Control systems sort out input parameters, disturbances and specifications of each piping system component to react and produce a desired output. System response is chiefly a function of the accuracy of control valves that respond to signals from the control system. Valve performance ratings provide information to the system designer that can be used to optimize control system response.

The premise of this project was not only to explore virtual valve testing, but to evaluate the practical and efficient use of CFD by the non-specialist design engineer. As a benchmark, the end user had no prior experience with the Autodesk software when the project initiated, no formal training in the software, and he was depended on the included tutorials, help utility, thorough documentation to produce good results and good data.


Figure 6: Application Domain. The control valve restriction components or “trim” reduces the annular area as the cavity profile on the right moves to the left. The location of highest velocity is indicated in red.

One of the benefits for the end-user was that cloud computing enabled accessing a large amount of computing power in a cost effective way. Rather than owning the hardware and software licenses, engineers can pay for what they need when they need it, rather than making a substantial upfront investment. In this project, over 200 simulations were run in the cloud. Given the runtimes involved and allowing for data download upon completion of the runs, it is possible for all of these simulations to be solved within a day. For an engineer with 1 simulation license on a single workstation, this would have required 800 hours (approximately 30 days) to complete if the simulations were running nonstop one after another. Table 1 compares the approximate time and investment that would be required for various solving approaches.

Table : Comparison of Desktop, Cloud, and HPC Solving Options

Simulation Solving Approach

Approx Time to Complete

Investment Required

Local Desktop Machine

800 hours

(1 month)

Engineering Workstation +Simulation SW License
Local Desktop Machine + Cloud Computing

24 hours

(1 day)

Engineering Workstation +Simulation SW License +

$1200 Cloud Compute Fee

Local Desktop Machine + Private HPC Cluster + Multiple Solver Licenses

24 hours

(1 day)

Engineering Workstation +Simulation SW License +

30 Node Compute Cluster +

30 Simulation Solver Licenses


More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

The 2nd UberCloud Compendium with HPC Cloud Case Studies is intended as a resource for engineers, scientists, managers and executives who believe in the strategic importance of HPC as a Service, in the Cloud. It’s a collection of selected HPC Cloud case studies from the participants in Rounds 3 & 4 of the UberCloud Experiment. Among these case studies are candid descriptions of challenges encountered, problems solved, lessons learned, and recommendations. This second UberCloud Compendium can be downloaded here.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Researchers Scale COSMO Climate Code to 4888 GPUs on Piz Daint

October 17, 2017

Effective global climate simulation, sorely needed to anticipate and cope with global warming, has long been computationally challenging. Two of the major obstacles are the needed resolution and prolonged time to compute Read more…

By John Russell

Student Cluster Competition Coverage New Home

October 16, 2017

Hello computer sports fans! This is the first of many (many!) articles covering the world-wide phenomenon of Student Cluster Competitions. Finally, the Student Cluster Competition coverage has come to its natural home: H Read more…

By Dan Olds

UCSD Web-based Tool Tracking CA Wildfires Generates 1.5M Views

October 16, 2017

Tracking the wildfires raging in northern CA is an unpleasant but necessary part of guiding efforts to fight the fires and safely evacuate affected residents. One such tool – Firemap – is a web-based tool developed b Read more…

By John Russell

HPE Extreme Performance Solutions

Transforming Genomic Analytics with HPC-Accelerated Insights

Advancements in the field of genomics are revolutionizing our understanding of human biology, rapidly accelerating the discovery and treatment of genetic diseases, and dramatically improving human health. Read more…

Exascale Imperative: New Movie from HPE Makes a Compelling Case

October 13, 2017

Why is pursuing exascale computing so important? In a new video – Hewlett Packard Enterprise: Eighteen Zeros – four HPE executives, a prominent national lab HPC researcher, and HPCwire managing editor Tiffany Trader Read more…

By John Russell

Student Cluster Competition Coverage New Home

October 16, 2017

Hello computer sports fans! This is the first of many (many!) articles covering the world-wide phenomenon of Student Cluster Competitions. Finally, the Student Read more…

By Dan Olds

Intel Delivers 17-Qubit Quantum Chip to European Research Partner

October 10, 2017

On Tuesday, Intel delivered a 17-qubit superconducting test chip to research partner QuTech, the quantum research institute of Delft University of Technology (TU Delft) in the Netherlands. The announcement marks a major milestone in the 10-year, $50-million collaborative relationship with TU Delft and TNO, the Dutch Organization for Applied Research, to accelerate advancements in quantum computing. Read more…

By Tiffany Trader

Fujitsu Tapped to Build 37-Petaflops ABCI System for AIST

October 10, 2017

Fujitsu announced today it will build the long-planned AI Bridging Cloud Infrastructure (ABCI) which is set to become the fastest supercomputer system in Japan Read more…

By John Russell

HPC Chips – A Veritable Smorgasbord?

October 10, 2017

For the first time since AMD's ill-fated launch of Bulldozer the answer to the question, 'Which CPU will be in my next HPC system?' doesn't have to be 'Whichever variety of Intel Xeon E5 they are selling when we procure'. Read more…

By Dairsie Latimer

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Intel Debuts Programmable Acceleration Card

October 5, 2017

With a view toward supporting complex, data-intensive applications, such as AI inference, video streaming analytics, database acceleration and genomics, Intel i Read more…

By Doug Black

OLCF’s 200 Petaflops Summit Machine Still Slated for 2018 Start-up

October 3, 2017

The Department of Energy’s planned 200 petaflops Summit computer, which is currently being installed at Oak Ridge Leadership Computing Facility, is on track t Read more…

By John Russell

US Exascale Program – Some Additional Clarity

September 28, 2017

The last time we left the Department of Energy’s exascale computing program in July, things were looking very positive. Both the U.S. House and Senate had pas Read more…

By Alex R. Larzelere

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

US Coalesces Plans for First Exascale Supercomputer: Aurora in 2021

September 27, 2017

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, in Arlington, Va., yesterday (Sept. 26), it was revealed that the "Aurora" supercompute Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

GlobalFoundries Puts Wind in AMD’s Sails with 12nm FinFET

September 24, 2017

From its annual tech conference last week (Sept. 20), where GlobalFoundries welcomed more than 600 semiconductor professionals (reaching the Santa Clara venue Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Leading Solution Providers

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

Intel Launches Software Tools to Ease FPGA Programming

September 5, 2017

Field Programmable Gate Arrays (FPGAs) have a reputation for being difficult to program, requiring expertise in specialty languages, like Verilog or VHDL. Easin Read more…

By Tiffany Trader

IBM Advances Web-based Quantum Programming

September 5, 2017

IBM Research is pairing its Jupyter-based Data Science Experience notebook environment with its cloud-based quantum computer, IBM Q, in hopes of encouraging a new class of entrepreneurial user to solve intractable problems that even exceed the capabilities of the best AI systems. Read more…

By Alex Woodie

Intel, NERSC and University Partners Launch New Big Data Center

August 17, 2017

A collaboration between the Department of Energy’s National Energy Research Scientific Computing Center (NERSC), Intel and five Intel Parallel Computing Cente Read more…

By Linda Barney

  • arrow
  • Click Here for More Headlines
  • arrow
Share This