The UberCloud Experiment Continues…

By Wolfgang Gentzsch and Burak Yenier

June 16, 2014

Improved product quality, faster time to market and ultimately increased ROI have long been associated with the adoption of HPC tools. The benefits that engineers and scientists can expect from using technical computing in their research, design and development processes can be huge. In spite of this, there is still unmet need with relatively few scientists and manufacturers using servers or clusters when designing and developing their products on computers. The majority of virtual prototyping and large-scale data modeling is still performed on desktop or laptop computers. It’s not surprising then that many of these users face problems stemming from the lack of performance of their machines. Obtaining additional computing power means either purchasing a server outright or using a cloud-based offering.

Many system vendors have developed a complete set of products, solutions and services for “high performance computing” (HPC) and buying an HPC server for an SME is no longer out of reach. Owning an HPC server, however, is not necessarily the best idea in terms of cost-efficiency or maintenance, because the Total Cost of Ownership (TCO) of such a machine is pretty high, and maintaining such a system requires additional manpower and expertise.

The other option today is to use an HPC Cloud solution that allows engineers and scientists to keep using their workstation for daily design and development work, and to “burst” larger, more complex jobs into the Cloud when needed. Thus, users have access to quasi infinite computing resources that offer higher-quality results. A Cloud solution helps reduce capital expenditure, offers greater business agility by dynamically scaling resources up and down as needed, and is only paid for when used.

The UberCloud Experiment Accelerating HPC Cloud Adoption

Since July 2012 the UberCloud Experiment has attracted more than 1500 organizations from 72 countries. We were able to build 152 teams in CFD, FEM, Biology, Finance, and other domains, and to publish many case studies reporting about different applications, experience, and lessons learned. UberCloud TechTalk has been founded providing educational lectures for our community. And the UberCloud Exhibit offers a Cloud services catalogue where community members can exhibit their Cloud related services or select the services which they want to use for their team experiment or for their daily work. Sponsored by Intel and HPCwire in 2013, the first Compendium with 25 HPC Cloud case studies received the HPCwire Readers Choice Award for the best HPC Cloud implementation. Now, in June 2014, the second Compendium of UberCloud case studies has been published, sponsored by Intel and HPCwire, which can be downloaded for free from here.

The UberCloud Experiment provides a platform for scientists and engineers to explore, learn and understand the end-to-end process of accessing and using HPC Cloud resources, and to identify and resolve the roadblocks. End-users, software providers, resource providers, and computing experts collaborate in teams to jointly solve the end-users’ application in the Cloud. Let’s start by defining what roles each stakeholder plays to make service-based high performance technical computing in the cloud come together:

End-user
– A typical example is a small or medium size manufacturer in the process of designing, prototyping and developing its next-generation product. These users are prime candidates for HPC-as-a-Service when in-house computation on workstations has become too lengthy and acquiring additional computing power in the form of a server is too cumbersome.

Application software provider
– These are software owners of all stripes, including ISVs, public domain software organizations and individual developers. The UberCloud Experiment prefers rock-solid software, which has the potential to be used on a wider scale.

Resource provider
– This pertains to anyone who owns HPC resources networked to the outside world. A classic HPC center would fall into this category as well as a standard datacenter used to handle batch jobs, or a cluster-owning commercial entity that is willing to offer up cycles to run non-competitive workloads during periods of low CPU-utilization.

Computing experts
– This group includes individuals and companies with HPC expertise, e.g. in areas like cluster management and software porting and optimizing. It also encompasses PhD-level domain specialists with in-depth application knowledge. These experts work as team leaders, with end-users, computer centers and software providers to help glue the pieces together.

For example, suppose the end-user is in need of additional compute resources to increase the quality of a product design or to speed up a product design cycle – say for simulating more sophisticated geometries or physics or for running many more simulations for a higher quality result. That suggests a specific software stack, domain expertise and even hardware configuration. The general idea is to look at the end-user’s tasks and software and select the appropriate resources and expertise that match certain requirements.

Cloud Case Studies from the UberCloud Experiment

As a glimpse into the wealth of practical use cases, we chose four HPC Cloud projects out of the 152 UberCloud Experiments: in life sciences teams 61 and 89, and in CAE teams 118 and 142.

Team 61: Molecular dynamics of the mutant PI3Kα protein

The end-user of this team was Zoe Cournia from the Biomedical Research Foundation of the Academy of Athens. Resource provider was the GRNET-Okeanos IaaS Cloud service, represented by Vangelis Floros and Stefanos Gerangelos from the Greek Research and Technology Network S.A. And HPC expert was Dominique Dehareng from the Center for Protein Engineering at the University of Liège. Open source software Gromacs 4.6.1 was used for the simulations.

Cancer is a leading cause of death worldwide, accounting for 7.6 million deaths in 2008 according to the World Health Organization. This number is expected to increase to 21 million by 2030. One of the signaling pathways which, when deregulated, becomes centrally involved in several types of cancers, like colon, mammary, and endometrial tumorigenesis, is served by phosphoinositide-3-kinase alpha (PI3Kα). The importance of PI3Kα in cancer is highlighted by the discovery that PIK3CA, the gene encoding the catalytic p110α subunit of PI3Kα, is frequently mutated in human malignancies. The goal of this project was to gain insights into the oncogenic mechanism of two commonly expressed PI3Kα mutants by studying their conformational changes with Molecular Dynamics (MD) simulations, in comparison with the PI3Kα wild-type (normal, non-cancerous) protein. The utilization of cloud computing in performing MD simulations of mutant PI3Kα with Gromacs was examined in this case study.

UberCloud_Fig1

Figure 1: The protein PI3Ka is depicted in ribbons and is placed in a water box, shown as red dots.

The cloud computing service that was provided was a free and very efficient service. Cloud computing proved to be an extremely easy process as compared to building and maintaining your own cluster. The team was provided with a VM with 8 cores and the possibility of building a cluster of up to 64 cores connected via Ethernet network. In the 8 cores, the MD simulation of the mutant PI3Kα protein scaled linearly and did run faster when compared to in-house simulations. Team communications through BaseCamp were very efficient. Collaborative work over the Internet, using on-line resources like cloud computing hardware and open source software such as Gromacs, is an efficient alternative to in-person meetings and in-house calculation servers.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 89:
Performance Analysis of GROMACS Molecular Dynamics for Simulating Enzyme Substrate in the Cloud

The end-users of this team were Pravin Kumar R, Thirupathi Jatti, Soma Ghosh, Satish Nandagond, and Naveen Kulkarni from Polyclone Bioservices. Patrice Calegari and Marc Levrier were from resource provider Bull ServiWare providing eXtreme factory cloud services. Jaap Flohil from Foldyne was the team expert and Dennis Nagy from BeyondCAE the UberCloud mentor of the team.

The team focused on evaluating the performance of double precision MPI-enabled GROMACS 4.6.3 on 25 Bullx 510 blades (each 16-core Intel SNB, total 400 cores) for peta-scaling molecular dynamics simulations. The activities were organized into three tasks: install and optimize GROMACS performance on the Bull extreme factory cluster; install accessory tools to analyze simulation data; and test different multi-scale molecular systems involving enzyme substrate complexes on the cluster. The starting point of eEF (enzyme engineering framework), Polyclones’ framework for enzyme engineering, is to conduct molecular dynamics (MD) studies and calculate different parameters using the MD trajectories.

UberCloud_Fig2

Figure 2: Protein enclosed in a box of water and ions. We can study proteins in atomic detail, down to the movements of individual water molecules (red/white balls and stick model), ions (purple), and the protein itself shown in the surface model (blue and green represents macromolecular dimer form of the protein).

Using 400 cores for MD studies was extremely helpful in estimating the time and resource for the partial completion of an enzyme engineering project using eEF. Bandwidth and latency were excellent. The remote visualization system was used to analyze huge (100 GB) simulation data. The web interface from the eXtreme factory cloud helped in shooting the jobs, allocating the cores for different jobs simultaneously, and also organizing the output results. The XF team installed and built a web interface for Gromacs, which can be used seamlessly to run and analyze molecular dynamics.

UberCloud_Fig3

Figure 3: Performance trend as the number of cores increases for a system with ~120K particles.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 118:
Coupling In-house FE Code with ANSYS Fluent CFD

The end user was Hubert Dengg from Rolls-Royce Deutschland, software providers were Wim Slagter and René Kapa from ANSYS, resource providers and team experts were Thomas Gropp and Alexander Heine from CPU 24/7, and Marius Swoboda from Rolls-Royce Deutschland acted as HPC/CAE expert.

In the present test case, a jet engine high pressure compressor assembly was the subject of a transient aerothermal analysis using FEA/CFD coupling technique. Coupling is achieved through an iterative loop with smooth exchange of information between the FEA and CFD simulations at each time step, ensuring consistency of temperature and heat flux on the coupled interfaces between the metal and the fluid domains. The aim of the HPC Experiment was to link ANSYS Fluent with an in-house FEA code. This was done by extracting heat flux profiles from the Fluent CFD model and applying them to the FE model. The FE model provides metal temperatures in the solid domain.

This conjugate heat transfer process is very consuming in terms of computing power, especially when 3D CFD models with more than 10 million cells are required. As a consequence, we thought that using cloud resources would have a beneficial effect regarding computing time.

The computation was performed on the 32 cores of two nodes with dual Intel Xeon processors. The calculation was done in cycles in which the FE code and Fluent CFD ran alternating, exchanging their results.

UberCloud_Fig4

Figure 4: Contours of total temperature for a jet engine component.

Outsourcing of the computational workload to an external cluster allowed the end user to distribute computing power in an efficient way – especially when the in-house computing resources were already at their limit.

Bigger models usually give more detailed insights into the physical behavior of the system.

In addition, the end user benefited from the HPC provider’s knowledge of how to setup a cluster, run applications in parallel based on MPI, create a host file, handle licenses, and prepare everything needed for turn-key access to the cluster.

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

Team 142:
Virtual Testing of Severe Service Control Valve

End User was Mark A. Lobo from Lobo Engineering.
Autodesk provided Simulation CFD 360 (SimCFD) and the supporting cloud infrastructure. And the HPC/CAE application experts were Jon den Hartog and Heath Houghton from Autodesk.

UberCloud_Fig5

Figure 5: a control valve model with idealized flow path was used to minimize effects of a complex body cavity and trim design.

Flow control valve specifications include performance ratings in order for a valve to be properly applied in fluid management systems. Control systems sort out input parameters, disturbances and specifications of each piping system component to react and produce a desired output. System response is chiefly a function of the accuracy of control valves that respond to signals from the control system. Valve performance ratings provide information to the system designer that can be used to optimize control system response.

The premise of this project was not only to explore virtual valve testing, but to evaluate the practical and efficient use of CFD by the non-specialist design engineer. As a benchmark, the end user had no prior experience with the Autodesk software when the project initiated, no formal training in the software, and he was depended on the included tutorials, help utility, thorough documentation to produce good results and good data.

UberCloud_Fig6

Figure 6: Application Domain. The control valve restriction components or “trim” reduces the annular area as the cavity profile on the right moves to the left. The location of highest velocity is indicated in red.

One of the benefits for the end-user was that cloud computing enabled accessing a large amount of computing power in a cost effective way. Rather than owning the hardware and software licenses, engineers can pay for what they need when they need it, rather than making a substantial upfront investment. In this project, over 200 simulations were run in the cloud. Given the runtimes involved and allowing for data download upon completion of the runs, it is possible for all of these simulations to be solved within a day. For an engineer with 1 simulation license on a single workstation, this would have required 800 hours (approximately 30 days) to complete if the simulations were running nonstop one after another. Table 1 compares the approximate time and investment that would be required for various solving approaches.

Table : Comparison of Desktop, Cloud, and HPC Solving Options

Simulation Solving Approach

Approx Time to Complete

Investment Required

Local Desktop Machine

800 hours

(1 month)

Engineering Workstation +Simulation SW License
Local Desktop Machine + Cloud Computing

24 hours

(1 day)

Engineering Workstation +Simulation SW License +

$1200 Cloud Compute Fee

Local Desktop Machine + Private HPC Cluster + Multiple Solver Licenses

24 hours

(1 day)

Engineering Workstation +Simulation SW License +

30 Node Compute Cluster +

30 Simulation Solver Licenses

 

More details about this case study can be found in the second UberCloud Compendium which can be downloaded here for free.

The 2nd UberCloud Compendium with HPC Cloud Case Studies is intended as a resource for engineers, scientists, managers and executives who believe in the strategic importance of HPC as a Service, in the Cloud. It’s a collection of selected HPC Cloud case studies from the participants in Rounds 3 & 4 of the UberCloud Experiment. Among these case studies are candid descriptions of challenges encountered, problems solved, lessons learned, and recommendations. This second UberCloud Compendium can be downloaded here.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

AMD Charges Back into the Datacenter and HPC Workflows with EPYC Processor

June 20, 2017

AMD is charging back into the enterprise datacenter and select HPC workflows with its new EPYC 7000 processor line, code-named Naples, announced today at a “global” launch event in Austin TX. In many ways it was a fu Read more…

By John Russell

Hyperion: Deep Learning, AI Helping Drive Healthy HPC Industry Growth

June 20, 2017

To be at the ISC conference in Frankfurt this week is to experience deep immersion in deep learning. Users want to learn about it, vendors want to talk about it, analysts and journalists want to report on it. Deep learni Read more…

By Doug Black

OpenACC Shows Growing Strength at ISC

June 19, 2017

OpenACC is strutting its stuff at ISC this year touting expanding membership, a jump in downloads, favorable benchmarks across several architectures, new staff members, and new support by key HPC applications providers, Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major shakeups -- China still has the top two spots locked with th Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

Creating a Roadmap for HPC Innovation at ISC 2017

In an era where technological advancements are driving innovation to every sector, and powering major economic and scientific breakthroughs, high performance computing (HPC) is crucial to tackle the challenges of today and tomorrow. Read more…

ISC: Extreme-Scale Requirements to Push the Frontiers of Deep Learning

June 17, 2017

Deep learning is the latest and most compelling technology strategy to take aim at the decades-old “drowning in data/starving for insight” problem. But contrary to the commonly held notion, deep learning is more than Read more…

By Doug Black

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD, Cray, Hewlett Packard Enterprise (HPE), IBM, Intel, and NV Read more…

By John Russell

OpenSuCo: Advancing Open Source Supercomputing at ISC

June 15, 2017

As open source hardware gains traction, the potential for a completely open source supercomputing system becomes a compelling proposition, one that is being investigated by the International Workshop on Open Source Super Read more…

By Tiffany Trader

Mellanox Raises SHIELD for Enhanced Network Resiliency

June 15, 2017

High-performance networking company Mellanox is announcing a new protocol it’s calling SHIELD, an acronym for “Self Healing Interconnect Enhancement for inteLligent Datacenters.” Enabled within Mellanox’s 100G ED Read more…

By Tiffany Trader

AMD Charges Back into the Datacenter and HPC Workflows with EPYC Processor

June 20, 2017

AMD is charging back into the enterprise datacenter and select HPC workflows with its new EPYC 7000 processor line, code-named Naples, announced today at a “g Read more…

By John Russell

Hyperion: Deep Learning, AI Helping Drive Healthy HPC Industry Growth

June 20, 2017

To be at the ISC conference in Frankfurt this week is to experience deep immersion in deep learning. Users want to learn about it, vendors want to talk about it Read more…

By Doug Black

OpenACC Shows Growing Strength at ISC

June 19, 2017

OpenACC is strutting its stuff at ISC this year touting expanding membership, a jump in downloads, favorable benchmarks across several architectures, new staff Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

ISC: Extreme-Scale Requirements to Push the Frontiers of Deep Learning

June 17, 2017

Deep learning is the latest and most compelling technology strategy to take aim at the decades-old “drowning in data/starving for insight” problem. But cont Read more…

By Doug Black

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

OpenSuCo: Advancing Open Source Supercomputing at ISC

June 15, 2017

As open source hardware gains traction, the potential for a completely open source supercomputing system becomes a compelling proposition, one that is being inv Read more…

By Tiffany Trader

Mellanox Raises SHIELD for Enhanced Network Resiliency

June 15, 2017

High-performance networking company Mellanox is announcing a new protocol it’s calling SHIELD, an acronym for “Self Healing Interconnect Enhancement for int Read more…

By Tiffany Trader

Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

March 21, 2017

For a technology that’s usually characterized as far off and in a distant galaxy, quantum computing has been steadily picking up steam. Just how close real-wo Read more…

By John Russell

Trump Budget Targets NIH, DOE, and EPA; No Mention of NSF

March 16, 2017

President Trump’s proposed U.S. fiscal 2018 budget issued today sharply cuts science spending while bolstering military spending as he promised during the cam Read more…

By John Russell

HPC Compiler Company PathScale Seeks Life Raft

March 23, 2017

HPCwire has learned that HPC compiler company PathScale has fallen on difficult times and is asking the community for help or actively seeking a buyer for its a Read more…

By Tiffany Trader

Google Pulls Back the Covers on Its First Machine Learning Chip

April 6, 2017

This week Google released a report detailing the design and performance characteristics of the Tensor Processing Unit (TPU), its custom ASIC for the inference Read more…

By Tiffany Trader

CPU-based Visualization Positions for Exascale Supercomputing

March 16, 2017

In this contributed perspective piece, Intel’s Jim Jeffers makes the case that CPU-based visualization is now widely adopted and as such is no longer a contrarian view, but is rather an exascale requirement. Read more…

By Jim Jeffers, Principal Engineer and Engineering Leader, Intel

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Nvidia’s Mammoth Volta GPU Aims High for AI, HPC

May 10, 2017

At Nvidia's GPU Technology Conference (GTC17) in San Jose, Calif., this morning, CEO Jensen Huang announced the company's much-anticipated Volta architecture a Read more…

By Tiffany Trader

Facebook Open Sources Caffe2; Nvidia, Intel Rush to Optimize

April 18, 2017

From its F8 developer conference in San Jose, Calif., today, Facebook announced Caffe2, a new open-source, cross-platform framework for deep learning. Caffe2 is the successor to Caffe, the deep learning framework developed by Berkeley AI Research and community contributors. Read more…

By Tiffany Trader

Leading Solution Providers

MIT Mathematician Spins Up 220,000-Core Google Compute Cluster

April 21, 2017

On Thursday, Google announced that MIT math professor and computational number theorist Andrew V. Sutherland had set a record for the largest Google Compute Engine (GCE) job. Sutherland ran the massive mathematics workload on 220,000 GCE cores using preemptible virtual machine instances. Read more…

By Tiffany Trader

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

US Supercomputing Leaders Tackle the China Question

March 15, 2017

Joint DOE-NSA report responds to the increased global pressures impacting the competitiveness of U.S. supercomputing. Read more…

By Tiffany Trader

Groq This: New AI Chips to Give GPUs a Run for Deep Learning Money

April 24, 2017

CPUs and GPUs, move over. Thanks to recent revelations surrounding Google’s new Tensor Processing Unit (TPU), the computing world appears to be on the cusp of Read more…

By Alex Woodie

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

DOE Supercomputer Achieves Record 45-Qubit Quantum Simulation

April 13, 2017

In order to simulate larger and larger quantum systems and usher in an age of “quantum supremacy,” researchers are stretching the limits of today’s most advanced supercomputers. Read more…

By Tiffany Trader

Messina Update: The US Path to Exascale in 16 Slides

April 26, 2017

Paul Messina, director of the U.S. Exascale Computing Project, provided a wide-ranging review of ECP’s evolving plans last week at the HPC User Forum. Read more…

By John Russell

Knights Landing Processor with Omni-Path Makes Cloud Debut

April 18, 2017

HPC cloud specialist Rescale is partnering with Intel and HPC resource provider R Systems to offer first-ever cloud access to Xeon Phi "Knights Landing" processors. The infrastructure is based on the 68-core Intel Knights Landing processor with integrated Omni-Path fabric (the 7250F Xeon Phi). Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This