Helix Nebula: from Grid to Cloud and Lessons Learned So Far

By Tiffany Trader

July 17, 2012

European cloud computing is taking off as can be seen in the progress of Helix Nebula – the Science Cloud, a collaboration between select service providers and three of Europe’s most prominent research centers, CERN, the European Space Agency (ESA), and the European Molecular Biology Laboratory (EMBL). The Helix Nebula project announced last week that they were on the verge of moving from the initial proof of concept phase to the start of the two-year pilot phase, which involves expanded proofs of concepts and perhaps some additional demand side partners.

Helix NebulaThe three flagship applications (one from each research institution) have been deployed to cloud resources provided by Atos, CloudSigma and T-Systems.

Michael Higgins, Chief Enterprise Solutions Officer at CloudSigma, makes the point that CERN and the other research institutions are not really customers, not yet at least. Currently, they are all partners exploring the feasibility of migrating workloads from the grid to the cloud. Higgins further explains that Helix Nebula brings together two sets of consortia, the demand side and the supply side, which is comprised of both public and private cloud providers.

In the initial proof-of-concept phase, commercial terms are not imposed by the cloud providers on a pay-per-use basis, but instead involve agreed-upon bulk-payment monetary contribution from each of the demand side participants, based on each vendor’s ability to deliver the proofs of concept.

CERN and the Worldwide LHC Computing Grid

Project participant CERN generates a huge amount of data on its Large Hadron Collider. The LHC generally produces about 15 petabytes (15 million gigabytes) of data annually, but this year, they’re on track to reach 30 petabytes as the search for the Higgs boson particle has picked up steam. To analyze all this data, research partners from around the world rely on the Worldwide LHC Computing Grid (WLCG), a global grid network of more than 150 computing centers.

When asked what role if any Helix Nebula played in the preliminary Higgs boson discovery, the response from WLCG Project Leader Ian Bird was a qualified none:

“We did succeed in running some simulation work in production, and I dare say some of that resulting simulation was used in the analysis of the data that led to the announcement last week, but this was a very tiny fraction compared to the huge amount of data that had to be processed.”

Helix Nebula has gone from its initial stages of technology review to the point now where CloudSigma has completed all three proofs of concept for the flagship workloads. Based on that success, they’re now moving toward the next phase, which is to expand the proofs of concept and to begin to refine the commercial terms.

“They’re not only expanding the original proofs of concept, but opening the door to more demand-side flagship projects. Up to now, the researchers have not been overly pressed to understand TCO – total cost of ownership – so this may be something they’re struggling with,” suggests Higgins. “Like at CERN, IT doesn’t pay for electricity, so they would not know how much to factor in for their in-house server electricity costs.”

Higgins makes the case that cloud bursting is more suitable for science than grid or on-premise systems because very little science occurs 24/7, around the clock. There are times when ATLAS is not generating data and other times when there is an backlog of work. Oversubscription and underutilization are often the norm with designated resources, but bursting allows researchers to use only the resources they need when they need it.

Institutions are facing funding issues, explains Higgins, which means there is less hardware to evergreen or purchase new. Every day there is more compute demand, and the resources are strained. They have to look to the cloud for cost-savings or they have to find new sources of capital investment. Running workloads in the cloud on a pay-per-use basis erases the problem of buying a $10,000 platform and running it two weeks out of four.

While these arguments make sense, applying virtualization and cloud technologies to current grid resources is another avenue for boosting utilization rates and creating elasticity and scalability, and CERN is exploring these options in addition to the public cloud. Still, Bird notes that simply having a private cloud won’t work either, because the research depends on a federated connected cloud.

Grid Versus Cloud

Asked what will happen to the Worldwide LHC grid as cloud ramps up, Bird says that it will remain. He uses the opportunity to discuss current grid developments. They are running virtual machines on some of the sites, and they are in the process of deploying OpenStack. These projects are designed to improve their internal efficiency as well as the way they run services and provide services, and will also give them additional opportunities to interact with cloud sites.

Bird points to an important distinction between grid and cloud which is one of federation. Grid, despite being a networked collection of distributed computing systems, has evolved to become a highly-unified computing resource. Whereas using multiple cloud providers essentially means you have a collection of disparate resources that are difficult to integrate, even when they’re working together as with Helix Nebula. In addition to the API headaches, there are a myriad of standards and integration pain points to contend with, he says, elaborating further on the grid/cloud dichotomy:

The reason why we used the grid in the first place is because the computing resources that we have access to which are provided by the science funding agencies are physically distributed around the world and we have to have a way of putting these together, so that we did with grid technologies. So for us, the grid is a way of sharing resources and collaborating, while the cloud isn’t really that, it’s more to do with economies of scale. It’s distributed in the way it’s remote from you, but it’s really a different concept. One of the interesting things is how much of that [cloud] technology can we use to improve the way we run our own computer centers simply by not having to support grid infrastructure, but switching that to some cloud technology and how much can we do by buying resources from commercial resources?

As for comparisons between the grid I/O problem versus the cloud I/O problem, Bird observes that while these are similar, this is an area that has received a lot of investment on the grid side. Over the years, the partner institutions have developed dedicated optical private networks between the servers and their large compute centers and they also make significant use of specialized academic networks. When asked if he sees similar developments happening in cloud, Bird is doubtful they’ll happen in the near term, and points to another question for Helix Nebula, which is what is the connectivity of these partners and can we reach them over the academic networks? He says these are among the types of issues they want to pin down in the next couple of years.

On the positive side, Bird notes that networking has changed a lot since the deployment of grid. Prices have come down and data management techniques have become more effective. These developments will be applied to the Helix Nebula project.

Regarding the more specific process of transferring workloads from the grid to the cloud, CloudSigma’s Higgins explains that some of them ported over without too much work, while others have required more extensive retooling due to the numerous changes in the software design practices and machine architectures since these applications were first written. As an illustration, many of the apps in use today were written prior to JAVA and NoSQL databases.

Bird has a somewhat different take on the nitty-gritty details of the cloud migration, saying that they did not need to change the code at all. He explains that the LHC codes fall into the “high-throughput computing” model, where the different pieces of the running code do not need to communicate with each other. The grid resource and cloud resource are basically the same, he notes, i.e., a big cluster of Linux machines. The main difference is how you access this resource and how you move data in and out, however “at the level of real-code running on the machine, it’s the same,” he says.

A Cloud is Not a Cloud

From Bird’s point of view, CERN saw the successful completion of its three proofs of concept. The process entailed running the same simulation workload with the three different cloud providers. The conclusion Bird draws from this, is that while successful, “a cloud is not a cloud is not a cloud.” You cannot just write-once, run-anywhere; there are integration headaches.

According to the grid expert, they absolutely will need an adapter layer that knows how to talk to the different providers. This is essential if they want to use these resources in a dynamic way that involves moving between cloud providers. When asked about a possible performance penalty, he responds that since this is mainly a way to get data into the cloud, any overhead would be likely be negligible. He adds these so-called cloud broker solutions already exist in the open source domain; Deltacloud and libcloud are examples. While this layer adds complexity and could even introduce faults, it’s unavoidable at this stage of the game if you value transparency and interoperability.

Up until now, CERN has been running cloud-friendly workloads with little network I/O dependency. When asked about the HPC cloud bandwidth issue, i.e., the limitations of getting data in and out of the cloud, Bird said this absolutely could be a problem. Their normal data processing workloads involve transferring petabytes of data. During the two-year pilot phase, they will address several issues related to data movement: whether they can move data in and out of the cloud at this scale, whether they can afford to do this, and possible policy issues involved with moving academic data into the commercial domain.

Bird returns to the bottom line, which is cost: “Can we afford to move data in and out and can we afford to store data in the cloud?” he asks.

“There are many different use cases. I think we can overcome the technical issues; the most interesting question is what’s the real cost of doing this and how does it compare with the infrastructure that we have currently?”

Bird gives the impression that while cloud migration has potential, it’s not a “sure thing” and by no means a panacea. How do you get a collection of cloud providers to behave as a federated resource? The first steps involve supplementing the existing grid resources with a few cloud providers, says Bird, which allows those involved to begin the process of learning how to integrate the various pieces.

New Science

CloudSigma’s Higgins is most excited about the new science that can be enabled by the cloud as more and more science databases are migrated over. Right now, there are three databases that do not combine to support any practical implications, but the possibilities for meta-analysis are intriguing. For example, with ESA’s earth observation data stored in the cloud, a researcher could ask the World Health Organization for the mosquito outbreak data. This would create a platform where both databases would be available, allowing scientists to expand their research horizons.

This kind of integration requires a lot of work because each database has its own schema. Right now CloudSigma is working on an initiative that is attempting to create master schemas. For example, the earth observation data is linked to latitude and longitude, whereas the mosquito outbreak data is based on distance from a known point on a compass bearing. So the question then is how to marry these distinct data points. The new effort is hammering out a global schema which can make sense of these disparate units so that it becomes a useful tool for researchers. At that point, a scientist could answer queries such as “six miles north of Nairobi, how wet is it and how many mosquitos are we expecting to break out?”

Higgins is confident that creating a rich ecosystem of multiple scientific databases will draw new researchers to the cloud.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Scientists Conduct First Quantum Simulation of Atomic Nucleus

May 23, 2018

OAK RIDGE, Tenn., May 23, 2018—Scientists at the Department of Energy’s Oak Ridge National Laboratory are the first to successfully simulate an atomic nucleus using a quantum computer. The results, published in Ph Read more…

By Rachel Harken, ORNL

Pattern Computer – Startup Claims Breakthrough in ‘Pattern Discovery’ Technology

May 23, 2018

If it weren’t for the heavy-hitter technology team behind start-up Pattern Computer, which emerged from stealth today in a live-streamed event from San Francisco, one would be tempted to dismiss its claims of inventing Read more…

By John Russell

Intel, Micro Debut Quad-Level Cell NAND Flash

May 22, 2018

Chipmakers continue to gear designs toward AI and other demanding cloud workloads that take advantage of datacenter flash storage capacity. To that end, memory specialist Micron Technology Inc. began shipping compact sol Read more…

By George Leopold

HPE Extreme Performance Solutions

HPC and AI Convergence is Accelerating New Levels of Intelligence

Data analytics is the most valuable tool in the digital marketplace – so much so that organizations are employing high performance computing (HPC) capabilities to rapidly collect, share, and analyze endless streams of data. Read more…

IBM Accelerated Insights

Mastering the Big Data Challenge in Cognitive Healthcare

Patrick Chain, genomics researcher at Los Alamos National Laboratory, posed a question in a recent blog: What if a nurse could swipe a patient’s saliva and run a quick genetic test to determine if the patient’s sore throat was caused by a cold virus or a bacterial infection? Read more…

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combined peak computing capacity, the new systems will extend the a Read more…

By Tiffany Trader

Pattern Computer – Startup Claims Breakthrough in ‘Pattern Discovery’ Technology

May 23, 2018

If it weren’t for the heavy-hitter technology team behind start-up Pattern Computer, which emerged from stealth today in a live-streamed event from San Franci Read more…

By John Russell

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combine Read more…

By Tiffany Trader

ASC18: Final Results Revealed & Wrapped Up

May 17, 2018

It was an exciting week at ASC18 in Nanyang, China. The student teams braved extreme heat, extremely difficult applications, and extreme competition in order to cross the cluster competition finish line. The gala awards ceremony took place on Wednesday. The auditorium was packed with student teams, various dignitaries, the media, and other interested parties. So what happened? Read more…

By Dan Olds

Spring Meetings Underscore Quantum Computing’s Rise

May 17, 2018

The month of April 2018 saw four very important and interesting meetings to discuss the state of quantum computing technologies, their potential impacts, and th Read more…

By Alex R. Larzelere

Quantum Network Hub Opens in Japan

May 17, 2018

Following on the launch of its Q Commercial quantum network last December with 12 industrial and academic partners, the official Japanese hub at Keio University is now open to facilitate the exploration of quantum applications important to science and business. The news comes a week after IBM announced that North Carolina State University was the first U.S. university to join its Q Network. Read more…

By Tiffany Trader

Democratizing HPC: OSC Releases Version 1.3 of OnDemand

May 16, 2018

Making HPC resources readily available and easier to use for scientists who may have less HPC expertise is an ongoing challenge. Open OnDemand is a project by t Read more…

By John Russell

PRACE 2017 Annual Report: Exascale Aspirations; Industry Collaboration; HPC Training

May 15, 2018

The Partnership for Advanced Computing in Europe (PRACE) today released its annual report showcasing 2017 activities and providing a glimpse into thinking about Read more…

By John Russell

US Forms AI Brain Trust

May 11, 2018

Amid calls for a U.S. strategy for promoting AI development, the Trump administration is forming a senior-level panel to help coordinate government and industry research efforts. The Select Committee on Artificial Intelligence was announced Thursday (May 10) during a White House summit organized by the Office of Science and Technology Policy (OSTP). Read more…

By George Leopold

MLPerf – Will New Machine Learning Benchmark Help Propel AI Forward?

May 2, 2018

Let the AI benchmarking wars begin. Today, a diverse group from academia and industry – Google, Baidu, Intel, AMD, Harvard, and Stanford among them – releas Read more…

By John Russell

How the Cloud Is Falling Short for HPC

March 15, 2018

The last couple of years have seen cloud computing gradually build some legitimacy within the HPC world, but still the HPC industry lies far behind enterprise I Read more…

By Chris Downing

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 19, 2018

Petaflop per second deep learning training performance on the NERSC (National Energy Research Scientific Computing Center) Cori supercomputer has given climate Read more…

By Rob Farber

AI Cloud Competition Heats Up: Google’s TPUs, Amazon Building AI Chip

February 12, 2018

Competition in the white hot AI (and public cloud) market pits Google against Amazon this week, with Google offering AI hardware on its cloud platform intended Read more…

By Doug Black

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Lenovo Unveils Warm Water Cooled ThinkSystem SD650 in Rampup to LRZ Install

February 22, 2018

This week Lenovo took the wraps off the ThinkSystem SD650 high-density server with third-generation direct water cooling technology developed in tandem with par Read more…

By Tiffany Trader

Leading Solution Providers

HPC and AI – Two Communities Same Future

January 25, 2018

According to Al Gara (Intel Fellow, Data Center Group), high performance computing and artificial intelligence will increasingly intertwine as we transition to Read more…

By Rob Farber

Google Chases Quantum Supremacy with 72-Qubit Processor

March 7, 2018

Google pulled ahead of the pack this week in the race toward "quantum supremacy," with the introduction of a new 72-qubit quantum processor called Bristlecone. Read more…

By Tiffany Trader

HPE Wins $57 Million DoD Supercomputing Contract

February 20, 2018

Hewlett Packard Enterprise (HPE) today revealed details of its massive $57 million HPC contract with the U.S. Department of Defense (DoD). The deal calls for HP Read more…

By Tiffany Trader

CFO Steps down in Executive Shuffle at Supermicro

January 31, 2018

Supermicro yesterday announced senior management shuffling including prominent departures, the completion of an audit linked to its delayed Nasdaq filings, and Read more…

By John Russell

Deep Learning Portends ‘Sea Change’ for Oil and Gas Sector

February 1, 2018

The billowing compute and data demands that spurred the oil and gas industry to be the largest commercial users of high-performance computing are now propelling Read more…

By Tiffany Trader

Nvidia Ups Hardware Game with 16-GPU DGX-2 Server and 18-Port NVSwitch

March 27, 2018

Nvidia unveiled a raft of new products from its annual technology conference in San Jose today, and despite not offering up a new chip architecture, there were still a few surprises in store for HPC hardware aficionados. Read more…

By Tiffany Trader

Hennessy & Patterson: A New Golden Age for Computer Architecture

April 17, 2018

On Monday June 4, 2018, 2017 A.M. Turing Award Winners John L. Hennessy and David A. Patterson will deliver the Turing Lecture at the 45th International Sympo Read more…

By Staff

Part One: Deep Dive into 2018 Trends in Life Sciences HPC

March 1, 2018

Life sciences is an interesting lens through which to see HPC. It is perhaps not an obvious choice, given life sciences’ relative newness as a heavy user of H Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This