Learning from Clouds Past: A Look Back at Magellan

By Tiffany Trader

February 1, 2012

In 2009, the US Department of Energy (DOE) launched a bold experiment, a $32 million program to assess the benefit of cloud computing to the scientific community. A distributed testbed infrastructure, named Magellan, was established at the Argonne Leadership Computing Facility (ALCF) and the National Energy Research Scientific Computing Center (NERSC) to provide a tool for computational science in a cloud environment. Magellan, with funding from the American Recovery and Reinvestment Act, was to help major research organizations answer the classic cloud question: is it better to rent or buy?

“What we’re exploring is the question of whether the DOE or other government agencies should be buying their own clusters … or whether those kinds of purchases should be done in a more consolidated way,” said NERSC Director Kathy Yelick in a previous article.

Despite high-hopes and community support, in late 2011, we learned that the Magellan project was being discontinued, leaving many wondering what happened. Now we have some answers in the form of a 169-page report, sponsored by the Department of Energy’s Office of Advanced Scientific Computing Research (ASCR), which funded the study to assess what Magellan tells us about the the role of cloud computing for scientific applications.

Since industry was already benefiting from the cloud model, from the economies of scale generated by a shared pool of network-accessible resources, the Magellan team members initially set out to determine if cloud would hold the same potential for science. As stated in the executive summary:

The goal of Magellan, a project funded through the U.S. Department of Energy (DOE) Office of Advanced Scientific Computing Research (ASCR), was to investigate the potential role of cloud computing in addressing the computing needs for the DOE Office of Science (SC), particularly related to serving the needs of mid-range computing and future data-intensive computing workloads. A set of research questions was formed to probe various aspects of cloud computing from performance, usability, and cost.

Specifically, Magellan was tasked with addressing the following questions:

  • Are the open source cloud software stacks ready for DOE HPC science?
  • Can DOE cyber security requirements be met within a cloud?
  • Are the new cloud programming models useful for scientific computing?
  • Can DOE HPC applications run efficiently in the cloud? What applications are suitable for clouds?
  • How usable are cloud environments for scientific applications?
  • When is it cost effective to run DOE HPC science in a cloud?

It should be noted that Magellan was not a typical commercial cloud, rather this “science cloud” was purpose-built for the special requirements of scientific computing. Magellan was based on the IBM iDataplex chassis using Intel processor cores for a theoretical peak performance of over 100 teraflop/s. Other components include:

  • High bandwidth, low-latency node interconnects (InfiniBand).
  • High-bin processors tuned for performance.
  • Preinstalled scientific applications, compilers, debuggers, math libraries and other tools.
  • High-bandwidth parallel file system.
  • High-capacity data archive.

During Magellan’s two-year run, the staff at NERSC and Argonne National Laboratory examined how different aspects of cloud computing infrastructure and technologies could be harnessed by various scientific applications. They evaluated cloud models such as Infrastructure as a Service (IaaS) and Platform as a Service (Paas), virtual software stacks, MapReduce and open-source implementation (Hadoop), as well as resource provider and user perspectives.

Using a wide-range of applications as benchmarks, the researchers compared the Magellan cloud with various other architectures, including a Cray XT4 supercomputer, a Dell cluster system, and Amazon’s EC2 commercial cloud offering. Despite the testbed moniker, a lot of important production science took place, contributing to advances in particle physics, climate research, quantum chemistry, plasma physics and astrophysics.

Science workloads, by their nature, tend to be cloud-challenged, although to varying degrees. The report outlines the three major classifications of computational models, beginning with large-scale tightly-coupled science codes, which require the power of traditional supercomputers and take a big penalty working in a virtualized cloud environment. Then, there are the mid-range tightly-coupled applications, which run at a smaller scale and tend to be good candidates for cloud, although there is some performance loss. The final category, high-throughput workloads, usually involve asynchronous, independent computations, and in the past relied on desktop and small clusters for processing. But due to an explosion in sensor data, cloud is a good fit, especially when you factor in the fact that these high-throughput and data-intensive workloads do not fit into current scheduling and allocation policies.

The two-year Magellan project led to these key findings:

  • Scientific applications have special requirements that require cloud solutions that are tailored to these needs.
  • Scientific applications with minimal communication and I/O are best suited for clouds.
  • Clouds require significant programming and system administration support.
  • Significant gaps and challenges exist in current open-source virtualized cloud software stacks for production science use.
  • Clouds expose a different risk model requiring different security practices and policies.
  • MapReduce shows promise in addressing scientific needs, but current implementations have gaps and challenges.
  • Public clouds can be more expensive than in-house large systems. Many of the cost benefits from clouds result from the increased consolidation and higher average utilization.
  • DOE supercomputing centers already achieve energy efficiency levels comparable to commercial cloud centers.
  • Cloud is a business model and can be applied at DOE supercomputing centers.

From this list, it is apparent that cloud was unable to measure up to a centralized supercomputer system in many ways, but the delivery model does have its place. According to the report, “users with applications that have more dynamic or interactive needs could benefit from on-demand, self-service environments and rapid elasticity through the use of virtualization technology, and the MapReduce programming model to manage loosely coupled application runs.”

In other words, cloud excels when it comes to flexibility and responsiveness. In fact, the report found that “for users who need the added flexibility offered by the cloud computing model, additional costs may be more than offset by the increased flexibility. Furthermore, in some cases the potential for more immediate access to compute resources could directly translate into cost savings.”

However, when it comes to the potential cost savings of using a public cloud versus the costs of hardware acquisition, the report makes the point that DOE procurement costs are often significantly discounted, which offsets some of the potential savings:

Existing DOE centers already achieve many of the benefits of cloud computing since these centers consolidate computing across multiple program offices, deploy at large scales, and continuously refine and improve operational efficiency. Cost analysis shows that DOE centers are cost competitive, typically 3-7x less expensive, when compared to commercial cloud providers. Because the commercial sector constantly innovates, DOE labs and centers should continue to benchmark their computing cost against public clouds to ensure they are providing a competitive service.

“Cloud computing is ultimately a business model,” state the authors. “But cloud models often provide additional capabilities and flexibility that are helpful to certain workloads. DOE labs and centers should consider adopting and integrating these features of cloud computing into their operations in order to support more diverse workloads and further enable scientific discovery, without sacrificing the productivity and effectiveness of computing platforms that have been optimized for science over decades of development and refinement.”

The authors further suggest that when an integrated approach is not sufficient, a private cloud solution should be considered based on its ability to provide many of the benefits of commercial clouds while avoiding some of the pitfalls, such as security, data management, and performance penalties.

To recap: cloud services are a good complement to centralized computing resources, but not a replacement. This should not come as a surprise to our community. This is HPC, high-performance computing, and whenever you add additional layers, i.e., virtualization, the application takes a performance hit. However, as the report makes clear, there are good use cases for cloud services, such as “scientific groups needing support for on-demand access to resources, sudden surges in resource needs, customized environments, periodic predictable resource needs (e.g., monthly processing of genome data, nightly processing of telescope data), or unpredictable events such as computing for disaster recovery.” The report goes on to note that “cloud services essentially provide a differentiated service model that can cater to these diverse needs, allowing users to get a virtual private cluster with a certain guaranteed level of service.”

Magellan was billed as an exploratory project, set to go for two years. In fact, the project was named Magellan in honor of the Portuguese explorer Fernão de Magalhães, the first person to lead an expedition across the Pacific. The original “clouds of Magellan” refers to two small galaxies in the southern sky. The current-day Magellan, as the first major scientific cloud testbed, also navigated uncharted waters and documented the journey for the benefit of future generations.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

IBM Launches Commercial Quantum Network with Samsung, ORNL

December 14, 2017

In the race to commercialize quantum computing, IBM is one of several companies leading the pack. Today, IBM announced it had signed JPMorgan Chase, Daimler AG, Samsung and a number of other corporations to its IBM Q Net Read more…

By Tiffany Trader

TACC Researchers Test AI Traffic Monitoring Tool in Austin

December 13, 2017

Traffic jams and mishaps are often painful and sometimes dangerous facts of life. At this week’s IEEE International Conference on Big Data being held in Boston, researchers from TACC and colleagues will present a new Read more…

By HPCwire Staff

AMD Wins Another: Baidu to Deploy EPYC on Single Socket Servers

December 13, 2017

When AMD introduced its EPYC chip line in June, the company said a portion of the line was specifically designed to re-invigorate a single socket segment in what has become an overwhelmingly two-socket landscape in the d Read more…

By John Russell

HPE Extreme Performance Solutions

Explore the Origins of Space with COSMOS and Memory-Driven Computing

From the formation of black holes to the origins of space, data is the key to unlocking the secrets of the early universe. Read more…

Microsoft Wants to Speed Quantum Development

December 12, 2017

Quantum computing continues to make headlines in what remains of 2017 as several tech giants jockey to establish a pole position in the race toward commercialization of quantum. This week, Microsoft took the next step in Read more…

By Tiffany Trader

IBM Launches Commercial Quantum Network with Samsung, ORNL

December 14, 2017

In the race to commercialize quantum computing, IBM is one of several companies leading the pack. Today, IBM announced it had signed JPMorgan Chase, Daimler AG, Read more…

By Tiffany Trader

AMD Wins Another: Baidu to Deploy EPYC on Single Socket Servers

December 13, 2017

When AMD introduced its EPYC chip line in June, the company said a portion of the line was specifically designed to re-invigorate a single socket segment in wha Read more…

By John Russell

Microsoft Wants to Speed Quantum Development

December 12, 2017

Quantum computing continues to make headlines in what remains of 2017 as several tech giants jockey to establish a pole position in the race toward commercializ Read more…

By Tiffany Trader

HPC Iron, Soft, Data, People – It Takes an Ecosystem!

December 11, 2017

Cutting edge advanced computing hardware (aka big iron) does not stand by itself. These computers are the pinnacle of a myriad of technologies that must be care Read more…

By Alex R. Larzelere

IBM Begins Power9 Rollout with Backing from DOE, Google

December 6, 2017

After over a year of buildup, IBM is unveiling its first Power9 system based on the same architecture as the Department of Energy CORAL supercomputers, Summit a Read more…

By Tiffany Trader

Microsoft Spins Cycle Computing into Core Azure Product

December 5, 2017

Last August, cloud giant Microsoft acquired HPC cloud orchestration pioneer Cycle Computing. Since then the focus has been on integrating Cycle’s organization Read more…

By John Russell

GlobalFoundries, Ayar Labs Team Up to Commercialize Optical I/O

December 4, 2017

GlobalFoundries (GF) and Ayar Labs, a startup focused on using light, instead of electricity, to transfer data between chips, today announced they've entered in Read more…

By Tiffany Trader

HPE In-Memory Platform Comes to COSMOS

November 30, 2017

Hewlett Packard Enterprise is on a mission to accelerate space research. In August, it sent the first commercial-off-the-shelf HPC system into space for testing Read more…

By Tiffany Trader

US Coalesces Plans for First Exascale Supercomputer: Aurora in 2021

September 27, 2017

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, in Arlington, Va., yesterday (Sept. 26), it was revealed that the "Aurora" supercompute Read more…

By Tiffany Trader

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

AMD Showcases Growing Portfolio of EPYC and Radeon-based Systems at SC17

November 13, 2017

AMD’s charge back into HPC and the datacenter is on full display at SC17. Having launched the EPYC processor line in June along with its MI25 GPU the focus he Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Japan Unveils Quantum Neural Network

November 22, 2017

The U.S. and China are leading the race toward productive quantum computing, but it's early enough that ultimate leadership is still something of an open questi Read more…

By Tiffany Trader

GlobalFoundries Puts Wind in AMD’s Sails with 12nm FinFET

September 24, 2017

From its annual tech conference last week (Sept. 20), where GlobalFoundries welcomed more than 600 semiconductor professionals (reaching the Santa Clara venue Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Leading Solution Providers

IBM Begins Power9 Rollout with Backing from DOE, Google

December 6, 2017

After over a year of buildup, IBM is unveiling its first Power9 system based on the same architecture as the Department of Energy CORAL supercomputers, Summit a Read more…

By Tiffany Trader

Perspective: What Really Happened at SC17?

November 22, 2017

SC is over. Now comes the myriad of follow-ups. Inboxes are filled with templated emails from vendors and other exhibitors hoping to win a place in the post-SC thinking of booth visitors. Attendees of tutorials, workshops and other technical sessions will be inundated with requests for feedback. Read more…

By Andrew Jones

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Tensors Come of Age: Why the AI Revolution Will Help HPC

November 13, 2017

Thirty years ago, parallel computing was coming of age. A bitter battle began between stalwart vector computing supporters and advocates of various approaches to parallel computing. IBM skeptic Alan Karp, reacting to announcements of nCUBE’s 1024-microprocessor system and Thinking Machines’ 65,536-element array, made a public $100 wager that no one could get a parallel speedup of over 200 on real HPC workloads. Read more…

By John Gustafson & Lenore Mullin

Flipping the Flops and Reading the Top500 Tea Leaves

November 13, 2017

The 50th edition of the Top500 list, the biannual publication of the world’s fastest supercomputers based on public Linpack benchmarking results, was released Read more…

By Tiffany Trader

Intel Launches Software Tools to Ease FPGA Programming

September 5, 2017

Field Programmable Gate Arrays (FPGAs) have a reputation for being difficult to program, requiring expertise in specialty languages, like Verilog or VHDL. Easin Read more…

By Tiffany Trader

HPC Chips – A Veritable Smorgasbord?

October 10, 2017

For the first time since AMD's ill-fated launch of Bulldozer the answer to the question, 'Which CPU will be in my next HPC system?' doesn't have to be 'Whichever variety of Intel Xeon E5 they are selling when we procure'. Read more…

By Dairsie Latimer

  • arrow
  • Click Here for More Headlines
  • arrow
Share This