Cloud Lends Power to Next Generation Martian Missions

By Jose Luis Vazquez-Poletti

October 29, 2012

Martian rover Curiosity successfully performed an on-site validation of an application essential to the next generation of Martian missions. Cloud computing, specifically Amazon Web Services, played an important role.

An SUV on Mars

Launched from Cape Canaveral on November 26, Curiosity became the latest guest of the Martian surface when it made an epic touchdown – considering the complex landing procedure – on Gale Crater on August 6.

As a part of NASA’s Mars Science Laboratory mission (MSL), the rover’s goals include researching the climate and geology of the Red Planet. This mission represents another giant leap for humanity on its road to a manned expedition.

Last September 13, Curiosity put the breaks on its 1,980 lb mass and pointed its Mast Camera (MastCam) to the Sun in order to take a set of truly awesome pictures.

Curiosity self-portrait

Curiosity’s famous self-portrait

Dozens of Martian meteorological stations

Our story begins with the Mars MetNet Mission, which aims to go where no other Mars missions have gone before, at least in terms of the way it will gather and process data. This mission to Mars will be based on the power of a new type of dandelion seed-shaped, semi-hard landing vehicle called the MetNet Lander.

The main idea behind these vehicles is that by using a state-of-the-art inflatable entry and descent systems (instead of rigid heat shields and parachutes like those from the earlier semi-hard landing devices) the ratio of the payload mass to the overall mass is optimized. This means that more mass and volume resources are spared for the science payload and more probes (dozens) are to be deployed.

The scientific payload of the Mars MetNet Mission encompasses separate instrument packages for the Martian surface operation phase. At the Martian surface, the lander will take panoramic pictures and will also perform observations of pressure, temperature, humidity, magnetism, as well as atmospheric optical depth.

The network of MetNet landers will provide valuable scientific data, indispensable for studying the Martian atmosphere and its phenomena. Leading the project are the countries of Finland (Finnish Meteorological Institute – FMI), Russia (Lavochkin Space Association and Russian Space Institute) and Spain (Instituto Nacional de Técnica Aerospacial).

Mars MetNet

The Mars MetNet mission logo

The Mars MetNet collaboration also includes the Distributed Systems Architecture Research Group of the Universidad Complutense de Madrid. Led by Prof. Ignacio M. Llorente, the group is involved in using cloud computing to support scientific research. The MetNet effort is also dedicated to using cloud computing for boosting all possible applications pertaining to the Mars mission, as will be explained in greater detail.

The important role of a non-so-distant moon

One challenge of this mission is that the specified landing area is not known until two hours before touchdown. Working toward a solution, the Meiga-MetNet team developed an application for tracing Phobos (Mars’ biggest moon). The resulting Phobos cyclogram, as its known, describes the trajectory of the Martian moon using coordinates, dates and time intervals as an input [1]. The MetNet lander would achieve its exact location on the Martian surface by comparing the position of Phobos and the cyclogram, which is sent to the probe in advance of the landing procedure. This solution will spare not only time but also the cost of using a Martian orbiter for downlinking the calibration.

If you are complaining about the cost of your cell phone data plan, think about the cost of relying on other agencies’ communication satellites to transmit data in outer space.

We performed an initial parallelization of the application so that the complete set of coordinates pertaining to the approximated landing area can be processed with a desired grain. This process of profiling brought us to the conclusion that the needed hardware could be too expensive for executing this HPC application only twice a year. We had no way of even knowing if there would be other uses for this costly hardware either.

For this reason we turned to Amazon EC2, the de facto standard public cloud, attractive for its on-demand deployment and its “pay-as-you-go” basis. From all the possible setups that Amazon EC2 offered, we crafted and validated an execution model for the application, taking into account time, cost and a metric involving both. This way, the optimal infrastructure could be obtained given a problem size.

The setup we looked at would be akin to getting 37 nodes of the latest HP Proliant DL170 G6 Server (as of year 2010). With a retail price of $4,909 per node, purchasing such machines outright would cost $181,633 without considering any other expenses like shipping or insurance. And what about electricity? Administrator’s salary? Startup time? Even more, would we be able to use this infrastructure at full power in a 24×7 fashion? Probably not.

On the other hand and according to our model, Amazon EC2 provides the needed infrastructure for only $7.50!

results

Results of the execution model by means of performance, cost and a metric which interrelates both for different instance types offered by Amazon

And, by the way, we turned over a public cloud for Martian business just one year before NASA did. By December 2010, NASA was processing large satellite images for the ATLETHE vehicle (to be used in future manned missions) while we were presenting our first paper in June of the same year [2]. The first executions of our application took place exactly six months before.

Martian Computational Archeology

Data coming from the Mars MetNet probes is to be processed using different applications. The more probes that will be literally nailed onto the Martian surface, the more data will be generated.

Although this process is to be done in a regular basis, it won’t be continuous. For this reason, cloud computing on public infrastructures appears again as a valid solution.

Not having data from the Mars MetNet probes yet, the Finnish Meteorological Institute (FMI) started an interesting approach: use legacy data from NASA’s Mars Viking missions.

Legacy data, indeed! The first Viking was launched in 1975 and all the data and programs for sorting and analysis (including the processing environment) were optimized for a PRIME computer built in the late 70s.

Carl Sagan
The great astronomer Carl Sagan with a full-size model of a Viking lander

The FMI conducted a great effort for analyzing this data again in order to identify instrument failures and instrument calibration changes. When this analysis is ready the full Viking meteorological data set will be available for the scientific community for the first time.

The MEIGA-MetNet project developed the Phobos eclipse application in order to help explain certain uncommon readings. The final computing framework will also be used for processing fresh Martian data, so the Viking data represents a great tuning opportunity [3].

Thank you, Curiosity

Returning to September 13 (on the mission’s 37th sol), Curiosity made a stop in its path to take pictures of a beautiful Phobos eclipse. The objective was to confirm that our predictions of the eclipse were accurate.

These predictions, based on the execution of the application described before, have been just published by the Monthly Notices of the Royal Astronomical Society journal in a paper entitled “Opportunities to observe solar eclipses by Phobos with the Mars Science Laboratory” [4].

The precision achieved was within 1 second. This was also the first on-site validation of our application.

The Mars Science Laboratory represents the newest milestone in the exploration of humanity’s closest “final frontier.” Curiosity is not only bringing us incredible data from the Red Planet but it’s also paving the way for the next generation of Martian missions thanks to cloud computing, an essential tool for space exploration.

Scientific References

[1] P. Romero, G. Barderas, J.L. Vázquez-Poletti and I.M. Llorente: Chronogram to detect Phobos Eclipses on Mars with the MetNet Precursor Lander. Planetary and Space Science, vol. 59, n. 13, 2011, pp. 1542-1550.

[2] J.L. Vázquez-Poletti, G. Barderas, I.M. Llorente and P. Romero: A Model for Efficient Onboard Actualization of an Instrumental Cyclogram for the Mars MetNet Mission on a Public Cloud Infrastructure. PARA2010: State of the Art in Scientific and Parallel Computing, Reykjavík (Iceland), June 2010. Proceedings published in Lecture Notes in Computer Science (LNCS). Volume 7133, pp. 33-42, 2012. Springer Verlag.

[3] A.-M. Harri, W. Schmidt, P. Romero, L. Vazquez, G. Barderas, O. Kemppinen, C. Aguirre, J.L. Vazquez-Poletti, I.M. Llorente and H. Haukka: Phobos Eclipse Detection on Mars, Theory and Practice. Finnish Meteorological Institute Research Report 2012:2, Finland, 2012.

[4] G. Barderas, P. Romero, L. Vazquez, J.L. Vazquez-Poletti and I.M. Llorente: Opportunities to observe solar eclipses by Phobos with the Mars Science Laboratory. Monthly Notices of the Royal Astronomical Society, 2012, Volume 426, Number 4, pp. 3195–3200. Wiley.

About the Author

Dr. Jose Luis Vazquez-Poletti is Assistant Professor in Computer Architecture at Complutense University of Madrid (UCM, Spain), and a Cloud Computing Researcher at the Distributed Systems Architecture Research Group. He is (and has been) directly involved in EU funded projects, such as EGEE (grid computing) and 4CaaSt (PaaS Cloud), as well as many Spanish national initiatives.

From 2005 to 2009 his research focused in application porting onto grid computing infrastructures, activity that let him be “where the real action was.” These applications pertained to a wide range of areas, from fusion physics to bioinformatics. During this period he achieved the abilities needed for profiling applications and making them benefit of distributed computing infrastructures. Additionally, he shared these abilities in many training events organized within the EGEE Project and similar initiatives.

Since 2010 his research interests lie in different aspects of cloud computing, but always having real life applications in mind, especially those pertaining to the high Performance computing domain.

Website: http://dsa-research.org/jlvazquez/

Linkedin: http://www.linkedin.com/in/jlvazquezpoletti/

Jose Luis Vazquez-Poletti

The author with Curiosity’s soviet grandfather, Lunokhod (Луноход, moon walker in Russian), at the Russian Space Research Institute (IKI).

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Which Schools Produce the Top Coders in the World?

December 8, 2016

Ever wonder which universities worldwide produce the best coders? The answers may surprise you, at least as judged by the results of a competition posted yesterday on the HackerRank blog. Read more…

By John Russell

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. The pilots, supported in part by DOE exascale funding, not only seek to do good by advancing cancer research and therapy but also to advance deep learning capabilities and infrastructure with an eye towards eventual use on exascale machines. Read more…

By John Russell

DDN Enables 50TB/Day Trans-Pacific Data Transfer for Yahoo Japan

December 6, 2016

Transferring data from one data center to another in search of lower regional energy costs isn’t a new concept, but Yahoo Japan is putting the idea into transcontinental effect with a system that transfers 50TB of data a day from Japan to the U.S., where electricity costs a quarter of the rates in Japan. Read more…

By Doug Black

Infographic Highlights Career of Admiral Grace Murray Hopper

December 5, 2016

Dr. Grace Murray Hopper (December 9, 1906 – January 1, 1992) was an early pioneer of computer science and one of the most famous women achievers in a field dominated by men. Read more…

By Staff

Ganthier, Turkel on the Dell EMC Road Ahead

December 5, 2016

Who is Dell EMC and why should you care? Glad you asked is Jim Ganthier’s quick response. Ganthier is SVP for validated solutions and high performance computing for the new (even bigger) technology giant Dell EMC following Dell’s acquisition of EMC in September. In this case, says Ganthier, the blending of the two companies is a 1+1 = 5 proposition. Not bad math if you can pull it off. Read more…

By John Russell

AWS Embraces FPGAs, ‘Elastic’ GPUs

December 2, 2016

A new instance type rolled out this week by Amazon Web Services is based on customizable field programmable gate arrays that promise to strike a balance between performance and cost as emerging workloads create requirements often unmet by general-purpose processors. Read more…

By George Leopold

AWS Launches Massive 100 Petabyte ‘Sneakernet’

December 1, 2016

Amazon Web Services now offers a way to move data into its cloud by the truckload. Read more…

By Tiffany Trader

Weekly Twitter Roundup (Dec. 1, 2016)

December 1, 2016

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. The pilots, supported in part by DOE exascale funding, not only seek to do good by advancing cancer research and therapy but also to advance deep learning capabilities and infrastructure with an eye towards eventual use on exascale machines. Read more…

By John Russell

Ganthier, Turkel on the Dell EMC Road Ahead

December 5, 2016

Who is Dell EMC and why should you care? Glad you asked is Jim Ganthier’s quick response. Ganthier is SVP for validated solutions and high performance computing for the new (even bigger) technology giant Dell EMC following Dell’s acquisition of EMC in September. In this case, says Ganthier, the blending of the two companies is a 1+1 = 5 proposition. Not bad math if you can pull it off. Read more…

By John Russell

AWS Launches Massive 100 Petabyte ‘Sneakernet’

December 1, 2016

Amazon Web Services now offers a way to move data into its cloud by the truckload. Read more…

By Tiffany Trader

Lighting up Aurora: Behind the Scenes at the Creation of the DOE’s Upcoming 200 Petaflops Supercomputer

December 1, 2016

In April 2015, U.S. Department of Energy Undersecretary Franklin Orr announced that Intel would be the prime contractor for Aurora: Read more…

By Jan Rowell

Seagate-led SAGE Project Delivers Update on Exascale Goals

November 29, 2016

Roughly a year and a half after its launch, the SAGE exascale storage project led by Seagate has delivered a substantive interim report – Data Storage for Extreme Scale. Read more…

By John Russell

Nvidia Sees Bright Future for AI Supercomputing

November 23, 2016

Graphics chipmaker Nvidia made a strong showing at SC16 in Salt Lake City last week. Read more…

By Tiffany Trader

HPE-SGI to Tackle Exascale and Enterprise Targets

November 22, 2016

At first blush, and maybe second blush too, Hewlett Packard Enterprise’s (HPE) purchase of SGI seems like an unambiguous win-win. SGI’s advanced shared memory technology, its popular UV product line (Hanna), deep vertical market expertise, and services-led go-to-market capability all give HPE a leg up in its drive to remake itself. Bear in mind HPE came into existence just a year ago with the split of Hewlett-Packard. The computer landscape, including HPC, is shifting with still unclear consequences. One wonders who’s next on the deal block following Dell’s recent merger with EMC. Read more…

By John Russell

Intel Details AI Hardware Strategy for Post-GPU Age

November 21, 2016

Last week at SC16, Intel revealed its product roadmap for embedding its processors with key capabilities and attributes needed to take artificial intelligence (AI) to the next level. Read more…

By Alex Woodie

Why 2016 Is the Most Important Year in HPC in Over Two Decades

August 23, 2016

In 1994, two NASA employees connected 16 commodity workstations together using a standard Ethernet LAN and installed open-source message passing software that allowed their number-crunching scientific application to run on the whole “cluster” of machines as if it were a single entity. Read more…

By Vincent Natoli, Stone Ridge Technology

IBM Advances Against x86 with Power9

August 30, 2016

After offering OpenPower Summit attendees a limited preview in April, IBM is unveiling further details of its next-gen CPU, Power9, which the tech mainstay is counting on to regain market share ceded to rival Intel. Read more…

By Tiffany Trader

AWS Beats Azure to K80 General Availability

September 30, 2016

Amazon Web Services has seeded its cloud with Nvidia Tesla K80 GPUs to meet the growing demand for accelerated computing across an increasingly-diverse range of workloads. The P2 instance family is a welcome addition for compute- and data-focused users who were growing frustrated with the performance limitations of Amazon's G2 instances, which are backed by three-year-old Nvidia GRID K520 graphics cards. Read more…

By Tiffany Trader

Think Fast – Is Neuromorphic Computing Set to Leap Forward?

August 15, 2016

Steadily advancing neuromorphic computing technology has created high expectations for this fundamentally different approach to computing. Read more…

By John Russell

The Exascale Computing Project Awards $39.8M to 22 Projects

September 7, 2016

The Department of Energy’s Exascale Computing Project (ECP) hit an important milestone today with the announcement of its first round of funding, moving the nation closer to its goal of reaching capable exascale computing by 2023. Read more…

By Tiffany Trader

HPE Gobbles SGI for Larger Slice of $11B HPC Pie

August 11, 2016

Hewlett Packard Enterprise (HPE) announced today that it will acquire rival HPC server maker SGI for $7.75 per share, or about $275 million, inclusive of cash and debt. The deal ends the seven-year reprieve that kept the SGI banner flying after Rackable Systems purchased the bankrupt Silicon Graphics Inc. for $25 million in 2009 and assumed the SGI brand. Bringing SGI into its fold bolsters HPE's high-performance computing and data analytics capabilities and expands its position... Read more…

By Tiffany Trader

ARM Unveils Scalable Vector Extension for HPC at Hot Chips

August 22, 2016

ARM and Fujitsu today announced a scalable vector extension (SVE) to the ARMv8-A architecture intended to enhance ARM capabilities in HPC workloads. Fujitsu is the lead silicon partner in the effort (so far) and will use ARM with SVE technology in its post K computer, Japan’s next flagship supercomputer planned for the 2020 timeframe. This is an important incremental step for ARM, which seeks to push more aggressively into mainstream and HPC server markets. Read more…

By John Russell

IBM Debuts Power8 Chip with NVLink and Three New Systems

September 8, 2016

Not long after revealing more details about its next-gen Power9 chip due in 2017, IBM today rolled out three new Power8-based Linux servers and a new version of its Power8 chip featuring Nvidia’s NVLink interconnect. Read more…

By John Russell

Leading Solution Providers

Vectors: How the Old Became New Again in Supercomputing

September 26, 2016

Vector instructions, once a powerful performance innovation of supercomputing in the 1970s and 1980s became an obsolete technology in the 1990s. But like the mythical phoenix bird, vector instructions have arisen from the ashes. Here is the history of a technology that went from new to old then back to new. Read more…

By Lynd Stringer

US, China Vie for Supercomputing Supremacy

November 14, 2016

The 48th edition of the TOP500 list is fresh off the presses and while there is no new number one system, as previously teased by China, there are a number of notable entrants from the US and around the world and significant trends to report on. Read more…

By Tiffany Trader

Intel Launches Silicon Photonics Chip, Previews Next-Gen Phi for AI

August 18, 2016

At the Intel Developer Forum, held in San Francisco this week, Intel Senior Vice President and General Manager Diane Bryant announced the launch of Intel's Silicon Photonics product line and teased a brand-new Phi product, codenamed "Knights Mill," aimed at machine learning workloads. Read more…

By Tiffany Trader

CPU Benchmarking: Haswell Versus POWER8

June 2, 2015

With OpenPOWER activity ramping up and IBM’s prominent role in the upcoming DOE machines Summit and Sierra, it’s a good time to look at how the IBM POWER CPU stacks up against the x86 Xeon Haswell CPU from Intel. Read more…

By Tiffany Trader

Beyond von Neumann, Neuromorphic Computing Steadily Advances

March 21, 2016

Neuromorphic computing – brain inspired computing – has long been a tantalizing goal. The human brain does with around 20 watts what supercomputers do with megawatts. And power consumption isn’t the only difference. Fundamentally, brains ‘think differently’ than the von Neumann architecture-based computers. While neuromorphic computing progress has been intriguing, it has still not proven very practical. Read more…

By John Russell

Dell EMC Engineers Strategy to Democratize HPC

September 29, 2016

The freshly minted Dell EMC division of Dell Technologies is on a mission to take HPC mainstream with a strategy that hinges on engineered solutions, beginning with a focus on three industry verticals: manufacturing, research and life sciences. "Unlike traditional HPC where everybody bought parts, assembled parts and ran the workloads and did iterative engineering, we want folks to focus on time to innovation and let us worry about the infrastructure," said Jim Ganthier, senior vice president, validated solutions organization at Dell EMC Converged Platforms Solution Division. Read more…

By Tiffany Trader

Container App ‘Singularity’ Eases Scientific Computing

October 20, 2016

HPC container platform Singularity is just six months out from its 1.0 release but already is making inroads across the HPC research landscape. It's in use at Lawrence Berkeley National Laboratory (LBNL), where Singularity founder Gregory Kurtzer has worked in the High Performance Computing Services (HPCS) group for 16 years. Read more…

By Tiffany Trader

Micron, Intel Prepare to Launch 3D XPoint Memory

August 16, 2016

Micron Technology used last week’s Flash Memory Summit to roll out its new line of 3D XPoint memory technology jointly developed with Intel while demonstrating the technology in solid-state drives. Micron claimed its Quantx line delivers PCI Express (PCIe) SSD performance with read latencies at less than 10 microseconds and writes at less than 20 microseconds. Read more…

By George Leopold

  • arrow
  • Click Here for More Headlines
  • arrow
Share This