Cloud Lends Power to Next Generation Martian Missions

By Jose Luis Vazquez-Poletti

October 29, 2012

Martian rover Curiosity successfully performed an on-site validation of an application essential to the next generation of Martian missions. Cloud computing, specifically Amazon Web Services, played an important role.

An SUV on Mars

Launched from Cape Canaveral on November 26, Curiosity became the latest guest of the Martian surface when it made an epic touchdown – considering the complex landing procedure – on Gale Crater on August 6.

As a part of NASA’s Mars Science Laboratory mission (MSL), the rover’s goals include researching the climate and geology of the Red Planet. This mission represents another giant leap for humanity on its road to a manned expedition.

Last September 13, Curiosity put the breaks on its 1,980 lb mass and pointed its Mast Camera (MastCam) to the Sun in order to take a set of truly awesome pictures.

Curiosity self-portrait

Curiosity’s famous self-portrait

Dozens of Martian meteorological stations

Our story begins with the Mars MetNet Mission, which aims to go where no other Mars missions have gone before, at least in terms of the way it will gather and process data. This mission to Mars will be based on the power of a new type of dandelion seed-shaped, semi-hard landing vehicle called the MetNet Lander.

The main idea behind these vehicles is that by using a state-of-the-art inflatable entry and descent systems (instead of rigid heat shields and parachutes like those from the earlier semi-hard landing devices) the ratio of the payload mass to the overall mass is optimized. This means that more mass and volume resources are spared for the science payload and more probes (dozens) are to be deployed.

The scientific payload of the Mars MetNet Mission encompasses separate instrument packages for the Martian surface operation phase. At the Martian surface, the lander will take panoramic pictures and will also perform observations of pressure, temperature, humidity, magnetism, as well as atmospheric optical depth.

The network of MetNet landers will provide valuable scientific data, indispensable for studying the Martian atmosphere and its phenomena. Leading the project are the countries of Finland (Finnish Meteorological Institute – FMI), Russia (Lavochkin Space Association and Russian Space Institute) and Spain (Instituto Nacional de Técnica Aerospacial).

Mars MetNet

The Mars MetNet mission logo

The Mars MetNet collaboration also includes the Distributed Systems Architecture Research Group of the Universidad Complutense de Madrid. Led by Prof. Ignacio M. Llorente, the group is involved in using cloud computing to support scientific research. The MetNet effort is also dedicated to using cloud computing for boosting all possible applications pertaining to the Mars mission, as will be explained in greater detail.

The important role of a non-so-distant moon

One challenge of this mission is that the specified landing area is not known until two hours before touchdown. Working toward a solution, the Meiga-MetNet team developed an application for tracing Phobos (Mars’ biggest moon). The resulting Phobos cyclogram, as its known, describes the trajectory of the Martian moon using coordinates, dates and time intervals as an input [1]. The MetNet lander would achieve its exact location on the Martian surface by comparing the position of Phobos and the cyclogram, which is sent to the probe in advance of the landing procedure. This solution will spare not only time but also the cost of using a Martian orbiter for downlinking the calibration.

If you are complaining about the cost of your cell phone data plan, think about the cost of relying on other agencies’ communication satellites to transmit data in outer space.

We performed an initial parallelization of the application so that the complete set of coordinates pertaining to the approximated landing area can be processed with a desired grain. This process of profiling brought us to the conclusion that the needed hardware could be too expensive for executing this HPC application only twice a year. We had no way of even knowing if there would be other uses for this costly hardware either.

For this reason we turned to Amazon EC2, the de facto standard public cloud, attractive for its on-demand deployment and its “pay-as-you-go” basis. From all the possible setups that Amazon EC2 offered, we crafted and validated an execution model for the application, taking into account time, cost and a metric involving both. This way, the optimal infrastructure could be obtained given a problem size.

The setup we looked at would be akin to getting 37 nodes of the latest HP Proliant DL170 G6 Server (as of year 2010). With a retail price of $4,909 per node, purchasing such machines outright would cost $181,633 without considering any other expenses like shipping or insurance. And what about electricity? Administrator’s salary? Startup time? Even more, would we be able to use this infrastructure at full power in a 24×7 fashion? Probably not.

On the other hand and according to our model, Amazon EC2 provides the needed infrastructure for only $7.50!

results

Results of the execution model by means of performance, cost and a metric which interrelates both for different instance types offered by Amazon

And, by the way, we turned over a public cloud for Martian business just one year before NASA did. By December 2010, NASA was processing large satellite images for the ATLETHE vehicle (to be used in future manned missions) while we were presenting our first paper in June of the same year [2]. The first executions of our application took place exactly six months before.

Martian Computational Archeology

Data coming from the Mars MetNet probes is to be processed using different applications. The more probes that will be literally nailed onto the Martian surface, the more data will be generated.

Although this process is to be done in a regular basis, it won’t be continuous. For this reason, cloud computing on public infrastructures appears again as a valid solution.

Not having data from the Mars MetNet probes yet, the Finnish Meteorological Institute (FMI) started an interesting approach: use legacy data from NASA’s Mars Viking missions.

Legacy data, indeed! The first Viking was launched in 1975 and all the data and programs for sorting and analysis (including the processing environment) were optimized for a PRIME computer built in the late 70s.

Carl Sagan
The great astronomer Carl Sagan with a full-size model of a Viking lander

The FMI conducted a great effort for analyzing this data again in order to identify instrument failures and instrument calibration changes. When this analysis is ready the full Viking meteorological data set will be available for the scientific community for the first time.

The MEIGA-MetNet project developed the Phobos eclipse application in order to help explain certain uncommon readings. The final computing framework will also be used for processing fresh Martian data, so the Viking data represents a great tuning opportunity [3].

Thank you, Curiosity

Returning to September 13 (on the mission’s 37th sol), Curiosity made a stop in its path to take pictures of a beautiful Phobos eclipse. The objective was to confirm that our predictions of the eclipse were accurate.

These predictions, based on the execution of the application described before, have been just published by the Monthly Notices of the Royal Astronomical Society journal in a paper entitled “Opportunities to observe solar eclipses by Phobos with the Mars Science Laboratory” [4].

The precision achieved was within 1 second. This was also the first on-site validation of our application.

The Mars Science Laboratory represents the newest milestone in the exploration of humanity’s closest “final frontier.” Curiosity is not only bringing us incredible data from the Red Planet but it’s also paving the way for the next generation of Martian missions thanks to cloud computing, an essential tool for space exploration.

Scientific References

[1] P. Romero, G. Barderas, J.L. Vázquez-Poletti and I.M. Llorente: Chronogram to detect Phobos Eclipses on Mars with the MetNet Precursor Lander. Planetary and Space Science, vol. 59, n. 13, 2011, pp. 1542-1550.

[2] J.L. Vázquez-Poletti, G. Barderas, I.M. Llorente and P. Romero: A Model for Efficient Onboard Actualization of an Instrumental Cyclogram for the Mars MetNet Mission on a Public Cloud Infrastructure. PARA2010: State of the Art in Scientific and Parallel Computing, Reykjavík (Iceland), June 2010. Proceedings published in Lecture Notes in Computer Science (LNCS). Volume 7133, pp. 33-42, 2012. Springer Verlag.

[3] A.-M. Harri, W. Schmidt, P. Romero, L. Vazquez, G. Barderas, O. Kemppinen, C. Aguirre, J.L. Vazquez-Poletti, I.M. Llorente and H. Haukka: Phobos Eclipse Detection on Mars, Theory and Practice. Finnish Meteorological Institute Research Report 2012:2, Finland, 2012.

[4] G. Barderas, P. Romero, L. Vazquez, J.L. Vazquez-Poletti and I.M. Llorente: Opportunities to observe solar eclipses by Phobos with the Mars Science Laboratory. Monthly Notices of the Royal Astronomical Society, 2012, Volume 426, Number 4, pp. 3195–3200. Wiley.

About the Author

Dr. Jose Luis Vazquez-Poletti is Assistant Professor in Computer Architecture at Complutense University of Madrid (UCM, Spain), and a Cloud Computing Researcher at the Distributed Systems Architecture Research Group. He is (and has been) directly involved in EU funded projects, such as EGEE (grid computing) and 4CaaSt (PaaS Cloud), as well as many Spanish national initiatives.

From 2005 to 2009 his research focused in application porting onto grid computing infrastructures, activity that let him be “where the real action was.” These applications pertained to a wide range of areas, from fusion physics to bioinformatics. During this period he achieved the abilities needed for profiling applications and making them benefit of distributed computing infrastructures. Additionally, he shared these abilities in many training events organized within the EGEE Project and similar initiatives.

Since 2010 his research interests lie in different aspects of cloud computing, but always having real life applications in mind, especially those pertaining to the high Performance computing domain.

Website: http://dsa-research.org/jlvazquez/

Linkedin: http://www.linkedin.com/in/jlvazquezpoletti/

Jose Luis Vazquez-Poletti

The author with Curiosity’s soviet grandfather, Lunokhod (Луноход, moon walker in Russian), at the Russian Space Research Institute (IKI).

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire