A Big Data Journey While Seeking to Catalog our Universe

By James Reinders

January 16, 2019

It turns out, astronomers have lots of photos of the sky but seek knowledge about what the photos mean. Sound familiar? Big data problems are often characterized as transforming data into insights – which is exactly what some ambitious scientists are working to do with “Sky Survey” data. A Sky Survey is essentially astronomer speak for “lots and lots of images taken by telescopes, along with information of when and where they were taken.”

The Celeste collaboration is a group of scientists who have worked to catalog the visible universe in a way never before accomplished. They seek to create and refine a catalog which can detail the placement and characteristics (such as brightness and rotation) of every visible object in the sky.

Along the way, the Celeste collaboration has already proven that one high productive language (Julia) can offer high performance “at scale” (using hundreds of thousands of processor cores for compute), and their success certainly indicates that we will see more “at scale” big data work.

Journey of the Photons

No amount of effort to design an amazing telescope can overcome the effects that a very long journey has had upon the photons. Putting a telescope into orbit might cut out the last few hundred miles through our atmosphere, but that is just the tip of the iceberg when it comes to figuring out what each photo means. The techniques being developed by the Celeste collaboration are applicable to data regardless of whether it is earth-based or space-based.  So far, the earth-based data has supplied plenty of work to do.

Aside from inherent limitations of any sensing device in a telescope, the final image we get from a telescope is imperfect on account of point spread from the atmosphere, diffraction spikes from the telescope, and gravitational lensing that has occurred along the journey, among other causes. The Celeste collaboration has plugged away at addressing such challenges in their quest to build their meaningful catalog. As I have learned more about all they have done, I have been both amazed with the magnitude of their accomplishments and in awe of the enormous scope of future work that is possible. A truly big data project, Celeste has an insatiable appetite for more data, and for more sophisticated analysis work.

Lots of Compute, and Lots of (High Productivity) Programming

Collecting all known data about the visible universe into a meaningful model certainly is a big data problem. Celeste collaborators’ computational work has landed in the petascale world, meaning they have performed computations at a rate exceeding a thousand million million (1015) floating-point operations per second. They did this with over nine thousand CPUs, a high productivity language called Julia, and a 178 terabyte dataset representing 188 million stars and galaxies. Processing also involved intensive I/O due to the multiple passes over the dataset processed during a 14.6-minute run on the Cori supercomputer.

They did not use FORTRAN or C++ as the language for this task. Instead, they choose a high productivity language out of MIT known as Julia, and used it to very efficiently utilize Intel processors at a petascale. Specifically, they used 1.3 million threads on 9,300 Intel Xeon Phi processors (650,000 cores) to achieve 1.54 petaflops peak performance. This was the first showing of Julia at petascale, and it certainly will not be the last.

The Julia programming language developers explain Julia by saying: “Julia excels at numerical computing. Julia was designed from the beginning for high performance. Its syntax is great for math, many numeric datatypes are supported, and parallelism is available out of the box. Julia’s multiple dispatch is a natural fit for defining number and array-like datatypes.”

Keys to High-Performance Julia

The developers of the Celeste code have a few Julia-specific tips for making sure Julia is competitive with other compiled languages for high performance. Their tips were:

  1. Follow the performance tips given with Julia (no global state/eval/etc. in hotspots).
  2. Type stability (dynamic re-typing might seem cool, but it kills performance).
  3. Minimize dynamic memory allocations; use memory profiles to find allocations to reduce (double benefit: less time allocating also means less time doing garbage collection).

The final tip may be especially important in languages with garbage collection, but it is a great suggestion for programmers in all languages. Similarly, avoiding global state (the first tip) has enormous merit outside Julia as well.

Finally, the developers stress the need to profile to find and optimize hotspots. Hardly a Julia specific tip!  All in all, the experience of the developers with Julia mostly resembled the experience of any HPC programmer using C, C++, and Fortran. They would say that Julia offers a more productive programming environment, but also offers performance you would not find with other high productive languages such as Python. Despite some solid accelerated Python capabilities that are out there, no Python application has shown anything close to petaflops performance.

It seems that making Julia scale to petaflops performance involves the same thinking as effective parallel programming in any high-performance language.

The Data: SDSS

Irénée du Pont Telescope at Las Campanas Observatory. (credit: Krzysztof Ulaczyk, CC BY-SA 4.0)

In 1998, the Apache Point Observatory in New Mexico began imaging every visible object from over 35 percent of the sky in a project known as the Sloan Digital Sky Survey. Today, data is also collected from the Irénée du Pont Telescope at Las Campanas Observatory in Chile (APOGEE-2S). The Sloan Digital Sky Survey (SDSS) has been one of the most successful surveys in the history of astronomy. After a decade of design and construction, the SDSS began regular survey operations in 2000. It has progressed through several phases, SDSS-I (2000-2005), SDSS-II (2005-2008), SDSS-III (2008-2014), and SDSS-IV (2014+). Each phase has involved multiple surveys with interlocking science goals. This project proudly shares that they have already created the most detailed three-dimensional maps of the Universe ever made, with deep multi-color images of one third of the sky, and spectra for more than three million astronomical objects. The project has released fourteen data versions of their datasets thus far. They continue to release new data sets annually. The dataset scheduled for the end of this year will include spectral data across the face of the nearest ten thousand galaxies, instead of the previous surveys which obtained spectra only at the centers of target galaxies. The SDSS team calls this work “Mapping Nearby Galaxies at APO (MaNGA).” The dataset in 2019 will include information from the Apache Point Observatory Galaxy Evolution Experiment (APOGEE-2) to observe the “archaeological” record embedded in hundreds of thousands of stars to explore the assembly history and evolution of the Milky Way. You could say that the details as to how the Galaxy evolved are preserved today in the motions and chemical compositions of its stars.

It’s not hard to image that these ever-expanding datasets will offer even more opportunities for the Celeste collaboration in their analysis work.

Version 1.0

Prior work focused on non-statistical models. The Celeste collaboration focused on a statistical model, a fully generative model to be precise. Over the course of their first three years, the Celeste collaboration developed a new parallel computing method that was used to process the dataset (about 178 terabytes) and produce the most accurate catalog of 188 million astronomical objects in just 14.6 minutes with state-of-the-art point and uncertainty estimates.

In addition to creating a catalog, an important objective of this work was to identify promising galaxies for spectrograph targeting with the hope of better understanding dark energy and the geometry of the universe.

A key design objective of Celeste is to help be an extensible model and inference procedure for use by the astronomical community. This will allow more computation to be applied selectively if deeper understanding of any particular object is desired (e.g., brightness, rotation). Other applications might include finding supernovas or detecting near-Earth asteroids. The teams see enormous potential in the framework they have built. An hour-long presentation offers many more details of the work of Celeste 1.0 and is available for viewing online.

To help grasp the processing being done, here is a sample (using a synthetic image) of processing being done by an early prototype for Celeste 2.0. The synthetic image (the “input” to an autoencoder) is first, then the recon_mean is the mean of the approximation we find to the “output” of an autoencoder. The fact that it appears the same as the input is exactly what is desired! In Celeste 2.0, the recon_mean is formed by summing the four images to the right – which are the “deblended” images. These four images are hopefully useful to astronomers.

Envisioning Version 2.0

They first reported their petascale results last year, and they’ve been busy since then envisioning and developing “Celeste 2.0.” The collaboration is focused on moving to a more sophisticated inference model to replace the purely graphical model approach of Celeste 1.0, which was quite successful in its own right using only conventional variable inference. A key objective of this work is not only more accurate placement and features, but also more accurate uncertainties (“error bars”) for these as well.

Celeste 2.0 utilizes an autoencoder (variable) with a recurrent neural network (RNN), that also employs bayesian inference, and adds a gravitational lensing capability. The Bayesian inference technique is commonly associated with big data and machine learning projects, and typically  gets sharper predictions from data than other techniques. Bayesian inference effectively aims to inject some common sense (bias based on additional knowledge) into an otherwise sterile statistical analysis. In the case of Celeste 2.0, the newer techniques capture meaning from the vast dataset more accurately.

Bayesian models are composable, meaning that they work well as add-ons. This enables work on using Bayesian models to create a new gravitational lensing capability to undo the distortions which have occurred by the time it reaches a telescope. This is an area of active development, which promises to further refine the catalog of visible objects.

Endless Possibilities

Of course, I’m guessing work will not end with Celeste 2.0. They’ve opened up the challenge of building a catalog of the universe, and like all big data problems it has an insatiable appetite for more data. The continually growing sources of data in the SDSS offers many opportunities for the analysis work of the Celeste collaboration[1]. One day, perhaps gravitational wave data from the newest source of astronomy data can be incorporated? By then, we might also be able to offer them a data feed from a telescope sitting on Mars. It will happen.

In the meantime, the Celeste collaboration continues to make excellent use of the Intel processors in the Cori supercomputer with the Julia language. And this provides a wealth of encouragement for all big data projects looking to scale.

[1] The key contributors to the Celeste collaboration have been: Jeffrey Regier, Bryan Liu and Jon McAuliffeat of UC Berkeley ; Andy Miller and Ryan Adams of Harvard; David Schlegel of LBL Physics; and Prabhat of NERSC.

James Reinders is an HPC enthusiast and author of eight books with more than 30 years of industry experience, including 27 years at Intel Corporation (retired June 2016).

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

University of Chicago Researchers Generate First Computational Model of Entire SARS-CoV-2 Virus

January 15, 2021

Over the course of the last year, many detailed computational models of SARS-CoV-2 have been produced with the help of supercomputers, but those models have largely focused on critical elements of the virus, such as its Read more…

By Oliver Peckham

Pat Gelsinger Returns to Intel as CEO

January 14, 2021

The Intel board of directors has appointed a new CEO. Intel alum Pat Gelsinger is leaving his post as CEO of VMware to rejoin the company that he parted ways with 11 years ago. Gelsinger will succeed Bob Swan, who will remain CEO until Feb. 15. Gelsinger previously spent 30 years... Read more…

By Tiffany Trader

Roar Supercomputer to Support Naval Aircraft Research

January 14, 2021

One might not think “aircraft” when picturing the U.S. Navy, but the military branch actually has thousands of aircraft currently in service – and now, supercomputing will help future naval aircraft operate faster, Read more…

By Staff report

DOE and NOAA Extend Computing Partnership, Plan for New Supercomputer

January 14, 2021

The National Climate-Computing Research Center (NCRC), hosted by Oak Ridge National Laboratory (ORNL), has been supporting the climate research of the National Oceanic and Atmospheric Administration (NOAA) for the last 1 Read more…

By Oliver Peckham

Using Micro-Combs, Researchers Demonstrate World’s Fastest Optical Neuromorphic Processor for AI

January 13, 2021

Neuromorphic computing, which uses chips that mimic the behavior of the human brain using virtual “neurons,” is growing in popularity thanks to high-profile efforts from Intel and others. Now, a team of researchers l Read more…

By Oliver Peckham

AWS Solution Channel

Now Available – Amazon EC2 C6gn Instances with 100 Gbps Networking

Amazon EC2 C6gn instances powered by AWS Graviton2 processors are now available!

Compared to C6g instances, this new instance type provides 4x higher network bandwidth, 4x higher packet processing performance, and 2x higher EBS bandwidth. Read more…

Intel® HPC + AI Pavilion

Intel Keynote Address

Intel is the foundation of HPC – from the workstation to the cloud to the backbone of the Top500. At SC20, Intel’s Trish Damkroger, VP and GM of high performance computing, addresses the audience to show how Intel and its partners are building the future of HPC today, through hardware and software technologies that accelerate the broad deployment of advanced HPC systems. Read more…

Honing In on AI, US Launches National Artificial Intelligence Initiative Office

January 13, 2021

To drive American leadership in the field of AI into the future, the National Artificial Intelligence Initiative Office has been launched by the White House Office of Science and Technology Policy (OSTP). The new agen Read more…

By Todd R. Weiss

Pat Gelsinger Returns to Intel as CEO

January 14, 2021

The Intel board of directors has appointed a new CEO. Intel alum Pat Gelsinger is leaving his post as CEO of VMware to rejoin the company that he parted ways with 11 years ago. Gelsinger will succeed Bob Swan, who will remain CEO until Feb. 15. Gelsinger previously spent 30 years... Read more…

By Tiffany Trader

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

By John Russell

Intel ‘Ice Lake’ Server Chips in Production, Set for Volume Ramp This Quarter

January 12, 2021

Intel Corp. used this week’s virtual CES 2021 event to reassert its dominance of the datacenter with the formal roll out of its next-generation server chip, the 10nm Xeon Scalable processor that targets AI and HPC workloads. The third-generation “Ice Lake” family... Read more…

By George Leopold

Researchers Say It Won’t Be Possible to Control Superintelligent AI

January 11, 2021

Worries about out-of-control AI aren’t new. Many prominent figures have suggested caution when unleashing AI. One quote that keeps cropping up is (roughly) th Read more…

By John Russell

AMD Files Patent on New GPU Chiplet Approach

January 5, 2021

Advanced Micro Devices is accelerating the GPU chiplet race with the release of a U.S. patent application for a device that incorporates high-bandwidth intercon Read more…

By George Leopold

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

By Tracey Bryant

Intel Touts Optane Performance, Teases Next-gen “Crow Pass”

January 5, 2021

Competition to leverage new memory and storage hardware with new or improved software to create better storage/memory schemes has steadily gathered steam during Read more…

By John Russell

Farewell 2020: Bleak, Yes. But a Lot of Good Happened Too

December 30, 2020

Here on the cusp of the new year, the catchphrase ‘2020 hindsight’ has a distinctly different feel. Good riddance, yes. But also proof of science’s power Read more…

By John Russell

Esperanto Unveils ML Chip with Nearly 1,100 RISC-V Cores

December 8, 2020

At the RISC-V Summit today, Art Swift, CEO of Esperanto Technologies, announced a new, RISC-V based chip aimed at machine learning and containing nearly 1,100 low-power cores based on the open-source RISC-V architecture. Esperanto Technologies, headquartered in... Read more…

By Oliver Peckham

Azure Scaled to Record 86,400 Cores for Molecular Dynamics

November 20, 2020

A new record for HPC scaling on the public cloud has been achieved on Microsoft Azure. Led by Dr. Jer-Ming Chia, the cloud provider partnered with the Beckman I Read more…

By Oliver Peckham

NICS Unleashes ‘Kraken’ Supercomputer

April 4, 2008

A Cray XT4 supercomputer, dubbed Kraken, is scheduled to come online in mid-summer at the National Institute for Computational Sciences (NICS). The soon-to-be petascale system, and the resulting NICS organization, are the result of an NSF Track II award of $65 million to the University of Tennessee and its partners to provide next-generation supercomputing for the nation's science community. Read more…

Is the Nvidia A100 GPU Performance Worth a Hardware Upgrade?

October 16, 2020

Over the last decade, accelerators have seen an increasing rate of adoption in high-performance computing (HPC) platforms, and in the June 2020 Top500 list, eig Read more…

By Hartwig Anzt, Ahmad Abdelfattah and Jack Dongarra

Aurora’s Troubles Move Frontier into Pole Exascale Position

October 1, 2020

Intel’s 7nm node delay has raised questions about the status of the Aurora supercomputer that was scheduled to be stood up at Argonne National Laboratory next year. Aurora was in the running to be the United States’ first exascale supercomputer although it was on a contemporaneous timeline with... Read more…

By Tiffany Trader

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

By John Russell

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

By Tracey Bryant

Leading Solution Providers

Contributors

Top500: Fugaku Keeps Crown, Nvidia’s Selene Climbs to #5

November 16, 2020

With the publication of the 56th Top500 list today from SC20's virtual proceedings, Japan's Fugaku supercomputer – now fully deployed – notches another win, Read more…

By Tiffany Trader

Texas A&M Announces Flagship ‘Grace’ Supercomputer

November 9, 2020

Texas A&M University has announced its next flagship system: Grace. The new supercomputer, named for legendary programming pioneer Grace Hopper, is replacing the Ada system (itself named for mathematician Ada Lovelace) as the primary workhorse for Texas A&M’s High Performance Research Computing (HPRC). Read more…

By Oliver Peckham

At Oak Ridge, ‘End of Life’ Sometimes Isn’t

October 31, 2020

Sometimes, the old dog actually does go live on a farm. HPC systems are often cursed with short lifespans, as they are continually supplanted by the latest and Read more…

By Oliver Peckham

Nvidia and EuroHPC Team for Four Supercomputers, Including Massive ‘Leonardo’ System

October 15, 2020

The EuroHPC Joint Undertaking (JU) serves as Europe’s concerted supercomputing play, currently comprising 32 member states and billions of euros in funding. I Read more…

By Oliver Peckham

Gordon Bell Special Prize Goes to Massive SARS-CoV-2 Simulations

November 19, 2020

2020 has proven a harrowing year – but it has produced remarkable heroes. To that end, this year, the Association for Computing Machinery (ACM) introduced the Read more…

By Oliver Peckham

Nvidia-Arm Deal a Boon for RISC-V?

October 26, 2020

The $40 billion blockbuster acquisition deal that will bring chipmaker Arm into the Nvidia corporate family could provide a boost for the competing RISC-V architecture. As regulators in the U.S., China and the European Union begin scrutinizing the impact of the blockbuster deal on semiconductor industry competition and innovation, the deal has at the very least... Read more…

By George Leopold

Intel Xe-HP GPU Deployed for Aurora Exascale Development

November 17, 2020

At SC20, Intel announced that it is making its Xe-HP high performance discrete GPUs available to early access developers. Notably, the new chips have been deplo Read more…

By Tiffany Trader

HPE, AMD and EuroHPC Partner for Pre-Exascale LUMI Supercomputer

October 21, 2020

Not even a week after Nvidia announced that it would be providing hardware for the first four of the eight planned EuroHPC systems, HPE and AMD are announcing a Read more…

By Oliver Peckham

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This