Satellite Advances, NSF Computation Power Rapid Mapping of Earth’s Surface

By Ken Chiacchia and Tiffany Jolley

July 13, 2017

New satellite technologies have completely changed the game in mapping and geographical data gathering, reducing costs and placing a new emphasis on time series and timeliness in general, according to Paul Morin, director of the Polar Geospatial Center at the University of Minnesota.

In the second plenary session of the PEARC conference in New Orleans on July 12, Morin described how access to the DigitalGlobe satellite constellation, the NSF XSEDE network of supercomputing centers and the Blue Waters supercomputer at the National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign have enabled his group to map Antarctica—an area of 5.4 million square miles, compared with the 3.7 million square miles of the “lower 48” United States—at 1-meter resolution in two years. Nine months later, then-president Barack Obama announced a joint White House initiative involving the NSF and the National Geospatial Intelligence Agency (NGIA) in which Morin’s group mapped a similar area in the Arctic including the entire state of Alaska in two years.

“If I wrote this story in a single proposal I wouldn’t have been able to write [any proposals] afterward,” Morin said. “It’s that absurd.” But the leaps in technology have made what used to be multi-decadal mapping projects—when they could be done at all—into annual events, with even more frequent updates soon to come.

The inaugural Practice and Experience in Advanced Research Computing (PEARC) conference—with the theme Sustainability, Success and Impact—stresses key objectives for those who manage, develop and use advanced research computing throughout the U.S. and the world. Organizations supporting this new HPC conference include the Advancing Research Computing on Campuses: Best Practices Workshop (ARCC), the Extreme Science and Engineering Development Environment (XSEDE), the Science Gateways Community Institute, the Campus Research Computing (CaRC) Consortium, the Advanced CyberInfrastructure Research and Education Facilitators (ACI-REF) consortium, the National Center for Supercomputing Applications’ Blue Waters project, ESnet, Open Science Grid, Compute Canada, the EGI Foundation, the Coalition for Academic Scientific Computation (CASC) and Internet2.

Follow the Poop

One project made possible with the DigitalGlobe constellation—a set of Hubble-like multispectral orbiting telescopes “pointed the other way”—was a University of Minnesota census of emperor penguin populations in Antarctica.

“What’s the first thing you do if you get access to a bunch of sub-meter-resolution [orbital telescopes covering] Antarctica?” Morin asked. “You point them at penguins.”

Thanks in part to a lack of predators the birds over-winter on the ice, huddling in colonies for warmth. Historically these colonies were discovered by accident: Morin’s project enabled the first continent-wide survey to find and estimate the population size of all the colonies.

The researchers realized that they had a relatively easy way to spot the colonies in the DigitalGlobe imagery: Because the penguins eat beta-carotene-rich krill, their excrement stains the ice red.

“You can identify their location by looking for poo,” Morin said. The project enabled the first complete population count of emperor penguins: 595,000 birds, +14%

“We started to realize we were onto something,” he added. His group began to wonder if they could leverage the sub-meter-resolution, multispectral, stereo view of the constellation’s WorldView I, II and III satellites to derive the topography of the Antarctic, and later the Arctic. One challenge, he knew, would be finding the computational power to extract topographic data from the stereo images in a reasonable amount of time. He found his answer at the NSF and the NGIA.

“We proposed to a science agency and a combat support agency that we were going to map the topography of 30 degrees of the globe in 24 months.”

Blue Waters on the Ice

Morin and his collaborators found themselves in the middle of a seismic shift in topographic technology.

“Eight years ago, people were doing [this] from the ground,” with a combination of land-based surveys and accurate but expensive LIDAR mapping from aircraft, he said. These methods made sense in places where population and industrial density made the cost worthwhile. But it had left the Antarctic and Arctic largely unmapped.

Deriving topographic information from the photographs posed a computational problem well beyond the capabilities of a campus cluster. The group did initial computations at the Ohio Supercomputer Center, but needed to expand for the final data analysis. In 2014 XSEDE Project Director John Towns offered XSEDE’s help in tackling the massive scale of data that would come from an array of satellites collecting topographic images. From 2014 to 2015, Morin used XSEDE resources, most notably Gordon at San Diego Supercomputer Center and XSEDE’s Extended Collaborative Support Service to carry out his initial computations. XSEDE then helped his group acquire an allocation on Blue Waters, an NSF-funded Cray Inc. system at Illinois and NCSA with 49,000 CPUs and a peak performance of 13.3 petaflops.

Collecting the equivalent area of California daily, a now-expanded group of subject experts made use of the polar-orbiting satellites and Blue Waters to derive elevation data. They completed a higher-resolution map of Alaska—the earlier version of which had taken the U.S. Geological Survey 50 years—in a year. While the initial images are licensed for U.S. government use only, the group was able to release the resulting topographic data for public use.

Mapping Change

Thanks to the one-meter resolution of their initial analysis, the group quickly found they could identify many man-made structures on the surface. They could also spot vegetation changes such as clearcutting. They could even quantify vegetation regrowth after replanting.

“We’re watching individual trees growing here.”

Another set of images he showed in his PEARC17 presentation were before-and-after topographic maps of Nuugaatsiaq, Greenland, which was devastated by a tsunami last month. The Greenland government is using the images, which show both human structures and the landslide that caused the 10-meter tsunami, to plan recovery efforts.

The activity of the regions’ ice sheets was a striking example of the technology’s capabilities.

“Ice is a mineral that flows,” Morin said, and so the new topographic data offer much more frequent information about ice-sheet changes driven by climate change than previously available. “We not only have an image of the ice but we know exactly how high it is.”

Morin also showed an image of the Larsen Ice Shelf revealing a crack that had appeared in the glacier. The real news, though, was that the crack—which created an iceberg the size of the big island of Hawaii—was less than 24 hours old. It had appeared sometime after midnight on July 12.

“We [now] have better topography for Siberia than we have for Montana,” he noted.

New Directions

While the large, high-resolution satellites have already transformed the field, innovations are already coming that could create another shift, Morin said.

“This is not your father’s topography,” he noted. “Everything has changed; everything is time sensitive; everything is on demand.” In an interview later that morning, he added, “XSEDE, Blue Waters and NSF have changed how earth science happens now.”

One advance won’t require new technology: just a little more time. While the current topographic dataset is at 1-meter resolution, the data can go tighter with more computation. The satellite images actually have a 30-centimeter resolution, which would allow for the project to shift from imaging objects the size of automobiles to those the size of a coffee table.

At that point, he said, “instead of [just the] presence or absence of trees we’ll be able to tell what species of tree. It doesn’t take recollection of imagery; it just takes reprocessing.”

The new, massive constellation of CubeSats such as the Planet company’s toaster-sized Dove satellites now being launched promises an even more disruptive advance. A swarm of these satellites will provide much more frequent coverage of the entire Earth’s surface than possible with the large telescopes.

Click to expand

“The quality isn’t as good, but right now we’re talking about coverage,” Morin said. His group’s work has taken advantage of a system that allows mapping of a major portion of the Earth in a year. “What happens when we have monthly coverage?”

Feature image caption: Buildings in Juneau, Alaska, as shown in the University of Minnesota topographic survey of the Arctic region. The airport runway can be seen at the bottom.

Authors

Ken Chiacchia, Senior Science Writer, Pittsburgh Supercomputing Center

Tiffany Jolley Content Producer, National Center for Supercomputing Applications

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

University of Chicago Researchers Generate First Computational Model of Entire SARS-CoV-2 Virus

January 15, 2021

Over the course of the last year, many detailed computational models of SARS-CoV-2 have been produced with the help of supercomputers, but those models have largely focused on critical elements of the virus, such as its Read more…

By Oliver Peckham

Pat Gelsinger Returns to Intel as CEO

January 14, 2021

The Intel board of directors has appointed a new CEO. Intel alum Pat Gelsinger is leaving his post as CEO of VMware to rejoin the company that he parted ways with 11 years ago. Gelsinger will succeed Bob Swan, who will remain CEO until Feb. 15. Gelsinger previously spent 30 years... Read more…

By Tiffany Trader

Roar Supercomputer to Support Naval Aircraft Research

January 14, 2021

One might not think “aircraft” when picturing the U.S. Navy, but the military branch actually has thousands of aircraft currently in service – and now, supercomputing will help future naval aircraft operate faster, Read more…

By Staff report

DOE and NOAA Extend Computing Partnership, Plan for New Supercomputer

January 14, 2021

The National Climate-Computing Research Center (NCRC), hosted by Oak Ridge National Laboratory (ORNL), has been supporting the climate research of the National Oceanic and Atmospheric Administration (NOAA) for the last 1 Read more…

By Oliver Peckham

Using Micro-Combs, Researchers Demonstrate World’s Fastest Optical Neuromorphic Processor for AI

January 13, 2021

Neuromorphic computing, which uses chips that mimic the behavior of the human brain using virtual “neurons,” is growing in popularity thanks to high-profile efforts from Intel and others. Now, a team of researchers l Read more…

By Oliver Peckham

AWS Solution Channel

Now Available – Amazon EC2 C6gn Instances with 100 Gbps Networking

Amazon EC2 C6gn instances powered by AWS Graviton2 processors are now available!

Compared to C6g instances, this new instance type provides 4x higher network bandwidth, 4x higher packet processing performance, and 2x higher EBS bandwidth. Read more…

Intel® HPC + AI Pavilion

Intel Keynote Address

Intel is the foundation of HPC – from the workstation to the cloud to the backbone of the Top500. At SC20, Intel’s Trish Damkroger, VP and GM of high performance computing, addresses the audience to show how Intel and its partners are building the future of HPC today, through hardware and software technologies that accelerate the broad deployment of advanced HPC systems. Read more…

Honing In on AI, US Launches National Artificial Intelligence Initiative Office

January 13, 2021

To drive American leadership in the field of AI into the future, the National Artificial Intelligence Initiative Office has been launched by the White House Office of Science and Technology Policy (OSTP). The new agen Read more…

By Todd R. Weiss

Pat Gelsinger Returns to Intel as CEO

January 14, 2021

The Intel board of directors has appointed a new CEO. Intel alum Pat Gelsinger is leaving his post as CEO of VMware to rejoin the company that he parted ways with 11 years ago. Gelsinger will succeed Bob Swan, who will remain CEO until Feb. 15. Gelsinger previously spent 30 years... Read more…

By Tiffany Trader

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

By John Russell

Intel ‘Ice Lake’ Server Chips in Production, Set for Volume Ramp This Quarter

January 12, 2021

Intel Corp. used this week’s virtual CES 2021 event to reassert its dominance of the datacenter with the formal roll out of its next-generation server chip, the 10nm Xeon Scalable processor that targets AI and HPC workloads. The third-generation “Ice Lake” family... Read more…

By George Leopold

Researchers Say It Won’t Be Possible to Control Superintelligent AI

January 11, 2021

Worries about out-of-control AI aren’t new. Many prominent figures have suggested caution when unleashing AI. One quote that keeps cropping up is (roughly) th Read more…

By John Russell

AMD Files Patent on New GPU Chiplet Approach

January 5, 2021

Advanced Micro Devices is accelerating the GPU chiplet race with the release of a U.S. patent application for a device that incorporates high-bandwidth intercon Read more…

By George Leopold

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

By Tracey Bryant

Intel Touts Optane Performance, Teases Next-gen “Crow Pass”

January 5, 2021

Competition to leverage new memory and storage hardware with new or improved software to create better storage/memory schemes has steadily gathered steam during Read more…

By John Russell

Farewell 2020: Bleak, Yes. But a Lot of Good Happened Too

December 30, 2020

Here on the cusp of the new year, the catchphrase ‘2020 hindsight’ has a distinctly different feel. Good riddance, yes. But also proof of science’s power Read more…

By John Russell

Esperanto Unveils ML Chip with Nearly 1,100 RISC-V Cores

December 8, 2020

At the RISC-V Summit today, Art Swift, CEO of Esperanto Technologies, announced a new, RISC-V based chip aimed at machine learning and containing nearly 1,100 low-power cores based on the open-source RISC-V architecture. Esperanto Technologies, headquartered in... Read more…

By Oliver Peckham

Azure Scaled to Record 86,400 Cores for Molecular Dynamics

November 20, 2020

A new record for HPC scaling on the public cloud has been achieved on Microsoft Azure. Led by Dr. Jer-Ming Chia, the cloud provider partnered with the Beckman I Read more…

By Oliver Peckham

NICS Unleashes ‘Kraken’ Supercomputer

April 4, 2008

A Cray XT4 supercomputer, dubbed Kraken, is scheduled to come online in mid-summer at the National Institute for Computational Sciences (NICS). The soon-to-be petascale system, and the resulting NICS organization, are the result of an NSF Track II award of $65 million to the University of Tennessee and its partners to provide next-generation supercomputing for the nation's science community. Read more…

Is the Nvidia A100 GPU Performance Worth a Hardware Upgrade?

October 16, 2020

Over the last decade, accelerators have seen an increasing rate of adoption in high-performance computing (HPC) platforms, and in the June 2020 Top500 list, eig Read more…

By Hartwig Anzt, Ahmad Abdelfattah and Jack Dongarra

Aurora’s Troubles Move Frontier into Pole Exascale Position

October 1, 2020

Intel’s 7nm node delay has raised questions about the status of the Aurora supercomputer that was scheduled to be stood up at Argonne National Laboratory next year. Aurora was in the running to be the United States’ first exascale supercomputer although it was on a contemporaneous timeline with... Read more…

By Tiffany Trader

Google Hires Longtime Intel Exec Bill Magro to Lead HPC Strategy

September 18, 2020

In a sign of the times, another prominent HPCer has made a move to a hyperscaler. Longtime Intel executive Bill Magro joined Google as chief technologist for hi Read more…

By Tiffany Trader

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

By John Russell

Leading Solution Providers

Contributors

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

By Tracey Bryant

Top500: Fugaku Keeps Crown, Nvidia’s Selene Climbs to #5

November 16, 2020

With the publication of the 56th Top500 list today from SC20's virtual proceedings, Japan's Fugaku supercomputer – now fully deployed – notches another win, Read more…

By Tiffany Trader

European Commission Declares €8 Billion Investment in Supercomputing

September 18, 2020

Just under two years ago, the European Commission formalized the EuroHPC Joint Undertaking (JU): a concerted HPC effort (comprising 32 participating states at c Read more…

By Oliver Peckham

Texas A&M Announces Flagship ‘Grace’ Supercomputer

November 9, 2020

Texas A&M University has announced its next flagship system: Grace. The new supercomputer, named for legendary programming pioneer Grace Hopper, is replacing the Ada system (itself named for mathematician Ada Lovelace) as the primary workhorse for Texas A&M’s High Performance Research Computing (HPRC). Read more…

By Oliver Peckham

At Oak Ridge, ‘End of Life’ Sometimes Isn’t

October 31, 2020

Sometimes, the old dog actually does go live on a farm. HPC systems are often cursed with short lifespans, as they are continually supplanted by the latest and Read more…

By Oliver Peckham

Nvidia and EuroHPC Team for Four Supercomputers, Including Massive ‘Leonardo’ System

October 15, 2020

The EuroHPC Joint Undertaking (JU) serves as Europe’s concerted supercomputing play, currently comprising 32 member states and billions of euros in funding. I Read more…

By Oliver Peckham

Gordon Bell Special Prize Goes to Massive SARS-CoV-2 Simulations

November 19, 2020

2020 has proven a harrowing year – but it has produced remarkable heroes. To that end, this year, the Association for Computing Machinery (ACM) introduced the Read more…

By Oliver Peckham

Nvidia-Arm Deal a Boon for RISC-V?

October 26, 2020

The $40 billion blockbuster acquisition deal that will bring chipmaker Arm into the Nvidia corporate family could provide a boost for the competing RISC-V architecture. As regulators in the U.S., China and the European Union begin scrutinizing the impact of the blockbuster deal on semiconductor industry competition and innovation, the deal has at the very least... Read more…

By George Leopold

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This