Preparing for the Arrival of Aurora with CPU-based Interactive Visualization

By Rob Farber

October 30, 2018

In preparation for the arrival of Aurora, slated to be the first U.S. exascale supercomputer, Argonne National Laboratory is actively working to make techniques such as in situ and in transit visualization and analysis available to their user community plus the HPC community at large. The result is a DOE multi-institutional effort that includes Argonne, private companies, and other national labs to leverage SENSEI, a portable framework that enables in situ, in transit, and traditional batch visualization workflows that can use either ray tracing or triangle-based rendering back ends, for analysis and scalable interactive rendering.

In situ visualization has been identified as a key technology to enable science at the exascale[i]. In situ visualization means that the visualization occurs on the same nodes that perform the computation. In transit visualization is not as directly coupled to the simulation, and can help load-balance by using more nodes to support computationally expensive simulations like LAMMPs. Unlike in situ, in transit does incur some overhead when moving data across the communications fabric between nodes. Both methods keep the data in memory and avoid writing to storage.

Joseph Insley (visualization and analysis team lead at the Argonne Leadership Computing Facility) points out, “With SENSEI, users can utilize in situ and in transit techniques to address the widening gap between Flop/s and I/O capacity which is making full-resolution, I/O-intensive post hoc analysis prohibitively expensive, if not impossible.” Silvio Rizzi (assistant computer scientist, Argonne) highlights portability when he states, “the idea behind SENSEI is to write once and use anywhere.”

The Argonne team led by Nicola Ferrier as PI has adapted the popular LAMMPS (e.g. the Large-scale Atomic/Molecular Massively Parallel Simulator) code to demonstrate the benefits of the SENSEI framework. The integration of SENSEI made use of existing mechanisms in LAMMPS for coupling with other simulation codes.[ii]

Understanding the choice of LAMMPS as a SENSEI testbed

Paul Navrátil, director of Visualization at the Texas Advanced Computing Center (TACC), helps us understand the meaning and importance of in situ and in transit visualization to the general HPC community as well as the choice of LAMMPS by the ALCF team.

Just as Argonne will host the fastest U.S. supercomputer with Aurora, TACC will be home to Frontera, which will become the fastest academic supercomputer in the United States when it becomes operational in 2019.

Navrátil notes, “We expect in situ workflows to become increasingly necessary on Frontera and across all large-scale simulation science.” He believes that, “In transit analysis will also play an increasing role as simulations improve support for loosely-coupled in situ frameworks. With an in transit pathway, the simulation resources do not need to be shared for analysis tasks, which is favorable when analysis is compute-intensive, or when the simulation requires all available resources itself.”

LAMMPS is a compute intensive application plus it is a very popular simulation code, which makes it a natural testbed for SENSEI as it lets large numbers of users explore the benefits of in situ visualization plus the load balancing benefits of in transit visualization and analysis. SENSEI is also being used in multiple science domains, including molecular dynamics and materials science.

An in transit workflow using SENSEI and OSPRay is shown below.

Figure 1: LAMMPS using SENSEI to execute an in transit visualization and analysis work flow. (Image courtesy taken from Usher, et.al.[iii])
Choosing the right rendering back end

SENSEI is very flexible and allows researchers to perform analysis and use either OpenGL rendering or create photo real images. Jim Jeffers (senior director and senior principle engineer, Intel Visualization Solutions) notes that the interactive performance delivered by the Intel Rendering Framework and photorealistic rendering with the freely available OSPRay library and viewer, “addresses the need and creates the want” for photorealistic rendering. Succinctly, interactive ray tracing with its inherent lighting capability lets scientists get more from their data. Jeffers’ is famous for stating, “a picture is worth an exabyte.”

The ALCF team provided the following figure to illustrate what is possible when instrumenting LAMMPS with SENSEI. They used the Intel OSPRay library that is part of the Intel Rendering Framework and the libIS, a lightweight, flexible library to create this in transit visualization. However, SENSEI was designed[iv] work with other libraries in place of libIS such as catalyst (part of ParaView), ADIOS (from Oak Ridge National Laboratory), and LibSim (part of VisIt), as well as GPU-based software to perform in transit visualizations.

Figure 2: Interactive in situ visualization of a 172k atom simulation of silicene formation [6] with 128 LAMMPS ranks sending to 16 renderer ranks, all executed on Theta. (Image from Usher, et. al[v])
SENSEI is not the first code to provide easy access to both OpenGL and ray tracing back end and analytic capabilities. Both the popular VisIt[vi] and ParaView viewers make it simple to switch between or even combine triangle-based OpenGL rendering with Intel OpenSWR and photorealistic ray-traced rendering with Intel OSPRay.

Understanding Software Defined Visualization (SDVis)

The foundation of CPU-based in situ and in transit visualization is Software Defined Visualization. The core functionality are the freely available, open-source Intel OSPRay, Embree, and OpenSWR libraries. These libraries have been incorporated into the Intel® Rendering Framework stack as shown below.

Figure 3: Scientific and Professional rendering stacks using the Intel Rendering Framework (Image courtesy Intel)

Using CPUs for rendering has taken the HPC community by storm. Rizzi summarizes the interest at Argonne by noting, “We want to enable visualization on our supercomputers which are CPU-based”. Navrátil highlights TACC’s commitment by pointing out that, “CPU-based SDVis will be our primary visual analysis mode on Frontera, leveraging the Intel Rendering Framework stack.”

Scaling and the ability to run efficiently are two key ideas behind the OSPRay ray tracing and the OpenSWR OpenGL SDVis renderer.

Kitware, for example, performed trillion triangle OpenGL visualizations using the LANL Trinity supercomputer. David DeMarle, (visualization luminary and engineer at Kitware) observes that, “CPU-based OpenGL performance does not trail off even when rendering meshes containing one trillion (10^12) triangles on the Trinity leadership class supercomputer. Further, we might see a 10-20 trillion triangle per second result as our current benchmark used only 1/19th of the machine.” The ability of the CPU to access large amounts of memory is key to realizing trillion triangle per second rendering capability.

Meanwhile, OSPRay users have demonstrated the ability to render and visualize large, photorealistic images on everything from cosmological data sets to molecules and complex scenes. No special hardware is required for rendering, which can achieve interactive photorealism on as few as eight Intel Xeon Scalable 8180 processors or scale to high-quality rendering for in situ nodes. [vii] [viii] [ix] [x]

Viewing the rendered images

The “visualize anywhere” nature of CPU-based SDVis means that visualizing locally or remotely is possible on devices that can display from memory. Extraordinary display flexibility without device dependencies makes “visualize anywhere” even better. HPC users appreciate how they can view results on their laptops and switch to display walls or a cave.

SENSEI also supports existing batched save-to-storage workflows.

Summary

The HPC community has always been about pressing the limits of computation. For this reason, in situ and in transit visualization frameworks have been created to work with CPU-based rendering to eliminate data movement. In this way, visualization can scale and keep pace with simulation as the HPC community runs on petascale and anticipates the next generation exascale supercomputers.

Rob Farber is a global technology consultant and author with an extensive background in HPC and in developing machine learning technology that he applies at national labs and commercial organizations. Rob can be reached at [email protected].


[i] https://science.energy.gov/~/media/ascr/pdf/program-documents/docs/Exascale-ASCR-Analysis.pdf

[ii] https://lammps.sandia.gov/doc/Howto_couple.html

[iii] Will Usher, Silvio Rizzi, Ingo Wald, Jefferson Amstutz, Joseph Insley, Venkatram Vishwanath, Nicola Ferrier, Michael E. Papka, and Valerio Pascucci. 2018. libIS: A Lightweight Library for Flexible In Transit Visualization. In ISAV: In Situ Infrastructures for Enabling Extreme-Scale Analysis and Visualization (ISAV ’18), November 12, 2018, Dallas, TX, USA. ACM, New York, NY, USA, 6 pages. https://doi.org/10.1145/3281464.3281466.

[v] https://doi.org/10.1145/3281464.3281466

[vi] https://tacc.github.io/visitOSPRay/

[vii] http://sdvis.org/

[viii] http://www.cgw.com/Press-Center/In-Focus/2018/Scalable-CPU-Based-SDVis-Enables-Interactive-Pho.aspx

[ix] https://www.ixpug.org/documents/1496440983IXPUG_insitu_S1_Jeffers.pdf

[x] http://www.techenablement.com/third-party-use-cases-illustrate-the-success-of-cpu-based-visualization/

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

IBM Quantum Update: Q System One Launch, New Collaborators, and QC Center Plans

January 10, 2019

IBM made three significant quantum computing announcements at CES this week. One was introduction of IBM Q System One; it’s really the integration of IBM’s existing 20-quibit platform into a more robust, self-contain Read more…

By John Russell

Intel at CES: Nervana; 10nm Server CPU; Cascade Lake

January 9, 2019

On the eve of the Consumer Electronics Show in Las Vegas this week, Intel staged a launch event that covered a new version of its Nervana AI processor and a demonstration of the next-generation Xeon 10nm chip. The Read more…

By Staff

IBM’s New Global Weather Forecasting System Runs on GPUs

January 9, 2019

Anyone who has checked a forecast to decide whether or not to pack an umbrella knows that weather prediction can be a mercurial endeavor. It is a Herculean task: the constant modeling of incredibly complex systems to a high degree of accuracy at a local level within very short spans of time. Read more…

By Oliver Peckham

HPE Extreme Performance Solutions

HPE Systems With Intel Omni-Path: Architected for Value and Accessible High-Performance Computing

Today’s high-performance computing (HPC) and artificial intelligence (AI) users value high performing clusters. And the higher the performance that their system can deliver, the better. Read more…

IBM Accelerated Insights

Data: The Key To Unlocking Modern Research

Research tackles the big questions, delving into uncharted territory in pursuit of knowledge that could change the world. Today’s research simulations are generating more data than ever before, a trend that shows no signs of slowing. Read more…

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourself – and you are the easiest person to fool.” This maxim Read more…

By Ben Criger

IBM Quantum Update: Q System One Launch, New Collaborators, and QC Center Plans

January 10, 2019

IBM made three significant quantum computing announcements at CES this week. One was introduction of IBM Q System One; it’s really the integration of IBM’s Read more…

By John Russell

IBM’s New Global Weather Forecasting System Runs on GPUs

January 9, 2019

Anyone who has checked a forecast to decide whether or not to pack an umbrella knows that weather prediction can be a mercurial endeavor. It is a Herculean task: the constant modeling of incredibly complex systems to a high degree of accuracy at a local level within very short spans of time. Read more…

By Oliver Peckham

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourse Read more…

By Ben Criger

The Deep500 – Researchers Tackle an HPC Benchmark for Deep Learning

January 7, 2019

How do you know if an HPC system, particularly a larger-scale system, is well-suited for deep learning workloads? Today, that’s not an easy question to answer Read more…

By John Russell

HPCwire Awards Highlight Supercomputing Achievements in the Sciences

January 3, 2019

In November at SC18 in Dallas, HPCwire Readers’ and Editors’ Choice awards program commemorated its 15th year of honoring achievement in HPC, with categories ranging from Best Use of AI to the Workforce Diversity Leadership Award and recipients across a wide variety of industrial and research sectors. Read more…

By the Editorial Team

White House Top Science Post Filled After Two-Year Vacancy

January 3, 2019

Half-way into Trump's term, the Senate has confirmed a director for the Office of Science and Technology Policy (OSTP), the agency that coordinates science poli Read more…

By Tiffany Trader

Batswana Gems

December 20, 2018

Most who work in the high-performance computing (HPC) industry agree; people problems are far more complicated than technical challenges. As I wrote in a 2015 HPCwire feature titled, “Women in HPC: Revelations and Reckoning,” diversity, or the lack thereof, is the HPC industry’s current grand challenge. Read more…

By Elizabeth Leake

HPC Reflections and (Mostly Hopeful) Predictions

December 19, 2018

So much ‘spaghetti’ gets tossed on walls by the technology community (vendors and researchers) to see what sticks that it is often difficult to peer through Read more…

By John Russell

Quantum Computing Will Never Work

November 27, 2018

Amid the gush of money and enthusiastic predictions being thrown at quantum computing comes a proposed cold shower in the form of an essay by physicist Mikhail Read more…

By John Russell

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

Summit Supercomputer is Already Making its Mark on Science

September 20, 2018

Summit, now the fastest supercomputer in the world, is quickly making its mark in science – five of the six finalists just announced for the prestigious 2018 Read more…

By John Russell

AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

House Passes $1.275B National Quantum Initiative

September 17, 2018

Last Thursday the U.S. House of Representatives passed the National Quantum Initiative Act (NQIA) intended to accelerate quantum computing research and developm Read more…

By John Russell

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourse Read more…

By Ben Criger

Contract Signed for New Finnish Supercomputer

December 13, 2018

After the official contract signing yesterday, configuration details were made public for the new BullSequana system that the Finnish IT Center for Science (CSC Read more…

By Tiffany Trader

Leading Solution Providers

SC 18 Virtual Booth Video Tour

Advania @ SC18 AMD @ SC18
ASRock Rack @ SC18
DDN Storage @ SC18
HPE @ SC18
IBM @ SC18
Lenovo @ SC18 Mellanox Technologies @ SC18
NVIDIA @ SC18
One Stop Systems @ SC18
Oracle @ SC18 Panasas @ SC18
Supermicro @ SC18 SUSE @ SC18 TYAN @ SC18
Verne Global @ SC18

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

HPE No. 1, IBM Surges, in ‘Bucking Bronco’ High Performance Server Market

September 27, 2018

Riding healthy U.S. and global economies, strong demand for AI-capable hardware and other tailwind trends, the high performance computing server market jumped 28 percent in the second quarter 2018 to $3.7 billion, up from $2.9 billion for the same period last year, according to industry analyst firm Hyperion Research. Read more…

By Doug Black

HPC Reflections and (Mostly Hopeful) Predictions

December 19, 2018

So much ‘spaghetti’ gets tossed on walls by the technology community (vendors and researchers) to see what sticks that it is often difficult to peer through Read more…

By John Russell

Intel Confirms 48-Core Cascade Lake-AP for 2019

November 4, 2018

As part of the run-up to SC18, taking place in Dallas next week (Nov. 11-16), Intel is doling out info on its next-gen Cascade Lake family of Xeon processors, specifically the “Advanced Processor” version (Cascade Lake-AP), architected for high-performance computing, artificial intelligence and infrastructure-as-a-service workloads. Read more…

By Tiffany Trader

Germany Celebrates Launch of Two Fastest Supercomputers

September 26, 2018

The new high-performance computer SuperMUC-NG at the Leibniz Supercomputing Center (LRZ) in Garching is the fastest computer in Germany and one of the fastest i Read more…

By Tiffany Trader

Houston to Field Massive, ‘Geophysically Configured’ Cloud Supercomputer

October 11, 2018

Based on some news stories out today, one might get the impression that the next system to crack number one on the Top500 would be an industrial oil and gas mon Read more…

By Tiffany Trader

Microsoft to Buy Mellanox?

December 20, 2018

Networking equipment powerhouse Mellanox could be an acquisition target by Microsoft, according to a published report in an Israeli financial publication. Microsoft has reportedly gone so far as to engage Goldman Sachs to handle negotiations with Mellanox. Read more…

By Doug Black

The Deep500 – Researchers Tackle an HPC Benchmark for Deep Learning

January 7, 2019

How do you know if an HPC system, particularly a larger-scale system, is well-suited for deep learning workloads? Today, that’s not an easy question to answer Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This