As Supercomputers Approach Exascale, Experts Wrestle with Big Data

By Dawn Levy

July 21, 2011

The Oak Ridge Leadership Computing Facility (OLCF) plans on delivering a system with peak performance of 10–20 petaflops at Oak Ridge National Laboratory (ORNL) in 2012. This system will begin the OLCF’s transition from petaflop to exaflop computing this decade. Providing an environment that balances computational speed, memory bandwidth, and the input/output of data — so no aspect bottlenecks the performance of scientific applications — will require dramatic advances in parallel I/O capabilities. These high performance computing systems will generate unprecedented amounts of data — and unprecedented challenges in managing it.

In this interview, Galen Shipman, who heads OLCF’s Technology Integration group, discusses the challenges of managing big data. Shipman’s group is charged with integrating advanced technologies in the areas of networks, file systems, and archival storage infrastructures into the OLCF’s HPC systems, such as Jaguar, a 2.33-petaflop Cray XT funded by the Department of Energy Office of Science. From the evolution of supercomputing architectures and storage systems to the development of software and tools, it is clear major change is afoot as computing accelerates toward the exascale.

HPCwire: What data management nightmares keep you up at night?

Shipman: In the short term, the challenge of providing high-performance, reliable, and scalable I/O systems to meet the needs of a growing number of users from broadening domains of science. Users of our computing systems often employ a number of different applications to support their research, often with distinct parallel I/O and data management requirements. As these applications are scaled to systems such as Jaguar, I/O techniques that may have worked adequately at a few thousand cores often prove unscalable at tens or even hundreds of thousands of cores. Educating users about how to optimize their I/O and assisting them in doing so has proved effective.

Once these applications are optimized for our computational environment, the next challenge, managing the data that these applications produce, comes to the forefront. The OLCF currently manages more than 220 million files and more than 22 petabytes of data stored across our high-performance Lustre file systems and our High Performance Storage System (HPSS) archive. Managing this “big-data” is truly a grand challenge and spans not only HPC environments but private industry as well.

To facilitate data management at this scale, the OLCF has developed a number of tools, including advanced search and discovery, metadata harvesting, parallel data movement, and system monitoring and administration for our large-scale data systems. While we have made great progress in this area, the sheer scope of this challenge will necessitate a multi-institutional approach that spans government research laboratories and private industry.

From a hardware technology perspective, a more fundamental challenge that the storage community is facing is the widening gap between the performance of traditional storage technologies and the amount of data they can store. Historically, disk bandwidth has improved by 20 percent for sequential I/O and 8 percent for random I/O annually, while disk drive densities have increased by nearly 50 percent per year. For traditional storage technologies, this trend is likely to continue, resulting in ever-increasing storage capacities with ever-dwindling accessibility of the data sets that reside on the storage media.

To illustrate the challenges we will face with traditional storage technologies, take for example the community-developed exascale roadmap. It points toward parallel I/O environments that support up to 60 terabytes per second of bandwidth to transfer data to and from persistent storage. Assuming historical rates of performance improvement continue, achieving this performance level with traditional storage technologies would require more than 200,000 disk drives with a projected cost of more than $200 million. Even at this level of investment, the performance delivered to most applications may be a small fraction of the goal as drive-latency improvements become increasingly harder to realize.

Within the HPC community, there is general consensus that, over the next decade, the traditional storage hierarchy will incorporate technologies such as non-volatile random access memory (NVRAM) to address this problem. Many point to flash-based storage as a likely candidate for inclusion in this deeper hierarchy, but I believe it is too early to make a call on a specific NVRAM technology. One thing is clear: A fundamental technological shift coupled with end-to-end optimization will be required to meet our performance requirements within a tractable price point.

HPCwire: What research enterprises are generating big data at the OLCF?

Shipman: The majority of our users — scientists and engineers in academia, government, and industry—come to us through the Innovative and Novel Computational Impact on Theory and Experiment (INCITE) program, through which Oak Ridge and Argonne national laboratories will deliver 1.7 billion processor hours on advanced supercomputers in 2011. While all OLCF users rely on a parallel I/O environment, a number of heavy hitters in terms of I/O have unique performance and data-management requirements. Of particular note are the climate science, fusion energy, and combustion communities.

Fusion energy simulations may generate up to 100 terabytes of data per day. This is really on the edge in terms of parallel I/O performance, dwarfing that of most other science domains. The climate science community is somewhat unique in terms of its need for structured data management to enable intermodel comparison. The Coupled Model Intercomparison Project Phase 5 will not only generate petabytes of data but also require sophisticated data management tools to facilitate scientific review and intercomparison of simulation results from climate modeling centers around the world.

As simulation continues to mature as a fundamental tool for scientific discovery, I expect other scientific domains to have similar data management requirements. The OLCF is well positioned to meet these requirements through its development of advanced data-management technologies.

HPCwire: How do you store such big data sets?

Shipman: Petascale datasets from all major OLCF platforms are accessible though our center-wide Spider file system, which has both short-term and longer-term storage for computational science teams. Spider is one of the world’s largest and fastest parallel I/O environments, with 10.7 petabytes of disk space and more than 240 gigabytes per second of aggregate throughput. In addition to providing a common center-wide file system to all of ORNL’s supercomputing platforms, the Spider parallel I/O environment provides connectivity to remote facilities such as Argonne National Laboratory and Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center via dedicated data-transfer nodes running GridFTP.

Most users generate large datasets as part of their scientific simulations either for improved resiliency through checkpoint/restart — saving the application state at predefined intervals — or for subsequent data analysis. Checkpoint/restart datasets are generally purged by the application itself or through a system-wide sweep of “expired” datasets and thus have a limited shelf life.

Datasets to support subsequent data analysis have longer-term value, particularly when a scientist is interested in analyzing and comparing the results of multiple simulations. The OLCF provides every INCITE project with a long-term archival area to store these datasets via HPSS.

HPCwire: What are the logistics of longer-term, or archival, storage?

Shipman: When users decide to move data from Spider into long-term storage, the data are archived on our HPSS, which offers tens of petabytes of capacity and more than 12 gigabytes per second of bandwidth. Incoming data are written to disk and later migrated to tape for long-term archiving in high-capacity, robotic libraries. As of June 2011, the OLCF has more than 18 petabytes stored in more than 20 million files in HPSS from researchers in climate science, nuclear fusion, combustion, astrophysics, materials science, and many other scientific fields. It is not uncommon for users from a single science domain to have well over a petabyte of data in archive. These datasets are of high value, often requiring millions of processor hours to generate. Recreating these datasets and the scientific insight they may provide is increasingly difficult as the demand from the scientific community for large-scale computing resources such as Jaguar is outpaces our ability to supply the needed processor hours.

For the foreseeable future, we expect tape technology to be in play for archival storage because of its low power consumption and relatively low cost per terabyte of capacity. While massive array of idle disk (MAID) technology will lower the power requirements of disk-only solutions, there are concerns with the reliability of drive technology compared with tape, particularly when drives are frequently spun-up/down as employed in MAID. I expect disk technologies will still be in play for quite some time as well, as the capacity and sequential bandwidth capabilities at a relatively competitive cost point will continue to make this technology compelling. For archiving, I expect disk and tape to continue to be the dominant storage technologies for scientific computing environments. Our primary challenge will be balancing the storage requirements of our users with the cost of storage capacity and bandwidth. In response, I expect the scientific computing community to increasingly adopt data-reduction strategies within their simulation and analysis workflows.

HPCwire: For analyzing and visualizing big data, what is the main challenge?

Shipman: From my experiences with the data-analysis and visualization communities, their primary bottleneck in time-to-solution is I/O performance. Our work with the climate community on the Ultrascale Visualization — Climate Data Analysis Tools (UV-CDAT) project is tackling this issue head-on from an end-to-end perspective. We are working on ultrascale visualization techniques with a number of researchers from Los Alamos and Lawrence Livermore national laboratories, New York University, and Kitware to optimize the infrastructure of the Visualization ToolKit, an open-source software system for 3D computer graphics, image processing, analysis, and visualization.

Early results are encouraging, resulting in linear speedup — scaling — of common visualization workloads on Jaguar using Spider. Through an end-to-end approach that encompassed system architecture experts, visualization researchers, and middleware software engineers, we are able to address one of the most challenging aspects of visualization workloads. I believe this coordinated multidisciplinary approach will be of increasing value as the I/O performance gap continues to widen in the future.

Another promising approach to overcoming the I/O performance bottleneck is “in situ” data analysis, moving from the traditional post-processing model of data analysis to integration of data analysis directly within the simulation. This strategy can result in significant data reduction for certain use cases, particularly when scientists are interested in well-understood phenomena within their simulations.

The OLCF provides a number of dedicated visualization and analysis resources to our computational scientists. Our 32-node Linux cluster, known as Lens, is a dedicated platform that enables analysis and visualization of simulation data generated on Jaguar, providing a conduit for large-scale scientific discovery.

The facility also features EVEREST, the Exploratory Visualization Environment for REsearch in Science and Technology, and its associated visualization cluster. EVEREST is 30 feet wide and 10 feet tall and features a 27-projector Powerwall to display 35 million pixels for extremely high-definition scientific visualizations. Coupled with a dedicated Lustre file system, EVEREST provides a compelling experience for scientists exploring data at extremely fine scales. ORNL’s Visualization Task Group, led by Sean Ahern, helps researchers gain a better understanding of their data through visualization techniques.

HPCwire: What are the big challenges in scalable I/O?

Shipman: There are a number of big challenges in scalable I/O, from the sheer component counts to support hundreds of gigabytes per second of bandwidth and the ever-growing demands on the scalability of the file system software, to the scalability of middleware libraries and the applications that use them. Over the past decade we have seen dramatic increases in the number of components deployed both within our archive and on our high-performance parallel I/O systems. The OLCF currently has more than 24,000 disk drives supporting our high-performance parallel I/O and archival systems with nearly 500 gigabytes per second of bandwidth and more than 15 petabytes of capacity. System reliability and resiliency are critical at this scale. We use a number of techniques to improve the reliability and resiliency of our systems, from hardware-level redundancy to advanced software-level resiliency. An eye toward engineering reliability into our large-scale systems has allowed us to maintain extremely high availability of these critical resources.

Over the past decade we have seen dramatic increases in the scale of the computational platforms that we deploy at the OLCF. Today we support more than 25,000 file system clients on our largest-scale parallel file systems, an increase of an order of magnitude since 2005. Much of our work in providing a scalable parallel I/O environment has been focused on the Lustre file system. Through collaborative development of the open-source Lustre file system with Cluster File Systems, Sun, Oracle, and most recently Whamcloud, we have successfully supported the OLCF’s transition from teraflop to petaflop computing over this relatively short time frame. Our efforts in forming Open Scalable File Systems, a non-profit mutual benefit organization, are aimed at building upon our successes in collaborative development to meet similar challenges in the future.

Although we have made significant strides in improving the scalability and resiliency of the underlying file and storage systems, achieving optimal performance and scalability at the application level can be elusive to all but a handful of I/O experts. To bridge this gap, we employ a number of I/O middleware technologies, including the Adaptable I/O System (ADIOS), HDF-5, MPI-I/O, NetCDF, Parallel NetCDF, and the Parallel Log-structured File System (PLFS). These I/O middleware technologies serve a variety of functions from structured data models as found in NetCDF and HDF-5, to I/O transformation techniques as found in ADIOS and PLFS. Much of the development of ADIOS is conducted here at ORNL and led by Scott Klasky, a member of our Scientific Computing Group. Using these I/O middleware technologies in concert with our large-scale parallel I/O environment, we have delivered unprecedented levels of I/O performance to our end-users.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Why HPC Storage Matters More Now Than Ever: Analyst Q&A

September 17, 2021

With soaring data volumes and insatiable computing driving nearly every facet of economic, social and scientific progress, data storage is seizing the spotlight. Hyperion Research analyst and noted storage expert Mark No Read more…

GigaIO Gets $14.7M in Series B Funding to Expand Its Composable Fabric Technology to Customers

September 16, 2021

Just before the COVID-19 pandemic began in March 2020, GigaIO introduced its Universal Composable Fabric technology, which allows enterprises to bring together any HPC and AI resources and integrate them with networking, Read more…

What’s New in HPC Research: Solar Power, ExaWorks, Optane & More

September 16, 2021

In this regular feature, HPCwire highlights newly published research in the high-performance computing community and related domains. From parallel programming to exascale to quantum computing, the details are here. Read more…

Cerebras Brings Its Wafer-Scale Engine AI System to the Cloud

September 16, 2021

Five months ago, when Cerebras Systems debuted its second-generation wafer-scale silicon system (CS-2), co-founder and CEO Andrew Feldman hinted of the company’s coming cloud plans, and now those plans have come to fruition. Today, Cerebras and Cirrascale Cloud Services are launching... Read more…

AI Hardware Summit: Panel on Memory Looks Forward

September 15, 2021

What will system memory look like in five years? Good question. While Monday's panel, Designing AI Super-Chips at the Speed of Memory, at the AI Hardware Summit, tackled several topics, the panelists also took a brief glimpse into the future. Unlike compute, storage and networking, which... Read more…

AWS Solution Channel

Supporting Climate Model Simulations to Accelerate Climate Science

The Amazon Sustainability Data Initiative (ASDI), AWS is donating cloud resources, technical support, and access to scalable infrastructure and fast networking providing high performance computing (HPC) solutions to support simulations of near-term climate using the National Center for Atmospheric Research (NCAR) Community Earth System Model Version 2 (CESM2) and its Whole Atmosphere Community Climate Model (WACCM). Read more…

ECMWF Opens Bologna Datacenter in Preparation for Atos Supercomputer

September 14, 2021

In January 2020, the European Centre for Medium-Range Weather Forecasts (ECMWF) – a juggernaut in the weather forecasting scene – signed a four-year, $89-million contract with European tech firm Atos to quintuple its supercomputing capacity. With the deal approaching the two-year mark, ECMWF... Read more…

GigaIO Gets $14.7M in Series B Funding to Expand Its Composable Fabric Technology to Customers

September 16, 2021

Just before the COVID-19 pandemic began in March 2020, GigaIO introduced its Universal Composable Fabric technology, which allows enterprises to bring together Read more…

Cerebras Brings Its Wafer-Scale Engine AI System to the Cloud

September 16, 2021

Five months ago, when Cerebras Systems debuted its second-generation wafer-scale silicon system (CS-2), co-founder and CEO Andrew Feldman hinted of the company’s coming cloud plans, and now those plans have come to fruition. Today, Cerebras and Cirrascale Cloud Services are launching... Read more…

AI Hardware Summit: Panel on Memory Looks Forward

September 15, 2021

What will system memory look like in five years? Good question. While Monday's panel, Designing AI Super-Chips at the Speed of Memory, at the AI Hardware Summit, tackled several topics, the panelists also took a brief glimpse into the future. Unlike compute, storage and networking, which... Read more…

ECMWF Opens Bologna Datacenter in Preparation for Atos Supercomputer

September 14, 2021

In January 2020, the European Centre for Medium-Range Weather Forecasts (ECMWF) – a juggernaut in the weather forecasting scene – signed a four-year, $89-million contract with European tech firm Atos to quintuple its supercomputing capacity. With the deal approaching the two-year mark, ECMWF... Read more…

Quantum Computer Market Headed to $830M in 2024

September 13, 2021

What is one to make of the quantum computing market? Energized (lots of funding) but still chaotic and advancing in unpredictable ways (e.g. competing qubit tec Read more…

Amazon, NCAR, SilverLining Team for Unprecedented Cloud Climate Simulations

September 10, 2021

Earth’s climate is, to put it mildly, not in a good place. In the wake of a damning report from the Intergovernmental Panel on Climate Change (IPCC), scientis Read more…

After Roadblocks and Renewals, EuroHPC Targets a Bigger, Quantum Future

September 9, 2021

The EuroHPC Joint Undertaking (JU) was formalized in 2018, beginning a new era of European supercomputing that began to bear fruit this year with the launch of several of the first EuroHPC systems. The undertaking, however, has not been without its speed bumps, and the Union faces an uphill... Read more…

How Argonne Is Preparing for Exascale in 2022

September 8, 2021

Additional details came to light on Argonne National Laboratory’s preparation for the 2022 Aurora exascale-class supercomputer, during the HPC User Forum, held virtually this week on account of pandemic. Exascale Computing Project director Doug Kothe reviewed some of the 'early exascale hardware' at Argonne, Oak Ridge and NERSC (Perlmutter), while Ti Leggett, Deputy Project Director & Deputy Director... Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer called Dojo to process truly vast amounts of video data. It’s a beast! … A truly useful exaflop at de facto FP32.” Read more…

Berkeley Lab Debuts Perlmutter, World’s Fastest AI Supercomputer

May 27, 2021

A ribbon-cutting ceremony held virtually at Berkeley Lab's National Energy Research Scientific Computing Center (NERSC) today marked the official launch of Perlmutter – aka NERSC-9 – the GPU-accelerated supercomputer built by HPE in partnership with Nvidia and AMD. Read more…

Esperanto, Silicon in Hand, Champions the Efficiency of Its 1,092-Core RISC-V Chip

August 27, 2021

Esperanto Technologies made waves last December when it announced ET-SoC-1, a new RISC-V-based chip aimed at machine learning that packed nearly 1,100 cores onto a package small enough to fit six times over on a single PCIe card. Now, Esperanto is back, silicon in-hand and taking aim... Read more…

Enter Dojo: Tesla Reveals Design for Modular Supercomputer & D1 Chip

August 20, 2021

Two months ago, Tesla revealed a massive GPU cluster that it said was “roughly the number five supercomputer in the world,” and which was just a precursor to Tesla’s real supercomputing moonshot: the long-rumored, little-detailed Dojo system. “We’ve been scaling our neural network training compute dramatically over the last few years,” said Milan Kovac, Tesla’s director of autopilot engineering. Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months after Red Hat deprecated its support for the widely popular, free CentOS server operating system. The Rocky Linux development effort... Read more…

Intel Completes LLVM Adoption; Will End Updates to Classic C/C++ Compilers in Future

August 10, 2021

Intel reported in a blog this week that its adoption of the open source LLVM architecture for Intel’s C/C++ compiler is complete. The transition is part of In Read more…

Google Launches TPU v4 AI Chips

May 20, 2021

Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I Read more…

AMD-Xilinx Deal Gains UK, EU Approvals — China’s Decision Still Pending

July 1, 2021

AMD’s planned acquisition of FPGA maker Xilinx is now in the hands of Chinese regulators after needed antitrust approvals for the $35 billion deal were receiv Read more…

Leading Solution Providers

Contributors

Hot Chips: Here Come the DPUs and IPUs from Arm, Nvidia and Intel

August 25, 2021

The emergence of data processing units (DPU) and infrastructure processing units (IPU) as potentially important pieces in cloud and datacenter architectures was Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

HPE Wins $2B GreenLake HPC-as-a-Service Deal with NSA

September 1, 2021

In the heated, oft-contentious, government IT space, HPE has won a massive $2 billion contract to provide HPC and AI services to the United States’ National Security Agency (NSA). Following on the heels of the now-canceled $10 billion JEDI contract (reissued as JWCC) and a $10 billion... Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Intel Unveils New Node Names; Sapphire Rapids Is Now an ‘Intel 7’ CPU

July 27, 2021

What's a preeminent chip company to do when its process node technology lags the competition by (roughly) one generation, but outmoded naming conventions make it seem like it's two nodes behind? For Intel, the response was to change how it refers to its nodes with the aim of better reflecting its positioning within the leadership semiconductor manufacturing space. Intel revealed its new node nomenclature, and... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire