As Supercomputers Approach Exascale, Experts Wrestle with Big Data

By Dawn Levy

July 21, 2011

The Oak Ridge Leadership Computing Facility (OLCF) plans on delivering a system with peak performance of 10–20 petaflops at Oak Ridge National Laboratory (ORNL) in 2012. This system will begin the OLCF’s transition from petaflop to exaflop computing this decade. Providing an environment that balances computational speed, memory bandwidth, and the input/output of data — so no aspect bottlenecks the performance of scientific applications — will require dramatic advances in parallel I/O capabilities. These high performance computing systems will generate unprecedented amounts of data — and unprecedented challenges in managing it.

In this interview, Galen Shipman, who heads OLCF’s Technology Integration group, discusses the challenges of managing big data. Shipman’s group is charged with integrating advanced technologies in the areas of networks, file systems, and archival storage infrastructures into the OLCF’s HPC systems, such as Jaguar, a 2.33-petaflop Cray XT funded by the Department of Energy Office of Science. From the evolution of supercomputing architectures and storage systems to the development of software and tools, it is clear major change is afoot as computing accelerates toward the exascale.

HPCwire: What data management nightmares keep you up at night?

Shipman: In the short term, the challenge of providing high-performance, reliable, and scalable I/O systems to meet the needs of a growing number of users from broadening domains of science. Users of our computing systems often employ a number of different applications to support their research, often with distinct parallel I/O and data management requirements. As these applications are scaled to systems such as Jaguar, I/O techniques that may have worked adequately at a few thousand cores often prove unscalable at tens or even hundreds of thousands of cores. Educating users about how to optimize their I/O and assisting them in doing so has proved effective.

Once these applications are optimized for our computational environment, the next challenge, managing the data that these applications produce, comes to the forefront. The OLCF currently manages more than 220 million files and more than 22 petabytes of data stored across our high-performance Lustre file systems and our High Performance Storage System (HPSS) archive. Managing this “big-data” is truly a grand challenge and spans not only HPC environments but private industry as well.

To facilitate data management at this scale, the OLCF has developed a number of tools, including advanced search and discovery, metadata harvesting, parallel data movement, and system monitoring and administration for our large-scale data systems. While we have made great progress in this area, the sheer scope of this challenge will necessitate a multi-institutional approach that spans government research laboratories and private industry.

From a hardware technology perspective, a more fundamental challenge that the storage community is facing is the widening gap between the performance of traditional storage technologies and the amount of data they can store. Historically, disk bandwidth has improved by 20 percent for sequential I/O and 8 percent for random I/O annually, while disk drive densities have increased by nearly 50 percent per year. For traditional storage technologies, this trend is likely to continue, resulting in ever-increasing storage capacities with ever-dwindling accessibility of the data sets that reside on the storage media.

To illustrate the challenges we will face with traditional storage technologies, take for example the community-developed exascale roadmap. It points toward parallel I/O environments that support up to 60 terabytes per second of bandwidth to transfer data to and from persistent storage. Assuming historical rates of performance improvement continue, achieving this performance level with traditional storage technologies would require more than 200,000 disk drives with a projected cost of more than $200 million. Even at this level of investment, the performance delivered to most applications may be a small fraction of the goal as drive-latency improvements become increasingly harder to realize.

Within the HPC community, there is general consensus that, over the next decade, the traditional storage hierarchy will incorporate technologies such as non-volatile random access memory (NVRAM) to address this problem. Many point to flash-based storage as a likely candidate for inclusion in this deeper hierarchy, but I believe it is too early to make a call on a specific NVRAM technology. One thing is clear: A fundamental technological shift coupled with end-to-end optimization will be required to meet our performance requirements within a tractable price point.

HPCwire: What research enterprises are generating big data at the OLCF?

Shipman: The majority of our users — scientists and engineers in academia, government, and industry—come to us through the Innovative and Novel Computational Impact on Theory and Experiment (INCITE) program, through which Oak Ridge and Argonne national laboratories will deliver 1.7 billion processor hours on advanced supercomputers in 2011. While all OLCF users rely on a parallel I/O environment, a number of heavy hitters in terms of I/O have unique performance and data-management requirements. Of particular note are the climate science, fusion energy, and combustion communities.

Fusion energy simulations may generate up to 100 terabytes of data per day. This is really on the edge in terms of parallel I/O performance, dwarfing that of most other science domains. The climate science community is somewhat unique in terms of its need for structured data management to enable intermodel comparison. The Coupled Model Intercomparison Project Phase 5 will not only generate petabytes of data but also require sophisticated data management tools to facilitate scientific review and intercomparison of simulation results from climate modeling centers around the world.

As simulation continues to mature as a fundamental tool for scientific discovery, I expect other scientific domains to have similar data management requirements. The OLCF is well positioned to meet these requirements through its development of advanced data-management technologies.

HPCwire: How do you store such big data sets?

Shipman: Petascale datasets from all major OLCF platforms are accessible though our center-wide Spider file system, which has both short-term and longer-term storage for computational science teams. Spider is one of the world’s largest and fastest parallel I/O environments, with 10.7 petabytes of disk space and more than 240 gigabytes per second of aggregate throughput. In addition to providing a common center-wide file system to all of ORNL’s supercomputing platforms, the Spider parallel I/O environment provides connectivity to remote facilities such as Argonne National Laboratory and Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center via dedicated data-transfer nodes running GridFTP.

Most users generate large datasets as part of their scientific simulations either for improved resiliency through checkpoint/restart — saving the application state at predefined intervals — or for subsequent data analysis. Checkpoint/restart datasets are generally purged by the application itself or through a system-wide sweep of “expired” datasets and thus have a limited shelf life.

Datasets to support subsequent data analysis have longer-term value, particularly when a scientist is interested in analyzing and comparing the results of multiple simulations. The OLCF provides every INCITE project with a long-term archival area to store these datasets via HPSS.

HPCwire: What are the logistics of longer-term, or archival, storage?

Shipman: When users decide to move data from Spider into long-term storage, the data are archived on our HPSS, which offers tens of petabytes of capacity and more than 12 gigabytes per second of bandwidth. Incoming data are written to disk and later migrated to tape for long-term archiving in high-capacity, robotic libraries. As of June 2011, the OLCF has more than 18 petabytes stored in more than 20 million files in HPSS from researchers in climate science, nuclear fusion, combustion, astrophysics, materials science, and many other scientific fields. It is not uncommon for users from a single science domain to have well over a petabyte of data in archive. These datasets are of high value, often requiring millions of processor hours to generate. Recreating these datasets and the scientific insight they may provide is increasingly difficult as the demand from the scientific community for large-scale computing resources such as Jaguar is outpaces our ability to supply the needed processor hours.

For the foreseeable future, we expect tape technology to be in play for archival storage because of its low power consumption and relatively low cost per terabyte of capacity. While massive array of idle disk (MAID) technology will lower the power requirements of disk-only solutions, there are concerns with the reliability of drive technology compared with tape, particularly when drives are frequently spun-up/down as employed in MAID. I expect disk technologies will still be in play for quite some time as well, as the capacity and sequential bandwidth capabilities at a relatively competitive cost point will continue to make this technology compelling. For archiving, I expect disk and tape to continue to be the dominant storage technologies for scientific computing environments. Our primary challenge will be balancing the storage requirements of our users with the cost of storage capacity and bandwidth. In response, I expect the scientific computing community to increasingly adopt data-reduction strategies within their simulation and analysis workflows.

HPCwire: For analyzing and visualizing big data, what is the main challenge?

Shipman: From my experiences with the data-analysis and visualization communities, their primary bottleneck in time-to-solution is I/O performance. Our work with the climate community on the Ultrascale Visualization — Climate Data Analysis Tools (UV-CDAT) project is tackling this issue head-on from an end-to-end perspective. We are working on ultrascale visualization techniques with a number of researchers from Los Alamos and Lawrence Livermore national laboratories, New York University, and Kitware to optimize the infrastructure of the Visualization ToolKit, an open-source software system for 3D computer graphics, image processing, analysis, and visualization.

Early results are encouraging, resulting in linear speedup — scaling — of common visualization workloads on Jaguar using Spider. Through an end-to-end approach that encompassed system architecture experts, visualization researchers, and middleware software engineers, we are able to address one of the most challenging aspects of visualization workloads. I believe this coordinated multidisciplinary approach will be of increasing value as the I/O performance gap continues to widen in the future.

Another promising approach to overcoming the I/O performance bottleneck is “in situ” data analysis, moving from the traditional post-processing model of data analysis to integration of data analysis directly within the simulation. This strategy can result in significant data reduction for certain use cases, particularly when scientists are interested in well-understood phenomena within their simulations.

The OLCF provides a number of dedicated visualization and analysis resources to our computational scientists. Our 32-node Linux cluster, known as Lens, is a dedicated platform that enables analysis and visualization of simulation data generated on Jaguar, providing a conduit for large-scale scientific discovery.

The facility also features EVEREST, the Exploratory Visualization Environment for REsearch in Science and Technology, and its associated visualization cluster. EVEREST is 30 feet wide and 10 feet tall and features a 27-projector Powerwall to display 35 million pixels for extremely high-definition scientific visualizations. Coupled with a dedicated Lustre file system, EVEREST provides a compelling experience for scientists exploring data at extremely fine scales. ORNL’s Visualization Task Group, led by Sean Ahern, helps researchers gain a better understanding of their data through visualization techniques.

HPCwire: What are the big challenges in scalable I/O?

Shipman: There are a number of big challenges in scalable I/O, from the sheer component counts to support hundreds of gigabytes per second of bandwidth and the ever-growing demands on the scalability of the file system software, to the scalability of middleware libraries and the applications that use them. Over the past decade we have seen dramatic increases in the number of components deployed both within our archive and on our high-performance parallel I/O systems. The OLCF currently has more than 24,000 disk drives supporting our high-performance parallel I/O and archival systems with nearly 500 gigabytes per second of bandwidth and more than 15 petabytes of capacity. System reliability and resiliency are critical at this scale. We use a number of techniques to improve the reliability and resiliency of our systems, from hardware-level redundancy to advanced software-level resiliency. An eye toward engineering reliability into our large-scale systems has allowed us to maintain extremely high availability of these critical resources.

Over the past decade we have seen dramatic increases in the scale of the computational platforms that we deploy at the OLCF. Today we support more than 25,000 file system clients on our largest-scale parallel file systems, an increase of an order of magnitude since 2005. Much of our work in providing a scalable parallel I/O environment has been focused on the Lustre file system. Through collaborative development of the open-source Lustre file system with Cluster File Systems, Sun, Oracle, and most recently Whamcloud, we have successfully supported the OLCF’s transition from teraflop to petaflop computing over this relatively short time frame. Our efforts in forming Open Scalable File Systems, a non-profit mutual benefit organization, are aimed at building upon our successes in collaborative development to meet similar challenges in the future.

Although we have made significant strides in improving the scalability and resiliency of the underlying file and storage systems, achieving optimal performance and scalability at the application level can be elusive to all but a handful of I/O experts. To bridge this gap, we employ a number of I/O middleware technologies, including the Adaptable I/O System (ADIOS), HDF-5, MPI-I/O, NetCDF, Parallel NetCDF, and the Parallel Log-structured File System (PLFS). These I/O middleware technologies serve a variety of functions from structured data models as found in NetCDF and HDF-5, to I/O transformation techniques as found in ADIOS and PLFS. Much of the development of ADIOS is conducted here at ORNL and led by Scott Klasky, a member of our Scientific Computing Group. Using these I/O middleware technologies in concert with our large-scale parallel I/O environment, we have delivered unprecedented levels of I/O performance to our end-users.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

2024 Winter Classic: Texas Two Step

April 18, 2024

Texas Tech University. Their middle name is ‘tech’, so it’s no surprise that they’ve been fielding not one, but two teams in the last three Winter Classic cluster competitions. Their teams, dubbed Matador and Red Read more…

2024 Winter Classic: The Return of Team Fayetteville

April 18, 2024

Hailing from Fayetteville, NC, Fayetteville State University stayed under the radar in their first Winter Classic competition in 2022. Solid students for sure, but not a lot of HPC experience. All good. They didn’t Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use of Rigetti’s Novera 9-qubit QPU. The approach by a quantum Read more…

2024 Winter Classic: Meet Team Morehouse

April 17, 2024

Morehouse College? The university is well-known for their long list of illustrious graduates, the rigor of their academics, and the quality of the instruction. They were one of the first schools to sign up for the Winter Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire