The Convergence of Big Data and Extreme-Scale HPC

By Rob Farber

August 31, 2018

As we are heading towards extreme-scale HPC coupled with data intensive analytics like machine learning, the necessary integration of big data and HPC is a current hot topic of research that is, as Rashid Mehmood notes, “still in its infancy”.[i] Mehmood is the Research Professor of Big Data Systems and the Director for Research, Training and Consultancy at the High Performance Computing Centre, King Abdulaziz University (KAU) in Saudi Arabia.

A driving force to incorporate big data into HPC, Mehmood observed in his presentation at the first Middle East meeting of the Intel Extreme Performance Users Group at KAUST (King Abdullah University of Science and Technology) that, “Increasingly more data is being produced by scientific experiments from areas such as bioscience, physics, and climate, and therefore, HPC needs to adopt data-driven paradigms.”

Mehmood is not alone in his observation. Over the past four years the Big Data and Exascale Computing (BDEC) project organized a series of five international workshops that explored ways in which new forms of data-centric discovery might be integrated with the established, simulation-centric paradigm of the high performance computing (HPC) community. [ii]

Looking toward the future of cyberinfrastructure for science and engineering, BDEC produced a whitepaper that highlights the critical problems involved in the diverse patterns of when, where, and how data is to be produced, transformed, shared, and analyzed.  We view the main points of the BDEC whitepaper in light of current efforts in the HPC community, such as the Wrangler data analytics supercomputer at the Texas Advanced Computing Center (TACC), the Argonne lab-wide data service, and data management efforts at NERSC.

Understanding the bifurcation between the two software ecosystems

Comparing HPC to High-end Data Analysis (HDA) people use a different vernacular and focus on different key concepts.

Those who work in HDA speak of the 4Vs of big data which are: volume (scale of the data), velocity (speed of intake particularly with streaming data), variety (different forms of data), and veracity (the uncertainty of the data). Meanwhile HPC scientists tend to speak in terms of performance, scaling, and the power efficiency of a computation.

This difference in focus is reflected in the representative big data and HPC software stacks as summarized by Reed and Dongarra. [iii]

Figure 1: Different software ecosystems for high-end Data Analytics and for traditional computational science stacks (Image source: BDEC white paper)

The BDEC committee attributes this bifurcation in software stacks to the natural evolution of the two separate communities (e.g. scientists vs. academics and commercial software developers) working to address their separate problem domains.

Working over the past four decades, the HPC scientific community focused in increasing the ability of scientists to model and simulate using numerical models. Meanwhile, the data analytics ecosystem has been rapidly developed over the past fifteen to process the torrents of business, industrial process, and social network data now being generated by consumer devices and the burgeoning Internet of Things. For the most part, the data analytics software ecosystem was not developed by the scientific computing community as they work to adapt to the massive increases in data that is being produced by new instruments and sensor systems.

Both paradigms are collapsing from the data deluge

The BDEC whitepaper observes that both HPC and HDA workflows are eroding, if not collapsing under the onslaught of an apparently ever-growing data deluge[iv]. The future, they advocate, is to stop thinking in terms of a “big machine” but rather focus on the many unsolved problems surrounding wide-area, multi-stage workflows.

Figure 2: Current problem of data logistics: The highest concentrations of computing power and storage are in the “center” (i.e., in commercial clouds or HPC Centers), but much of the rapid increase in data volumes and the dramatic proliferation of data generators is occurring in edge environments. (Image source: BDEC whitepaper)

Such workflows represent a remarkable reversal in thinking about data, where the issue is not connecting the edge via “the last mile”. Instead, these workflows present a multidimensional “first mile problem” that is not currently addressed by either cloud-based HDA or on-premises based HPC solutions.  The BDEC whitepaper authors state, “Arguably, the main cyberinfrastructure challenge of the Big Data era is to adapt or replace the legacy paradigm with a new type of distributed services platform (DSP), one that combines computing, communication, and buffer/storage resources in a data processing network that is far more integrated than anything hitherto available”.

Current efforts to address the HPC data challenge

Figure 3: The general problem with multiple high volume generators at the edge: Edge environments (i.e., across network from the centralized facilities) are, or soon will be, experiencing unprecedented increases of data rates from diverse and rapidly proliferating sources. (Image source: BDEC whitepaper)

Both vendors and the HPC community are working to address the big data challenge in a variety of ways – especially with the general acceptance of AI and its dependence on large data sets. One example is how Intel is working with the ecosystems to develop a reference platform to guide the development of future infrastructure to leverage the growing data and the power of HPC supercomputers.

Academic projects such as the ones listed below have shown remarkable success and have provided valuable “lessons learned” to the HPC community.

The Argonne lab-wide data service

At Argonne National Laboratory, researchers are preparing for the exascale era by exploring ways to improve collaboration, eliminate barriers to using next-generation systems like Aurora, and facilitate seamless workflows.

In one example, a team at Argonne’s Data Science and Learning Division is developing a lab-wide service that will make it easier to access, share, analyze, and reuse large-scale datasets.

“Our motivation,” Ian Foster (Argonne Data Science and Learning Division Director and Distinguished Fellow) explains, “is to create increasingly rich data services so people don’t just come to the ALCF for simulation but for simulation and data-centric activities.” Foster also observes that, “It’s becoming increasingly impractical for supercomputing facility users to move their data to their home institution’s system for analysis”.

Aimed at enabling more effective data capture and discovery, as well as association of machine learning models with data collections for improved reproducibility and simpler deployment at scale, the service leverages well-known tools including Globus for research data management and the Argonne’s Petrel storage system.

TACC Wrangler

The Texas Advanced Computing Center (TACC) Wrangler supercomputer is the first of its kind and the most powerful data analysis system allocated in the Extreme Science and Engineering Discovery Environment (XSEDE). [v]

The system is designed to support HDA in an HPC environment. It provides around a half a petabyte (0.5 PB) high speed flash storage system that can be used to handle data analysis and processing workflows not practical on other systems. TACC notes, “Wrangler’s unique architecture handles the many aspects of the volume, velocity, and variety that can make digital data research difficult to handle on standard high performance systems”. [vi]

Very importantly, the system is dynamically provisioned by the users to handle different data workflows, including databases (both relational database systems and the newer noSQL style databases), Hadoop/HDFS based workflows (including MapReduce and Spark), and more custom workflows leveraging the flash-based parallel file system.

The success of Wrangler can be seen in the several hundred projects in the TACC Wrangler Data Portal that range from Advanced 3D Microscopy to a Zebrafish map that identifies recessive mutations in Zebrafish.

Recent research shows TACC at the forefront of deep-learning with a new algorithm that speeds training on the Stampede 2 supercomputer so it only take 11 minutes to train ImageNet.

Addressing the challenge of the two paradigm splits

The end goal, according to the BDEC whitepaper is to, “define a new, common and open Distributed Services Platform (DSP), one that offers programmable access to shared processing, storage and communication resources, and that can serve as a universal foundation for the component interoperability that novel services and applications will require”.[vii]

The following schematic reflects this vision.

Figure 4: Design pattern for a converged HPC and HDA future[viii] [ix] (image courtesy KAUST)
As the future recipient of the nation’s first exascale supercomputer, Argonne National Laboratory is particularly vested in taking a leadership role in testing the wide-area, multi-stage workflows recommended by the BDEC whitepaper. The Argonne Petrel project appears to be a good start. In particular, the ability to ingest data from instruments and simulation as well as collaborate and publish data regardless of the size of the data set are particularly valuable. An experimental effort using Kubernetes containers may help to democratize the software stack as well as data by providing HDA and HPC convergence through applications containers. The ability to dynamically provision the machine is a “lesson learned” from TACC.

Summary

It makes sense to cross-fertilize as much as possible between the HDA and HPC software stacks for big data while looking ahead to an even bigger data future. There is much to be gained as we know that big data is here to stay and exascale supercomputers will certainly play an essential role in helping scientists use this data to make ground-breaking scientific discoveries.

Rob Farber is a global technology consultant and author with an extensive background in HPC and in developing machine learning technology that he applies at national labs and commercial organizations. Rob can be reached at [email protected]

[i]    Usman S., Mehmood R., Katib I. (2018) Big Data and HPC Convergence: The Cutting Edge and Outlook. In: Mehmood R., Bhaduri B., Katib I., Chlamtac I. (eds) Smart Societies, Infrastructure, Technologies and Applications. SCITA 2017. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 224, pp. 11–26. Springer, Cham. DOI: https://doi.org/10.1007/978-3-319-94180-6_4

[ii] See http://www.exascale.org/bdec/ and specifically the report which can be downloaded here: http://www.exascale.org/bdec/sites/www.exascale.org.bdec/files/whitepapers/bdec_pathways.pdf.

[iii] The freely available BDEC whitepaper credits Reed and Dongarra citing Daniel A. Reed and Jack Dongarra. Exascale computing and big data. Commun. ACM, 58(7):56–68, June 2015. ISSN 0001-0782. doi: 10.1145/2699414. URL http://doi.acm.org/10.1145/2699414.

[iv] ibid

[v] http://www.dailytexanonline.com/2016/05/04/new-tacc-supercomputer-wrangles-big-data

[vi] https://portal.wrangler.tacc.utexas.edu/

[vii] http://www.exascale.org/bdec/sites/www.exascale.org.bdec/files/whitepapers/bdec_pathways.pdf.

[viii] Usman S., Mehmood R., Katib I. (2018) Big Data and HPC Convergence: The Cutting Edge and Outlook. In: Mehmood R., Bhaduri B., Katib I., Chlamtac I. (eds) Smart Societies, Infrastructure, Technologies and Applications. SCITA 2017. Lecture Notes of the Institute for Computer Sciences, Social Informatics and Telecommunications Engineering, vol 224, pp. 11–26. Springer, Cham. DOI: https://doi.org/10.1007/978-3-319-94180-6_4

[ix] Sardar Usman, Rashid Mehmood and Iyad Katib HPC & Big Data Convergence: The Cutting Edge & Outlook Poster presented at the first Middle East meeting of the Intel Extreme Performance Users Group, Intel IXPUG, KAUST, April 2018 https://epostersonline.com/ixpug-me2018/node/19

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Is Amazon’s Plunge into Server Chips a Watershed Moment?

December 11, 2018

For several years now the big cloud providers – Amazon, Microsoft Azure, Google, et al – have been transforming from technology consumers into technology creators in hardware and software. The most recent example bei Read more…

By John Russell

Mellanox Uses Univa to Extend Silicon Design HPC Operation to Azure

December 11, 2018

Call it a corollary to Murphy’s Law: When a system is most in demand, when end users are most dependent on the system performing as required, when it’s crunch time – that’s when the system is most likely to blow up. Or make you wait in line to use it. Read more…

By Doug Black

Clemson’s Cautionary Cryptomining Tale

December 11, 2018

In some ways, the bigger the computer, the more vulnerable it is to cryptomining as Clemson University discovered after cryptominers dug into its Palmetto supercomputer. When a number of nodes on Clemson University’s P Read more…

By Staff

HPE Extreme Performance Solutions

AI Can Be Scary. But Choosing the Wrong Partners Can Be Mortifying!

As you continue to dive deeper into AI, you will discover it is more than just deep learning. AI is an extremely complex set of machine learning, deep learning, reinforcement, and analytics algorithms with varying compute, storage, memory, and communications needs. Read more…

IBM Accelerated Insights

Blurring the Lines Between HPC and AI @ SC18

The dominant topic at SC18 was the convergence of HPC and Artificial Intelligence (AI) with some of the biggest research and enterprise HPC users providing perspectives on how HPC and AI are moving closer together. Read more…

Data West Brings Technology Leaders to SDSC

December 6, 2018

Data and technology enthusiasts from around the world descended upon the San Diego Supercomputing Center (SDSC) for the third annual Data West conference, which is taking place this week on the campus of the University o Read more…

By Alex Woodie

Topology Can Help Us Find Patterns in Weather

December 6, 2018

Topology--–the study of shapes-- seems to be all the rage. You could even say that data has shape, and shape matters. Shapes are comfortable and familiar conc Read more…

By James Reinders

Zettascale by 2035? China Thinks So

December 6, 2018

Exascale machines (of at least a 1 exaflops peak) are anticipated to arrive by around 2020, a few years behind original predictions; and given extreme-scale performance challenges are not getting any easier, it makes sense that researchers are already looking ahead to the next big 1,000x performance goal post: zettascale computing. Read more…

By Tiffany Trader

Robust Quantum Computers Still a Decade Away, Says Nat’l Academies Report

December 5, 2018

The National Academies of Science, Engineering, and Medicine yesterday released a report – Quantum Computing: Progress and Prospects – whose optimism about Read more…

By John Russell

Revisiting the 2008 Exascale Computing Study at SC18

November 29, 2018

A report published a decade ago conveyed the results of a study aimed at determining if it were possible to achieve 1000X the computational power of the the Read more…

By Scott Gibson

AWS Debuts Lustre as a Service, Accelerates Data Transfer

November 28, 2018

From the Amazon re:Invent main stage in Las Vegas today, Amazon Web Services CEO Andy Jassy introduced Amazon FSx for Lustre, citing a growing body of applicati Read more…

By Tiffany Trader

AWS Launches First Arm Cloud Instances

November 28, 2018

AWS, a macrocosm of the emerging high-performance technology landscape, wants to be everywhere you want to be and offer everything you want to use (or at least Read more…

By Doug Black

Move Over Lustre & Spectrum Scale – Here Comes BeeGFS?

November 26, 2018

Is BeeGFS – the parallel file system with European roots – on a path to compete with Lustre and Spectrum Scale worldwide in HPC environments? Frank Herold Read more…

By John Russell

DOE Under Secretary for Science Paul Dabbar Interviewed at SC18

November 21, 2018

During the 30th annual SC conference in Dallas last week, SC18 hosted U.S. Department of Energy Under Secretary for Science Paul M. Dabbar. In attendance Nov. 13-14, Dabbar delivered remarks at the Top500 panel, met with a number of industry stakeholders and toured the show floor. He also met with HPCwire for an interview, where we discussed the role of the DOE in advancing leadership computing. Read more…

By Tiffany Trader

Quantum Computing Will Never Work

November 27, 2018

Amid the gush of money and enthusiastic predictions being thrown at quantum computing comes a proposed cold shower in the form of an essay by physicist Mikhail Read more…

By John Russell

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

IBM at Hot Chips: What’s Next for Power

August 23, 2018

With processor, memory and networking technologies all racing to fill in for an ailing Moore’s law, the era of the heterogeneous datacenter is well underway, Read more…

By Tiffany Trader

House Passes $1.275B National Quantum Initiative

September 17, 2018

Last Thursday the U.S. House of Representatives passed the National Quantum Initiative Act (NQIA) intended to accelerate quantum computing research and developm Read more…

By John Russell

Summit Supercomputer is Already Making its Mark on Science

September 20, 2018

Summit, now the fastest supercomputer in the world, is quickly making its mark in science – five of the six finalists just announced for the prestigious 2018 Read more…

By John Russell

CERN Project Sees Orders-of-Magnitude Speedup with AI Approach

August 14, 2018

An award-winning effort at CERN has demonstrated potential to significantly change how the physics based modeling and simulation communities view machine learni Read more…

By Rob Farber

AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

Leading Solution Providers

SC 18 Virtual Booth Video Tour

Advania @ SC18 AMD @ SC18
ASRock Rack @ SC18
DDN Storage @ SC18
HPE @ SC18
IBM @ SC18
Lenovo @ SC18 Mellanox Technologies @ SC18
NVIDIA @ SC18
One Stop Systems @ SC18
Oracle @ SC18 Panasas @ SC18
Supermicro @ SC18 SUSE @ SC18 TYAN @ SC18
Verne Global @ SC18

TACC’s ‘Frontera’ Supercomputer Expands Horizon for Extreme-Scale Science

August 29, 2018

The National Science Foundation and the Texas Advanced Computing Center announced today that a new system, called Frontera, will overtake Stampede 2 as the fast Read more…

By Tiffany Trader

HPE No. 1, IBM Surges, in ‘Bucking Bronco’ High Performance Server Market

September 27, 2018

Riding healthy U.S. and global economies, strong demand for AI-capable hardware and other tailwind trends, the high performance computing server market jumped 28 percent in the second quarter 2018 to $3.7 billion, up from $2.9 billion for the same period last year, according to industry analyst firm Hyperion Research. Read more…

By Doug Black

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

Germany Celebrates Launch of Two Fastest Supercomputers

September 26, 2018

The new high-performance computer SuperMUC-NG at the Leibniz Supercomputing Center (LRZ) in Garching is the fastest computer in Germany and one of the fastest i Read more…

By Tiffany Trader

Houston to Field Massive, ‘Geophysically Configured’ Cloud Supercomputer

October 11, 2018

Based on some news stories out today, one might get the impression that the next system to crack number one on the Top500 would be an industrial oil and gas mon Read more…

By Tiffany Trader

Intel Confirms 48-Core Cascade Lake-AP for 2019

November 4, 2018

As part of the run-up to SC18, taking place in Dallas next week (Nov. 11-16), Intel is doling out info on its next-gen Cascade Lake family of Xeon processors, specifically the “Advanced Processor” version (Cascade Lake-AP), architected for high-performance computing, artificial intelligence and infrastructure-as-a-service workloads. Read more…

By Tiffany Trader

Google Releases Machine Learning “What-If” Analysis Tool

September 12, 2018

Training machine learning models has long been time-consuming process. Yesterday, Google released a “What-If Tool” for probing how data point changes affect a model’s prediction. The new tool is being launched as a new feature of the open source TensorBoard web application... Read more…

By John Russell

The Convergence of Big Data and Extreme-Scale HPC

August 31, 2018

As we are heading towards extreme-scale HPC coupled with data intensive analytics like machine learning, the necessary integration of big data and HPC is a curr Read more…

By Rob Farber

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This