Say Hello to the New ADIOS

By Gregory Scott Jones

September 12, 2010

Updated I/O performance library improves ease of use and achieves even better performance

Big machines are one thing. Taking advantage of their full potential is quite another. Application performance has long been trailing hardware as supercomputers have sought, entered, and now surpassed petaflop performance.

One of the factors commonly affecting application performance is input/output (I/O). Researchers regularly find themselves having to choose between the performance of their applications and the amount and quality of the data they write.

It’s a problem familiar to the Oak Ridge Leadership Computing Facility’s (OLCF’s) Scott Klasky from his early years as a researcher with a team from Princeton Plasma Physics Laboratory using the Gyrokinetic Toroidal Code.

“We looked at the performance of how often we would like to write, and we were spending over 30 percent of the time writing the analysis files in a very popular file format. Thirty percent of all your computational time writing data to files is too much,” said Klasky. “The scientists eventually decided that unless it was a run that we definitely wanted to get some visualization out of, we weren’t going to write those because we were wasting our valuable computing time doing this.”

Klasky, along with a team of researchers (Qing Liu, Norbert Podhorszki, Jay Lofstead, Hasan Abbasi, Ron Oldfield, Matt Wolf, Fang Zheng, Ciprian Docan, Manish Parashar, Weikuan Yu, Yuan Tian, Nagiza Samatova, Sriram Lakshminarasimh, Todd Kordenbrock, and others) from Georgia Tech, the OLCF, Rutgers University, and Sandia National Laboratories are the developers of ADIOS, an open-source middleware with the primary goal of making the process of getting information in and out of a supercomputer easier and more effective.

Last week the team released ADIOS 1.2, the latest incarnation of one of computational science’s most effective I/O tools. So far ADIOS has helped researchers make huge strides in fusion, astrophysics and combustion. The new version features some interesting improvements that will doubtless aid researchers in taking full advantage of leading supercomputing platforms.

For starters, previous versions of ADIOS had users construct an external XML file that allowed them to organize their simulation variables into distinct groups and add important metadata to their output. With the new application programming interface (API), which allows for interaction between different software packages, users can now place the APIs directly into their code and interactively construct new variables during run time. This was especially important for adaptive mesh refinement (AMR) codes, such as Chombo, that can alter the variables placed on disk during run time. This new API makes ADIOS much more flexible and allows researchers to choose between defining the output in an external file for maximum flexibility or in their codes.

ADIOS also features a custom I/O method that writes data to subfiles and aggregates it into larger pieces for maximum performance on the leadership-class systems. This method has been shown to get near peak I/O performance for many codes, particularly S3D, on the Cray XT5 and Cray XT4 at the OLCF and Lawrence Berkeley National Laboratory’s National Energy Research Scientific Computing Center.

“We are now able to speed up applications such as S3D to near-peak I/O bandwidth through simple and easy-to-use ADIOS APIs,” said Qing Liu, a member of the ADIOS team at the OLCF. “We are also able to speed up S3D by a factor of more than 15. This is achieved by intelligently aggregating and writing data to storage targets in ADIOS.”

Now users who run on large systems can switch from running on P-processors and writing to P-files — or one file or M-files, transparently. ADIOS users can switch to the best method for individual systems, including the IBM Blue Gene/P at Argonne National Laboratory, where PhD student Yuan Tian, along with her advisor Weikuan Yu at Auburn University, has created a custom method to write more efficiently with ADIOS.

Version 1.2 also features further support for self-describing data in the output. Users can now write more statistics into their data and have more flexibility in their output. For example, users can automatically retrieve the average value, minimum, maximum, and standard deviation for all arrays at negligible computational cost. This feature allows users to take large files (terabytes) and automatically determine these parameters in less than 2 seconds when listing the contents of the data. Furthermore, users can get these statistics for each independent time step in the output.

Finally, version 1.2 features some new asynchronous transport methods, allowing even faster I/O. The trick is scheduling. I/O uses the network bandwidth, and by taking advantage of the downtime during communication between processors, researchers “can essentially get I/O for free,” said Klasky.

For example, both the DataTap and the Network Scalable Service Interface (NSSI) methods, from Georgia Tech and Sandia Labs respectively, send data to a user defined set of nodes (a staging area) and writes the data from these nodes, reducing the performance linkage between the file system and the application. Furthermore, the DataSpace method from Rutgers creates a PGAS environment in the staging area so that independently compiled codes with ADIOS can be used as services to efficiently couple them together.

“The focus for this release is broader compatibility and user convenience. The introduction of the API calls to replace the XML file addresses long-standing requests from a small but vocal part of our user community,” said team member Jay Lofstead. “The AMR-focused enhancements broaden the classes of application that can use ADIOS while maintaining 100 percent backward compatibility. Some additional changes smooth the user experience.”

Taken separately, all of ADIOS’s individual improvements represent significant advances toward more efficient simulations. Taken together, they embody a major innovation in the way computational science will be conducted.

“Working with Scott Klasky and his team has moved our research and our software, such as DataTap and Data Staging, from being interesting research prototypes to becoming artifacts that address the real needs of petascale simulations,” said Georgia Tech team member Karsten Schwan. “By then also interacting with the fusion, astrophysics and combustion modeling communities, we have not only found ways to alleviate their problems with I/O at scale, but we have also gained valuable information about ways to better organize data and quickly analyze it to help scientists understand the behavior of their petascale codes and gain the scientific insights they seek.”

There are few foreseeable limits to ADIOS’s potential. As it is expanded to additional platforms, simulating big science will become correspondingly simpler, allowing researchers to concentrate more on their results than the technical aspects of their simulations. And as high-performance computing becomes an increasingly powerful research tool, there will be no shortage of grateful scientists.

For more information on ADIOS and/or to download the source, check out the project’s Web page.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Exascale Computing Project Names Doug Kothe as Director

September 20, 2017

The Department of Energy’s Exascale Computing Project (ECP) has named Doug Kothe as its new director effective October 1. He replaces Paul Messina, who is stepping down after two years to return to Argonne National L Read more…

Takeaways from the Milwaukee HPC User Forum

September 19, 2017

Milwaukee’s elegant Pfister Hotel hosted approximately 100 attendees for the 66th HPC User Forum (September 5-7, 2017). In the original home city of Pabst Blue Ribbon and Harley Davidson motorcycles the agenda addresse Read more…

By Merle Giles

NSF Awards $10M to Extend Chameleon Cloud Testbed Project

September 19, 2017

The National Science Foundation has awarded a second phase, $10 million grant to the Chameleon cloud computing testbed project led by University of Chicago with partners at the Texas Advanced Computing Center (TACC), Ren Read more…

By John Russell

HPE Extreme Performance Solutions

HPE Prepares Customers for Success with the HPC Software Portfolio

High performance computing (HPC) software is key to harnessing the full power of HPC environments. Development and management tools enable IT departments to streamline installation and maintenance of their systems as well as create, optimize, and run their HPC applications. Read more…

NERSC Simulations Shed Light on Fusion Reaction Turbulence

September 19, 2017

Understanding fusion reactions in detail – particularly plasma turbulence – is critical to the effort to bring fusion power to reality. Recent work including roughly 70 million hours of compute time at the National E Read more…

Exascale Computing Project Names Doug Kothe as Director

September 20, 2017

The Department of Energy’s Exascale Computing Project (ECP) has named Doug Kothe as its new director effective October 1. He replaces Paul Messina, who is s Read more…

Takeaways from the Milwaukee HPC User Forum

September 19, 2017

Milwaukee’s elegant Pfister Hotel hosted approximately 100 attendees for the 66th HPC User Forum (September 5-7, 2017). In the original home city of Pabst Blu Read more…

By Merle Giles

Kathy Yelick Charts the Promise and Progress of Exascale Science

September 15, 2017

On Friday, Sept. 8, Kathy Yelick of Lawrence Berkeley National Laboratory and the University of California, Berkeley, delivered the keynote address on “Breakt Read more…

By Tiffany Trader

DARPA Pledges Another $300 Million for Post-Moore’s Readiness

September 14, 2017

The Defense Advanced Research Projects Agency (DARPA) launched a giant funding effort to ensure the United States can sustain the pace of electronic innovation vital to both a flourishing economy and a secure military. Under the banner of the Electronics Resurgence Initiative (ERI), some $500-$800 million will be invested in post-Moore’s Law technologies. Read more…

By Tiffany Trader

IBM Breaks Ground for Complex Quantum Chemistry

September 14, 2017

IBM has reported the use of a novel algorithm to simulate BeH2 (beryllium-hydride) on a quantum computer. This is the largest molecule so far simulated on a quantum computer. The technique, which used six qubits of a seven-qubit system, is an important step forward and may suggest an approach to simulating ever larger molecules. Read more…

By John Russell

Cubes, Culture, and a New Challenge: Trish Damkroger Talks about Life at Intel—and Why HPC Matters More Than Ever

September 13, 2017

Trish Damkroger wasn’t looking to change jobs when she attended SC15 in Austin, Texas. Capping a 15-year career within Department of Energy (DOE) laboratories, she was acting Associate Director for Computation at Lawrence Livermore National Laboratory (LLNL). Her mission was to equip the lab’s scientists and research partners with resources that would advance their cutting-edge work... Read more…

By Jan Rowell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

MIT-IBM Watson AI Lab Targets Algorithms, AI Physics

September 7, 2017

Investment continues to flow into artificial intelligence research, especially in key areas such as AI algorithms that promise to move the technology from speci Read more…

By George Leopold

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Leading Solution Providers

IBM Clears Path to 5nm with Silicon Nanosheets

June 5, 2017

Two years since announcing the industry’s first 7nm node test chip, IBM and its research alliance partners GlobalFoundries and Samsung have developed a proces Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

GlobalFoundries: 7nm Chips Coming in 2018, EUV in 2019

June 13, 2017

GlobalFoundries has formally announced that its 7nm technology is ready for customer engagement with product tape outs expected for the first half of 2018. The Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This