Visit additional Tabor Communication Publications
December 06, 2012
Every few years, the relevance of the TOP500 list is called into question. Community insiders understand that this single-metric test is not representative of most supercomputing workloads. Still no open science system had ever opted out of the list until NCSA made the bold move to keep Blue Waters' Linpack results private.
In an interview with HPCwire, Blue Waters Project Director Bill Kramer provided a detailed account of the NCSA's decision-making process and what they perceive to be the list's flaws. While some may consider their stance controversial, it has nevertheless sparked a dialogue, which is what Kramer was hoping for.
In response to that interview, we received a letter from Council on Competitiveness President and CEO Deborah L. Wince-Smith offering an alternative point-of-view. We've published that letter below (edited only for style).
The Council on Competitiveness has long championed high-performance computing (HPC) as a basis for increased innovation and productivity, and as a competitive advantage for the United States in the global marketplace. We closely follow developments in the advancement of high-end supercomputing; hence are somewhat concerned by the comments made by Blue Waters Project Director Bill Kramer in his recent interview with HPCwire, November 16th, 2012 ("Blue Waters Opts Out of TOP500").
The larger HPC community and we agree the Linpack benchmark used as the metric for the TOP500 list is limited as a measure of overall HPC system performance, but we came away from reading the article with the conclusion that those who choose to "opt in" to the TOP500 are less interested in sustained performance on mission critical applications than they are in achieving a high Linpack result.
From our external vantage point, sustained performance on scientific and engineering applications is clearly central to U.S. Department of Energy (DOE) academic and research missions, and is the primary driver for the development of ever more capable HPC platforms and solutions. As an example, we were pleased to see teams from Argonne and Lawrence Livermore each achieve truly astonishing performance of over 11 petaflops sustained on the 20 petaflops peak Sequoia system (TOP500 #2) on two very challenging, scientific applications, which HPCwire published on November 28th, 2012 ("DOE Labs Set Records with IBM Blue Gene/Q").
From the Council's perspective, the idea of the Sustained Petascale Performance test is interesting, especially if we had specifics on how this might be applied to the larger HPC community and industry in particular. Ongoing discussions on how best to evolve the standard that TOP500 represents are welcome, but at the same time, we should not deprive the community of this very important benchmark.
The fleeting glory of a high TOP500 ranking is not what motivates the HPC community, but these benchmarks do focus attention and conversation on how we are doing and where we might improve, particularly in light of growing off-shore competition for HPC leadership.
Deborah L. Wince-Smith
President and CEO
Council on Competitiveness
Contributing commentator, Andrew Jones, offers a break in the news cycle with an assessment of what the national "size matters" contest means for the U.S. and other nations...
Today at the International Supercomputing Conference in Leipzing, Germany, Jack Dongarra presented on a proposed benchmark that could carry a bit more weight than its older Linpack companion. The high performance conjugate gradient (HPCG) concept takes into account new architectures for new applications, while shedding the floating point....
Not content to let the Tianhe-2 announcement ride alone, Intel rolled out a series of announcements around its Knights Corner and Xeon Phi products--all of which are aimed at adding some options and variety for a wider base of potential users across the HPC spectrum. Today at the International Supercomputing Conference, the company's Raj....
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?
Join our webinar to learn how IT managers can migrate to a more resilient, flexible and scalable solution that grows with the data center. Mellanox VMS is future-proof, efficient and brings significant CAPEX and OPEX savings. The VMS is available today.