Visit additional Tabor Communication Publications
November 18, 2010
Results released at Supercomputing Conference 2010
ALBUQUERQUE, NM, Nov. 18 -- Nine supercomputers have been tested, validated and ranked by the new "Graph500" challenge, first introduced this week by an international team led by Sandia National Laboratories. The list of submitters and the order of their finish was released Nov. 17 at the supercomputing conference SC10 meeting in New Orleans.
The machines were tested for their ability to solve complex problems involving random-appearing graphs, rather than for their speed in solving a basic numerical problem, today's popular method for ranking top systems.
"Some, whose supercomputers placed very highly on simpler tests like the Linpack, also tested them on the Graph500, but decided not to submit results because their machines would shine much less brightly," said Sandia computer scientist Richard Murphy, a lead researcher in creating and maintaining the test.
Murphy developed the Graph500 Challenge with researchers at the Georgia Institute of Technology, University of Illinois at Urbana-Champaign, and Indiana University, among others.
Complex problems involving huge numbers of related data points are found in the medical world where large numbers of medical entries must be correlated, in the analysis of social networks with their huge numbers of electronically related participants, or in international security where huge numbers of containers on ships roaming the world and their ports of call must be tracked.
Such problems are solved by creating large, complex graphs with vertices that represent the data points -- say, people on Facebook -- and edges that represent relations between the data points -- say, friends on Facebook. These problems stress the ability of computing systems to store and communicate large amounts of data in irregular, fast-changing communication patterns, rather than the ability to perform many arithmetic operations. The Graph500 benchmarks are indicative of the ability of supercomputers to handle such complex problems.
The Graph500 benchmarks present problems in different input sizes. These are described as huge, large, medium, small, mini and toy. No machine proved capable of handling problems in the huge or large categories.
"I consider that a success," Murphy said. "We posed a really hard challenge and I think people are going to have to work to do 'large' or 'huge' problems in the available time." More memory, he said, might help.
The abbreviations "GE/s" and "ME/s" represented in the table below describe each machine's capabilities in giga-edges per second and mega-edges per second -- a billion and million edges traversed in a second, respectively.
Competitors were ranked first by the size of the problem attempted and then by edges per second.
The rankings were:
Rank #1 – Intrepid, Argonne National Laboratory – 6.6 GE/s on scale 36 (Medium)
Rank #2 – Franklin, National Energy Research Scientific Computing Center – 5.22 GE/s on Scale 32 (Small)
Rank #3 – cougarxmt, Pacific Northwest National Laboratory – 1.22 GE/s on Scale 29 (Mini)
Rank #4 – graphstorm, Sandia National Laboratories' – 1.17 GE/s on Scale 29 (Mini)
Rank #5 – Endeavor, Intel Corporation, 533 ME/s on Scale 29 (Mini)
Rank #6 – Erdos, Oak Ridge National Laboratory – 50.5 ME/s on Scale 29 (Mini)
Rank #7 – Red Sky, Sandia National Laboratories – 477.5 ME/s on Scale 28 (Toy++)
Rank #8 – Jaguar, Oak Ridge National Laboratory – 800 ME/s on Scale 27 (Toy+)
Rank #9 – Endeavor, Intel Corporation – 615.8 ME/s on Scale 26 (Toy)
A more detailed description of the Graph500 benchmark and additional results are available at graph500.org. Any organization may participate in the ratings. The next Graph500 Challenge list is expected to be released at the International Supercomputing Conference 2011 next summer, and then at SC 2011 again in the fall.
About Sandia National Laboratories
Sandia National Laboratories is a multiprogram laboratory operated and managed by Sandia Corporation, a wholly owned subsidiary of Lockheed Martin Corporation, for the U.S. Department of Energy's National Nuclear Security Administration. With main facilities in Albuquerque, NM, and Livermore, Calif., Sandia has major R&D responsibilities in national security, energy and environmental technologies, and economic competitiveness.
Source: Sandia National Laboratories
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.