Visit additional Tabor Communication Publications
November 29, 2012
Although Lawrence Livermore Lab's Sequoia supercomputer got knocked of its TOP500 perch a few weeks ago, the DOE machine hosted by National Nuclear Security Administration (NNSA) is proving its worth in the world of real applications.
According to the National Nuclear Security Administration, Sequoia, the world's largest IBM Blue Gene/Q system, delivered nearly 14 petaflops on the recently developed Hardware/Hybrid Accelerated Cosmology Codes (HACC), a software framework that simulates the behavior of galaxies on a cosmological scale. Its purpose is to help scientists to reveal the nature of dark matter and dark energy. While that might seem a little tangential to NNSA's primary mission of managing the nation's nuclear arsenal, it does demonstrate the power of Blue Gene platform.
|Sequoia supercomputer; Photo credit: Bob Hirschfeld/LLNL|
In fact, 14 petaflops is just a couple of petaflops shy of Sequoia's Linpack mark, and just four petaflops off its peak performance number. According the DOE press release: "The HACC framework is designed for extreme performance in the weak scaling limit (high levels of memory utilization) by integrating innovative algorithms, as well as programming paradigms, in a way that easily adapts to different computer architectures."
Applications that exhibit weak scaling (the ability to increase the problem size by applying more processors) are good candidates to use the full capability these petascale supers since they rely on high levels of compute parallelism. This is especially true of the Blue Gene architecture, which uses large numbers of relatively slow CPUs (1.6 GHz, in this case) to achieve high aggregate performance. Sequoia, with more than 1.5 million PowerPC A2 CPUs, is perhaps the most extreme example of this.
Although these results were obtained in the NNSA's shop at LLNL, the team conducting the work came from the Argonne National Lab (ANL), a DOE facility devoted to open science and engineering. They will be running this same application on the 10-petaflop Mira supercomputer, another Blue Gene/Q system, installed at ANL.
Blue Gene systems haven't cornered the market on petascale apps though. Titan, the new Cray XK7 supercomputer at Oak Ridge, recently debuted with a 10-petaflop run of WL-LSMS, a material science code that performs thermodynamic calculations. Titan relies on NVIDIA GPUs of the Kepler persuasion for 24 of its 27 peak petaflops, so this represents a much different architecture than that of the CPU-only Sequoia.
As multi-petaflops supercomputers start to fill in the TOP500 list, applications that can sustain this level of computing will start to proliferate as well. In three years, all of the top 500 supercomputers are expected to be a petaflop or better, offering a much wider array of machines for such computing. The real era of petascale supercomputing has just begun.
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
Although Horst Simon was named Deputy Director of Lawrence Berkeley National Laboratory, he maintains his strong ties to the scientific computing community as an editor of the TOP500 list and as an invited speaker at conferences.
Supercomputing veteran, Bo Ewald, has been neck-deep in bleeding edge system development since his twelve-year stint at Cray Research back in the mid-1980s, which was followed by his tenure at large organizations like SGI and startups, including Scale Eight Corporation and Linux Networx. He has put his weight behind quantum company....
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.