Visit additional Tabor Communication Publications
September 23, 2009
After the 1992 moratorium on underground testing of nuclear weapons in the US went into effect, the Department of Energy's National Nuclear Security Administration's (NNSA) was tasked to maintain the country's nuclear weapon deterrent via computing simulations. As a result, Lawrence Livermore National Laboratory (LLNL) and its two sister labs at Los Alamos and Sandia became the recipients of some of the most muscular computing hardware in the world. Today these institutions are at the forefront of supercomputing expertise, both hardware and software.
Because the weapons simulation applications are always looking to achieve higher resolution, higher fidelity, and full-system modeling, there is an ongoing demand for ever-more powerful capability-class supercomputers. Today, Los Alamos houses what is ostensibly the world's most powerful computer -- Roadrunner -- which clocks in at over a petaflop. In a couple of years, LLNL is slated to deploy "Sequoia," a 20-petaflop IBM Blue Gene/Q machine, and a likely contender for the top supercomputer in 2011. Sequoia's predecessor, "Dawn," is a 500 teraflop Blue Gene/P machine installed earlier this year at Livermore.
But according to Mike McCoy, who heads Livermore's Scientific Computing and Communications Department, it's not all about these elite capability machines. He says 10 to 30 percent of the computational resources at the lab are devoted to capacity systems, that is, commodity HPC Linux clusters. The reason is simple. There is a lot of computing to be done, and time on the expensive capability systems is dear. By necessity a lot of application work has to be developed and tested on these smaller, less expensive machines as a way to contain costs.
There is also quite a bit of unclassified science work performed at the lab in the areas of climate, biology, molecular dynamics, and energy research. Some of this basic science supports the weapons programs, but the remainder is just part of the NNSA's larger mission of furthering national security. The unclassified work also serves to nurture the lab's scientists, and without them, there is no weapons program. In any case, the vast majority of this class of computing takes place on vanilla Linux clusters, albeit very large ones.
Today at Livermore, capacity clusters account for 404 teraflops of computing power, while the capability machines deliver 1,324 teraflops. Another 205 teraflops are available in visualization and collaboration systems. The most powerful capability system at the facility is the half-petaflop Dawn, while the largest capacity cluster is Juno, which weighs in at 167 teraflops.
Livermore has relied on a number of cluster computer vendors over the years. In 2002, the now-defunct Linux Networx installed a the MCR cluster, which delivered a 7.6 teraflops, a performance level that earned it the number three spot on the TOP500 list in June 2003. A more recent vendor is Appro, who won the Peloton contract in 2006 and then the subsequent Tri-Lab Linux Capacity Cluster (TLCC) deal, which served all three NNSA labs.
Today Lawrence Livermore appears to be grooming Dell for some major deployments. Up until last year, the only Dell machines at the lab were sitting on people's desks. But in November 2008, the company became the cluster partner on the Hyperion project, a testbed system to be used to develop system and application software for HPC. The idea was to provide a platform for developers to build and test codes at scale before they are deployed on larger production systems. That effort has produced some early results including simulating the file system and I/O rates of the future Sequoia system using Hyperion's InfiniBand and Ethernet SANs.
Last week, Michael Dell met with LLNL officials at Livermore to get a sense of what the NNSA is expecting from its future cluster system. The agency's goal is to maintain at least a 1:10 performance ratio between capacity systems and capability systems. Today that means you need roughly a 100 teraflop cluster to match up with the purpose-built one-petaflop supers. With Sequoia coming online in 2011, the folks at LLNL are already thinking about clusters in the two-petaflop range. Beyond that the lab see the need for 100-teraflop commodity machines in 2018, in anticipation of capability machines hitting the exaflop mark. That means vendors need to scale today's commodity clusters by a factor of 10 over the next 9 years.
Recently Dell installed "Coastal," an 88.5 teraflop system that is being used by the Lawrence Livermore's National Ignition Facility to help with fusion research. Next year, with Dell's help, the lab will be more than doubling the performance of the 90 teraflop Hyperion system with "Sierra," a new cluster that is spec'ed to reach 220 teraflops.
Michael Dell is hoping that's just the beginning. From his point of view, designing systems pushing the envelope of scalability and technology dovetails nicely with the company's other big server segments, namely web services infrastructure and cloud computing. For example, the inclusion of SSD technology to increase I/O performance in the Livermore's Coastal cluster also turned out to be a good solution for Dell servers deployed for a Web search provider in China (presumably Baidu). He sees the demand for these super-sized machines inside and outside of HPC as two sides of the same hyperscale coin. And, he says, the technology transfer travels in both directions. "You always learn from your best customers," says Dell.
Jun 17, 2013 |
The advent of low-power mobile processors and cloud delivery models is changing the economics of computing. But just as an economy car is good at different things than a full size truck, an HPC workload still has certain computing demands that neither the fastest smartphone nor the most elastic cloud cluster can fulfill.
Jun 14, 2013 |
For all the progress we've made in IT over the last 50 years, there's one area of life that has steadfastly eluded the grasp of computers: understanding human language. Now, researchers at the Texas Advanced Computing Center (TACC) are utilizing a Hadoop cluster on its Longhorn supercomputer to move the state of the art of language processing a little bit further.
Jun 13, 2013 |
Titan, the Cray XK7 at the Oak Ridge National Lab that debuted last fall as the fastest supercomputer in the world with 17.59 petaflops of sustained computing power, will rely on its previous LINPACK test for the upcoming edition of the Top 500 list.
Jun 12, 2013 |
At 31 petaflops of sustained LINPACK capacity, the new Chinese Tianhe-2 supercomputer will be the fastest supercomputer in the world when this month's Top 500 list comes out, as we reported previously in HPCwire.
Jun 12, 2013 |
HPC system makers are lining up to announce compatibility with the new fourth generation Intel Core processor, codenamed "Haswell." The new Iris GPUs based on the Haswell architecture are giving Intel new credibility in the graphics processing department.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?
Join our webinar to learn how IT managers can migrate to a more resilient, flexible and scalable solution that grows with the data center. Mellanox VMS is future-proof, efficient and brings significant CAPEX and OPEX savings. The VMS is available today.