Visit additional Tabor Communication Publications
June 23, 2006
The HPC community has been reaching towards systems capable of petaflops performance ever since the teraflops barrier was conquered back in December of 1996. Today systems that will execute a quadrillion floating point operations per second are close to becoming a reality. In fact, one such system may already exist. While HPC people-in-the-know realize that it's just a number, the petaflops metric does seem to capture the imagination of the public whenever it's mentioned in the media. So I'm mentioning it.
And I'm not the only one. Just this last week, EETimes reported that Riken, a Japanese government-funded science and technology research organization, has a developed a supercomputer that achieves a theoretical peak performance of one petaflops. The system, called MDGrape-3, was apparently designed as a dedicated computing system to run molecular dynamics simulations. The details of the hardware are a little sketchy, but basically it looks like a very large cluster of Intel Xeon based-servers with some additional proprietary MDGrape-3 processors thrown into the mix.
Peter Ungaro, CEO of Cray, told me that has some familiarity with the MDGrape technology and he believes the Riken claim is probably true (although the Japanese team at Riken admit that they haven't run Linpack on the system so no direct comparison with other benchmarked machines can be made). Ungaro also reminded me that the Japanese machine was specifically built to run one type of application, so it really shouldn't be compared with more general-purpose supercomputers.
Like, for instance, Cray machines. The Seattle supercomputer company made its own peta-splash late last week, when it announced that it will deliver a petaflops-speed computer to Oak Ridge National Laboratory (ORNL) in late 2008. In a conversation I had with Jan Silverman, Cray's senior VP of Corporate Strategy and Business Development, he described the significance of the ORNL machine and how it will fit into the company's future product line. And although we talked about the peak petaflops milestone, that particular attribute of the system seemed much less important than what the new machine represents to both Cray and ORNL. Read the article in this week's issue for the complete story.
And Cray is not the only one playing in the deep end of the pool this summer. IBM and the National Nuclear Security Administration (NNSA) claimed a new world record for a scientific application by achieving a sustained performance of 207.3 teraflops. IBM and NNSA have managed to scale the "Qbox" code so that it can take better advantage of the 131,072 PowerPC processors in the Blue Gene/L system at Lawrence Livermore National Laboratory. Qbox is an application used to conduct materials science simulations critical to national security.
"This is an important step on the path to performing predictive simulations of nuclear weapons, and these simulations are vital to ensuring the safety and reliability of our nuclear weapons stockpile." said Dimitri Kusnezov, head of NNSA's ASC Program. For more about the ASC program, read our own interview with Kusnezov in this week's issue.
The IBM Blue Gene/L record of 207.3 sustained teraflops record is only a fifth of the way to a sustained petaflops. But this is the metric that really matters -- sustained performance on real live applications. Achieving this level of capability will require a more advanced hardware/software model than that required for peak performance alone.
Our own High-End Crusader(HEC) tackles this issue himself in this week's feature article, "Heterogeneous Processing Needs Software Revolutions." It's a long read, but I encourage anyone interested in the future of HPC system software to take a look. HEC argues that a much more sophisticated approach to thread migration will be required for our next-generation heterogeneous supercomputers. Be forewarned -- HEC is not pulling his punches with this one. Here's just a slice of what's inside:
"Perhaps the U.S. has such expertise in principle, but it clearly has not yet worked the issues of developing system software to manage, operate, and exploit next-generation heterogeneous systems. More generally, in the area of heterogeneous processing for high-end computing, there is a total absence of leadership---from vendors, from academia, from government. Whether we consider heterogeneous system architectures, hardware technology, language processing, etc., etc., there are no private-sector computer architects or government agencies who either can, or are willing to, assume a leadership role; there are no compelling visions of heterogeneous processing from which to choose. We are lacking even a simple roadmap for heterogeneous processing."
Last week's first annual TeraGrid Conference in Indianapolis offered up some great presentations from individuals who are interested in building and using cyberinfrastructure. Check out our special section of the TeraGrid 2006 Conference in this week's issue, right beneath our Features section. You'll find additional coverage of the event in this Monday's issue of GRIDtoday. And for even more information about the conference, visit http://www.teragrid.org/events/2006conference/, where you can download the individual presentations.
It seems like as soon as one HPC conference ends, another one begins. Next week, on June 27-30, the 21st Annual International Supercomputing Conference (ISC) is taking place in Dresden, Germany. ISC is the premier supercomputing conference in Europe and provides an international forum for the global HPC community. This year, we'll be providing special coverage of the event featuring interviews, articles, news and commentary. Visit our live coverage at http://www.hpcwire.com/ starting on June 28.
As always, comments about HPCwire are welcomed and encouraged. Write to me, Michael Feldman, at firstname.lastname@example.org.
Posted by Michael Feldman - June 22, 2006 @ 9:00 PM, Pacific Daylight Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.