Visit additional Tabor Communication Publications
April 07, 2006
In this special issue of HPCwire, all of our feature articles are devoted to DARPA's High Productivity Computer Systems (HPCS) program. The program is designed to take supercomputing to the petascale level and increase overall system productivity ten-fold by the end of this decade.
In 2001, DARPA initiated the HPCS program to drive the development of a new generation of economically viable, high productivity supercomputing systems for government and industry. As Phase II of the HPCS program draws to a close in the middle of this year, the three competing vendors -- Cray, IBM and Sun Microsystems -- are finalizing their R&D for their respective entries.
The new generation of supercomputers will scale to 10's to 100's of thousands of processors, connected by ultra-fast interconnects, and be able to access petabytes of memory. There are substantial challenges to construct such systems, but the technology that provide these capabilities is available now, or soon will be. The real challenge will be to build such systems in a cost effective way -- a stated goal of HPCS is to produce commercially viable systems.
But the HPCS program is about more than just petaflops. The real value of an HPC system is measured by a variety of factors including cost (both up-front and lifecycle), performance, robustness, portability and programmability. Taken together, these factors represent the productivity of the system. One of the main goals of the HPCS program is to increase overall HPC productivity by a factor of ten compared to current technology. However, the concept of system productivity is quite complex and being able to define and measure it is, itself, a major focus of the HPCS effort.
Increasing productivity means creating more powerful software models. Recent surveys of supercomputing centers found that most HPC applications are being implemented with legacy programming language -- FORTRAN, C, or C++ -- usually with a mix of MPI or OpenMP to facilitate parallelism. It is widely believed that this software model will be inadequate for fully exploiting petascale systems. Even looking at today's terascale systems, the gap between hardware and software capabilities is already uncomfortably large. Programmability -- the "time to solution" -- is a big problem for HPC applications and is about to become bigger.
To address this, new programming languages and new development tools are also being researched. Each HPCS vendor has proposed a programming language specifically targeted for high performance computing -- Chapel (Cray), X10 (IBM), and Fortress (Sun). New compiler and run-time technology will needed provide a portable, abstract programming model for highly parallelized applications. Debugging tools that can deal with thousand of threads will also need to be developed. At the same time, the software environment must also support the legacy HPC applications that have already been built with current toolsets. The software challenges are formidable.
Our first HPCS feature article is an excerpt of a recent interview we did with Douglass Post, chief scientist at the DoD High Performance Computing Modernization Program. He gives us his impressions of the HPCS program and talks about its significance to the HPC community. His considerable expertise in HPC modernization provides some interesting insights.
The remaining three feature articles were provided by the vendors, themselves. In them, they discuss their individual HPCS designs -- Cray's Cascade, IBM's PERCS and Sun's Hero. Some of the design details are left to our imaginations, but all the vendors have presented a compelling vision of their next generation systems.
In Cray's Cascade, they are looking to implement their adaptive computing strategy, which they announced just last month. In this model, software transparently maps application code to the most appropriate processor in an integrated heterogeneous architecture (scalar, vector, multithreading and hardware accelerators ). Cray summarizes their design as "an HPC datacenter in a box."
IBM's PERCS (Productive, Easy-to-Use, Reliable Computing System) design is, not surprisingly, based on their own POWER architecture. For IBM, this makes sense. They've got a big investment in the architecture and it is the basis of a lot of their HPC offerings, including the Blue Gene technology. The company also plans to leverage their expertise in semiconductor manufacturing to produce cutting-edge performance and reliability.
The Sun motto of "The Network is the Computer" is evident in their HPCS entry, named Hero. The company plans to use emerging interconnect technologies to provide dramatic increases in interprocessor communication. Proximity communications technology will be used to increase chip-to-chip data transfers, while silicon photonics will be used to speed inter-module communications. The Hero design also integrates object-based storage to provide a "smart" storage system.
Sometime this summer, DARPA will be selecting one (or possibly two) vendors to be funded for Phase III of HPCS, prototype development. This decision will have a direct effect on the direction of capability-class HPC systems for the next five to ten years and more far-reaching effects on overall HPC technology.
I hope you enjoy this special issue of HPCwire. As always, I appreciate any feedback. Mail me at email@example.com.
-- Michael Feldman
Posted by Michael Feldman - April 06, 2006 @ 9:00 PM, Pacific Daylight Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
Contributing commentator, Andrew Jones, offers a break in the news cycle with an assessment of what the national "size matters" contest means for the U.S. and other nations...
Today at the International Supercomputing Conference in Leipzing, Germany, Jack Dongarra presented on a proposed benchmark that could carry a bit more weight than its older Linpack companion. The high performance conjugate gradient (HPCG) concept takes into account new architectures for new applications, while shedding the floating point....
Not content to let the Tianhe-2 announcement ride alone, Intel rolled out a series of announcements around its Knights Corner and Xeon Phi products--all of which are aimed at adding some options and variety for a wider base of potential users across the HPC spectrum. Today at the International Supercomputing Conference, the company's Raj....
Jun 18, 2013 |
The world's largest supercomputers, like Tianhe-2, are great at traditional, compute-intensive HPC workloads, such as simulating atomic decay or modeling tornados. But data-intensive applications--such as mining big data sets for connections--is a different sort of workload, and runs best on a different sort of computer.
Jun 18, 2013 |
Researchers are finding innovative uses for Gordon, the 285 teraflop supercomputer housed at the San Diego Supercomputer Center (SDSC) that has a unique Flash-based storage system. Since going online, researchers have put the incredibly fast I/O to use on a wide variety of workloads, ranging from chemistry to political science.
Jun 17, 2013 |
The advent of low-power mobile processors and cloud delivery models is changing the economics of computing. But just as an economy car is good at different things than a full size truck, an HPC workload still has certain computing demands that neither the fastest smartphone nor the most elastic cloud cluster can fulfill.
Jun 14, 2013 |
For all the progress we've made in IT over the last 50 years, there's one area of life that has steadfastly eluded the grasp of computers: understanding human language. Now, researchers at the Texas Advanced Computing Center (TACC) are utilizing a Hadoop cluster on its Longhorn supercomputer to move the state of the art of language processing a little bit further.
Jun 13, 2013 |
Titan, the Cray XK7 at the Oak Ridge National Lab that debuted last fall as the fastest supercomputer in the world with 17.59 petaflops of sustained computing power, will rely on its previous LINPACK test for the upcoming edition of the Top 500 list.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?
Join our webinar to learn how IT managers can migrate to a more resilient, flexible and scalable solution that grows with the data center. Mellanox VMS is future-proof, efficient and brings significant CAPEX and OPEX savings. The VMS is available today.