Visit additional Tabor Communication Publications
November 02, 2011
Sandia National Labs made a bit of HPC history this week when it announced it had installed the first HPC cluster outfitted with AMD's Accelerated Processing Units (APUs), based on the chipmakers 'Fusion' processor design. The chip integrates x86 CPU cores and ATI GPU cores onto the same processor.
This is an experimental system only, to be used as a platform to evaluate the CPU-GPU heterogenous processor model of supercomputing. Currently, AMD's Fusion chips are designed for personal computing platforms and are not intended for servers. But apparently Sandia's Scalable Computer Architectures group was eager enough to get their hands on an APU cluster to contract Penguin and AMD to be build them a system.
According to the press release, the Penguin system, known as the Altus 2A00, was specifically designed with AMD APUs in mind. The QDR InfiniBand-based cluster is comprised of 104 servers, which houses an unspecified number AMD A8-3850 APUs.
Designed for desktop PC duty, the A8-3850 a quad-core x86 design integrated with 400 Radeon (HD 6550D) cores. The CPU cores run at 2.9 GHz, while the GPU side runs at a more modest 600 MHz. TDP is a respectable 100 watts.
The whole cluster delivers 59.6 peak teraflops, but the idea here is not to break any performance records. The Sandia researchers will use the machine to explore programming models for integrated CPU-GPU platforms, relying primarily on OpenCL and MPI. The draw here is to be able to access the considerable vector capabilities of the GPU within the same memory space as the CPU.
The rationale for such research is that heterogeneous processors such as these APUs will be the basis for future exascale computers. In a roundtable discussion organized by AMD on Wednesday, AMD Technology Group CTO Chuck Moore noted that while their x86 server CPUs, like their new Interlagos Opterons, continue to make strides, those designs will likely not be the path to exaflop performance by the end of this decade.
"Instead what we need to do is to make really good use of heterogenous computing," said Moore. "By combining the best of CPU and GPU technology, presumably on the same chip, we believe we can build computing nodes that are on the order of 10 teraflops in that timeframe."
According to him, AMD has plans in the works for an "HPC APU that would utilize even a larger GPU and fewer x86 core" than the current desktop chips. He thinks the company can build HPC APUs that run about 150 watts and would be powerful enough to power an exaflop computer that consumes no more than 20 MW. And, thanks to the integrated CPU-GPU memory space, these machines would be reasonably easy to program.
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.