Visit additional Tabor Communication Publications
December 17, 2008
For those of you who thought Intel was angling for an HPC play with its upcoming Larrabee processor family, think again. In case you're not a regular reader of this publication, Larrabee is Intel's manycore x86 GPU-like processor scheduled to debut in late 2009 or early 2010. With Larrabee, Intel is gearing up to challenge NVIDIA and AMD for GPU leadership, but doesn't appear interested in exploiting the chip for GPGPU.
Although the company has engaged in some mixed messaging with regard to Larrabee, recent conversations with Richard Dracott, the general manager of the Intel's high performance computing business unit, and Stephen Wheat, Intel's senior director of HPC, has convinced me that Larrabee will be targeted only for standard graphics applications and the more generalized visual computing realm. The latter does overlap into the HPC space, inasmuch as visual computing encompasses applications like video rendering and other types of compute-intensive image processing. But for general-purpose HPC, Intel has other designs in mind.
My first hint that Larrabee was being more narrowly targeted came from a recent chat with Dracott at the Wall Street on HPC Conference, where he talked about how the CPU will eventually prevail over accelerator offload engines (GPUs, FPGAs, Cell-type processors and FP ASICs), which I wrote about back in September. His basic point was that CPUs offered a better long-term value proposition than accelerators because industry standard processors like the x86 offer lower and more predictable software development costs. Also, any performance advantage demonstrated by an accelerator would eventually erode as CPUs continued to evolve -- a dubious assumption, I thought, since GPUs and FPGAs are evolving at least as quickly as CPUs. From his perspective, though, a relatively small number of HPC users would continue to experiment with acceleration over the next several years, but would eventually return to the comfort of the CPU.
What I didn't mention from that conversation is that while Dracott was trash talking accelerators, he also managed to diss Larrabee -- at least as a scientific computing architecture. Although Larrabee would at least offer an x86 compatible ISA, the problem, he said, was that the implementation contains some of the same drawbacks as the traditional GPU for scientific computing -- namely the lack of ECC memory to protect against soft errors and a shortage of double precision floating point capability. From his perspective that would prevent Larrabee or GPUs to be deployed more generally in high performance computing, "But," Dracott added, "we are working on products that will meet that need."
At this point, the nature of those products is a mystery. But Dracott did offer that "it will be feasible to have products that combine small IA cores and large IA cores that would be more suited for some HPC workloads." To speculate a bit, the architecture will likely have many of the attributes of Larrabee, that is, a manycore x86 architecture with fully coherent L1 and L2 caches. At least some of those cores will contain wide SIMD units similar to the Larrabee design, but those units will support lots of 64-bit floating point horsepower and be fully IEEE 754 compliant. Also, instead of GDDR-type memory, ECC-capable DDR memory will be supported. In short, it would be an x86 vector processor.
When I talked with Stephen Wheat at the Supercomputing Conference in Austin last month, he reiterated Dracott's (and Intel's) CPU-centric view of the universe. Wheat suggested that if accelerator features, like vector processing, become more widely used, they will naturally migrate onto the CPU. There is historical precedence for such a position. Before the 80486 chip, floating point operations were performed in external coprocessors (the x87 series). In the late 80s Intel decided FP operations were general-purpose enough to warrant transistor space on the CPU.
The rationale behind feature integration is that it's difficult to have different silicon engines move forward individually and maintain a balanced performance profile without being tied to a unified architecture. This dovetails nicely with Intel's business strategy, in which the x86 CPU ISA is the common denominator in all IT sectors -- desktop, mobile, enterprise, HPC, and, if Larrabee is successful, gaming and visualization.
So is now the time for heavy-duty vector processing to be integrated as well? "It's not that far in the future where the things that make these [accelerators] attractive can become an integral part of the processor itself," offered Wheat.
The market for this family of chips is not necessarily going to be driven by server-based HPC. Intel has a whole strategy around model-based computing, called RMS (PDF), which requires teraflops a-plenty. RMS encompasses a range of advanced apps from real-time rendering to portfolio management, and some of these will be running on desktop and mobile platforms.
The company is already starting to build a software foundation for such applications with Intel Parallel Studio. Released as a beta this week, Intel Parallel Studio is a suite of tools and libraries for Windows C/C++ developers that supports Threading Building Blocks (TBB), OpenMP, auto-vectorization and other parallel programming goodies. It is built around the concept of "forward scaling," Intel's nomenclature for automatic application scaling for multicore/manycore architectures.
The idea is that the same programs that are initially built for quad-core Nehalems can transparently scale up to 8-core Nehalems, manycore Larrabees, and all their descendents. Intel's research language for manycore throughput computing, Ct, will probably get productized into the company's software offerings at some point as these manycore products start to hit the streets. If all goes according to plan, by 2015 manycore x86 will be the dominant processor species and parallel programming will be the norm.
Posted by Michael Feldman - December 16, 2008 @ 9:00 PM, Pacific Standard Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.