Visit additional Tabor Communication Publications
September 24, 2009
In information technology, there's really no such thing as a product unveiling anymore. Preparing the market for new hardware or software starts way before the products are rolled out. It's a drawn-out process that begins with PowerPoint presentations and continues up to the point of commercial release. I call this process "unveilation" (literally, the process of unveiling). It can take years, and often does. I'm not saying this is a bad thing. Given the complexity of technology, it's almost a necessity.
Nowhere was the process more evident than at this week's Intel Developer Forum (IDF) in San Francisco, where company execs extolled the virtues of chips yet born. There they talked up a number of Intel's upcoming microprocessors, all in various stages of unveilation. Of particular interest to the HPC crowd was the first demo of the GPU-ish Larrabee chip, an update on Nehalem-EX, a refinement of the Westmere roadmap, and the invention of ultra-low-power Xeons for a new "microserver" category.
First up is Larrabee, a chip that is in the midst of an extended unveilation. After first floating the idea of a high performance CPU-GPU hybrid chip back in 2007, Intel finally gave Larrabee followers their first demo of the silicon in action. Although Intel insists that the initial product line is strictly geared for traditional graphics and visualization apps, I'm convinced that later versions, or derivatives thereof, are being groomed for general-purpose HPC. The first products are expected to hit the streets sometime next year.
I've embedded the Larrabee demo below, showing how the chip manages a real-time ray-tracing application.
Almost at the end of its unveilation is Nehalem-EX, the Xeon that will go into servers with four, eight, or more sockets. It's really the first time Intel will have a competitive multi-socket (i.e., more than two sockets) offering for x86 servers. Nehalem-EX will support a number of RAS features, including Machine Check Architecture (MCA) recovery, which allows the CPU to right itself after encountering certain kinds of system errors. The chip is expected to go into production later this year.
Speaking about Nehalem-EX, Sean Maloney, executive VP and GM of the Intel Architecture Group, said they currently know of over 15 eight-socket-plus designs from eight different OEMs. Some of these are certainly destined for HPC duty. Even a relatively modest four-socket machine will support up to 64 threads and a terabyte of memory. A couple of these four-socket systems have already been announced: one, the IBM BladeCenter EX; the other, a Supermicro 1U box, specifically targeted at HPC. To hammer home the HPC theme, Maloney pointed to a quote from Mark Seager, who leads the advanced computing group at Lawrence Livermore National Lab: "Nehalem EX represents a new SMP on a chip super-node that can help us improve our predictive science and simulation capabilities without having to invest in a vast rewrite of our applications."
Meanwhile, Westmere, the 32 nm shrink of the Nehalem microarchitecture, is apparently running a little ahead of schedule. The first Xeon implementation (for dual-socket servers), Westmere-EP, is poised for release in the first half of 2010. And in late 2010, Westmere-EX will take the hand-off from Nehalem-EX for multi-socket platforms. Due to the process shrink from 45 to 32 nanometers, lower power consumption and/or faster clocks are in the offing, although no specific numbers were forthcoming at IDF. For the security-minded, Intel has added Advanced Encryption Standard (AES) instructions to enable faster encryption and decryption.
One of the most interesting announcements had to do with the entry-level Xeon 3400 processors. Low-power variants of these chips have been developed for what Intel is calling "microservers" -- essentially mini-blades, that take up much less space and use much less power than standard hardware. The chipmaker has come up with a reference design that packs 16 hot-swappable microserver modules into a 5U rack. Intel is planning to release a 45-watt version of the 3400 later this year and a 30-watt model in early 2010.
The idea, of course, is to be able to build extremely dense machines that are inexpensive to both buy and run. The big target market is large-scale and "containerized" datacenters, where power consumption and floor space are enemies number one and two, respectively.
If the performance per watt numbers prove out for technical computing apps, these low-power 3400s could make their way into HPC. SGI will almost certainly make these Intel parts available in its newly announced Octane III personal super and in its CloudRack product line. Other HPC OEMs may follow suit.
Keep in mind that these single-socket 3400 chips are the antithesis of the multi-socket EX Xeon processors. But sometimes scaling out is much more preferable than scaling up. (There are a number of HPC system architects who think hyperscale designs using extremely low-power CPUs is the way to go if exascale computing is to be made practical.) In any case, Intel is making sure it is covering all its bases, and is willing to let the applications decide which computing model fits best.
Posted by Michael Feldman - September 24, 2009 @ 5:31 PM, Pacific Daylight Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.