Visit additional Tabor Communication Publications
July 24, 2008
Planting CUDA Seeds
NVIDIA is continuing to push hard on CUDA, the company's C-based software environment for GPU computing. With last month's announcement of the first CUDA Center of Excellence at the University of Illinois at Urbana-Champaign, NVIDIA said it donated half a million dollars to the school.
The announcement also noted that "[u]niversities wishing to become CUDA Centers of Excellence must teach a CUDA class and use CUDA technology in their research, usually across several labs. In return, NVIDIA supports the school through funding and equipment donations, including help to set up a GPU computing cluster." This is the same general idea behind the Sony, Toshiba and IBM (STI) Center of Competence program for Cell technology. The STI Center was established at Georgia Tech last year. Can a Larrabee Center of Distinction be far behind?
Big Time Protein Folding
On the hardware side, even NVIDIA's non-Tesla gear is starting to show up in HPC applications. On Thursday, the company announced the big impact that the GeForce GPUs are having on Stanford's Folding@Home project. The project aggregates donated computer cycles on desktop systems to calculate how different types of proteins fold. The idea is to help understand protein behavior as it relates to cancer, cystic fibrosis, Parkinson's and other diseases.
Powered by CUDA-enabled software, the GeForce GPUs are doing yeoman's work for the folders. According to a recent ExtremeTech report, there are currently over 7,000 GPUs running the Folding@Home program, yielding around 840 teraflops of application performance. The article goes on to say:
That's somewhere around 110 gigaflops per GPU, on average. To put that in perspective, the regular windows CPU client is about one gigaflop per client (it's a mix of the single-threaded client and the multi-core SMP version). The PS3 looks like it leads the pack with a total of 1,358 teraflops, but that's from over 48,000 active PS3s. Each PS3 is actually delivering about 28 gigaflops apiece.
Those are all 32-bit floating point performance numbers, but if you can cure cancer with single precision, that's fine with me.
Argonne's Blue Gene/P Gets a Visual Buddy
On Wednesday, Argonne National Laboratory announced its plans to add a NVIDIA Quadro-based data analytics/visualization system to pair up with its Blue Gene/P supercomputer. The visualization system, named Eureka, will turn the torrents of data produced by applications running on Blue Gene into pretty pictures that make sense to mere mortals.
Apparently, 208 NVIDIA Quadro GPUs will be used to construct the system, which is being built by GraphStream, Inc. The hardware will consist of four racks of 1U boxes, with each box containing four Quadro graphics cards. According to a Dr. Dobbs article, the Eureka server building block is the SuperMicro 6015-UR. Each GPU box is hooked to two SuperMicro servers so that each compute server drives two GPUs. Visualization is driven by data that comes from a very large storage array, which is also hooked up to the Blue Gene machine.
The Manycore War of Words
Meanwhile, NVIDIA and Intel continued to spar on the GPU-Larrabee match-up. An article in Custom PC recorded Andy Keane's reaction to Pat Gelsinger's recent comments disparaging the CUDA technology. Keane is general manager of NVIDIA's GPU computing group and took offense at some off-handed remarks made by the Intel exec earlier this month. Gelsinger, a senior vice president and co-general manager of Intel's Digital Enterprise, started the brouhaha by claiming that GPGPU languages like NVIDIA's CUDA will one day be nothing more than "interesting footnotes in the history of computing annals," adding that Larrabee, Intel's upcoming manycore processor, will be the solution that succeeds in the long term.
If you're a self-respecting Nvidian, those are fighting words. From the article, here is the gist of Keane's reaction:
[T]he high level of interest in CUDA "is causing Larrabee. Larrabee's the reaction." He then added that "these comments from Gelsinger; if we were not making a lot of headway do you think he'd even give us a moment's notice? No. It's because he sees a lot of this activity. The strategy is to try to position it [CUDA] as something scary and unique, and it's really not; it's something that's very accessible."
Next month, Intel may release a lot more details on Larrabee at SIGGRAPH and the Intel Developer Forum. An article published Tuesday in The Inquirer says the Larrabee developer boards are being shipped in November. If true, NVIDIA is going to have a much better idea what it's up against real soon.
NVIDIA for Sale?
Finally, Simon Brew in UK's IT PRO speculates whether NVIDIA could be bought out. The reasoning behind the speculation is the AMD-ATI merger, which was designed to synergize two successful technologies into a greater whole -- or into a greater hole, as the case may be. Even Brew admits, "things haven’t quite gone to plan."
His real argument is that in the long term, NVIDIA's toughest competition will be Intel, not AMD. The implication is that the GPU vendor will have to be a lot bigger and brawnier to go up against the chip giant, especially in the expanding mobile graphics market, where Intel dominates. That's a valid point. But so far, NVIDIA has been more nimble than Intel in the graphics arena, and it's got a big head start at the high end of the market. I'm guessing NVIDIA figures it can still outrun the competition. We'll see...
Posted by Michael Feldman - July 23, 2008 @ 9:00 PM, Pacific Daylight Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.