Visit additional Tabor Communication Publications
October 27, 2006
HPC and modern computing in general has a seemingly insatiable demand for more performance, better efficiency and scalability. Now with the expansion of computing to practically every commercial and non-commercial endeavor, an additional requirement is to apply these attributes to a much wider range of applications.
"So there's a need there for more types of processing," says Jeff Jussel, vice president of marketing and general manager of the Americas for Celoxica. "There's a number of ways the industry is addressing that -- with massively parallel processing (MPP) and with all sorts of different types of co-processors. At Celoxica, we believe that the FPGA represents a huge opportunity for co-processing, because it can deliver the massive parallelization, with the advantages of custom hardware, but in a way that is programmable."
But there are three things that you need in order for FPGAs to really take off:
"That's where the tools that Celoxica provides comes in," says Jussel. "That's our mission in life -- to make that FPGA programming transparent. And in doing so, enable FPGA use for the high performance computing market."
As part of this strategy, this week Celoxica announced a new off-the-shelf hardware and software compiler design bundle for high performance computing using HyperTransport (HTX) slots. The HTX bundle combines an intellectual property (IP) core for HTX connectivity, an FPGA-based HTX acceleration card and a software programming environment. The solution is designed to allow users to accelerate applications in Opteron-based computing systems with FPGA co-processing and HyperTransport technology. The bundle provides compilers that map C code onto FPGA hardware, a run-time OS (RTOS) for FPGA computing, and FPGA hardware that plugs into a host server system.
The hardware consists of the RCHTX acceleration card, which includes two Xilinx Virtex 4 FPGAs devices (in the future it will support more advanced Virtex 5 devices), 24 MB of QDR SRAM, and a range of I/O. The main co-processor FPGA is a 16 million gate device that is meant to run the user algorithms. The second FPGA is configured as a bridge, containing an HTX IP core developed by Celoxica. The bridge FPGA and IP provide the HyperTransport interconnect between the FPGA co-processor and the host processor system and memory space.
The software component consists of the DK Design Suite, which includes a C compiler for programming the FPGA co-processor, a board support package (BSP) and data communications drivers for the RCHTX card, a basic floating point library (single and double precision) and the software API which provides the interfaces.
The idea is not for the user to port their whole application to the FPGAs, just the compute-intensive algorithms that represent the workload bottlenecks. For example FFT calculations, a Black Scholes algorithm or wave migration calculations can be offloaded to the FPGA to take advantage of the parallel hardware resources.
The user replaces the algorithm loop in the original FORTRAN or C source with a Celoxica API call, which calls the C code that will be compiled into the FPGA. Jussel says the original algorithm needs to be "tweaked" somewhat to insert parallelism, but they've tried to simplify this as much as possible. The FPGA C compiler brings in the appropriate run-time pieces to make it work in its new hardware environment. At execution time, the data communication between the host processor and the FPGA is done across the HyperTransport connection, but this is transparent to the user.
Jussel notes that the product announced this week represents the first FPGA solutions that uses the HTX slot. DRC Computer Corporation has a somewhat similar solution, where its FPGA uses an Opteron socket to directly connect to HyperTransport. As it turns out, DRC is an OEM partner with Celoxica and makes use of the same C compiler technology.
Celoxica's current (beta) customers for the HTX solution are in the financial services, oil & gas, and life sciences industries. With this particular product, users have achieved a 200X performance improvement for the application (offloading a Black Scholes algorithm to the FPGA).
"We've done enough with the finance industry to know that the metric we need to hit is about a 10X price-performance benefit," says Jussel. "If we hit that 10X factor then it's worth it for them to invest in new technology. And we've been able to show quite a bit greater than 10X for all these applications."
Even though this HTX product includes the FPGA card, hardware is not Celoxica's main focus. The company's real goal is to be the leader in compilers and the run-time support for FPGAs. Jussel says that compared to other FPGA compiler companies, Celoxica is quite a bit larger and more established, having developed and matured its software technology over the past 10 years. It provides the compiler for the SGI RASC RC100 system as well as Cray FPGA systems, not to mention its large customer base in the embedded computing space -- still the majority of their business. But because FPGAs have this unique aspect of reconfigurability and high performance, the company believes that these devices will become ubiquitous throughout computing. And Celoxica wants to be there with their software.
Says Jussel: "We really want to be the Microsoft of FPGA computing. We want to provide the compilers and RTOS for that solution."
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.