Visit additional Tabor Communication Publications
May 12, 2010
GPGPU Development Workshop offered May 25th at OSC
COLUMBUS, Ohio, May 12 -- A new, advanced service offered by the Ohio Supercomputer Center leverages the unique computing properties of the graphics processing unit (GPU) to provide a robust visualization environment to researchers in fields as diverse as biomedicine, electrosciences and the animation arts.
OSC recently completed deployment of the "Csuri" Advanced GPU environment, increasing the Center's capabilities for advanced large-scale remote visualization and batch-rendering applications, as well as GPGPU applications. This powerful computing environment is now available to all Ohio college and university faculty, students and staff, as is an upcoming workshop planned to provide insights on the effective use of the service.
The high performance computing (HPC) community is aggressively exploring general-purpose GPU (GPGPU) computing by using them as "many-core processors" to solve scientific problems. Modern, commercially-available central processing units (CPUs) are multicore processors with 2, 4 or 8 independent processors on a single chip. Many-core processors, on the other hand, have hundreds or thousands of processors that are more tightly connected.
"For the right kind of problems, GPGPU computing can provide revolutionary performance advantages," said David Hudak, Ph.D., program director of cyberinfrastructure and software development at OSC. "The Csuri platform is designed to support the development of both GPGPU and advanced visualization solutions. We look forward to working with our user communities to develop codes and evaluate GPU-enabled third party applications."
One example of the work that can benefit from the Csuri Advanced GPU environment is the work of its namesake, Charles "Chuck" Csuri, whose sophisticated digital art involves giant rendering of thousands of frames.
Csuri is best known as the father of computer graphics, computer animation and digital fine art, creating the first computer art in the 1960s. In addition to being recognized by the Smithsonian Magazine, Csuri is seen as a pioneer of computer animation by the Museum of Modern Art (MOMA) and the Association for Computing Machinery Special Interest Group Graphics (ACM-SIGGRAPH).
Csuri became interested in the digital computer as a means of imaging in 1964, when he saw a computer-generated face in a publication from the university's department of electrical engineering. While a senior professor at The Ohio State University (OSU), Csuri founded the Computer Graphics Research Group, the OSC Graphics Project and accad, an academic unit dedicated to the development of digital art and computer animation.
During the beta testing of the GPU system, Umit Catalyurek, Ph.D., associate professor in the Department of Biomedical Informatics and Department of Electrical and Computer Engineering at OSU, used the nodes to develop a component-based runtime system for various biomedical image analyses and synthetic aperture radar image formations.
"One of student researchers is using the new GPU system to finish his experiments for a project on automatic tuning of radar signal processing on emergent architectures," said Catalyurek. "We are now developing software systems that will enable applications to easily scale from a single CPU or GPU to a cluster of GPUs and multicore CPUs."
With matching funding from DaytaOhio, a Wright Center of Innovation, OSC was able to develop the Csuri Advanced GPU Environment with the latest NVIDIA visualization technology, providing researchers with 18 NVIDIA Quadro Plex 2200 S4 units installed in the Center's flagship production cluster, the Glenn IBM 1350, and 16 nVidia C1060 cards installed in the BALE research cluster. The GPU environment also features the CUDA toolkit, including the CUDA SDK, a collection of example programs illustrating various aspects of CUDA and GPGPU usage.
OSC's Csuri Advanced GPU Environment Configuration
36 GPU-accelerated nodes on the Glenn IBM 1350 cluster, connected to 18 NIVIDA Quadro Plex S4's for a total of 72 CUDA-enabled graphics devices. Each node has access to two Quadro FX 5800-level graphics cards.
Each Quadro Plex S4 includes:
In addition, each of the 36 compute nodes contain:
Supported software includes:
GPU Orientation Workshop Offered
In order to orient researchers to the new computing environment, OSC will host a GPGPU Development Workshop from 9:30 a.m. -5 p.m., on May 25, 2010, in OSC's BALE Theatre. The workshop will provide an overview of the GPGPU resources available at OSC, summarize some of the key potential benefits and provide an overview of the most popular programming toolkits and techniques currently in use.
More information is available by visiting the workshop registration page on the OSC Web site at www.osc.edu/supercomputing/training/register/gpgpuregister.shtml or by emailing email@example.com.
As in prior summers, OSC also is a training site for the Virtual School for Computational Science, with a session that covers many-core processors and GPGPU programming. More information is available at www.vscse.org/summerschool/2010/
About the Ohio Supercomputer Center
The Ohio Supercomputer Center is a catalytic partner of Ohio universities and industries that provides a reliable high performance computing infrastructure for a diverse statewide/regional community. Funded by the Ohio Board of Regents, OSC promotes and stimulates computational research and education in order to act as a key enabler for the state's aspirations in advanced technology, information systems, and advanced industries. For additional information, visit http://www.osc.edu.
Source: Ohio Supercomputer Center
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.