Visit additional Tabor Communication Publications
November 14, 2011
Directives-based programming makes accelerating applications using CPUs and GPUs dramatically easier than modifying underlying code
SEATTLE, Wash., Nov. 14 -- In an effort to make it easier for programmers to take advantage of parallel computing, NVIDIA, Cray Inc., the Portland Group (PGI), and CAPS enterprise announced today a new parallel-programming standard, known as OpenACC.
Initially developed by PGI, Cray, and NVIDIA, with support from CAPS, OpenACC is a new open parallel programming standard designed to enable the millions of scientific and technical programmers to easily take advantage of the transformative power of heterogeneous CPU/GPU computing systems.
OpenACC allows parallel programmers to provide simple hints, known as "directives," to the compiler, identifying which areas of code to accelerate, without requiring programmers to modify or adapt the underlying code itself. By exposing parallelism to the compiler, directives allow the compiler to do the detailed work of mapping the computation onto the accelerator.
OpenACC is anticipated to benefit a broad range of programmers working in chemistry, biology, physics, data analytics, weather and climate, intelligence, and many fields others. Existing compilers from Cray, PGI and CAPS are expected to provide initial support for the OpenACC standard beginning in the first quarter of 2012. The OpenACC standard is fully compatible and interoperable with the NVIDIA CUDA parallel programming architecture, which is designed to allow detailed control over the accelerator for maximum performance tuning.
Directives provide a common code base that is multi-platform and multi-vendor compatible, offering an ideal way to preserve investment in legacy applications by enabling an easy migration path to accelerated computing. Based on recent data, the majority of developers have reported 2x to 10x increases in application performance in as little as two weeks when using existing directive-based compilers.
"Compiler directives are integral as we continue the build-out and deployment of the Titan GPU-accelerated supercomputer," said Buddy Bland, Titan project director at Oak Ridge National Laboratory, referring to a system expected to be the world's fastest supercomputer. "Our ultimate goal is to have all Titan supercomputing code run on hybrid CPU/GPU nodes, and OpenACC will enable programmers to develop portable applications that maximize the performance and power efficiency benefits of this architecture."
"OpenACC represents a major development for the scientific community," said Jeffrey Vetter, joint professor in the Computational Science and Engineering School of the College of Computing at Georgia institute of Technology. "Programming models for open science by definition need to be flexible, open and portable across multiple platforms; OpenACC is well designed to fill this need. It provides a valuable new tool to empower the vast numbers of domain scientists who could benefit from application acceleration, but who may not have the funding or expertise to port their code to emerging architectures."
"I am enthusiastic about the future of accelerator technologies," said Michael Wong, CEO of the OpenMP Architecture Review Board. "The OpenACC announcement highlights the technically impressive initiative undertaken by members of the OpenMP Working Group on Accelerators. I look forward to working with all four companies within the OpenMP organization to merge OpenACC with other ideas to create a common specification which extends OpenMP to support accelerators. We look forward to incorporating accelerator support with the full support of all OpenMP members in a future version of the OpenMP specification."
More information about OpenACC, as well as the OpenACC specification, is available today at www.OpenACC-standard.org. OpenACC is based largely on the PGI and Cray accelerator programming models. NVIDIA, Cray, PGI, and CAPS are members of the OpenMP subcommittee on accelerators, and intend to work within that organization to come to a common standard.
Developers interested in trying directives-based programming can obtain a one-month free trial of the PGI Accelerator Fortran and C compilers at NVIDIA's new 2x in 4 weeks web site, or by contacting Cray.
About Cray Inc.
As a global leader in supercomputing, Cray provides highly advanced supercomputers and world-class services and support to government, industry and academia. Cray technology is designed to enable scientists and engineers to achieve remarkable breakthroughs by accelerating performance, improving efficiency and extending the capabilities of their most demanding applications. Cray's Adaptive Supercomputing vision is focused on delivering innovative next-generation products that integrate diverse processing technologies into a unified architecture, allowing customers to surpass today's limitations and meeting the market's continued demand for realized performance. Go to www.cray.com for more information.
About The Portland Group (PGI)
The Portland Group, a wholly-owned subsidiary of STMicroelectronics (NYSE: STM), is the premier supplier of high-performance parallel Fortran, C, and C++ compilers and tools for workstations, servers, and clusters based on x64 processors from Intel and AMD, and GPUs accelerators from NVIDIA. Further information on The Portland Group products can be found at www.pgroup.com, by calling Sales at (503) 682-2806, or by email to email@example.com
About CAPS enterprise, The Many-Core Programming Company
CAPS is a major supplier of solutions dedicated to application migration and deployment on many-core processors. CAPS global solutions for many-core lead the developer to performance by providing top-of-the-range technology (HMPP hybrid compiler and wizard), code porting methodology and ecosystem. Its directive-based & multi-target HMPP compiler enables developers to safely move to hybrid CPU / GPU model and leverage the computing power of stream processors without the pain associated to GPU programming. HMPP is offered within CAPS DevDeck package: an ALL-IN-ONE multi-level suite for many-core application definition, porting and optimization with tools, methodology and resources.
NVIDIA (NASDAQ: NVDA) awakened the world to computer graphics when it invented the GPU in 1999. Today, its processors power a broad range of products from smart phones to supercomputers. NVIDIA's mobile processors are used in cell phones, tablets and auto infotainment systems. PC gamers rely on GPUs to enjoy spectacularly immersive worlds. Professionals use them to create visual effects in movies and design everything from golf clubs to jumbo jets. And researchers utilize GPUs to advance the frontiers of science with high-performance computing. The company holds more than 2,100 patents worldwide, including ones covering ideas essential to modern computing. For more information, see www.nvidia.com.
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.