Visit additional Tabor Communication Publications
November 15, 2011
SEATTLE, Nov. 15 -- The University of Tennessee's National Institute for Computational Sciences (NICS) announced today at SC11 that it has entered a multi-year strategic engagement with Intel Corporation to pursue development of next-generation, high-performance computing solutions based on the Intel Many Integrated Core (Intel MIC) architecture and the design of scientific applications emphasizing a sustainable approach for both performance and productivity.
Funded by the National Science Foundation (NSF) and located at Oak Ridge National Laboratory, NICS manages the Cray XT5 "Kraken," the NSF's most productive supercomputer. Intel is the world's leader in silicon innovation and the designer of the Intel MIC architecture. Together, NICS and Intel make a formidable team for high-performance computing (HPC) development.
"As a leading center for high-performance computing, NICS is working with Intel to ensure that the Intel MIC architecture and its software environment are developed to meet the needs of the scientific supercomputing community," explained Glenn Brook, director of the Application Acceleration Center of Excellence and head of the Intel MIC architecture initiative at NICS.
"Intel is pleased to work closely with NICS to help move highly parallel scientific codes from multicore to many core technology," said Raj Hazra, general manager of the Technical Computing Group at Intel Corporation. "Our partnership with NICS seeks to help scientists reduce application development time with the benefit of faster time to results and insight."
Next-generation systems based on the Intel MIC architecture will combine multicore Intel Xeon processors with Intel MIC co-processors. Such systems are expected to propel high-end computing from its current status at the petascale level to its future at the exascale level, when machines will be capable of a thousand trillion calculations per second.
Constructed around manycore processors that utilize the venerable x86 instruction set, the Intel MIC architecture will prove more versatile and serviceable than architectures based on graphics processing units (GPUs). Users will be able to reuse their existing codes and programming knowledge, allowing for immediate scientific discovery and increased productivity. Further, users will be able to use standardized approaches and tools such as OpenMP and Intel Parallel Studio to optimize their codes to achieve high performance on both host processors and co-processors.
"NICS has provided insight to Intel regarding the technology requirements of the scientific computing community," added Joe Curley, director of marketing for Intel's Technical Computing Group. "The impact of our partnership can be seen in the focus of our Intel MIC 'Knights Ferry' software development platform on extending well understood, high-level, standard programming languages and models."
NICS is committed to assisting NSF users with their transition to the Intel MIC architecture by researching parallelization techniques on the Intel MIC platform and by porting key NSF applications to the Intel MIC architecture in advance of its commercial release. NICS will also offer training on the Intel MIC architecture following its commercial debut.
"We want to help users prepare for future architectures so they can immediately make effective and efficient use of new technologies as they become commercially available," stated Brook.
NICS has already successfully ported millions of lines of code -- full applications using MPI and OpenMP, not just kernels -- from a variety of scientific and engineering disciplines to the Intel MIC architecture. Now, NICS is preparing to expand its efforts by developing partnerships with NSF research teams to port and optimize key NSF research codes. To facilitate this activity, NICS is working with Intel to provide a cluster based on the Intel MIC architecture for use by NSF project teams early next year.
"Increasing problem sizes and solution times associated with groundbreaking science drive the need for new high-performance architectures," said Brook. "We are committed to maximizing returns from the Intel MIC architecture while minimizing investments of time for development and training. We see the Intel MIC architecture as the fundamental technology that will drive the next generation of supercomputers to the heights of new scientific discoveries."
The National Institute for Computational Sciences (NICS) is a joint effort of the University of Tennessee and Oak Ridge National Laboratory that is funded by the National Science Foundation (NSF). Located on the campus of Oak Ridge National Laboratory, NICS is a major partner in NSF's Extreme Science and Engineering Discovery Environment (XSEDE).
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
Although Horst Simon was named Deputy Director of Lawrence Berkeley National Laboratory, he maintains his strong ties to the scientific computing community as an editor of the TOP500 list and as an invited speaker at conferences.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
May 08, 2013 |
For engineers looking to leverage high-performance computing, the accessibility of a cloud-based approach is a powerful draw, but there are costs that may not be readily apparent.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.