Visit additional Tabor Communication Publications
September 29, 2006
University of Texas, Arizona State University, Cornell University and Sun Microsystems to deploy the world's most powerful general-purpose computing system on the TeraGrid
The National Science Foundation (NSF) has made a five-year, $59 million award to the Texas Advanced Computing Center (TACC) at The University of Texas at Austin to acquire, operate and support a high performance computing system that will provide unprecedented computational power to the nation's research scientists and engineers.
"This is a very valuable resource for the scientific community and society in general," said William Powers Jr., president of the university. "This award confirms that The University of Texas at Austin is an innovative leader in high performance computing and research." The award is the largest NSF award ever to The University of Texas at Austin.
The University of Texas at Austin project team is led by Dr. Jay Boisseau, director of TACC, and includes leading researchers from TACC and the Institute for Computational Engineering & Sciences (ICES). UT Austin, in collaboration with Sun Microsystems, Arizona State University and Cornell Theory Center (CTC) at Cornell University, submitted the proposal in response to the NSF's High Performance Computing System Acquisition Program's inaugural competition. The program is designed to deploy and support world-class high performance computing systems with tremendous capacity and capability to empower the U.S. research community. The award covers the acquisition and deployment of the new Sun system and four years of operations and support to the national community to enhance leading research programs. TACC will be the lead partner, with assistance from ICES, ASU and CTC in the areas of applications optimization, large-scale data management, software tools evaluation and testing, and user training and education.
High performance computing has become a vital investigative tool in many science and engineering disciplines. It enables testing and validation of theories and analysis of vast volumes of experimental data generated by modern scientific instruments, such as the very high-energy particle accelerators in the United States and Europe. HPC makes it possible for researchers to conduct experiments that would otherwise be impossible -- studying the dynamics of the Earth's climate in the distant past, for example, investigating how the universe developed, or discovering how complex biological molecules mediate the processes that sustain life. In industry, high performance computing is used in everything from aircraft design and improvement of automobile crash-worthiness, to the creation of breath-taking animations in the cinema and production of snack food.
The NSF Office of Cyberinfrastructure (OCI) coordinates and supports the acquisition, development and provision of state-of-the-art cyberinfrastructure resources, tools and services essential to 21st century science and engineering research and education, including HPC systems. The TeraGrid, sponsored by OCI, integrates a distributed set of high capability computational, data management and visualization resources to enable and accelerate discovery in science and engineering research, making research in the United States more productive. The new Sun HPC system at TACC will become the most powerful computational resource in the TeraGrid.
Juan Sanchez, vice president for research at UT Austin, said the new supercomputer will enable a new wave of research and researchers. "The Texas Advanced Computing Center is highly qualified to manage this powerful system, which will have a deep impact on science," Sanchez said. "The scale of the hardware and its scientific potential will influence technology research and development in many areas, and the results and possibilities will contribute to increasing public awareness of high performance computing. In addition, the project team is deeply committed to training the next generation of researchers for using HPC resources."
TACC is partnering with Sun Microsystems to deploy a supercomputer system specifically developed to support very large science and engineering computing requirements. In its final configuration in 2007, the supercomputer will have a peak performance in excess of 400 teraflops, making it one of the most powerful supercomputer systems in the world. It will also provide over 100 terabytes of memory and 1.7 petabytes of disk storage. The system is based on Sun Fire x64 (x86, 64-bit) servers and Sun StorageTek disk and tape storage technologies, and will use over 13,000 of AMD's forthcoming quad-core processors. It will be housed in TACC's new building on the J.J. Pickle Research Campus in Austin, Texas.
This system marks Sun's largest HPC installation to-date. "Sun's new supercomputer and storage technologies create a powerful combination that will allow TACC to build and operate a supercomputer delivering more than 400 teraflops," said Marc Hamilton, director of HPC Solutions, Sun Microsystems. "We are excited about extending our long standing relationship with TACC with this system, making it possible for scientists and engineers to reap the benefits of one of the world's most powerful supercomputers." Kevin Knox, AMD's vice president for worldwide commercial business, said, "The design and performance of the AMD Opteron processor and our planned quad-core processor roadmap have been integral in supplying the best option for high performance computing deployments to customers such as Sun to provide to businesses, universities and government research centers."
"The new Sun system will provide unmatched capability and capacity for scientific discovery for the open research community," Boisseau said. "The technologies in the new Sun systems will enable breakthrough performance on important science problems." Added Tommy Minyard, assistant director for advanced computational systems at TACC and the team project manager, "With tremendous and balanced processor, memory, disk, and interconnect capabilities, this powerful system will enable both numerically-intensive and large scale data applications in many scientific disciplines."
Under the agreement with the NSF, five percent of the computer's processing time will be allocated to industrial research and development through TACC's Science & Technology Affiliates for Research (STAR) program. "High performance computing is essential to innovation, in business as well as in science," said Melyssa Fratkin, TACC's industrial affiliates program manager. "We anticipate collaborations with a wide range of companies that will take advantage of this powerful computing system, to achieve the breakthrough insights they need to maintain a competitive edge in the global marketplace."
Another five percent will be allocated to other Texas academic institutions. "This resource will help Texas academic researchers provide answers to some of the most perplexing scientific questions," said Dr. Mark Yudof, chancellor of the University of Texas System.
The initial configuration of this system will go into production on June 1, 2007, with the final configuration in operation by October 2007. User training will begin shortly before deployment to help researchers utilize this resource. "Our Virtual Workshop technology will help researchers across the US rapidly come up to speed on using the new system," said Dave Lifka, CTC's director of high performance & innovative computing. Added Dan Stanzione, director of the ASU High Performance Computing Initiative, "Effectively training and supporting a national community will be just as important to addressing the most important scientific challenges as making the hardware available."
HPC systems are enabling researchers to address important problems in nearly all fields of science. From understanding the 3D structure and function of proteins to predicting severe weather events, HPC resources have become indispensable to knowledge discovery in life sciences, geosciences, social sciences and engineering, producing results that have direct bearing on society and quality of life. Furthermore, HPC resources are required for basic research across disciplines, from understanding the synthesis of all heavy elements via supernova explosions to mapping the evolutionary history of all organisms throughout the history of life on Earth.
"The new TACC/Sun system has great potential for advancing the study of quantum chromodynamics," said Bob Sugar, a research professor in the department of physics at the University of California. Sugar and his colleagues study the fundamental forces of nature to obtain a deeper understanding of the laws of physics -- electromagneticism, weak interactions, and quantum chromodynamics (QCD), the theory of the strong interactions. They also study the properties of matter under extreme conditions of temperature and density, such as those that existed immediately after the Big Bang.
"Our research requires highly capable computers," Sugar continued. "This system will lead to major advances in our work and that of many other high energy physicists. I expect to see important progress on problems that are presently beyond our reach."
As the head of the Theoretical and Computational Biophysics Group at the University of Illinois at Urbana-Champaign, Klaus Schulten conducts groundbreaking research in computational life science, investigating how cells in all organisms synthesize new proteins from genetic instructions and how plants convert sunlight into chemical energy. Schulten also assists bioengineers in developing medical nanodevices.
"TACC is a major provider of supercomputer power to U.S. researchers," Schulten said. "The new TACC/Sun system, combined with our group's award-winning parallel molecular dynamics code, promises to simulate the largest structures yet of living cells. This will turn the TACC/Sun system into a new type of microscope that shows how viruses infect human cells, how obesity is fought through the cell's own proteins, and how nature harvests sunlight to fuel all life on Earth," Schulten concluded.
"TeraGrid users will be able to conduct simulations that are currently impossible, and researchers from diverse fields of science will develop entirely new applications for scientific discovery," said Omar Ghattas of ICES, the project's chief applications scientist. Ghattas, Karl Schulz of TACC, and Giri Chukkapalli of Sun will lead the high-level collaborations activities with leading researchers across the US such as Sugar and Schulten to ensure that the Sun system is used most effectively on important and strategic research challenges.
"This Sun system will enable scientific codes to achieve greater performance on vastly larger problems, with higher resolution and accuracy, than ever before. It is no exaggeration to say it will be one of the most important scientific instruments in the world," concluded Boisseau.
Source: Texas Advanced Computing Center
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.