Visit additional Tabor Communication Publications
September 30, 2010
Carnegie Mellon to provide expertise as a global leader in the field
PITTSBURGH, Sept. 30 -- The National Science Foundation (NSF) has announced a $10 million award to the New Mexico Consortium (NMC) at the Los Alamos National Laboratory (LANL), Carnegie Mellon University and the University of Utah to build and operate the Parallel Reconfigurable Observational Environment (PRObE), a one-of-a-kind computer systems research center.
This innovative concept utilizes decommissioned supercomputing systems from the Department of Energy (DOE) facilities to provide a very large-scale systems research capability. Targeted at both high-performance and data-intensive, or cloud, computing, the center will allow systems software researchers to have dedicated access to 1,000 computer clusters and control all application and operating system software down to and including the lowest-level hardware control systems. The PRObE center is the only one of its kind in the United States, and possibly the world.
"The need to expand research and educational opportunities for the systems research community is critical," said Garth Gibson, professor of computer science at Carnegie Mellon and respected thought leader in data storage and in data-intensive computing. "No sooner have computer systems such as LANL's Roadrunner achieved sustained petascale performance, capable of a trillion or more floating-point calculations per second, than we have recognized the need for exascale systems, which will be a thousand times faster," he said. "Designing exascale systems will be a tremendous challenge and one that will be difficult for the computer science community to meet without a resource such as PRObE."
"Computing researchers need to be able to test system-level innovations at scale," said Ed Lazowska, chair of the Computing Community Consortium and professor of computer science and engineering at the University of Washington. "This is the big gap. Nothing currently available fills it."
Academic researchers across the field recognize this need. Michael Dahlin, professor of computer science at the University of Texas at Austin, said, "Computer systems researchers need large-scale clusters to have any hope of doing much of the work we should be doing."
"About three years ago, we began to work on a way to re-utilize open/unclassified decommissioned supercomputers," said Gary Grider, co-PI and deputy division leader from LANL's High Performance Computing (HPC) Division. "We noticed that when new supercomputers are installed, there is a mad rush to get them into production with a focus on getting science applications to run quickly and well."
In the early phase of commissioning a new supercomputer, a significant amount of work goes into software development. The people that develop software at the systems level only get a chance to try new things out for a relatively short period of time while new large computers are brought online. "This presents an issue," Grider said, "as there is no large-scale resource for these systems-level people to utilize for long periods of time to develop new concepts and functions."
The DOE continually decommissions large supercomputers, some of which are open/unclassified resources. These systems can be used for high-performance and data-intensive computing systems research, however funding is needed to house, power and air condition the systems and to provide systems support people.
"NSF seemed like the natural government sponsor for such a concept," Grider said. "Also, to be flexible enough to be able to support this kind of research, it seemed appropriate to have universities involved."
PRObE builds on an existing partnership between the LANL and the NMC to support educational and research collaborations with universities. Carnegie Mellon provides expertise as a leader in computer systems research. The University of Utah will adapt software developed for its network emulation testbed -- Emulab -- to PRObE.
The Emulab software has been developed over the past decade by the Flux Research Group, part of the School of Computing at the University of Utah. It is widely used in the systems research community: it powers over three dozen testbeds used around the world by thousands of researchers and educators.
PRObE will be the largest-scale Emulab installation to date. "We are excited to be part of the PRObE effort," said Robert Ricci of the University of Utah, "because we believe it addresses and important gap in the public research infrastructure."
"PRObE may be built from recycled supercomputers, but because the hardware is not exotic, the same hardware will support data-intensive computing," said CMU's Gibson, who led the DOE's Petascale Data Storage Institute. "In CMU's experience this hardware will be excellent at running data analytics for eScience or Internet service applications using open source software such as Hadoop. This will allow PRObE to serve both styles of large-scale computing, high-performance computing and data-intensive computing."
"It's good to see the NSF outsource the construction and support of a flexible large-scale experimental datacenter to an organization designed to do just that," said Margo Seltzer, a professor of engineering and applied sciences at Harvard University. "Let's not spend university research resources replicating engineering that is better done by others."
In addition to providing the large-scale systems research environment, PRObE will conduct an innovative summer school to train university students in how to build and manage very large high-performance computing environments. Selected top students from the summer school will be invited to intern at the PRObE Center and the LANL.
The NMC is an independent non-profit corporation managed and led by the three New Mexico research universities -- University of New Mexico, New Mexico Institute of Mining and Technology, and New Mexico State University. Within the LANL, the PRObE partnership includes the High Performance Computing Division, the Institutional Computing Program, and the DOE/National Nuclear Security Administration Advanced Supercomputing Program. Additional support comes from the principal associate director for Science, Technology and Engineering and the associate director for Theory, Simulation and Computation.
See the PRObE website for more information (http://newmexicoconsortium.org/probe) or contact firstname.lastname@example.org.
About Carnegie Mellon University
Carnegie Mellon (www.cmu.edu) is a private, internationally-ranked research university with programs in areas ranging from science, technology and business, to public policy, the humanities and the arts. More than 11,000 students in the university's seven schools and colleges benefit from a small student-to-faculty ratio and an education characterized by its focus on creating and implementing solutions for real problems, interdisciplinary collaboration and innovation. A global university, Carnegie Mellon's main campus in the United States is in Pittsburgh, Pa. It has campuses in California's Silicon Valley and Qatar, and programs in Asia, Australia, Europe and Mexico. The university is in the midst of a $1 billion fundraising campaign, titled "Inspire Innovation: The Campaign for Carnegie Mellon University," which aims to build its endowment, support faculty, students and innovative research, and enhance the physical campus with equipment and facility improvements.
Source: Carnegie Mellon University
Contributing commentator, Andrew Jones, offers a break in the news cycle with an assessment of what the national "size matters" contest means for the U.S. and other nations...
Today at the International Supercomputing Conference in Leipzing, Germany, Jack Dongarra presented on a proposed benchmark that could carry a bit more weight than its older Linpack companion. The high performance conjugate gradient (HPCG) concept takes into account new architectures for new applications, while shedding the floating point....
Not content to let the Tianhe-2 announcement ride alone, Intel rolled out a series of announcements around its Knights Corner and Xeon Phi products--all of which are aimed at adding some options and variety for a wider base of potential users across the HPC spectrum. Today at the International Supercomputing Conference, the company's Raj....
Jun 19, 2013 |
Supercomputer architectures have evolved considerably over the last 20 years, particularly in the number of processors that are linked together. One aspect of HPC architecture that hasn't changed is the MPI programming model.
Jun 18, 2013 |
The world's largest supercomputers, like Tianhe-2, are great at traditional, compute-intensive HPC workloads, such as simulating atomic decay or modeling tornados. But data-intensive applications--such as mining big data sets for connections--is a different sort of workload, and runs best on a different sort of computer.
Jun 18, 2013 |
Researchers are finding innovative uses for Gordon, the 285 teraflop supercomputer housed at the San Diego Supercomputer Center (SDSC) that has a unique Flash-based storage system. Since going online, researchers have put the incredibly fast I/O to use on a wide variety of workloads, ranging from chemistry to political science.
Jun 17, 2013 |
The advent of low-power mobile processors and cloud delivery models is changing the economics of computing. But just as an economy car is good at different things than a full size truck, an HPC workload still has certain computing demands that neither the fastest smartphone nor the most elastic cloud cluster can fulfill.
Jun 14, 2013 |
For all the progress we've made in IT over the last 50 years, there's one area of life that has steadfastly eluded the grasp of computers: understanding human language. Now, researchers at the Texas Advanced Computing Center (TACC) are utilizing a Hadoop cluster on its Longhorn supercomputer to move the state of the art of language processing a little bit further.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
Join HPCwire Editor Nicole Hemsoth and Dr. David Bader from Georgia Tech as they take center stage on opening night at Atlanta's first Big Data Kick Off Week, filmed in front of a live audience. Nicole and David look at the evolution of HPC, today's big data challenges, discuss real world solutions, and reveal their predictions. Exactly what does the future holds for HPC?
Join our webinar to learn how IT managers can migrate to a more resilient, flexible and scalable solution that grows with the data center. Mellanox VMS is future-proof, efficient and brings significant CAPEX and OPEX savings. The VMS is available today.