Visit additional Tabor Communication Publications
October 21, 2010
Here is a collection of highlights from this week's news stream as reported by HPCwire.
UC Santa Barbara-Led Team Developing Next-Generation Ethernet
Sony Equips PCs with World Community Grid Software
Allinea Software Signs Further Collaboration Agreement with CEA
Netlist Demonstrates 100 VMs on a Single Standard Server Using HyperCloud Memory at Interop
IBM Reports 12 Percent Increase in Net Income for Q3
Voltaire Grows adVantage Partner Program to More Than 50 Members
Fusion-io Creates New Technology Alliance Program to Drive Innovation Through Collaboration
HyperWorks Partner Alliance Adds RAMSIS by Human Solutions
Coalition of High Performance Computing Leaders Form Community-Based Open-Source File System Alliance
SDSC Celebrates Its 25th Year
Supermicro Showcases HPC Servers at SEG 2010
AMAX Introduces Petabyte-Scale NAS Clustered Storage Solutions for Oil and Gas Exploration
Rocky Mountain Supercomputing Centers Introduces M.O.R.E. POWER Service
CANARIE, Ciena Demo 100G Network
Mellanox InfiniBand Switch Systems Selected by IBM
Philip E. Bourne Wins Microsoft's 2010 Jim Gray eScience Award
XtreemOS Consortium Announces Public Access to Open Test Bed
NSF Grant to Study National Energy Policy and Technology Impacts
New Algorithm Reduces Linear Equation Runtimes
Computer scientists at Carnegie Mellon University have developed a ground-breaking algorithm that can solve systems of linear equations used in important applications, including image processing, logistics and scheduling problems, and recommendation systems. The new algorithm is incredibly efficient and may make it possible for a desktop workstation to solve systems with a billion variables in just a few seconds.
Linear systems are used to model real-world systems, such as transportation, energy, telecommunications and manufacturing, which often include millions, or even billions, of variables. Solving such complex systems is time-consuming on even the fastest systems and has confounded computer scientists and stymied research goals for a long long time. In fact, solving simultaneous equations quickly and accurately is truly an age old mathematical problem. One of the classic algorithms for solving linear systems, which is today dubbed Gaussian elimination, was first published by Chinese mathematicians 2,000 years ago.
Researchers from Carnegie Mellon's Computer Science Department have experienced a breakthrough, one that has great practical potential. The algorithm that they've devised relies on new tools from graph theory, randomized algorithms and linear algebra to greatly speed the time to completion for these linear system problems, with runtimes up to a billion times faster than with Gaussian elimination.
The algorithm applies to a class of problems known as symmetric diagonally dominant (SDD) systems, which have gained prominence in recent years. Recommendation systems, like that used by Netflix, use SSD to compare the preferences of an individual to those of millions of other customers. Image processing, logistics, and engineering are other key uses cases for SSD.
The press release highlights the importance of this achievement:
"The new linear system solver of Koutis, Miller and Peng is wonderful both for its speed and its simplicity," said Spielman, a professor of applied mathematics and computer science at Yale. "There is no other algorithm that runs at even close to this speed. In fact, it's impossible to design an algorithm that will be too much faster."
The work will be presented at the annual IEEE Symposium on Foundations of Computer Science (FOCS 2010), Oct. 23-36 in Las Vegas, and the group's research paper, "Approaching Optimality for Solving SDD Linear Systems," can be downloaded at http://www.cs.cmu.edu/~glmiller/Publications/Papers/KoutisApproaching-2010.pdf.
University of Queensland Deploys SGI Supercomputer
This week the University of Queensland increased its technical computing prowess with a high performance computing (HPC) solution from SGI. The SGI Rackable half-depth servers will be used to support a broad range of research from the fields of bioinformatics, computational chemistry, finite element analysis, computational fluid dynamics, earth sciences, market economics and image processing.
According to Professor Max Lu, deputy vice-chancellor or research at the University of Queensland, "These computers will strengthen an important part of the University's research capacity. Tasks such as processing enormous amounts of biological data generated through techniques such as genome-sequencing, micro-arrays and imaging cannot be done on standard desktop computers."
This will be one of the biggest deployments in Australia. The new SGI system boasts 3,144 processor cores, specifically Intel Xeon 5500 and 7500 series processors, with 11.52 TB memory and 249 TB of disk storage. Other specifications include InfiniBand QDR interconnect with Voltaire Grid Director 4700 switches and Unified Fabric Manager switching technology, and a Panasas file system. DC-based racks and innovative cooling techniques were selected for their energy-efficiency. The design offers flexible configurations to suit the university's current and future requirements. The university opted for SGI Professional Services to provide project management, installation services, datacenter services, training, as well as ongoing consultation and maintenance.
The new machine will be put to work handling the complex research and data needs of universities in Queensland and partner organizations, such as the Queensland Cyber Infrastructure Foundation (QCIF), Commonwealth Scientific and Industrial Research Organisation (CSIRO) and Bioplatforms Australia. Additionally, the infrastructure will be hosting several projects, including the National Computational Infrastructure (NCI) Specialised Facility in Bioinformatics and the European Molecular Biology Laboratory (EMBL) Australia, European Bioinformatics Institute (EBI) Mirror project.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
May 08, 2013 |
For engineers looking to leverage high-performance computing, the accessibility of a cloud-based approach is a powerful draw, but there are costs that may not be readily apparent.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.