Sandia National Labs decommissions legendary supercomputer.
A researcher at the New Mexico State University is modeling new ways to address common challenges with data-intensive, graph-based problems.
The Weekly Top Five features the five biggest HPC stories of the week, condensed for your reading pleasure. This week, we cover the Cray/Sandia partership to found a knowledge institute; RenderStream’s FireStream-based workstations and servers; NVIDIA’s latest CUDA centers; Reservoir Labs and Intel’s extreme scale ambitions; and Jülich Supercomputing Centre’s new hybrid cluster.
The Weekly Top Five features the five biggest HPC stories of the week, condensed for your reading pleasure. This week, we cover the TeraGrid effort to support the Japanese research community; NNSA’s ‘Supercomputing Week’ coverage; Mellanox’s new double-duty switch silicon; Platform’s latest Symphony; and the Oracle Sun Server-based Sandia Red Sky/Red Mesa supercomputer upgrades.
Data-intensive applications are quickly emerging as a significant new class of HPC workloads. For this class of applications, a new kind of supercomputer, and a different way to assess them, will be required. That is the impetus behind the Graph 500, a set of benchmarks that aim to measure the suitability of systems for data-intensive analytics applications.
Researchers virtualize Sandia’s Red Storm supercomputer; and Princeton University announces plans for new research center. We recap those stories and more in our weekly wrapup.