In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they’re turning an eye to solving data-intensive challenges. We spoke with the agency’s Irene Qualters and Barry Schneider about..
<img src=”http://media2.hpcwire.com/hpcwire/light-speed.gif” alt=”” width=”94″ height=”94″ />The Intelligence Advanced Research Projects Activity (IARPA) is putting out some RFI feelers in hopes of pushing new boundaries with an HPC program. However, at the core of their evaluation process is an overt dismissal of benchmarks, including floating operations per second (FLOPS).
UV 2 system can create heat maps of tweets during hurricanes and elections.
<img style=”float: left;” src=”http://media2.hpcwire.com/hpcwire/Convey_boards.jpg” alt=”” width=”98″ height=”85″ />Last week at SC12 in Salt Lake Convey pulled the lid off its MX big data-driven architecture designed to shine against graph analytics problems, which were at the heart of the show’s unmistakable data-intensive computing thrust this year. The new MX line is designed to exploit massive degrees of parallelism while efficiently handling hard-to-partition big data applications.
<img style=”float: left;” src=”http://media2.hpcwire.com/hpcwire/bigdatagraphic_132x.jpg” alt=”” width=”75″ height=”105″ />Big data is all the rage these days. It is the subject of a recent Presidential Initiative, has its own news portal, and, in the guise of Watson, is a game show celebrity. Big data has also caused concern in some circles that it might sap interest and funding from the exascale computing initiative. So, is big data distinct from HPC – or is it just a new aspect of our evolving world of high-performance computing?
<img style=”float: left;” src=”http://media2.hpcwire.com/hpcwire/ESnet_logo.jpg” alt=”” width=”114″ height=”119″ />In order to help research institutions capitalize on the growing availability of high-bandwidth networks to manage their growing data sets, the DOE’s Energy Sciences Network, known as ESnet, is working with the scientific community to encourage the use of a network design model called the “Science DMZ.” Leading the development of this effort is Eli Dart, a network engineer with previous experience at Sandia National Laboratories and the National Energy Research Scientific Computing Center. In this interview, Dart talks about the nature of the project and explains how such an architecture can help researchers.
<img style=”float: left;” src=”http://media2.hpcwire.com/hpcwire/Convey_HC-2.bmp” alt=”” width=”141″ height=”61″ />Convey Computer has launched its newest x86-FPGA “hybrid-core” server. Dubbed HC-2, it represents the first major upgrade of the system since the company introduced the HC-1 product back in 2008. The new offering promises much better performance, but with a similar price range as the original system.
Gordon, the largest flash memory-based computer on the planet, was officially launched at a ceremony that took place on Monday at the San Diego Supercomputer Center (SDSC). Two years in the making, and backed by a $20 million Track 2 grant from the National Science Foundation (NSF), Gordon represents the first really big purpose-built supercomputer for data-intensive applications.
Indiana University’s Scinet Research Sandbox entry sets new records, renews promise of cloud for data-intensive science workloads.
One of the most promising use cases for “Big Data” is to help advance climate research. At SC11, Reinhard Budich (Max Planck Institute for Meteorology), John Feo (Pacific Northwest National Laboratory) Tobias Weigel (DKRZ) and Per Nyberg (Cray) will co-host the second Climate Knowledge Discovery (CKD) workshop to explore new data-intensive methods. HPCwire talked about this with Budich and Feo.