November 2, 2010
There is a growing feeling that merely taking the latest processor offerings from Intel, AMD or IBM will not get us to exascale in a reasonable time frame, cost budget, and power constraint. One avenue to explore is designing and building more specialized systems, aimed at the types of problems seen in HPC, or at least at the problems seen in some important subset of HPC. Of course, such a strategy loses the advantages we've enjoyed over the past two decades of commoditization in HPC; however, a more special purpose design may be wise, or necessary. Read more…
Did you miss out on Supercomputing 2017? Did you attend, but were stuck in meetings the whole time without an opportunity to walk the show floor and see what new announcements were being made? HPCwire's got you covered, we visited some of the hottest booths in the exhibit hall and spoke with their top executives to get the scoop on the latest solutions, partnerships, and product announcements.
From mismatches between compute and storage capabilities to colossal data volumes, data storage presents a number of challenges for scientific research. And as silos pop up and challenges expand, the pace of research often suffers.
As genomic data becomes ubiquitous, infrastructure bottlenecks for life sciences organizations are narrowing. But speedy analysis and real-time decision making don't have to remain out of reach: modern end-to-end systems are emerging as flexible solutions for a competitive edge.
© HPCwire. All Rights Reserved. A Tabor Communications Publication
Reproduction in whole or in part in any form or medium without express written permission of Tabor Communications, Inc. is prohibited.