Visit additional Tabor Communication Publications
September 01, 2011
Panasas made news in 2008 when it was selected as the storage system to power the world’s first petascale system and they are looking ahead to 2018—the projected year that exascale computing will arrive.
This week the company’s co-founder and CTO, Garth Gibson, hit on some of the specific challenges on the storage front that will need to be addressed for the era of 70 TB/s bandwidth required for exascale computing.
Gibson said that most estimates contend that the storage system for checkpointing will need to deliver 60-70 TB/s of storage, which means his company is facing a lofty target. Panasas and other storage vendors with their eye on the 2018 prize will be looking at approximately 20% per year disk improvement to get to that level—and that is, as Gibson simply put it, “a lot of disks.”
He claims that the most cost effective way to reach these goals is to stage through SSD, not so much because of the non-volatility issue, but because it provides cheap memory capability and “decent megabytes per second per dollar.”
Gibson detailed the SSD slant to Panasas’ exascale strategy:
"I think there’s a good likelihood that we’ll see a checkpoint restart package evolve that will take data from the main memories of the nodes, drop it into SSD then go back to the compute and dribble to the storage—that will probably give us a factor of ten as the ratio of the amount of time we spend in capturing checkpoints versus the amount of time before the next checkpoint, so we’ll probably get to storage in the neighborhood of six terabytes per second.
Now that’s a much more achievable number in terms of the cost of the storage system but the disks aren’t going away because they’re three to thirty times the capacity of main memory and you don’t want to try to provision all of that in any other solid state technology because of cost."
In such conversations the issue of high-density capabilities always tends to arise. Gibson addressed this by pointing to new movements in the drive industry, which he says is finally realizing that even though it’s been possible for some time, there is value in creating read heads that are thinner than write heads, which creates an opportunity to build higher density drives by taking the write head and shifting it half a track over, thus half overwriting the last track since the read track can still see the half of the track that wasn’t overwritten.
While he notes that this isn’t “perfect” he does say that it mirrors current technologies already. In his words, “The big problem is that when you go to do a rewrite of a sector, you would trash the downstream sectors along the way, which is going to force a change in behavior—but that’s what we do with SSDs anyway."
Full story at Panasas
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.