ESG Labs Evaluates the Panasas® ActiveStor™ 12, the World’s Fastest HPC Storage System

By Nicole Hemsoth

March 28, 2011

Companies and institutions in the bioscience, energy, government, finance, manufacturing, and other research and development sectors are increasingly realizing the significant benefits of scalable clustered computing. As cluster sizes continue to grow, scaling storage I/O performance from large numbers of clients to the same file system—or even the same file—requires deploying high performance parallel storage designed specifically to address this need.
Panasas ActiveStor 12 (PAS 12) brings plug-and-play simplicity to large-scale storage deployments and is well known for achieving blazingly fast performance levels without sacrificing reliability or manageability. PAS 12 more than doubles performance over the prior generation, scaling to 150GB/s of throughput from a single 4PB file system.

ESG documents the results of its hands-on testing of the performance scalability of PAS 12. ESG also evaluated PAS 12 with an eye on validating the claims of appliance-like simplicity in support of large scale storage environments with demanding HPC performance requirements.

Click here to download the full ESG report.

According to ESG, data growth has outstripped e-mail- and database-driven structured data growth for some time now. The drivers are diverse: general purpose office data, richer file formats, ubiquitous photo and video, online communities, collaboration tools, 3-D modeling, and 4D imaging to name a few. 

Economic conditions have led businesses to emphasize both CAPEX, and OPEX reductions now more than ever; file growth has resulted in higher costs in terms of additional storage, complex management, and data center floor space and energy. As a result, the ability to scale out—that is, independently scale and tune bandwidth, processing, and storage capacity on the fly while managing a single, global namespace—is extremely popular for increasing efficiency and saving money.
Adoption of scale-out storage solutions is driven by their  ability to address multiple challenges. ESG survey respondents report selecting scale-out storage for a number of reasons: better scalability with easier management, improved performance of both I/O and throughput, faster storage provisioning, and reduced infrastructure cost.

But traditional file storage is only part of the challenge. Large enterprises have been using HPC to garner a competitive advantage for a long time. From technology development, to financial modeling, to seismic analysis and bio/pharmaceutical applications, HPC is critical in the development of the products and services that drive revenue in the enterprise.

HPC has a special set of requirements that traditional file storage solutions cannot address, including support for very large file systems and mount points and the ability for a large number of systems to read from and write to not only the same file share, but the same file with high throughput, to name two. Storage performance has a significant impact on HPC application performance as well: if compute nodes in an HPC cluster are waiting to access data needed for processing, the job takes longer to complete, leading to delays in time to market and potentially lost revenue.

Click here to download the full ESG report.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

NREL ‘Eagle’ Supercomputer to Advance Energy Tech R&D

August 14, 2018

The U.S. Department of Energy (DOE) National Renewable Energy Laboratory (NREL) has contracted with HPE for a new 8-petaflops (peak) supercomputer that will be used to advance early-stage R&D on energy technologies s Read more…

By Tiffany Trader

Training Time Slashed for Deep Learning

August 14, 2018

Fast.ai, an organization offering free courses on deep learning, claimed a new speed record for training a popular image database using Nvidia GPUs running on public cloud infrastructure. A pair of researchers trained Read more…

By George Leopold

CERN Project Sees Orders-of-Magnitude Speedup with AI Approach

August 14, 2018

An award-winning effort at CERN has demonstrated potential to significantly change how the physics based modeling and simulation communities view machine learning. The CERN team demonstrated that AI-based models have the Read more…

By Rob Farber

HPE Extreme Performance Solutions

Introducing the First Integrated System Management Software for HPC Clusters from HPE

How do you manage your complex, growing cluster environments? Answer that big challenge with the new HPC cluster management solution: HPE Performance Cluster Manager. Read more…

IBM Accelerated Insights

Super Problem Solving

You might think that tackling the world’s toughest problems is a job only for superheroes, but at special places such as the Oak Ridge National Laboratory, supercomputers are the real heroes. Read more…

Rigetti Eyes Scaling with 128-Qubit Architecture

August 10, 2018

Rigetti Computing plans to build a 128-qubit quantum computer based on an equivalent quantum processor that leverages emerging hybrid computing algorithms used to test programs and potential applications. Founded in 2 Read more…

By George Leopold

NREL ‘Eagle’ Supercomputer to Advance Energy Tech R&D

August 14, 2018

The U.S. Department of Energy (DOE) National Renewable Energy Laboratory (NREL) has contracted with HPE for a new 8-petaflops (peak) supercomputer that will be Read more…

By Tiffany Trader

CERN Project Sees Orders-of-Magnitude Speedup with AI Approach

August 14, 2018

An award-winning effort at CERN has demonstrated potential to significantly change how the physics based modeling and simulation communities view machine learni Read more…

By Rob Farber

Intel Announces Cooper Lake, Advances AI Strategy

August 9, 2018

Intel's chief datacenter exec Navin Shenoy kicked off the company's Data-Centric Innovation Summit Wednesday, the day-long program devoted to Intel's datacenter Read more…

By Tiffany Trader

SLATE Update: Making Math Libraries Exascale-ready

August 9, 2018

Practically-speaking, achieving exascale computing requires enabling HPC software to effectively use accelerators – mostly GPUs at present – and that remain Read more…

By John Russell

Summertime in Washington: Some Unexpected Advanced Computing News

August 8, 2018

Summertime in Washington DC is known for its heat and humidity. That is why most people get away to either the mountains or the seashore and things slow down. H Read more…

By Alex R. Larzelere

NSF Invests $15 Million in Quantum STAQ

August 7, 2018

Quantum computing development is in full ascent as global backers aim to transcend the limitations of classical computing by leveraging the magical-seeming prop Read more…

By Tiffany Trader

By the Numbers: Cray Would Like Exascale to Be the Icing on the Cake

August 1, 2018

On its earnings call held for investors yesterday, Cray gave an accounting for its latest quarterly financials, offered future guidance and provided an update o Read more…

By Tiffany Trader

Google is First Partner in NIH’s STRIDES Effort to Speed Discovery in the Cloud

July 31, 2018

The National Institutes of Health, with the help of Google, last week launched STRIDES - Science and Technology Research Infrastructure for Discovery, Experimen Read more…

By John Russell

Leading Solution Providers

SC17 Booth Video Tours Playlist

Altair @ SC17

Altair

AMD @ SC17

AMD

ASRock Rack @ SC17

ASRock Rack

CEJN @ SC17

CEJN

DDN Storage @ SC17

DDN Storage

Huawei @ SC17

Huawei

IBM @ SC17

IBM

IBM Power Systems @ SC17

IBM Power Systems

Intel @ SC17

Intel

Lenovo @ SC17

Lenovo

Mellanox Technologies @ SC17

Mellanox Technologies

Microsoft @ SC17

Microsoft

Penguin Computing @ SC17

Penguin Computing

Pure Storage @ SC17

Pure Storage

Supericro @ SC17

Supericro

Tyan @ SC17

Tyan

Univa @ SC17

Univa

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This