Data-centric Organizations need both Performance and Flexible Data Management

January 27, 2020

Converged AI and HPC have arrived in 2020 and commercial IT organizations are putting data at the core of their business model. Just in the HPC market alone, Intersect360[1] found that around 90% of HPC organizations were either running machine learning (56%) or investigating/planning its introduction. Gartner[2] claims that 60% of organizations are adapting their business model to AI and a 2019 ESG[3] survey showed that 49% of IT organizations state that “data is their business”, with another 31% expecting to offer datacentric products in the next 2 years.  Yet, how to reliably accomplish that goal often remains elusive.  The same Gartner survey reported a 50% failure rate for these projects.

Performance, as ever, is important and DDN’s EXAScaler systems continue to outperform forthcoming competitor offerings with over 63 Million IOPs per Rack with DDN’s AI400X and over 140GB/s per rack in HDD-only sequential performance. But delivering this performance potential to today’s workloads is the real target. Datacentric organizations are investing heavily in more powerful compute platforms, often relying on GPUs. Keeping systems operating at 100% utilization is not an easy proposition without at-scale experience and concerted engineering effort in optimizing for containerized workloads, AI frameworks, GPU platforms, extreme-scale CPU and fast networks. Underutilized infrastructure is a recipe for failure in a world where competitive stakes are high, investments are large and the amount of data to be analyzed is vast. A data management vendor must deliver performance no matter how large the requirement nor how diverse the applications.

Modern organizations shifting to a data-centric strategy need very large scale data-in-place analytics environments that allow data to be accessible by many different methods for ingestion, labeling, processing, etc. In this rapidly changing and converging ecosystem of AI, Big Data and HPC, fast yet flexible access to all data is a differentiator. Accessing and manipulating that data from workstations and cloud applications require protocols like NFS, SMB, and S3 as well as super-fast Native access. With the demands for data mobility (getting the right data in the right place at the right time), the ability to copy, move or sync data between systems and to third party cloud providers is equally important.

There is a clear priority for organizations to build the right infrastructure for data-centric services and selecting the right vendor partnerships will improve their path to success. Some suppliers struggle with the ability to manage data easily at petabyte scale. Many vendors have neither the focus on scalable data processing nor familiarity with those types of environments. Other vendors can’t deliver consistent performance or reliable large scale solutions. Whilst Data Lakes and pure Big Data approaches are declining due to inflexibility and inefficiency, there is a surge in implementations of true parallel filesystems to address AI, Big Data and HPC challenges. DDN’s EXAScaler is regularly implemented at scales of 100s of Petabytes without the downsides of data duplication and with far higher performance delivered to analytics and AI applications than any object store. EXAScaler brings a simpler experience and a powerful in-built data management engine to ensure businesses can stay agile with their data as new analytics models evolve and their data sets broaden.

The IT infrastructure market is transforming when it comes to data-centric requirements. No longer consigned to a small research group or the high performance computing team, the ability to acquire, analyze, manipulate and distribute data is core to IT operations. Similarly, the landscape of vendors to consider is undergoing significant change, whether through mergers and acquisitions or other market realities. Selecting a proven, stable and trusted partner with a track record of innovation to meet at-scale needs is an important first step to becoming a data-driven organization or informing the next market-changing AI application.

About DDN

DDN ranks #1 for HPC Storage deployments[4] at scale and holds a leadership position in AI and HPC with over 10 EXAbytes of shipped capacity and 150+ patents in data management. Recently awarded “Top 5 vendors to watch”[5] for our A3I product line and capability advances in the EXA5 platforms DDN continues to invest in innovation focussed on our customer’s forward path. As stated by IDC[6], “Increasingly, parallelization is the preferred approach, with AI infrastructure starting to resemble HPC infrastructure”. With a market leading 16% of revenues invested back into our R&D, DDN’s accelerated cadence of innovation in our parallel filesystems for AI and HPC is second to none.

Why is DDN EXA5 your undisputed filesystem of choice for the new converged world of AI and HPC?

#1 DDN Delivers the best customer experience. DDN is a long-term dependable partner with the technical expertise that improves time to value and accelerates your productivity.

#2 EXA5 Performance and Scale Leadership – DDN platforms attain the highest performance efficiency.

#3 EXA5 Capability – EXA5 brings the capabilities needed for converged AI and HPC – MultiCloud, Data Aware Intelligence, Global Collaboration and Strong Security including Multi Tenancy support

The DDN Customer Experience

DDN is different because our focus is different. That focus is wholly on our 7000 customers and their data challenges. Our customers benefit by partnering with DDN for storage alongside compute vendors, bringing DDN’s unique expertise in storage. With a DDN storage strategy, businesses can escape from monolithic proprietary systems, mitigating against spiralling license costs and retaining their control in getting the best return on investment.

“With the goal of expanding the boundaries of science as we know it today, we are excited about the arrival of advanced new technology that can dramatically increase the performance at scale of our systems, and specifically of our new Top 10 supercomputer, Frontera. DDN’s new EXA5 has the power to provide the best I/O performance our users have ever experienced, and greatly reduce the I/O bottlenecks in large scale computation. We believe that EXA5 will play a role in many of the ground breaking discoveries scientists will make with Frontera.” – Dan Stanzione, executive director, Texas Advanced Computing Center.

DDN brings the skills of more than 1000 technical staff, all focussed on storage and data management. With filesystem engineering teams that contribute tens of thousands of lines of code each year[7] to our open filesystems, field teams that deliver the largest tiered storage into top 10 supercomputers[8][9]and global operations that support billions of dollars of infrastructure, DDN is the largest privately held storage company in the world. Now with DDN’s rapid expansion into Enterprise[10] we have dramatically added our ability to accelerate your entire infrastructure.


[1] Intersect360, HPC User Site Census: Storage Suppliers, May 2019

[2] https://www.idc.com/getdoc.jsp?containerId=prUS45344519

[3] ESG Master Survey Results, 2019 Data Storage Trends, November 2019

[4] Intersect360 Research, ibid.

[5] https://www.ddn.com/press-releases/ddn-honored-top-five-vendor-watch-annual-datanami-awards/

[6] https://www.idc.com/getdoc.jsp?containerId=DR2019_T2_PR

[7] http://cdn.opensfs.org/wp-content/uploads/2019/07/LUG2019-Community-Release-Update-Jones.pdf

[8] https://www.ddn.com/blog/tacc-frontera-helping-make-impossible-possible/

[9] http://lustrefs.cn/wp-content/uploads/2018/04/Lustre-Persistent-Cache-on-Client-for-AI-Machine-Learning-and-Bigdata-Processing.pdf

[10] https://www.nextplatform.com/2019/10/09/ddn-uses-acquisitions-to-grow-in-the-enterprise/

Shares
found that around 90% of HPC organizations were either running machine learning (56%) or investigating/planning its introduction. Read more…

" share_counter=""]
Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

With New Owner and New Roadmap, an Independent Omni-Path Is Staging a Comeback

July 23, 2021

Put on a shelf by Intel in 2019, Omni-Path faced a uncertain future, but under new custodian Cornelis Networks, OmniPath is looking to make a comeback as an independent high-performance interconnect solution. A "significant refresh" – called Omni-Path Express – is coming later this year according to the company. Cornelis Networks formed last September as a spinout of Intel's Omni-Path division. Read more…

PEARC21 Panel Reviews Eight New NSF-Funded HPC Systems Debuting in 2021

July 23, 2021

Over the past few years, the NSF has funded a number of HPC systems to further supply the open research community with computational resources to meet that community’s changing and expanding needs. A review of these systems at the PEARC21 conference (July 19-22) highlighted... Read more…

Chameleon’s HPC Testbed Sharpens Its Edge, Presses ‘Replay’

July 22, 2021

“One way of saying what I do for a living is to say that I develop scientific instruments,” said Kate Keahey, a senior fellow at the University of Chicago and a computer scientist at Argonne National Laboratory, as s Read more…

PEARC21 Plenary Session: AI for Innovative Social Work

July 21, 2021

AI analysis of social media poses a double-edged sword for social work and addressing the needs of at-risk youths, said Desmond Upton Patton, senior associate dean, Innovation and Academic Affairs, Columbia University. S Read more…

Summer Reading: “High-Performance Computing Is at an Inflection Point”

July 21, 2021

At last month’s 11th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies (HEART), a group of researchers led by Martin Schulz of the Leibniz Supercomputing Center (Munich) presented a “position paper” in which they argue HPC architectural landscape... Read more…

AWS Solution Channel

Accelerate innovation in healthcare and life sciences with AWS HPC

With Amazon Web Services, researchers can access purpose-built HPC tools and services along with scientific and technical expertise to accelerate the pace of discovery. Whether you are sequencing the human genome, using AI/ML for disease detection or running molecular dynamics simulations to develop lifesaving drugs, AWS has the infrastructure you need to run your HPC workloads. Read more…

PEARC21 Panel: Wafer-Scale-Engine Technology Accelerates Machine Learning, HPC

July 21, 2021

Early use of Cerebras’ CS-1 server and wafer-scale engine (WSE) has demonstrated promising acceleration of machine-learning algorithms, according to participants in the Scientific Research Enabled by CS-1 Systems panel Read more…

With New Owner and New Roadmap, an Independent Omni-Path Is Staging a Comeback

July 23, 2021

Put on a shelf by Intel in 2019, Omni-Path faced a uncertain future, but under new custodian Cornelis Networks, OmniPath is looking to make a comeback as an independent high-performance interconnect solution. A "significant refresh" – called Omni-Path Express – is coming later this year according to the company. Cornelis Networks formed last September as a spinout of Intel's Omni-Path division. Read more…

Chameleon’s HPC Testbed Sharpens Its Edge, Presses ‘Replay’

July 22, 2021

“One way of saying what I do for a living is to say that I develop scientific instruments,” said Kate Keahey, a senior fellow at the University of Chicago a Read more…

Summer Reading: “High-Performance Computing Is at an Inflection Point”

July 21, 2021

At last month’s 11th International Symposium on Highly Efficient Accelerators and Reconfigurable Technologies (HEART), a group of researchers led by Martin Schulz of the Leibniz Supercomputing Center (Munich) presented a “position paper” in which they argue HPC architectural landscape... Read more…

PEARC21 Panel: Wafer-Scale-Engine Technology Accelerates Machine Learning, HPC

July 21, 2021

Early use of Cerebras’ CS-1 server and wafer-scale engine (WSE) has demonstrated promising acceleration of machine-learning algorithms, according to participa Read more…

15 Years Later, the Green500 Continues Its Push for Energy Efficiency as a First-Order Concern in HPC

July 15, 2021

The Green500 list, which ranks the most energy-efficient supercomputers in the world, has virtually always faced an uphill battle. As Wu Feng – custodian of the Green500 list and an associate professor at Virginia Tech – tells it, “noone" cared about energy efficiency in the early 2000s, when the seeds... Read more…

Frontier to Meet 20MW Exascale Power Target Set by DARPA in 2008

July 14, 2021

After more than a decade of planning, the United States’ first exascale computer, Frontier, is set to arrive at Oak Ridge National Laboratory (ORNL) later this year. Crossing this “1,000x” horizon required overcoming four major challenges: power demand, reliability, extreme parallelism and data movement. Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

ExaWind Prepares for New Architectures, Bigger Simulations

July 10, 2021

The ExaWind project describes itself in terms of terms like wake formation, turbine-turbine interaction and blade-boundary-layer dynamics, but the pitch to the Read more…

AMD Chipmaker TSMC to Use AMD Chips for Chipmaking

May 8, 2021

TSMC has tapped AMD to support its major manufacturing and R&D workloads. AMD will provide its Epyc Rome 7702P CPUs – with 64 cores operating at a base cl Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Berkeley Lab Debuts Perlmutter, World’s Fastest AI Supercomputer

May 27, 2021

A ribbon-cutting ceremony held virtually at Berkeley Lab's National Energy Research Scientific Computing Center (NERSC) today marked the official launch of Perlmutter – aka NERSC-9 – the GPU-accelerated supercomputer built by HPE in partnership with Nvidia and AMD. Read more…

Ahead of ‘Dojo,’ Tesla Reveals Its Massive Precursor Supercomputer

June 22, 2021

In spring 2019, Tesla made cryptic reference to a project called Dojo, a “super-powerful training computer” for video data processing. Then, in summer 2020, Tesla CEO Elon Musk tweeted: “Tesla is developing a [neural network] training computer called Dojo to process truly vast amounts of video data. It’s a beast! … A truly useful exaflop at de facto FP32.” Read more…

Google Launches TPU v4 AI Chips

May 20, 2021

Google CEO Sundar Pichai spoke for only one minute and 42 seconds about the company’s latest TPU v4 Tensor Processing Units during his keynote at the Google I Read more…

CentOS Replacement Rocky Linux Is Now in GA and Under Independent Control

June 21, 2021

The Rocky Enterprise Software Foundation (RESF) is announcing the general availability of Rocky Linux, release 8.4, designed as a drop-in replacement for the soon-to-be discontinued CentOS. The GA release is launching six-and-a-half months after Red Hat deprecated its support for the widely popular, free CentOS server operating system. The Rocky Linux development effort... Read more…

CERN Is Betting Big on Exascale

April 1, 2021

The European Organization for Nuclear Research (CERN) involves 23 countries, 15,000 researchers, billions of dollars a year, and the biggest machine in the worl Read more…

Iran Gains HPC Capabilities with Launch of ‘Simorgh’ Supercomputer

May 18, 2021

Iran is said to be developing domestic supercomputing technology to advance the processing of scientific, economic, political and military data, and to strengthen the nation’s position in the age of AI and big data. On Sunday, Iran unveiled the Simorgh supercomputer, which will deliver.... Read more…

Leading Solution Providers

Contributors

HPE Launches Storage Line Loaded with IBM’s Spectrum Scale File System

April 6, 2021

HPE today launched a new family of storage solutions bundled with IBM’s Spectrum Scale Erasure Code Edition parallel file system (description below) and featu Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

GTC21: Nvidia Launches cuQuantum; Dips a Toe in Quantum Computing

April 13, 2021

Yesterday Nvidia officially dipped a toe into quantum computing with the launch of cuQuantum SDK, a development platform for simulating quantum circuits on GPU-accelerated systems. As Nvidia CEO Jensen Huang emphasized in his keynote, Nvidia doesn’t plan to build... Read more…

Microsoft to Provide World’s Most Powerful Weather & Climate Supercomputer for UK’s Met Office

April 22, 2021

More than 14 months ago, the UK government announced plans to invest £1.2 billion ($1.56 billion) into weather and climate supercomputing, including procuremen Read more…

Q&A with Jim Keller, CTO of Tenstorrent, and an HPCwire Person to Watch in 2021

April 22, 2021

As part of our HPCwire Person to Watch series, we are happy to present our interview with Jim Keller, president and chief technology officer of Tenstorrent. One of the top chip architects of our time, Keller has had an impactful career. Read more…

Quantum Roundup: IBM, Rigetti, Phasecraft, Oxford QC, China, and More

July 13, 2021

IBM yesterday announced a proof for a quantum ML algorithm. A week ago, it unveiled a new topology for its quantum processors. Last Friday, the Technical Univer Read more…

Senate Debate on Bill to Remake NSF – the Endless Frontier Act – Begins

May 18, 2021

The U.S. Senate today opened floor debate on the Endless Frontier Act which seeks to remake and expand the National Science Foundation by creating a technology Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire