Cray XT5m Midrange Supercomputer Builds Market Momentum

By Nicole Hemsoth

September 28, 2009

In March 2009, Cray announced the Cray XT5m system, a compatible midrange extension of the high-end Cray XT5 product line that was first to break the sustained petaflop performance barrier on real-world applications. HPCwire asked Barry Bolding, Cray’s vice president of scalable systems, for an update on the midrange product.

Cray XT5HPCwire: How is the Cray XT5m midrange supercomputer doing in the market place?

Bolding: We disclosed one customer name with the March announcement, the High Performance Computing Center Stuttgart, where it’s being used for automotive, academic and public sector work. We have also publically announced a second customer, the Finnish Meteorological Institute (FMI) in the production weather forecasting segment. The Cray XT5m has been meeting our sales goals since then. Granted, our expectations for the first six months were fairly modest because the midrange is a new space for us and we viewed this as a learning period. The good news is that we have validated that the XT5m can be a leader in the mid-range. We have won customers in several of our key target segments and our pricing is very competitive in this price sensitive space so we expect higher volumes going forward. Today we have a large number of bids outstanding in academia, automotive, aerospace, pharma and weather forecasting, as well as in government and R&D labs.

HPCwire: Why did Cray decide to develop this product?

Bolding: There were several reasons. We were consistently running into procurements that were in the midrange and needed a more price-competitive system for that space, which we view as a growth area for us. According to IDC, Cray is the current global market share leader for HPC systems priced at $3 million and up, but we have a very small presence in the midrange segment. Many of the characteristics of our largest machines are applicable in the midrange, but to succeed in this space we needed to develop a price-competitive product, turn around procurements faster, and provide a wider range of ISV applications. ISV applications are also important at the high end, but industry typically buys midrange systems and relies more heavily on third-party apps than government and academia do. The Cray XT5m is the initial product on our midrange roadmap.

HPCwire: Why would someone buy a Cray XT5m system instead of a same-size cluster? What are the key differences?

Bolding: Primarily they get Cray’s more capable SeaStar interconnect and interconnect roadmap. The Cray SeaStar has proven itself over the past several years as the industry-leading interconnect for MPI scalability.  With the Cray SeaStar interconnect, the Cray XT5m handles complex messaging traffic very efficiently.  You also get the entire Cray software stack that has been scaled and validated up to the petascale performance level, plus the network roadmap driving toward global addressability and high-performance UPC and Co-Array Fortran, along with MPI of course. So, Cray XT5m customers are buying into a broader portfolio than just a typical midrange system.  They’re buying into the demonstrated petascale scalability of the XT5 architecture. To achieve true scaling today, even at midrange size, one needs a full portfolio of network, software and infrastructure support, which Cray provides in the XT5m. And with multicore processors becoming prevalent, users will soon be facing the need to scale substantially higher even with midrange systems, especially starting in 2010.  Our midrange systems are designed to benefit from Cray’s high-end system development and this will continue. For example, the Cray XT5m line provides the same TCO benefits to our midrange customers as our largest supercomputer customers enjoy with the ECOphlex cooling technology we developed for petascale systems.

HPCwire: Assuming the Cray XT5m has a more capable network and memory subsystem than a standard cluster, with better bandwidth and latency characteristics, wouldn’t it be able to tackle a broader range of applications efficiently?

Bolding: The Cray XT5m today is a midrange industry leader from a network bandwidth and latency perspective. This gives users an important advantage over commodity InfiniBand networks and allows the Cray XT5m to handle a broader range of applications efficiently. The system is aimed at codes scaling to 256 cores and beyond, and in this range users typically see significant benefits from the overall system and software design.

HPCwire: Are customers using their Cray XT5m systems as their main HPC systems or for specific portions of their workloads?

Bolding: It varies. In production weather forecasting, we have a customer using it as their primary production system. At academic sites it varies from being primary system to one among multiple HPC systems. Some users are experimenting with the new functionalities, including the network features and the Cray software and compilers. It’s also allowing users to experiment with scalability in ways they haven’t been able to do before. So, the Cray XT5m is being used both as a development platform and a production platform.

HPCwire: Is there any customer who operates both a Cray XT5 and a Cray XT5m system?

Bolding: There are customers running applications across both Cray XT5 and Cray  XT5m systems, although they don’t have both systems in-house today.  By the end of this year, we’ll have sites that have both in their data centers.

HPCwire: The XT5m is a compatible downward extension of the Cray XT5 architecture, but uses a 2D torus interconnect instead of the XT5’s 3D torus. What’s the strategy behind this change?

Bolding: It’s a cost-saving strategy for the midrange scale. With a Cray XT5m system consisting of 1-6 cabinets, customers tend not to have applications that require the full 3D topology as much as with a larger, high-end Cray XT5. We right-sized the Cray XT5m for midrange requirements, including the interconnect, allowing us to provide a price-competitive product in this space. We’ve done extensive studies on application performance on the Cray XT5m, and there has been minimal performance impact at six cabinets and below. Above that size, you need the Cray XT5’s 3D torus to maintain scalability. For most apps in the 1-6 cabinet range, performance degradation due to the topology is less than 5 percent even for applications running across several thousand cores of a Cray XT5m.

HPCwire: If a site maxes out on their XT5m, what’s the upgrade path to a Cray XT5 system?

Bolding: It’s very simple. Today, it just involves replacing the network mezzanine card and adding more cables to transform the 2D torus into a 3D torus. We’ll make the upgrade path even simpler in the future of our midrange systems.

HPCwire: Do you expect some users to take advantage of non-MPI programming models that are available on the XT5m, such as SHMEM, UPC and Co-Array Fortran?

Bolding: We do. We port those in the software today and will be making announcements of enhancements to the hardware support for some of these features in the next 12 months. We are committed to making more innovations and remaining a leader in HPC, and this requires providing our customers with multiple, high performance programming models.

HPCwire: What is the Data Virtualization Service that comes with the XT5 and XT5m, and why is it important?

Bolding: DVS is an important part of Cray systems. DVS is a flexible virtualization layer that Cray plans use to expand our software functionality and performance . One feature of DVS is that it can allow Cray to project various file systems onto the compute nodes (which are diskless on Cray XT5 systems). This allows Cray systems to act more like a standard commodity cluster if it needs to.  We support IO and storage functionalities that we haven’t in the past. We can share file systems with high-performance file systems on platforms other than clusters. Customers such as NERSC and Oak Ridge are doing very innovative things with file systems and DVS can play a role in providing the flexibility they need.  So, DVS helps us both with compute and IO/storage.

HPCwire: Where is the Cray XT5m product line headed in the future?

Bolding: We are going to continue driving into the midrange market, which is a segment that has excellent growth opportunities for Cray. We want to build a substantial and sustainable market presence there, especially with customers focused on scalability of applications. To grow the XT5m line we need to continue to be competitive on cost while improving price/performance, network scalability,  software features and ISV availability, and we’ll be doing all those things. We’ll also be improving the processor roadmap with new technologies coming into market place. Cray’s innovation, combined with AMD’s strong roadmap is a winning combination for the next few years for the entire XT family of systems. 
HPCwire: What should we look for next?

Bolding: At SC09, we plan to talk more specifically about our future plans for the Cray XT5m product line.

For more information on how the Cray XT5m is making petaflops performance affordable, download the AMD white paper here.

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Advancing Modular Supercomputing with DEEP and DEEP-ER Architectures

February 24, 2017

Knowing that the jump to exascale will require novel architectural approaches capable of delivering dramatic efficiency and performance gains, researchers around the world are hard at work on next-generation HPC systems. Read more…

By Sean Thielen

Weekly Twitter Roundup (Feb. 23, 2017)

February 23, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

HPE Server Shows Low Latency on STAC-N1 Test

February 22, 2017

The performance of trade and match servers can be a critical differentiator for financial trading houses. Read more…

By John Russell

HPC Financial Update (Feb. 2017)

February 22, 2017

In this recurring feature, we’ll provide you with financial highlights from companies in the HPC industry. Check back in regularly for an updated list with the most pertinent fiscal information. Read more…

By Thomas Ayres

HPE Extreme Performance Solutions

Manufacturers Reaping the Benefits of Remote Visualization

Today’s manufacturers are operating in an ever-changing atmosphere, and finding new ways to boost productivity has never been more vital.

This is why manufacturers are ramping up their investments in high performance computing (HPC), a trend which has helped give rise to the “connected factory” and Industrial Internet of Things (IIoT) concepts that are proliferating throughout the industry today. Read more…

Rethinking HPC Platforms for ‘Second Gen’ Applications

February 22, 2017

Just what constitutes HPC and how best to support it is a keen topic currently. Read more…

By John Russell

HPC Technique Propels Deep Learning at Scale

February 21, 2017

Researchers from Baidu’s Silicon Valley AI Lab (SVAIL) have adapted a well-known HPC communication technique to boost the speed and scale of their neural network training and now they are sharing their implementation with the larger deep learning community. Read more…

By Tiffany Trader

IDC: Will the Real Exascale Race Please Stand Up?

February 21, 2017

So the exascale race is on. And lots of organizations are in the pack. Government announcements from the US, China, India, Japan, and the EU indicate that they are working hard to make it happen – some sooner, some later. Read more…

By Bob Sorensen, IDC

ExxonMobil, NCSA, Cray Scale Reservoir Simulation to 700,000+ Processors

February 17, 2017

In a scaling breakthrough for oil and gas discovery, ExxonMobil geoscientists report they have harnessed the power of 717,000 processors – the equivalent of 22,000 32-processor computers – to run complex oil and gas reservoir simulation models. Read more…

By Doug Black

Advancing Modular Supercomputing with DEEP and DEEP-ER Architectures

February 24, 2017

Knowing that the jump to exascale will require novel architectural approaches capable of delivering dramatic efficiency and performance gains, researchers around the world are hard at work on next-generation HPC systems. Read more…

By Sean Thielen

HPC Technique Propels Deep Learning at Scale

February 21, 2017

Researchers from Baidu’s Silicon Valley AI Lab (SVAIL) have adapted a well-known HPC communication technique to boost the speed and scale of their neural network training and now they are sharing their implementation with the larger deep learning community. Read more…

By Tiffany Trader

IDC: Will the Real Exascale Race Please Stand Up?

February 21, 2017

So the exascale race is on. And lots of organizations are in the pack. Government announcements from the US, China, India, Japan, and the EU indicate that they are working hard to make it happen – some sooner, some later. Read more…

By Bob Sorensen, IDC

TSUBAME3.0 Points to Future HPE Pascal-NVLink-OPA Server

February 17, 2017

Since our initial coverage of the TSUBAME3.0 supercomputer yesterday, more details have come to light on this innovative project. Of particular interest is a new board design for NVLink-equipped Pascal P100 GPUs that will create another entrant to the space currently occupied by Nvidia's DGX-1 system, IBM's "Minsky" platform and the Supermicro SuperServer (1028GQ-TXR). Read more…

By Tiffany Trader

Tokyo Tech’s TSUBAME3.0 Will Be First HPE-SGI Super

February 16, 2017

In a press event Friday afternoon local time in Japan, Tokyo Institute of Technology (Tokyo Tech) announced its plans for the TSUBAME3.0 supercomputer, which will be Japan’s “fastest AI supercomputer,” Read more…

By Tiffany Trader

Drug Developers Use Google Cloud HPC in the Fight Against ALS

February 16, 2017

Within the haystack of a lethal disease such as ALS (amyotrophic lateral sclerosis / Lou Gehrig’s Disease) there exists, somewhere, the needle that will pierce this therapy-resistant affliction. Read more…

By Doug Black

Azure Edges AWS in Linpack Benchmark Study

February 15, 2017

The “when will clouds be ready for HPC” question has ebbed and flowed for years. Read more…

By John Russell

Is Liquid Cooling Ready to Go Mainstream?

February 13, 2017

Lost in the frenzy of SC16 was a substantial rise in the number of vendors showing server oriented liquid cooling technologies. Three decades ago liquid cooling was pretty much the exclusive realm of the Cray-2 and IBM mainframe class products. That’s changing. We are now seeing an emergence of x86 class server products with exotic plumbing technology ranging from Direct-to-Chip to servers and storage completely immersed in a dielectric fluid. Read more…

By Steve Campbell

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

US, China Vie for Supercomputing Supremacy

November 14, 2016

The 48th edition of the TOP500 list is fresh off the presses and while there is no new number one system, as previously teased by China, there are a number of notable entrants from the US and around the world and significant trends to report on. Read more…

By Tiffany Trader

Lighting up Aurora: Behind the Scenes at the Creation of the DOE’s Upcoming 200 Petaflops Supercomputer

December 1, 2016

In April 2015, U.S. Department of Energy Undersecretary Franklin Orr announced that Intel would be the prime contractor for Aurora: Read more…

By Jan Rowell

D-Wave SC16 Update: What’s Bo Ewald Saying These Days

November 18, 2016

Tucked in a back section of the SC16 exhibit hall, quantum computing pioneer D-Wave has been talking up its new 2000-qubit processor announced in September. Forget for a moment the criticism sometimes aimed at D-Wave. This small Canadian company has sold several machines including, for example, ones to Lockheed and NASA, and has worked with Google on mapping machine learning problems to quantum computing. In July Los Alamos National Laboratory took possession of a 1000-quibit D-Wave 2X system that LANL ordered a year ago around the time of SC15. Read more…

By John Russell

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. Read more…

By John Russell

IBM Wants to be “Red Hat” of Deep Learning

January 26, 2017

IBM today announced the addition of TensorFlow and Chainer deep learning frameworks to its PowerAI suite of deep learning tools, which already includes popular offerings such as Caffe, Theano, and Torch. Read more…

By John Russell

Tokyo Tech’s TSUBAME3.0 Will Be First HPE-SGI Super

February 16, 2017

In a press event Friday afternoon local time in Japan, Tokyo Institute of Technology (Tokyo Tech) announced its plans for the TSUBAME3.0 supercomputer, which will be Japan’s “fastest AI supercomputer,” Read more…

By Tiffany Trader

HPC Startup Advances Auto-Parallelization’s Promise

January 23, 2017

The shift from single core to multicore hardware has made finding parallelism in codes more important than ever, but that hasn’t made the task of parallel programming any easier. Read more…

By Tiffany Trader

Leading Solution Providers

CPU Benchmarking: Haswell Versus POWER8

June 2, 2015

With OpenPOWER activity ramping up and IBM’s prominent role in the upcoming DOE machines Summit and Sierra, it’s a good time to look at how the IBM POWER CPU stacks up against the x86 Xeon Haswell CPU from Intel. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

Nvidia Sees Bright Future for AI Supercomputing

November 23, 2016

Graphics chipmaker Nvidia made a strong showing at SC16 in Salt Lake City last week. Read more…

By Tiffany Trader

TSUBAME3.0 Points to Future HPE Pascal-NVLink-OPA Server

February 17, 2017

Since our initial coverage of the TSUBAME3.0 supercomputer yesterday, more details have come to light on this innovative project. Of particular interest is a new board design for NVLink-equipped Pascal P100 GPUs that will create another entrant to the space currently occupied by Nvidia's DGX-1 system, IBM's "Minsky" platform and the Supermicro SuperServer (1028GQ-TXR). Read more…

By Tiffany Trader

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment firm International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

Dell Knights Landing Machine Sets New STAC Records

November 2, 2016

The Securities Technology Analysis Center, commonly known as STAC, has released a new report characterizing the performance of the Knight Landing-based Dell PowerEdge C6320p server on the STAC-A2 benchmarking suite, widely used by the financial services industry to test and evaluate computing platforms. The Dell machine has set new records for both the baseline Greeks benchmark and the large Greeks benchmark. Read more…

By Tiffany Trader

Is Liquid Cooling Ready to Go Mainstream?

February 13, 2017

Lost in the frenzy of SC16 was a substantial rise in the number of vendors showing server oriented liquid cooling technologies. Three decades ago liquid cooling was pretty much the exclusive realm of the Cray-2 and IBM mainframe class products. That’s changing. We are now seeing an emergence of x86 class server products with exotic plumbing technology ranging from Direct-to-Chip to servers and storage completely immersed in a dielectric fluid. Read more…

By Steve Campbell

What Knights Landing Is Not

June 18, 2016

As we get ready to launch the newest member of the Intel Xeon Phi family, code named Knights Landing, it is natural that there be some questions and potentially some confusion. Read more…

By James Reinders, Intel

  • arrow
  • Click Here for More Headlines
  • arrow
Share This