Fabric7 Weaves an Interesting Tale

By Michael Feldman

August 11, 2006

With Sun Microsystems' recent introduction of its new Opteron-based system, including the eight-socket, 16-core Sun Fire X4600, the industry witnessed a Tier 1 OEM significantly expand its commitment to the x64 (64-bit x86) server market. Sun wasn't the first vendor to venture into the eight-socket Opteron space. Silicon Valley start-up, Fabric7 Systems Inc., introduced its x64 eight-socket, 128 GB machine back in November of 2005. Fabric7 actually has two offerings — the Q80 server, their basic eight-socket model, and the Q160 server, the company's high-end machine, which scales to 14 Opteron processors and incorporates a built-in low-latency, high-bandwidth I/O fabric.

When the Q80 and Q160 were unveiled last year, Fabric7 had only enterprise server customers on their minds. The company was — and still is — primarily focused on businesses providing financial equity trading and financial services, media entertainment, content distribution, telecommunications and web services. Using the commodity components — the AMD Opteron processor and the Linux and Windows operating systems, Fabric7 is aiming for the sweet spot in the mid-range enterprise server market.

But the company's offerings are not just vanilla x64 servers. With its Q-Par hardware partitioning capability, Fabric7 systems can dynamically carve the eight-socket server into two four-socket servers or four two-socket servers. Until the company made hardware partitioning available on their x64 systems, this capability was restricted to high-end mainframes, typically RISC machines running some flavor of Unix. Hardware partitioning is an efficient mechanism for sharing computational resources within a single box and avoids the performance penalty associated with software virtualization. However, software virtualization can run on top of the hardware partitions, adding another layer of resource sharing.

Resource virtualization allows datacenters to consolidate hardware, while increasing flexibility. The growing popularity of this model is encouraging the IT industry to focus on the type of architecture that can best accommodate it. Customers are starting to wonder if they can get a different dimension of computational performance and I/O throughput with a smaller collection of fat nodes rather than with a larger collection of skinny nodes.

“Traditionally, or for the past several years, it's been a build-out of lots of little 1U pizza boxes,” explains Bryan Sweeley, VP of marketing at Fabric7. “Customers have been very successful taking their applications and breaking them down to run across hundreds or thousands of servers. Now we are starting to see the pendulum swinging back again — people are looking at larger SMP and larger memory footprints.”

Larger SMP systems can make sense for HPC workloads as well. The recent deployment of Tokyo Tech's 38-teraflops TSUBAME supercomputer, which consists of 655 Sun Fire X4600 machines, is an example of building a large high performance system from a relatively small number of fat nodes. The fewer the nodes, the less you have to rely on expensive I/O switches and comparatively slow system interconnects to shuttle data around. A smaller number of boxes also means fewer individual pieces of hardware and software to manage and maintain.

The Tokyo Tech machine has not escaped the notice of other supercomputer users who are exploring different ways of scaling out systems. Sweeley revealed that Fabric7 has been contacted by a number of unnamed HPC users wanting to know more about what the company can offer.
 
“We've had some calls from the HPC crowd and have a couple of prospects cooking,” says Sweeley. My sense is that there is a trend developing in HPC to move towards the larger SMP footprints. And so the conversations we're having with some of the supercomputing installations tend to be focused around that.”

A single eight-socket machine may be powerful enough to run some types of industrial HPC applications such as electronic design automation (EDA). One of Fabric7's customers, Magma Design Automation, needs the large memory space — in their case 64 GB — to support their EDA applications. The larger memory footprint is used to improve performance for their full chip design runs. Less demanding test runs may require a hardware partition of only two processors and this can be reconfigured on the fly in 10 minutes.

Though both the Sun and Fabric7 systems share the same eight-socket Opteron architecture, the resemblance basically ends there. Besides the hardware partitioning capability, the other big feature of Fabric7 machines is their high-performance network I/O, that is, the fabric. It uses four HyperTransport links for I/O, as compared to two for the Sun Fire X4600.

On the Q160, Fabric7's high-end machine, the native interconnect supports both Ethernet and Fibre Channel. According to the company, the fabric supports 128 Gbps of non-blocking, switched I/O that provides up to 40 Gbps bandwidth to the processor and memory complex. The switched I/O can be extended across the fabric to provide 30 Gbps of Ethernet (or 16 Gbps Fibre Channel) between chassis or across sites. A software switch can dynamically reconfigure the Fibre Channel and Ethernet bandwidth, creating a flexible interconnect. For many enterprise applications, this is probably more networking capability than they can use. But to the extent that I/O is important for HPC applications, Fabric7's implementation of a high performance, flexible network makes for an interesting differentiator.

“Our fabric is a blend of 10 Gigabit Ethernet with a low-latency InfiniBand-like technology, says Sweeley. “On top of that, we can run Fibre Channel over the same fabric. For the HPC folks, we have more networking capability than they've seen from anyone else.”

Both Sun and Fabric7 are following the AMD Opteron/HyperTransport technology roadmap pretty aggressively. These companies are pushing the envelope with their 16-core offering, although this represents only a fraction of what will be possible in the next year or two . When the Opterons go quad-core in 2007, OEMs will be able to build 32-core SMP servers. When another HyperTransport link is added, the number of supported sockets will go from 8 to 32, allowing for the possibility of a 128-core machine. With Sun scaling up its Opteron systems, Fabric7 will face some tough competition from a Tier 1 OEM. But the team at Fabric7 seems comfortable with the prospect.

“Right now Sun and Fabric7 are the only two vendors with an eight-socket Opteron [system],” says Sweeley. “That's a healthy competitive environment.”

As far as Fabric7 pursuing the high performance computing market more aggressively, that probably depends on how the larger SMP machines are perceived by HPC users over the next six to eighteen months. The company certainly seems open to the possibilities. When asked about what Fabric7 offers high performance computing, Sharad Mehrotra, the company's president, CEO and founder, had this to say:

“Fabric7 does think that the pendulum will swing back towards larger SMP servers. The Tokyo Tech example is the lighthouse that is making heads turn and prompting people to question the conventional wisdom of today. Fabric7's implementation of hardware partitioning across its entire product line, provides HPC users with the flexibility to move quickly from small SMP to large SMP configurations in a matter of minutes. Additionally, the switched, virtualized I/O capability available in our larger system, the Q160, provides customers the flexibility in network infrastructure that is required to keep up with the variable compute needs of the processing farm. We believe that the HPC world will shift from 'grid' to 'fabric' computing in the future as the benefits of our approach become more apparent with real-world deployments.”

—–

We will be providing an extended interview with Fabric7 CEO Sharad Mehrotra during our special coverage of the LinuxWorld Conference and Expo, which takes place August 14 – 17, in San Francisco, California.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Researchers Scale COSMO Climate Code to 4888 GPUs on Piz Daint

October 17, 2017

Effective global climate simulation, sorely needed to anticipate and cope with global warming, has long been computationally challenging. Two of the major obstacles are the needed resolution and prolonged time to compute Read more…

By John Russell

Student Cluster Competition Coverage New Home

October 16, 2017

Hello computer sports fans! This is the first of many (many!) articles covering the world-wide phenomenon of Student Cluster Competitions. Finally, the Student Cluster Competition coverage has come to its natural home: H Read more…

By Dan Olds

UCSD Web-based Tool Tracking CA Wildfires Generates 1.5M Views

October 16, 2017

Tracking the wildfires raging in northern CA is an unpleasant but necessary part of guiding efforts to fight the fires and safely evacuate affected residents. One such tool – Firemap – is a web-based tool developed b Read more…

By John Russell

HPE Extreme Performance Solutions

Transforming Genomic Analytics with HPC-Accelerated Insights

Advancements in the field of genomics are revolutionizing our understanding of human biology, rapidly accelerating the discovery and treatment of genetic diseases, and dramatically improving human health. Read more…

Exascale Imperative: New Movie from HPE Makes a Compelling Case

October 13, 2017

Why is pursuing exascale computing so important? In a new video – Hewlett Packard Enterprise: Eighteen Zeros – four HPE executives, a prominent national lab HPC researcher, and HPCwire managing editor Tiffany Trader Read more…

By John Russell

Student Cluster Competition Coverage New Home

October 16, 2017

Hello computer sports fans! This is the first of many (many!) articles covering the world-wide phenomenon of Student Cluster Competitions. Finally, the Student Read more…

By Dan Olds

Intel Delivers 17-Qubit Quantum Chip to European Research Partner

October 10, 2017

On Tuesday, Intel delivered a 17-qubit superconducting test chip to research partner QuTech, the quantum research institute of Delft University of Technology (TU Delft) in the Netherlands. The announcement marks a major milestone in the 10-year, $50-million collaborative relationship with TU Delft and TNO, the Dutch Organization for Applied Research, to accelerate advancements in quantum computing. Read more…

By Tiffany Trader

Fujitsu Tapped to Build 37-Petaflops ABCI System for AIST

October 10, 2017

Fujitsu announced today it will build the long-planned AI Bridging Cloud Infrastructure (ABCI) which is set to become the fastest supercomputer system in Japan Read more…

By John Russell

HPC Chips – A Veritable Smorgasbord?

October 10, 2017

For the first time since AMD's ill-fated launch of Bulldozer the answer to the question, 'Which CPU will be in my next HPC system?' doesn't have to be 'Whichever variety of Intel Xeon E5 they are selling when we procure'. Read more…

By Dairsie Latimer

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Intel Debuts Programmable Acceleration Card

October 5, 2017

With a view toward supporting complex, data-intensive applications, such as AI inference, video streaming analytics, database acceleration and genomics, Intel i Read more…

By Doug Black

OLCF’s 200 Petaflops Summit Machine Still Slated for 2018 Start-up

October 3, 2017

The Department of Energy’s planned 200 petaflops Summit computer, which is currently being installed at Oak Ridge Leadership Computing Facility, is on track t Read more…

By John Russell

US Exascale Program – Some Additional Clarity

September 28, 2017

The last time we left the Department of Energy’s exascale computing program in July, things were looking very positive. Both the U.S. House and Senate had pas Read more…

By Alex R. Larzelere

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

US Coalesces Plans for First Exascale Supercomputer: Aurora in 2021

September 27, 2017

At the Advanced Scientific Computing Advisory Committee (ASCAC) meeting, in Arlington, Va., yesterday (Sept. 26), it was revealed that the "Aurora" supercompute Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

GlobalFoundries Puts Wind in AMD’s Sails with 12nm FinFET

September 24, 2017

From its annual tech conference last week (Sept. 20), where GlobalFoundries welcomed more than 600 semiconductor professionals (reaching the Santa Clara venue Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Leading Solution Providers

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Delays, Smoke, Records & Markets – A Candid Conversation with Cray CEO Peter Ungaro

October 5, 2017

Earlier this month, Tom Tabor, publisher of HPCwire and I had a very personal conversation with Cray CEO Peter Ungaro. Cray has been on something of a Cinderell Read more…

By Tiffany Trader & Tom Tabor

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

Intel Launches Software Tools to Ease FPGA Programming

September 5, 2017

Field Programmable Gate Arrays (FPGAs) have a reputation for being difficult to program, requiring expertise in specialty languages, like Verilog or VHDL. Easin Read more…

By Tiffany Trader

IBM Advances Web-based Quantum Programming

September 5, 2017

IBM Research is pairing its Jupyter-based Data Science Experience notebook environment with its cloud-based quantum computer, IBM Q, in hopes of encouraging a new class of entrepreneurial user to solve intractable problems that even exceed the capabilities of the best AI systems. Read more…

By Alex Woodie

Intel, NERSC and University Partners Launch New Big Data Center

August 17, 2017

A collaboration between the Department of Energy’s National Energy Research Scientific Computing Center (NERSC), Intel and five Intel Parallel Computing Cente Read more…

By Linda Barney

  • arrow
  • Click Here for More Headlines
  • arrow
Share This