Rockport Networks Names New Co-CEO as It Closes on $48M Funding Round

By Tiffany Trader

December 14, 2021

Rockport Networks, a new entrant in the HPC networking space with its switchless fabric offering, today announced the appointment of Marc Sultzbaugh to co-CEO. An active Rockport board member since December 2020, Sultzbaugh will lead the company alongside Rockport Networks Co-Founder and Co-CEO Doug Carwardine.

Sultzbaugh previously spent 20 years at HPC networking company Mellanox as a sales and marketing executive. In his new role at Rockport, he will be responsible for all aspects of the company’s growth, including product management, marketing, sales, manufacturing and customer relationships. Carwardine will maintain the R&D and financial management of the company. The announcement comes a week after Rockport closed on a $48 million financing round, led by Northern Private Capital, bringing the company’s total funding to just under $100 million.

Marc Sultzbaugh

The latest funding will go toward growing sales and marketing as Rockport transitions from an engineering firm to commercializing their product: the 300 Gbps switchless networking architecture launched in October. Instead of relying on a traditional switch architecture, Rockport distributes network functions to each node, which are direct-connected to each other over passive cabling. Although the company’s technology is ‘switchless’ in the conventional sense of a hardware switch, the switching functionality is carried out at the node-level by Rockport NC1225 networking cards.

Sultzbaugh started out in engineering at Bell Labs some three decades ago but soon moved into business roles. He spent seven years at fabless chip company Brooktree, which went through a successful IPO. And he was with Mellanox from product launch all the way through the successful $6.9 billion handoff to Nvidia.

Sultzbaugh is confident he can lead a similar successful path at Rockport Networks, where he says the company is on a mission to help end users get the most utilization out of their networking. And that boils down to the way Rockport’s switchless architecture handles congestion, according to the company.

“Our architecture is absolutely about performance,” said Sultzbaugh, “But it just as importantly, if not more so, is about predictable performance. It doesn’t matter what kind of workloads are in the fabric simultaneously, it doesn’t matter how loaded your fabric is, you’re going to get very consistent runtimes.”

Expanding on that point, he said, “Datacenter operators go through all kinds of gyrations to try to either limit the scale at which they run their applications or limit the kinds of traffic that are mixed in the network in order to overcome [congestion], or simply, they don’t know why their applications are inconsistent or the performance isn’t what they expect. And so the answer from traditional switch vendors is, let’s just move you up the feeds and speeds. We’ll move you from 10 to 25, to 100, and now to 400 gigabits per second. But it really hasn’t addressed the core issue on the architecture, and all they really accomplished is that they create congestion much faster. And worse, as you move up the theoretical performance of those switches, your cost is going up linearly. It’s not like other technologies, where the costs tend to stay very neutral, if not go down.”

The company says it has a number of engagements across multiple vertical markets in HPC and AI/ML, including within financial services, life sciences, energy, as well as the traditional supercomputing centers.

Early customers include the Texas Advanced Computing Center (TACC), which has installed the networking technology across 396 nodes of its Frontera system as an exploratory testbed, as well as DiRAC/Durham University, which is also trying out the networking gear. The high-performance networking group at Ohio State is also engaged with Rockport, lending its expertise with standards support.

Traditional HPC sites are Rockport’s initial focus, but Sultzbaugh says there is very much an enterprise play. “There’s more and more work being done in enterprises around what we call performance compute applications, which includes traditional HPC, design and simulation and modeling, and also new workloads around AI, machine learning, etc.,” he tells HPCwire. “We will be 100 percent channel driven, but as you can imagine, with any kind of disruptive technology, we need to have high touch with those early adopters. And so we’ll be scaling out the salesforce, as well as scaling out our marketing efforts.”

Education is a key part of many of the conversations Rockport is having.

“Anytime you have a transition or a disruption to the market, you have to educate customers, and part of our education for the industry is that we need to stop looking at the theoretical performance speeds and speeds that come from the traditional network vendors, and really look at and measure what’s the impact in real-world applications when the network is loaded, when there’s multiple different types of workloads being being run on that network, and that’s really what matters to customers.”

To assist with its performance validations and demonstrations, Rockport has lined up two Centers of Excellence (one at TACC and one at Durham), and plans to have several channel partners that will also validate the solution in their own testbeds. Dell, the system-maker behind the TACC Frontera system, is a key partner, and Sultzbaugh mentioned an early engagement with Lenovo. Rockport is also working on building additional ‘lighthouse customers’ across multiple vertical markets.

Asked about competitive head-to-head latency comparisons, Sultzbaugh indicated that on the surface, the metrics can be misleading. Specs on nanoseconds of latency are usually cited with no data transfer, in which case it’s just theoretical capability, he said. “What’s important is how you utilize that bandwidth and how you permit predictable performance across the fabric.”

“On our roadmap today,” he continued, “we’re supplying 300 gigabits per second per node. You have much higher bandwidth switches today, but they get congested because they’re built in terms of the spine leaf architecture. And that’s fundamentally the issue. It’s that you have to go through multiple switch hops, and you aggregate those switch hops. And so even though they’re capable of having very low latency, because you have this tail of latency that exists because packets get stuck in the network, the whole infrastructure is waiting. And so you move from capability of nanoseconds of latency to milliseconds of latency at an application level. We need to find a way to really measure in a benchmark that [indicates] what you’re really going to get in terms of performance. And that’s our challenge.”

“I think a lot of the customer base today already understands it and knows it,” he added. “And then we have to educate the rest of them.”

Rockport faces a competitive landscape dominated by InfiniBand – supplied by Nvidia’s Mellanox division – and Ethernet, which is how HPE’s Slingshot networking can technically be categorized, as well as the proprietary OmniPath fabric, now under the control of startup Cornelis Networks. 

Mark Nossokoff, senior analyst with Hyperion Research, sees potential in the new switchless fabric offering. “I think there is definitely room in the networking space for some innovation in different approaches,” he told HPCwire. “InfiniBand and Ethernet are marching along on their evolutionary cycles, then you have HPE come in with Slingshot, which is Ethernet-like. Now comes the switchless approach, which is unique and innovative, and I think it may very well have some merit [by] addressing the ability to have shortened predictable latencies and minimizing the congestion that occurs on heavily utilized networks.”

Congestion issues are most pervasive on systems with lots of different types of heterogeneous jobs, said Nossokoff. “I believe the Rockport solution is aimed at large systems that have large numbers of users, and large numbers of heterogeneous types of jobs that are run on them. And by heterogenous, that’s about the size of the jobs, the length of how long the jobs take to get done, and, and the utilization of the system.The greater the utilization of the system, the more you’re going to see the benefit of the Rockport solution,” he said.

Shipping now, Rockport’s networking technology comprises three elements (described below):

  • The Rockport NC1225: This half height, half length PCI Express card is the only active component of the Rockport solution. It replaces standard NICs and plugs directly into servers or storage enclosures. The card has three elements: a standard Ethernet host interface; an FPGA, which enables upgradeable firmware; and embedded optics. The NC1225 implements the host interface as well as the distributed switching by running Rockport’s rNOS software in the embedded hardware. The NC1225 replaces standard network interface cards (NIC) and “adaptively aggregates the bandwidth of multiple parallel network paths, drawing from 300 Gbps of available network capacity,” according to the company.
  • The Rockport SHFL: The SHFL is a new passive cabling invention consisting of a box that is pre-wired to form a topology and distribute links correctly, either within the rack or across the racks. The cables themselves are commodity cables, specifically MTP/MPO-24 to the SHFL and MTP/MPO-32 between SHFLs. This is standard passive fiber optic cabling and all the optics are on the endpoint cards. To be clear, SHFLs have no have no power and no electronics. The model shown (below) has 24 ports, but Rockport offers different form factors for different network topologies at different scales. The maximum distance for a link is about 100 meters, but due to the mesh connectivity, the network will still operate even if some links exceed that distance.
  • The Rockport Autonomous Network Manager (ANM): The ANM provides high-level and granular insights into the active network. It continuously monitors all aspects of the network and stores historical data for 60 days with the most recent seven days collected and presented at high fidelity.

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire