Visit additional Tabor Communication Publications
December 02, 2010
The folks at CERN are upgrading their high performance network infrastructure and for the first time are bringing Brocade Communications into the mix in a big way. CERN's IT group is set to deploy Brocade's new MLXe Core Routers in order to support the voluminous and growing storage and compute capabilities needed to keep the organization's datacenters humming.
In case you've been on another planet for the past few years, CERN, aka the European Organization for Nuclear Research, has been busy trying to uncover the mysteries of particle physics in order to discover some of the fundamental laws of the universe. Using its Large Hadron Collider, the world's largest high-energy particle accelerator, physicists have been conducting experiments that are generating scads of data -- on the order of 15 petabytes per year -- which CERN is streaming to its own Tier 0 datacenters in Switzerland and France, and then on to Tier 1 partner centers around the world.
Although CERN drives the global LHC computing grid, it alone is responsible for the five Tier 0 datacenters, which get first crack at the raw data, and it is here where the Brocade routers will be deployed. Four of these are smaller facilities dedicated to performing the initial number crunching on the collider data. After this first stage of processing, the results are sent to the main CERN datacenter for storage and reprocessing, and subsequent distribution to Tier 1 partner sites. In the main center itself, there has 8,000 servers that encapsulate 40,000 cores, as well as 40 petabytes of disk and 45 petabytes of tape.
As you might imagine, processing and shuffling all this data around requires some serious networking prowess. At CERN's Tier 0 centers, the current non-blocking capacity of the network core is around 4.2 terabits per second (Tbps), and is projected to hit more than 10 Tbps in 2011. "We are basically doubling the capacity of the system roughly every two years, " explained Jean-Michel Jouanigot, communication systems group leader for CERN's IT Department.
To accommodate this kind of growth curve, CERN realized they would need core routers that would be 100GbE-capable. Jouanigot told HPCwire that the selection of a vendor that could deliver this kind of capability began last year. The process was fairly straightforward: CERN had an extensive list of requirements, the primary one being the ability to get to 10 Tbps with a year. After in-house testing of manufacturers' gear that could meet these demands, it came down to price. As Jouanigot put it, the Brocade proposal was "the least expensive technically acceptable solution." The company's MLXe gear will eventually displace the Force 10 routers that currently predominate at CERN.
Although the goal is 100GbE networking, the Brocade routers delivered to date sport 10GbE ports. The MLXe routers support the higher data rates, but Jouanigot is still waiting on the company to get the pricing info on the 100GbE version, which are not yet available in quantity.
He says the problem they currently have with their 10GbE network core is that each router is connected to every other router through 16 10GbE ports. This entails a lot of cabling and additional management, so going to 100GbE to simplify the networking setup just makes sense. But according to Jouanigot, it really depends on how Brocade decides to price the 100GbE ports; if it turns out to be more than 10 time the 10GbE price, that becomes a tougher decision, extra cabling and management notwithstanding.
Another challenge is that when they renew their compute servers, they need to decide if they want to move to 10GbE system interconnects. Today, CERN servers are predominately GbE connected, with just a few hundred 10GbE-attached. It is very likely that newer servers will be delivered with 10GbE NICs, Since these systems are hooked up to one another, this will further draw on the bandwidth of the core network.
The routers also drive the external links between the various datacenters, so going to 100GbE will allow them to accomplish faster inter-center networking without having to lay new optical fiber. That tends to be a very costly operation -- yet another motivator to up to 100GbE. Jouanigot is optimistic that he'll get his faster ports, though. "For us, I see the introduction of 100GbE at CERN very likely next year."
Posted by Michael Feldman - December 02, 2010 @ 4:40 PM, Pacific Standard Time
Michael Feldman is the editor of HPCwire.
No Recent Blog Comments
The Xeon Phi coprocessor might be the new kid on the high performance block, but out of all first-rate kickers of the Intel tires, the Texas Advanced Computing Center (TACC) got the first real jab with its new top ten Stampede system.We talk with the center's Karl Schultz about the challenges of programming for Phi--but more specifically, the optimization...
Although Horst Simon was named Deputy Director of Lawrence Berkeley National Laboratory, he maintains his strong ties to the scientific computing community as an editor of the TOP500 list and as an invited speaker at conferences.
Supercomputing veteran, Bo Ewald, has been neck-deep in bleeding edge system development since his twelve-year stint at Cray Research back in the mid-1980s, which was followed by his tenure at large organizations like SGI and startups, including Scale Eight Corporation and Linux Networx. He has put his weight behind quantum company....
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
May 10, 2013 |
Program provides cash awards up to $10,000 for the best open-source end-user applications deployed on 100G network.
May 09, 2013 |
The Japanese government has revealed its plans to best its previous K Computer efforts with what they hope will be the first exascale system...
May 08, 2013 |
For engineers looking to leverage high-performance computing, the accessibility of a cloud-based approach is a powerful draw, but there are costs that may not be readily apparent.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.