Visit additional Tabor Communication Publications
November 05, 2012
SALT LAKE CITY, Utah. – November 5, 2012 – This year, SC12 will not only showcase the next generation of high performance computing applications but it will also be home to seven innovative network research projects through a special program called the SCinet Research Sandbox (SRS).
SCinet is the primary high performance network infrastructure built each year for SC exhibitors to highlight their cutting edge computing applications and collaborations. As a key component of SCinet, the SRS is designed to enable researchers to experiment and demonstrate their ideas on innovative network architectures, applications and protocols in the unique live environment of the SCinet network. This year, the SRS will provide researchers with dedicated access to multiple 100 Gigabits per second wide area network links as well as a 10 Gigabit per second (Gbps) OpenFlow network testbed.
“In addition to supporting the extreme demands of the HPC-based demonstrations that have become the trademark of the conference, SCinet also seeks to foster and highlight developments in network research that will be necessary to support the next-generation of science applications,” said Brian Tierney, SRS co-chair for SC12 and head of ESnet’s Advanced Network Technologies Group. “Both 100 Gbps networking and OpenFlow have become some of the most influential networking technologies of this decade. SRS allows the community to showcase innovations on these platforms while in their infancy to demonstrate the impact they may have on the entire HPC community in the future.”
“Openflow has the potential to greatly improve applications that are necessary for advancements in HPC such as GridFTP and others. By allowing these applications to access the network in a highly intelligent and programmable manner, OpenFlow can significantly improve end-to-end network performance especially for bulk data transfers, which will be an increasingly difficult challenge in the era of data intensive science,” said Andrew Lee, SRS co-chair for SC12 and Principal Network Systems Engineer for the Global Research Network Operations Center at Indiana University. “Demonstrations like those being supported by the SCinet Research Sandbox are laying the groundwork for these critical advancements as well as demonstrating to the community - in a tangible way - the possibilities that Openflow provides."
Seven projects have been selected as part of the SRS program, all of which will be showcased as part of the Technical Program and will be demonstrating their research in several exhibit booths during the conference. For detailed information on the projects and their presentations visit: http://sc12.supercomputing.org/content/scinet-research-sandbox
2012 SRS projects include:
Efficient LHC Data Distribution across 100Gbps Networks
The analysis of data leading to the recent discoveries at the Large Hadron Collider produces data flows of more than 100 Petabytes per year, and increasingly relies on the efficient movement of data sets between the globally distributed computing sites.
The team will demonstrate the state-of-the-art data movement tools, as enabling technology for high-throughput data distribution over 100Gbps WAN circuits. The demo will interconnect 3 major LHC Tier-2 computing sites and the SC12 show floor (booth 809) using 100Gbps technology.
Collaborating organizations: University of Victoria, University of Michigan, California Institute of Technology, Vanderbilt University, Internet2, ESnet, CENIC, Starlight, PacWave/GLORIAD, KNU/KISTI as well as vendors: Alcatel Lucent, Ciena, Cisco, Juniper Networks, Mellanox. Dell-Force10, SuperMicro, ASA Micro Systems, Data Direct Network, Fusion-IO, PADTECH
Demonstration booth: 809
Presentation: November 15, 2012, 8:50am MT in Room 155-F2
Exploiting Network Parallelism for Improving Data Transfer Performance
The task of scientific bulk data movement, e.g. migrating collected results from the instrumentation to the processing and storage facilities, is hampered by a lack of available network resources. Traditional R&E connectivity can be congested on portions of an end-to-end path causing degradation of overall performance. This SRS project will explore dynamic network control to facilitate efficient bulk data movement, combining opportunistic use of "traditional" networks with dedicated reservations over virtual circuits and OpenFlow enabled resources. The GridFTP application has been instrumented with the eXtensible Session Protocol (XSP), an intelligent system capable of controlling programmable networks. The project intends to show end-to-end performance improvement between the SC12 conference and campuses involved in the DYNES project, through a combination of regular connectivity, dynamic bandwidth allocations, TCP acceleration, and operations using multiple paths.
Collaborating organizations: Indiana University, Lawrence Berkeley National Laboratory, Argonne National Laboratory and Internet2
Demonstration booths: 1042, 1343
Presentation: Thursday, November 15, 2012, 9:10am MT in Room 155-F2
Multipathing with MPTCP and OpenFlow
This demo shows several emerging network technologies and how these can be used in big data transfers between data centers. In this demonstration traffic is sent simultaneously across multiple OpenFlow controlled paths between Geneva and Salt Lake City. The congestion control mechanism of Multipath TCP (MPTCP) favours the least congested paths and ensures that the load balancing across the paths is always optimal.
Collaborating organizations SURFnet, SARA, iCAIR and California Institute of Technology
Demonstration booths: 2333, 809, 501
Presentation: November 15, 2012, 10:30am MT in Room 155-F2
Next Generation Science DMZ Model with OpenFlow
The emerging era of “Big Science” demands the highest possible network performance. End-to-end circuit automation and workflow-driven customization are two essential capabilities needed for networks to scale to meet this challenge. This demonstration showcases how combining software-defined networking techniques with virtual circuits capabilities can transform the network into a dynamic, customer-configurable virtual switch. In doing so, users are able to rapidly customize network capabilities to meet their unique workflows with little to no configuration effort. The demo also highlights how the network can be automated to support multiple collaborations in parallel.
Collaborating organizations: ESnet, Ciena Corporation
Demonstration booth: 2437
Presentation: November 15, 2012, 11:10am MT in Room 155-F2
OpenFlow Enabled Hadoop over Local and Wide Area Cluster
The Hadoop Distributed File Systems and Hadoop's implementation of MapReduce is one of the most widely used platforms for data intensive computing. The shuffle and sort phases of a MapReduce computation often saturate network links to nodes and the reduce phase of the computation must wait for data. This study explores the use of OpenFlow to the control network configuration for different flows to thereby provide different network characteristics for different categories of Hadoop traffic.
Collaborating organizations: University of Chicago
Demonstration booth: 501
Presentation: November 15, 2012, 10:50am MT in Room 155-F2
OpenFlow Services for Science: An International Experimental Research Network
Large-scale data intensive science requires global collaboration and sophisticated high capacity data management. The emergence of more flexible networking, for example, using techniques based on OpenFlow, provides opportunities to address these issues because these techniques enable a high degree of network customization and dynamic provisioning. These techniques enable large-scale facilities to be created that can be used to prototype new architecture, services, protocols, and technologies. A number of research organizations from several countries have designed and implemented a persistent international experimental research facility that can be used to prototype, investigate, and test network innovations for large-scale global science. For SC12, this international experimental network facility will be extended to from sites across the world to the conference showfloor, and it will be used to support several testbeds and to showcase a series of complementary demonstrations.
Collaborating organizations: International Center for Advanced Internet Research Northwestern University; National Center for High-Performance Computing, Taiwan; University of Applied Sciences, Taiwan; National Cheng-Kung University, Taiwan; SARA, The Netherlands, California Institute of Technology/CERN; SURFnet, The Netherlands.
Demonstration booths: 2333, 501, 843, 809
Presentation: November 15, 2012, 8:30am MT in Room 155-F2
Scalable Cyber-Security for Terabit Cloud Computing
Reservoir Labs will demonstrate R-Scope, a scalable, high-performance network packet inspection technology that forms the core of a new generation of Intrusion Detection Systems enabling the construction and deployment of cyber security infrastructures scaling to terabit per second ingest bandwidths. This scalability is enabled by the use of low- power and high-performance manycore network processors combined with Reservoir’s enhancements to Bro. The innovative R-Scope PACE-T appliance, implemented on a 1U Tilera TILExtreme-Gx platform, will demonstrate the capacity to perform cyber-security analysis at 80Gbps, by combining cyber-security aware front-end network traffic load balancing tightly coupled with the full back-end analytic power of Bro. This fully-programmable platform incorporates the full Bro semantics into the appliance’s load-balancing front-end and the back-end analytic nodes.
Collaborating organizations: Reservoir Labs, SCinet Security Team
Presentation: November 15, 2012, 9:30am MT in Room 155-F2
In quieter times, sounding the bell of funding big science with big systems tends to resonate further than when ears are already burning with sour economic and national security news. For exascale's future, however, the time could be ripe to instill some sense of urgency....
In a recent solicitation, the NSF laid out needs for furthering its scientific and engineering infrastructure with new tools to go beyond top performance, Having already delivered systems like Stampede and Blue Waters, they're turning an eye to solving data-intensive challenges. We spoke with the agency's Irene Qualters and Barry Schneider about..
Large-scale, worldwide scientific initiatives rely on some cloud-based system to both coordinate efforts and manage computational efforts at peak times that cannot be contained within the combined in-house HPC resources. Last week at Google I/O, Brookhaven National Lab’s Sergey Panitkin discussed the role of the Google Compute Engine in providing computational support to ATLAS, a detector of high-energy particles at the Large Hadron Collider (LHC).
May 23, 2013 |
The study of climate change is one of those scientific problems where it is almost essential to model the entire Earth to attain accurate results and make worthwhile predictions. In an attempt to make climate science more accessible to smaller research facilities, NASA introduced what they call ‘Climate in a Box,’ a system they note acts as a desktop supercomputer.
May 22, 2013 |
At some point in the not-too-distant future, building powerful, miniature computing systems will be considered a hobby for high schoolers, just as robotics or even Lego-building are today. That could be made possible through recent advancements made with the Raspberry Pi computers.
May 16, 2013 |
When it comes to cloud, long distances mean unacceptably high latencies. Researchers from the University of Bonn in Germany examined those latency issues of doing CFD modeling in the cloud by utilizing a common CFD and its utilization in HPC instance types including both CPU and GPU cores of Amazon EC2.
May 15, 2013 |
Supercomputers at the Department of Energy’s National Energy Research Scientific Computing Center (NERSC) have worked on important computational problems such as collapse of the atomic state, the optimization of chemical catalysts, and now modeling popping bubbles.
05/10/2013 | Cleversafe, Cray, DDN, NetApp, & Panasas | From Wall Street to Hollywood, drug discovery to homeland security, companies and organizations of all sizes and stripes are coming face to face with the challenges – and opportunities – afforded by Big Data. Before anyone can utilize these extraordinary data repositories, however, they must first harness and manage their data stores, and do so utilizing technologies that underscore affordability, security, and scalability.
04/15/2013 | Bull | “50% of HPC users say their largest jobs scale to 120 cores or less.” How about yours? Are your codes ready to take advantage of today’s and tomorrow’s ultra-parallel HPC systems? Download this White Paper by Analysts Intersect360 Research to see what Bull and Intel’s Center for Excellence in Parallel Programming can do for your codes.
In this demonstration of SGI DMF ZeroWatt disk solution, Dr. Eng Lim Goh, SGI CTO, discusses a function of SGI DMF software to reduce costs and power consumption in an exascale (Big Data) storage datacenter.
The Cray CS300-AC cluster supercomputer offers energy efficient, air-cooled design based on modular, industry-standard platforms featuring the latest processor and network technologies and a wide range of datacenter cooling requirements.