HUBzero Paving the Way for the Third Pillar of Science

By Michael McLennan and Greg Kline

February 28, 2011

What if researchers could access and share scientific simulation and modeling tools as easily as YouTube videos with the power of the cloud to drive it all? That’s the underlying premise for the HUBzero Platform for Scientific Collaboration, a cyberinfrastructure developed at Purdue University.

HUBzero was created to support, an online community for the Network for Computational Nanotechnology (NCN), which the U.S .National Science Foundation has funded since 2002 to connect the theorists who develop simulation tools with the experimentalists and educators who might use them.

Since 2007, HUBzero’s use has expanded to support more than 30 hubs — and growing — in fields ranging from microelectromechanical systems and volcanology to translating lab discoveries into new medical treatments and the development of assistive technologies for people with disabilities.

HUBzero is now supported by a consortium including Purdue, Indiana, Clemson and Wisconsin. Researchers at Rice, the State University of New York system, the University of Connecticut and Notre Dame use hubs. Purdue offers a hub-building and -hosting service and the consortium also supports an open source release, allowing people to build and host their own. HUBbub2010, the first of planned annual HUBzero conferences, drew more than 100 people from 33 institutions as far away as Korea, South Africa and Quebec, along with U.S. universities nationwide.

Although they serve different communities, the hubs all support collaborative development and dissemination of computational models running in an infrastructure that leverages cloud computing resources and makes it easier to take advantage of them. Meanwhile, built-in social networking features akin to Facebook create communities of researchers, educators and practitioners in almost any field or subject matter and facilitate communication and collaboration, distribution of research results, training and education.

“Contributors can structure their material and upload it without an inordinate amount of handholding; that’s really a key because you want people to contribute,” says Purdue chemical engineering Professor Gintaras Reklaitis. He’s the principal investigator for, a National Science Foundation-supported Virtual Engineering Organization for advancing the science and engineering of pharmaceutical product development and manufacturing.

One could cobble some of this functionality together with commercial Web software, but HUBzero integrates everything in a single package. Add the research tool-enabling features and research-oriented functions like tracking the use of tools (useful for quantifying outreach) and citation tracking and you have something quite different — and powerful.

HUBzero can be a prime tool for satisfying cyberinfrastructure requirements, such as data management and access, of granting agencies like the NSF. HUBzero’s emphasis on interdisciplinary collaboration only makes it more attractive in funding proposals. A hub is central to the Purdue-based Network for Earthquake Engineering Simulation (NEES), a $105 Million NSF program announced in 2009, the largest single award in Purdue history. Purdue’s PRISM Center for micro-electro-mechanical systems and C3Bio biofuels research center, both funded by the U.S. Department of Energy, are some other recent major award winners employing hubs.

Such an infrastructure can have an impact on scientific discovery, as clearly shows.

As of December 2010, NCN identified 719 citations in the scientific literature that referenced In addition, user registration information indicates that more than 480 classes at more than 150 institutions have utilized nanoHUB. Because the site is completely open and notification of classroom usage is voluntary, the actual classroom usage undoubtedly exceeds these numbers. There are users in the top 50 U.S. universities (per the listing by U.S. News and World Report) and in 18 percent of the 7,073 U.S. institutions carrying the .edu extension in their domain name. “Nano” is a tiny area in science and technology, but nanoHUB is big in many institutions.

The community Web site now has more than 740 contributors and 195 interactive simulation tools. In 2010 more than 9,800 users ran 372,000 simulations. In addition to online simulations, the site offers 52 courses on various nano topics as well as 2,300 seminars and other resources, which have propelled the annual user numbers to more than 170,000 people in 172 countries.

Likewise, the cancer care engineering hub, one of the early hubs following nanoHUB, has proven to be the linchpin in building an online data tracking, access and statistical modeling community aimed at advancing cancer prevention and care.

“We were looking for a solution for sample tracking and data storage that would not cost $5 million and it was a true logistical challenge needing a comprehensive cyberinfrastructure support system,” says Julie Nagel, managing director of the Oncological Sciences Center in Purdue’s Discovery Park. “The hub is the core of the CCE project and has brought the project forward so much faster than we could have if we had started from scratch.”

The success of is what attracted the attention of Noha Gaber, who was seeking a good way to facilitate collaboration in the environmental modeling field when she came across the thriving international resource for nanotechnology research and education.

HUBzero, the technology powering nanoHUB, could obviously be used to build a Web-based repository of models and related documentation for projecting the spread and impact of pollutants. It also had built-in features, such as wiki space, enabling environmental researchers to share ideas and information. But the ability to make the models operable online, right in a Web browser window, and allow researchers to collaborate virtually in developing and using models was the deal closer.

“It’s not just providing a library of models, but providing direct access to these tools,” says Gaber, executive director of the U.S. Environmental Protection Agency’s Council for Regulatory Environmental Modeling. She’s a driving force behind the new for integrated environmental modeling.

Under the hood, HUBzero is a software stack developed by Purdue (and being refined continuously by the consortium and hub users) and designed to work with open source software supported by active developer communities. This includes Debian GNU/Linux, Apache HTTP Server, LDAP, MySQL, PHP, Joomla and OpenVZ.

HUBzero’s middleware hosts the live simulation tool sessions and makes it easy to connect the tools to supercomputing clusters and cloud computing infrastructure to solve large computational problems. HUBzero’s Rappture tool kit helps turn research codes written in C/C++, Fortran, Java, MATLAB, and other languages into graphical, Web-enabled applications.

On the surface, the simulation tools look like simple Java applets embedded within the browser window, but they’re actually running on cluster or cloud hosts and projected to the user’s browser using virtual network computing (VNC). Each tool runs in a restricted lightweight virtual environment implemented using OpenVZ, which carefully controls access to file systems, networking, and other server processes. A hub can direct jobs to national resources such as the TeraGrid, Open Science Grid and Purdue’s DiaGrid as well as other cloud-style systems. This delivers substantial computing power to thousands of end users without requiring, for example, that they log into a head node or fuss with proxy certificates.

The tools on each hub come not from the core development team but from hundreds of other researchers scattered throughout the world. HUBzero supports the workflow for all of these developers and has a content management system for tool publication. Developers receive access to a special HUBzero “workspace,” which is a Linux desktop running in a secure execution environment and accessed via a Web browser (like any other hub tool). There, they create and test their tools in the same execution environment as the published tools, with access to the same visualization cluster and cloud resources for testing. HUBzero can scale to support hundreds of independent tool development teams, each publishing, modifying, and republishing their tool dozens of times per year.

If a tool already has a GUI that runs under Linux, it can be deployed as-is in a matter of hours. If not, tool developers can use HUBzero’s Rappture toolkit to create a GUI with little effort. Rappture reads an XML description of the tool’s inputs and outputs and then automatically generates a GUI. The Rappture library supports approximately two dozen objects — including numbers, Boolean values, curves, meshes, scalar/vector fields, and molecules — which can be used to represent each tool’s inputs and outputs. The input and output values are accessed within a variety of programming languages via an Application Programming Interface (API). Rappture supports APIs for C/C++, Fortran, Java, MATLAB, Python, Perl, Ruby, and Tcl, so it can accommodate various modeling codes. The results from each run are loaded back into the GUI and displayed in a specialized viewer created for each output type. Viewers for molecules, scalar and vector fields, and other complex types can be automatically connected to a rendering farm for hardware-accelerated 3-D data views.

HUBzero sites also provide ways for colleagues to work together. For example, because of the unique way the HUBzero middleware hosts tool sessions, a single session can be shared among any number of people. A group of people can look at the same session at the same time and discuss ideas over the phone or instant messaging. If some of the people aren’t online or available, they can access the session later from their My Hub page and follow up at their convenience. Some commercial collaboration tools, such as Adobe Presenter, also work within HUBzero (hub builders are required to license these).

As people are using the tools, questions arise and sometimes things go wrong. HUBzero supports many ways for users to find help and help one another and includes a built-in trouble report system. Users also can post questions in a community forum modeled after’s Askville or Yahoo! Answers. In practice, many tickets aren’t really problems but are actually requests for new features. HUBzero supports a wish list capability for collecting, prioritizing and acting on such requests. Users can post an idea to the wish list associated with each tool or to the general list associated with the hub itself.

HUBzero’s unique blend of simulation power and social networking seems to resonate across engineering and science communities. As hub use continues to grow, a goal is to develop new capabilities to connect related content so that tools published on one hub can be easily found on all others. Another goal is to improve tool interconnection, so that one tool’s output can be used as input to another, letting developers solve larger problems by connecting a series of models from independent authors.

Michael McLennan is the senior research scientist and hub technology architect at Purdue. Greg Kline is the science and technology writer for Information Technology at Purdue (ITaP).

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

New Exascale System for Earth Simulation Introduced

April 23, 2018

After four years of development, the Energy Exascale Earth System Model (E3SM) will be unveiled today and released to the broader scientific community this month. The E3SM project is supported by the Department of Energy Read more…

By Staff

RSC Reports 500Tflops, Hot Water Cooled System Deployed at JINR

April 18, 2018

RSC, developer of supercomputers and advanced HPC systems based in Russia, today reported deployment of “the world's first 100% ‘hot water’ liquid cooled supercomputer” at Joint Institute for Nuclear Research (JI Read more…

By Staff

New Device Spots Quantum Particle ‘Fingerprint’

April 18, 2018

Majorana particles have been observed by university researchers employing a device consisting of layers of magnetic insulators on a superconducting material. The advance opens the door to controlling the elusive particle Read more…

By George Leopold

HPE Extreme Performance Solutions

Hybrid HPC is Speeding Time to Insight and Revolutionizing Medicine

High performance computing (HPC) is a key driver of success in many verticals today, and health and life science industries are extensively leveraging these capabilities. Read more…

Cray Rolls Out AMD-Based CS500; More to Follow?

April 18, 2018

Cray was the latest OEM to bring AMD back into the fold with introduction today of a CS500 option based on AMD’s Epyc processor line. The move follows Cray’s introduction of an ARM-based system (XC-50) last November. Read more…

By John Russell

Cray Rolls Out AMD-Based CS500; More to Follow?

April 18, 2018

Cray was the latest OEM to bring AMD back into the fold with introduction today of a CS500 option based on AMD’s Epyc processor line. The move follows Cray’ Read more…

By John Russell

IBM: Software Ecosystem for OpenPOWER is Ready for Prime Time

April 16, 2018

With key pieces of the IBM/OpenPOWER versus Intel/x86 gambit settling into place – e.g., the arrival of Power9 chips and Power9-based systems, hyperscaler sup Read more…

By John Russell

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Cloud-Readiness and Looking Beyond Application Scaling

April 11, 2018

There are two aspects to consider when determining if an application is suitable for running in the cloud. The first, which we will discuss here under the title Read more…

By Chris Downing

Transitioning from Big Data to Discovery: Data Management as a Keystone Analytics Strategy

April 9, 2018

The past 10-15 years has seen a stark rise in the density, size, and diversity of scientific data being generated in every scientific discipline in the world. Key among the sciences has been the explosion of laboratory technologies that generate large amounts of data in life-sciences and healthcare research. Large amounts of data are now being stored in very large storage name spaces, with little to no organization and a general unease about how to approach analyzing it. Read more…

By Ari Berman, BioTeam, Inc.

IBM Expands Quantum Computing Network

April 5, 2018

IBM is positioning itself as a first mover in establishing the era of commercial quantum computing. The company believes in order for quantum to work, taming qu Read more…

By Tiffany Trader

FY18 Budget & CORAL-2 – Exascale USA Continues to Move Ahead

April 2, 2018

It was not pretty. However, despite some twists and turns, the federal government’s Fiscal Year 2018 (FY18) budget is complete and ended with some very positi Read more…

By Alex R. Larzelere

Nvidia Ups Hardware Game with 16-GPU DGX-2 Server and 18-Port NVSwitch

March 27, 2018

Nvidia unveiled a raft of new products from its annual technology conference in San Jose today, and despite not offering up a new chip architecture, there were still a few surprises in store for HPC hardware aficionados. Read more…

By Tiffany Trader

Inventor Claims to Have Solved Floating Point Error Problem

January 17, 2018

"The decades-old floating point error problem has been solved," proclaims a press release from inventor Alan Jorgensen. The computer scientist has filed for and Read more…

By Tiffany Trader

Researchers Measure Impact of ‘Meltdown’ and ‘Spectre’ Patches on HPC Workloads

January 17, 2018

Computer scientists from the Center for Computational Research, State University of New York (SUNY), University at Buffalo have examined the effect of Meltdown Read more…

By Tiffany Trader

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

How the Cloud Is Falling Short for HPC

March 15, 2018

The last couple of years have seen cloud computing gradually build some legitimacy within the HPC world, but still the HPC industry lies far behind enterprise I Read more…

By Chris Downing

Chip Flaws ‘Meltdown’ and ‘Spectre’ Loom Large

January 4, 2018

The HPC and wider tech community have been abuzz this week over the discovery of critical design flaws that impact virtually all contemporary microprocessors. T Read more…

By Tiffany Trader

How Meltdown and Spectre Patches Will Affect HPC Workloads

January 10, 2018

There have been claims that the fixes for the Meltdown and Spectre security vulnerabilities, named the KPTI (aka KAISER) patches, are going to affect applicatio Read more…

By Rosemary Francis

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Fast Forward: Five HPC Predictions for 2018

December 21, 2017

What’s on your list of high (and low) lights for 2017? Volta 100’s arrival on the heels of the P100? Appearance, albeit late in the year, of IBM’s Power9? Read more…

By John Russell

Leading Solution Providers

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 19, 2018

Petaflop per second deep learning training performance on the NERSC (National Energy Research Scientific Computing Center) Cori supercomputer has given climate Read more…

By Rob Farber

Lenovo Unveils Warm Water Cooled ThinkSystem SD650 in Rampup to LRZ Install

February 22, 2018

This week Lenovo took the wraps off the ThinkSystem SD650 high-density server with third-generation direct water cooling technology developed in tandem with par Read more…

By Tiffany Trader

AI Cloud Competition Heats Up: Google’s TPUs, Amazon Building AI Chip

February 12, 2018

Competition in the white hot AI (and public cloud) market pits Google against Amazon this week, with Google offering AI hardware on its cloud platform intended Read more…

By Doug Black

HPC and AI – Two Communities Same Future

January 25, 2018

According to Al Gara (Intel Fellow, Data Center Group), high performance computing and artificial intelligence will increasingly intertwine as we transition to Read more…

By Rob Farber

New Blueprint for Converging HPC, Big Data

January 18, 2018

After five annual workshops on Big Data and Extreme-Scale Computing (BDEC), a group of international HPC heavyweights including Jack Dongarra (University of Te Read more…

By John Russell

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Momentum Builds for US Exascale

January 9, 2018

2018 looks to be a great year for the U.S. exascale program. The last several months of 2017 revealed a number of important developments that help put the U.S. Read more…

By Alex R. Larzelere

Google Chases Quantum Supremacy with 72-Qubit Processor

March 7, 2018

Google pulled ahead of the pack this week in the race toward "quantum supremacy," with the introduction of a new 72-qubit quantum processor called Bristlecone. Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Share This