Larry Smarr: The Future of Distributed Computing Is Here

By Alex Woodie

August 25, 2020

Larry Smarr may have stepped back from full-time work in the Computer Science and Engineering Department at the University of California, San Diego, but that doesn’t mean he’s slowing down. In fact, since his schedule has opened up, he’s now free to do other things, such as evangelizing the benefits of distributed computing for local startup Kazuhm. Datanami recently caught up with Smarr to get his take on how new technologies like Kubernetes and 5G are changing the distributed computing game.

Smarr, who received his PhD. in physics from the University of Texas at Austin in 1975, has had the sort of academic career that most computer scientists can only dream of. In 1983, he famously submitted a short 10-page report to the National Science Foundation called “A Center for Scientific and Engineering Supercomputing.” Better known as the Black Proposal (for the color of the cover), it was the first unsolicited proposal accepted by the NSF, and led to the foundation of five supercomputer centers, at Cornell, Illinois, Princeton, San Diego, and Pittsburgh.

As the founding director of the National Center for Supercomputing Applications (NCSA) at the University of Illinois, Champaign-Urbana, Smarr played a role in ushering in the modern Web with the development of Mosaic, a Web browser released in 1993 that built on the HTTP and HTML protocols that Tim Berners-Lee created three years earlier. Smarr moved to California in 2000 and helped found the California Institute for Telecommunications and Information Technology (Calit2) at UCSD and then UC Irvine.

In June, Smarr stepped back from full-time duty at UCSD and serving as the director of Calit2 and took the title Distinguished Professor Emeritus. As the principal investigator on three large computing research proposals, however, Smarr has not entirely divorced himself from UCSD. In fact, the school has called him back to “active duty” to work on these projects (so long as it doesn’t consume in excess of 43% of the time, per the bureaucratic rule book).

Freed from the day-to-day requirements of his former job, Smarr signed up to be an evangelist for distributed computing with a local San Diego outfit called Kazhum. While he has no official position with the Analytic Ventures-funded startup, he clearly sees potential in the company, which develops a platform that utilizes Docker containers and Kubernetes to enable customers to distribute computing workloads to any device, including smaller devices on the IoT.

According to Smarr, the computing world is on the cusp of major changes, even if it doesn’t necessarily feel that way to many people.

“When these things come–and they come on in exponentials–people begin to hear the words, but they don’t get there’s going to be major transformation,” Smarr says. “Go back to file sharing with Napster. All the executives at the record companies were sending out lawyers to old ladies and the kids. They didn’t get that the world was going to completely change and they were going to become extinct in about three years. Or Blockbuster. Or Netflix.

“I’ve lived through so many of these transformations, and I just think the day of ubiquitous distributed computing is finally here,” he continues. “It isn’t like it hasn’t been here for decades. That’s what I, of course, have been doing for decades. But it’s getting to the point where it’s actually becoming mainstream.”

While many computing advances emerged from academia in decades past, we’re now seeing more technological breakthroughs from companies in the private sector, Smarr points out. Google, for instance, has been party to many of the advances since 2000, including the Google File System, which Doug Cutting and Mike Cafaraella used as the inspiration for the Hadoop Distributed File System; Big Table, a NoSQL database built atop the Google File System; and Spanner, a relational database built atop Big Table.

Smarr identified Kubernetes–a resource scheduler for containers created by Google to manage its internal computing resource and released as an open source project in 2014–as an especially disruptive technology, and one that will help to usher in the next generation of distributed computing.

“To me, the containers are just an extraordinary advance because what you always needed in a distributed system was a mobile way for your software to run around in distributed system, and then execute when it found a host that it could sit on,” Smarr explains. “In other words, you’re not doing RPC [remote procedure calls]. You’re not doing remote log-in on that computer to be able to use it, which is what we had to do forever. You let the software worry about that.”

For decades, companies have needed a system that could identify all the computing resources in the organization. But it’s only been in the past few years, with Docker and Kubernetes, such systems have been available and practical to use, according to Smarr.

“I think it’s the biggest change in distributed computing since Mosaic, and more importantly since Tim Berners-Lee and CERN wrote down the HTML and HTTP protocols,” Smarr says. “That unleashed an enormous capability that we didn’t know was there. Now 2 billion people per day update their Facebook page.”

Many companies that want to move their applications to the cloud have had to “containerize” their software–i.e. adapt it to run within a container (usually Docker). Those that have done that work have benefited from the increased flexibility that cloud computing brings (although it’s not necessarily cheaper as Smarr points out).

The cloud is a homogenous computing environment, with Linux on X86 being the norm, even for Microsoft Azure. But with tools like Kazuhm, which works with heterogenous devices spanning Linux, Windows, Mac, VMware, and other operating environments for servers, desktops, laptops, smart phones, tablets and IoT devices, companies now have a way to manage their entire IT infrastructure using containers and Kubernetes.

“It’s not obvious to me that Kubernetes is an appropriate way to orchestrate containers in a normal enterprise. It’s a wonderful thing if you go, say, between the clouds or anything else,” Smarr says. “But I’m also sort of struck by the number of folks who haven’t yet understood that containers change everything and they’re still trying to figure out, ‘You mean I’ve got to containerize? What the hell does that mean?’

“We’re on that part of the learning curve where there’s still an awful lot of very smart, savvy IT people who haven’t made the connection,” he continues. “One of the things that really helps people is you have something like Kazuhm where it’s a tool that you can go into your existing enterprise– you don’t have to change any of your stuff–and it will actually map out that enterprise that you own but you maybe don’t know.”

Smarr says we’re on the cusp of another breakthrough, around connected devices and the IoT. The rapid increase in the capabilities of system-on-a-chip designs are enabling companies to place more powerful devices in more locations in the real world. As the 5G rollout brings fast network access with low latencies, the potential to leverage all that previously isolated computing power will open up new business opportunities.

“One of the things that we think are not on people’s radar yet is just how rapidly IoT processors are gaining in capacity in speed and complexity,” Smarr says. “What I see coming is this partnership between the cloud and an exponentially larger set of machine intelligent sensors on the edge of the net.”

With Calit2, Smarr was involved with WIFIRE, a project that used remote sensors to monitor for wildfires in San Diego County. The project includes a high-speed wireless grid that connects 60 backcountry stations, where optical sensors use machine learning algorithms trained on large neural nets to identify smoke plumes. This is an example of how distributed computing applications will be deployed in the future, he says.

“You’re doing a bunch of training on a neural net for what does a smoke plume look like. You don’t want to do that on the camera. You want to do that on the cloud with a zillion instances of images, and then you want to deploy that trained, weighted net into a low-energy, fast device on the edge,” Smarr says.

“Well, if you actually look at those devices, how they’re executing, many of them are becoming quite powerful,” he continues. “Right now we think of the server room as the heavy weight IT. And then we have all these hundreds or thousands of laptops. That’s lightweight. Now imagine you’ve got tens to hundreds of thousands of IoT devices. That’s the next layer coming.”

With a wide deployment of 5G antennas and many connected devices with increasingly fast processors and useful memory footprints, companies will essentially have their own private cloud at their disposal. With Kubernetes and containers, and software like Kazuhm’s handling the low-level technical details of distribution of workloads and interrupts, companies will be equipped to use that private cloud to do useful work on the massive amount of data they’re collecting.

We’re in the midst of “an invisible rising tide of new computing parallelism,” Smarr says. “It’s like somehow we did a phase change and dissolved all the barriers between all of our computing devices with containers.”

Smarr has been on the forefront of these kind of phase changes before. He points out that, when he helped start Calit2 in 2000, the group’s byline was “The Internet moving to the physical world.” That was before Wi-Fi was widely available, he points out, and three years before users started accessing the Internet on their cell phones.

“It’s kind of scary to live long enough to see your prophecies coming true,” Smarr says. “I do thank the Lord that that I’ve lived this long so that I can see some of this happening. But it was inevitable. And as it happens, as you get these hugely disruptive things happening, whether it’s the Internet or the Web or social media, people are still stuck in the older world. As William Gibson said, the future is already here. It’s just unevenly distributed.”

This story originally appeared on sister publication Datanami.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Q&A with Altair CEO James Scapa, an HPCwire Person to Watch in 2021

May 14, 2021

Chairman, CEO and co-founder of Altair James R. Scapa closed several acquisitions for the company in 2020, including the purchase and integration of Univa and Ellexus. Scapa founded Altair more than 35 years ago with two Read more…

HLRS HPC Helps to Model Muscle Movements

May 13, 2021

The growing scale of HPC is allowing simulation of more and more complex systems at greater detail than ever before, particularly in the biological research spheres. Now, researchers at the University of Stuttgart are le Read more…

Behind the Met Office’s Procurement of a Billion-Dollar Microsoft System

May 13, 2021

The UK’s national weather service, the Met Office, caused shockwaves of curiosity a few weeks ago when it formally announced that its forthcoming billion-dollar supercomputer – expected to be the most powerful weather and climate-focused supercomputer in the world when it launches in 2022... Read more…

AMD, GlobalFoundries Commit to $1.6 Billion Wafer Supply Deal

May 13, 2021

AMD plans to purchase $1.6 billion worth of wafers from GlobalFoundries in the 2022 to 2024 timeframe, the chipmaker revealed today (May 13) in an SEC filing. In the face of global semiconductor shortages and record-high demand, AMD is renegotiating its Wafer Supply Agreement and bumping up capacity. Read more…

Hyperion Offers Snapshot of Quantum Computing Market

May 13, 2021

The nascent quantum computer (QC) market will grow 27 percent annually (CAGR) reaching $830 million in 2024 according to an update provided today by analyst firm Hyperion Research at the HPC User Forum being held this we Read more…

AWS Solution Channel

Numerical weather prediction on AWS Graviton2

The Weather Research and Forecasting (WRF) model is a numerical weather prediction (NWP) system designed to serve both atmospheric research and operational forecasting needs. Read more…

Hyperion: HPC Server Market Ekes 1 Percent Gain in 2020, Storage Poised for ‘Tipping Point’

May 12, 2021

The HPC User Forum meeting taking place virtually this week (May 11-13) kicked off with Hyperion Research’s market update, covering the 2020 period. Although the HPC server market had been facing a 6.7 percent COVID-re Read more…

Behind the Met Office’s Procurement of a Billion-Dollar Microsoft System

May 13, 2021

The UK’s national weather service, the Met Office, caused shockwaves of curiosity a few weeks ago when it formally announced that its forthcoming billion-dollar supercomputer – expected to be the most powerful weather and climate-focused supercomputer in the world when it launches in 2022... Read more…

AMD, GlobalFoundries Commit to $1.6 Billion Wafer Supply Deal

May 13, 2021

AMD plans to purchase $1.6 billion worth of wafers from GlobalFoundries in the 2022 to 2024 timeframe, the chipmaker revealed today (May 13) in an SEC filing. In the face of global semiconductor shortages and record-high demand, AMD is renegotiating its Wafer Supply Agreement and bumping up capacity. Read more…

Hyperion Offers Snapshot of Quantum Computing Market

May 13, 2021

The nascent quantum computer (QC) market will grow 27 percent annually (CAGR) reaching $830 million in 2024 according to an update provided today by analyst fir Read more…

Hyperion: HPC Server Market Ekes 1 Percent Gain in 2020, Storage Poised for ‘Tipping Point’

May 12, 2021

The HPC User Forum meeting taking place virtually this week (May 11-13) kicked off with Hyperion Research’s market update, covering the 2020 period. Although Read more…

IBM Debuts Qiskit Runtime for Quantum Computing; Reports Dramatic Speed-up

May 11, 2021

In conjunction with its virtual Think event, IBM today introduced an enhanced Qiskit Runtime Software for quantum computing, which it says demonstrated 120x spe Read more…

AMD Chipmaker TSMC to Use AMD Chips for Chipmaking

May 8, 2021

TSMC has tapped AMD to support its major manufacturing and R&D workloads. AMD will provide its Epyc Rome 7702P CPUs – with 64 cores operating at a base cl Read more…

Fast Pass Through (Some of) the Quantum Landscape with ORNL’s Raphael Pooser

May 7, 2021

In a rather remarkable way, and despite the frequent hype, the behind-the-scenes work of developing quantum computing has dramatically accelerated in the past f Read more…

IBM Research Debuts 2nm Test Chip with 50 Billion Transistors

May 6, 2021

IBM Research today announced the successful prototyping of the world's first 2 nanometer chip, fabricated with silicon nanosheet technology on a standard 300mm Read more…

AMD Chipmaker TSMC to Use AMD Chips for Chipmaking

May 8, 2021

TSMC has tapped AMD to support its major manufacturing and R&D workloads. AMD will provide its Epyc Rome 7702P CPUs – with 64 cores operating at a base cl Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

CERN Is Betting Big on Exascale

April 1, 2021

The European Organization for Nuclear Research (CERN) involves 23 countries, 15,000 researchers, billions of dollars a year, and the biggest machine in the worl Read more…

HPE Launches Storage Line Loaded with IBM’s Spectrum Scale File System

April 6, 2021

HPE today launched a new family of storage solutions bundled with IBM’s Spectrum Scale Erasure Code Edition parallel file system (description below) and featu Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Saudi Aramco Unveils Dammam 7, Its New Top Ten Supercomputer

January 21, 2021

By revenue, oil and gas giant Saudi Aramco is one of the largest companies in the world, and it has historically employed commensurate amounts of supercomputing Read more…

Quantum Computer Start-up IonQ Plans IPO via SPAC

March 8, 2021

IonQ, a Maryland-based quantum computing start-up working with ion trap technology, plans to go public via a Special Purpose Acquisition Company (SPAC) merger a Read more…

Leading Solution Providers

Contributors

AMD Launches Epyc ‘Milan’ with 19 SKUs for HPC, Enterprise and Hyperscale

March 15, 2021

At a virtual launch event held today (Monday), AMD revealed its third-generation Epyc “Milan” CPU lineup: a set of 19 SKUs -- including the flagship 64-core, 280-watt 7763 part --  aimed at HPC, enterprise and cloud workloads. Notably, the third-gen Epyc Milan chips achieve 19 percent... Read more…

Can Deep Learning Replace Numerical Weather Prediction?

March 3, 2021

Numerical weather prediction (NWP) is a mainstay of supercomputing. Some of the first applications of the first supercomputers dealt with climate modeling, and Read more…

Livermore’s El Capitan Supercomputer to Debut HPE ‘Rabbit’ Near Node Local Storage

February 18, 2021

A near node local storage innovation called Rabbit factored heavily into Lawrence Livermore National Laboratory’s decision to select Cray’s proposal for its CORAL-2 machine, the lab’s first exascale-class supercomputer, El Capitan. Details of this new storage technology were revealed... Read more…

African Supercomputing Center Inaugurates ‘Toubkal,’ Most Powerful Supercomputer on the Continent

February 25, 2021

Historically, Africa hasn’t exactly been synonymous with supercomputing. There are only a handful of supercomputers on the continent, with few ranking on the Read more…

GTC21: Nvidia Launches cuQuantum; Dips a Toe in Quantum Computing

April 13, 2021

Yesterday Nvidia officially dipped a toe into quantum computing with the launch of cuQuantum SDK, a development platform for simulating quantum circuits on GPU-accelerated systems. As Nvidia CEO Jensen Huang emphasized in his keynote, Nvidia doesn’t plan to build... Read more…

New Deep Learning Algorithm Solves Rubik’s Cube

July 25, 2018

Solving (and attempting to solve) Rubik’s Cube has delighted millions of puzzle lovers since 1974 when the cube was invented by Hungarian sculptor and archite Read more…

The History of Supercomputing vs. COVID-19

March 9, 2021

The COVID-19 pandemic poses a greater challenge to the high-performance computing community than any before. HPCwire's coverage of the supercomputing response t Read more…

Microsoft to Provide World’s Most Powerful Weather & Climate Supercomputer for UK’s Met Office

April 22, 2021

More than 14 months ago, the UK government announced plans to invest £1.2 billion ($1.56 billion) into weather and climate supercomputing, including procuremen Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire