Timesharing 2.0

By Steve Campbell

November 3, 2009

Cloud computing: Is there anything new to say? A fair question as it seems that hardly a week, or even a day, goes by without a new announcement about some new product or service for the “cloud.” When you read about cloud computing in The Economist, BusinessWeek or Forbes, you know something is really happening. Further evidence of this is the series of IBM prime time TV ads extolling the virtues of cloud computing. The technology has become mainstream.

One of the reasons business publications are writing about the cloud is because the technology is breaking out from its roots in high performance computing (HPC) and is being adopted for commercial applications. But is cloud computing today’s hot technology that promises to lower TCO, reduce energy costs, and enable dynamic or agile datacenters or is it just the latest hype? That is, will cloud computing really happen and will it deliver on its promises? And what does it mean for high performance computing?

Picture this: You’re sitting at a keyboard and you login to the system. Your ID is verified, which is good, and you begin to enter the data for your application need. When finished entering the data, the application begins executing your workload, along with many other users’ workloads. Eventually your workload completes and you receive the results together with a statement for CPU time, memory usage, disc I/O usage connect time, etc. A very comprehensive statement for all the services used. This method of access enables several other users to access the same system thus dramatically lowering the cost of computing, enabling organizations to use compute resources without owning them, and creating a development environment resulting in new applications being created.

Sound familiar? What I described was my experience using a computer system at a College in London, circa 1971. The era of timesharing had just begun. The computer system was in the datacenter (glass house) and utilized new technologies such virtualization, based on LPARs and domain, and workflow management.

In my mind, cloud computing today is Timesharing 2.0. What’s new? There are three basic differences 1) access, 2) standards, and 3) management/middleware software.

  1. Access today is from any Web-based device connected to the Internet; anytime, anywhere, any device has finally arrived.
  2. The use of standards-based software, connectivity, etc., enables heterogeneous systems to co-exist within the same cloud.
  3. Rich suites of management and middleware software and virtualization tools relieve the IT resource administration of the burden of managing this heterogeneous infrastructure and mapping workloads to infrastructure.

It’s that simple. Timesharing 2.0, better known as cloud computing, has arrived. Enough of the soapbox.

Cloud computing basics

Cloud computing is becoming ubiquitous and yet it is still evolving. Consequently, there is no accepted industry definition. Gartner defines cloud computing as “a style of computing in which scalable and elastic IT-enabled capabilities are delivered as a service to external customers using Internet technologies.”

Or try the Wikipedia definition:

Cloud computing is the provision of dynamically scalable and often virtualized resources as a service over the Internet on a utility basis. Users need not have knowledge of, expertise in, or control over the technology infrastructure in the “cloud” that supports them. Cloud computing services often provide common business applications online that are accessed from a web browser, while the software and data are stored on the servers.

The general consensus is that cloud computing has the following attributes:

  • Users can access their applications and data from any device connected to the Internet.
  • The concept generally incorporates a combination of the following:
    • Infrastructure as a Service (IaaS)
    • Platform as a Service (PaaS)
    • Software as a Service (SaaS)
  • It is frequently associated with virtualization and Web 2.0 technologies.
  • It exhibits elastic scaling – dynamic and fine grained.
  • Users can access large scale computing resources without making the heavy investment in IT infrastructure.
  • Users can access IT resources as utility service, pay-for-usage model — computing on demand.

The huge benefit of cloud computing is that companies can access the latest IT infrastructure for their workloads without having to make the huge investment in infrastructure; they can simply pay-for-usage. This is good for everyone, but for small and economically strapped firms, it is especially attractive.

One of the key software technologies is virtualization. This is significantly different from Timesharing 1.0 where virtualization was proprietary and built into the hardware. Today virtualization is a fundamental technology that enables cloud computing resource provisioning, for example, in a heterogeneous environment. Based on industry standards and utilizing the x86 VT instruction to enhanced the performance and supports multiple operating systems. Hypervisor technology is enhanced by rich set of tools for from resource provisioning to live migration.

Delivery models

Cloud computing architects are faced with many decisions and choices when developing cloud deployment models. There are several different models that are accepted in the industry today:

  • Private Cloud: Operated solely by and for the organization.
  • Public Cloud: Available to the general public on a pay-for-usage model.
  • Hybrid Cloud: A composition of private and public clouds.

There are infrastructure delivery models for seasonal fluctuations, for example, at tax time. In such models, companies with private clouds open up part of their infrastructure, creating public clouds to manage seasonal traffic.

Trends

IT vendors will continue to evolve their product lines and develop more “marketingware” as they strive for defining their uniqueness, value add and messaging. Many of them need a lot of help in differentiating themselves.

But there are a number of offerings from existing vendors that are worth watching:

The datacenter-in-a-box or container. This is a self contained IT datacenter that is delivered in a container such as Sun’s Modular Datacenter or Verari’s FOREST Container. These container-based datacenters can provide almost instant datacenter capacity for today’s cloud computing infrastructure. Designed to be eco-friendly, cost effective, and flexible.

The traditional approach. Solutions like IBM’s Cloudburst, based on IBM’s BladeCenter, or HP’s BladeSystem Matrix are conventional blade designs that can serve as cloud infrastructure. These datacenter-in-a-rack solutions can help organizations drive down the complexity and growing operating costs in particular reduce their OPEX utilities cost by delivering true green computing solutions.

Management and middleware software. Simplifying the deployment and operation of hardware (servers, storage, and networking) is the critical glue that makes the cloud model possible. The model is dependent upon this software to hide the complexity of the underlying infrastructure for the end user. For the IT organizations that are building and delivering cloud services the benefits of rich software tools will ease their task while reducing time to deploy services and simplify management.

Security. The protection of data and algorithms is perhaps the biggest concern end users have regarding cloud computing. Cybercrime is on the rise despite efforts to thwart the hackers. As consumer technology, social networking and Web 2.0 continue their rapid adoption in the workplace building secure cloud IT infrastructure is becoming more and more difficult. The best advice here is to design in security before you start building and deploying services. Don’t wait for a breach in security before taking action. Do your research.

Service. We’re starting to see third party compute cycle brokers emerge. Nimbis Services, for example, connects its clients through an industry wide brokerage and clearinghouse with 3rd party compute resources, commercial application software and expertise. The goal is to reduce risk and provide pay-as-you-go. Match users with resources.

Hybrid architectures. Over the past three or four decades HPC computing has seen many architecture to solve complex scientific workloads, we’ve seen the big SMP nodes, vector supercomputers such as Cray and mini-supercomputers such as Convex change the price performance dynamics of HPC. We have also seen numerous MPP systems. The rise of powerful commodity chipsets changed the market forever and gave birth to the distributed cluster and Grid architectures, connected via high speed network fabrics. The one architecture that survived is, Symmetric Multi-Processor (SMP), where multiple CPUs access a large shared memory, typically ccNUMA, with a single OS instance. Today that architecture is at the chip level with the x86 chipsets form Intel and AMD being multicore and 64-bit, they are SMP on a chip.

For example, Convey Computer’s server architecture combines the familiar world of x86 computing with hardware-based, application-specific instructions to accelerate certain HPC applications. Another approach to hybrid computing is that provided by vendors such as 3Leaf Systems and ScaleMP. These solutions enables a group of x86 servers to look like one big SMP system with a single pool of CPU processing and memory that can be dynamically allocated and/or repurposed to applications as needed. Essentially it turns a distributed architecture into a ccNUMA SMP.

Storage and networking. Most analysts confirm that storage is doubling every eighteen months. HPC workloads, in particular, have huge storage needs that can stress the system. There are developments such as the recent Panasas and Penguin partnership to provide high-performance parallel storage and on demand services designed specifically for high performance computing. Amazon S3 (Simple Storage Service) is an online storage web service offered by Amazon Web Services providing unlimited storage through a simple web services interface.

In the network arena, InfiniBand continues to increase its market penetration due to lower price points and a more mature software ecosystem. More interesting, however, is that several vendors are now building InfiniBand capabilities into their HPC-focused cloud solutions.

The increase demand for network performance is driven by HPC application demands and the new generations of x86 chips are able to fully utilize 10 Gigabit Ethernet (10GigE). Performance demand coupled with increased volumes of data creates the perfect storm for 10GigE adoption. One final comment on networking is the expected growth in converged network adapters (CNA) and Fiber Channel over Ethernet (FCoE). Both these offer the benefits of reduced costs and higher throughput.

How big is the opportunity?

For the vendors of products and services, the growth opportunity is large and growing rapidly. In some cases, it is hard to get any attention to your offerings if you do not have the name cloud associated with the product or service.

At the International Supercomputing Conference (ISC’09) in June 2009, Platform Computing surveyed IT executives who attended the conference. Over a quarter (28 percent) of IT executives surveyed are planning to deploy private clouds in 2009. Increased workload demand of applications and the need for IT to cut cost are cited as two major factors behind the planned adoption of HPC clouds.

The traditional analyst firms that specialize in market sizing and growth are predicting a bright future for IT infrastructure and services in the cloud. One of the most recent forecasts is in an October 2009 IDC Exchange blog titled IDC’s New IT Cloud Services Forecast: 2009-2013. In this post, IDC is forecasting that “the five year growth outlook remains strong, with a five-year annual growth rate of 26 percent — over six times the rate of traditional IT offerings.” Full details will be published in the upcoming IDC’s Cloud Services: Global Overview.

The HPC connection

For the high performance computing space, there are a growing number of companies and organizations providing services that target the special needs of this group of users. Our companion article encapsulates the vendors that are addressing this market today.

The HPC research community is also on board. In February of this year, UC Berkeley researchers released a report (PDF) discussing the impact and future directions of cloud computing. It served as a one of the first academic treatises on the subject. Eight months later, the US Department of Energy launched a five-year, $32 million program to study how scientific codes can make use of cloud technology. That work will take place at the DOE’s Argonne and Berkeley national laboratories.

Conclusion

Cloud computing is not new; it is largely an evolution of IT infrastructure. The pay-as-you-go model of cloud computing has its roots in the timesharing era of 1970s. As such, we are seeing cloud computing grow from a promising business concept to one of the fastest growing segments of the IT industry.

Organizations with challenging workload profiles or recession-hit companies are realizing they can access best-in-breed applications and infrastructure easily quickly and on a pay-for-usage basis. This now includes HPC users, who are looking to the cloud to maximize their FLOPS per dollar.

About the Author

Steve Campbell, an HPC Industry Consultant and HPC/Cloud Evangelist, has held senior VP positions in product management and product marketing for HPC and Enterprise vendors. Campbell has served in the vice president of marketing capacity for Hitachi, Sun Microsystems, FPS Computing and has also had lead marketing roles in Convex Computer Corporation and Scientific Computer Systems. Campbell has also served on the boards of and as interim CEO/CMO of several early-stage technology companies.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combined peak computing capacity, the new systems will extend the a Read more…

By Tiffany Trader

ASC18: Final Results Revealed & Wrapped Up

May 17, 2018

It was an exciting week at ASC18 in Nanyang, China. The student teams braved extreme heat, extremely difficult applications, and extreme competition in order to cross the cluster competition finish line. The gala awards ceremony took place on Wednesday. The auditorium was packed with student teams, various dignitaries, the media, and other interested parties. So what happened? Read more…

By Dan Olds

ASC18: Tough Applications & Tough Luck

May 17, 2018

The applications at the ASC18 Student Cluster Competition were tough. Tougher than the $3.99 steak special at your local greasy spoon restaurant. The apps are so tough that even Chuck Norris backs away from them slowly. Read more…

By Dan Olds

HPE Extreme Performance Solutions

HPC and AI Convergence is Accelerating New Levels of Intelligence

Data analytics is the most valuable tool in the digital marketplace – so much so that organizations are employing high performance computing (HPC) capabilities to rapidly collect, share, and analyze endless streams of data. Read more…

IBM Accelerated Insights

Mastering the Big Data Challenge in Cognitive Healthcare

Patrick Chain, genomics researcher at Los Alamos National Laboratory, posed a question in a recent blog: What if a nurse could swipe a patient’s saliva and run a quick genetic test to determine if the patient’s sore throat was caused by a cold virus or a bacterial infection? Read more…

Spring Meetings Underscore Quantum Computing’s Rise

May 17, 2018

The month of April 2018 saw four very important and interesting meetings to discuss the state of quantum computing technologies, their potential impacts, and the technology challenges ahead. These discussions happened in Read more…

By Alex R. Larzelere

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combine Read more…

By Tiffany Trader

ASC18: Final Results Revealed & Wrapped Up

May 17, 2018

It was an exciting week at ASC18 in Nanyang, China. The student teams braved extreme heat, extremely difficult applications, and extreme competition in order to cross the cluster competition finish line. The gala awards ceremony took place on Wednesday. The auditorium was packed with student teams, various dignitaries, the media, and other interested parties. So what happened? Read more…

By Dan Olds

Spring Meetings Underscore Quantum Computing’s Rise

May 17, 2018

The month of April 2018 saw four very important and interesting meetings to discuss the state of quantum computing technologies, their potential impacts, and th Read more…

By Alex R. Larzelere

Quantum Network Hub Opens in Japan

May 17, 2018

Following on the launch of its Q Commercial quantum network last December with 12 industrial and academic partners, the official Japanese hub at Keio University is now open to facilitate the exploration of quantum applications important to science and business. The news comes a week after IBM announced that North Carolina State University was the first U.S. university to join its Q Network. Read more…

By Tiffany Trader

Democratizing HPC: OSC Releases Version 1.3 of OnDemand

May 16, 2018

Making HPC resources readily available and easier to use for scientists who may have less HPC expertise is an ongoing challenge. Open OnDemand is a project by t Read more…

By John Russell

PRACE 2017 Annual Report: Exascale Aspirations; Industry Collaboration; HPC Training

May 15, 2018

The Partnership for Advanced Computing in Europe (PRACE) today released its annual report showcasing 2017 activities and providing a glimpse into thinking about Read more…

By John Russell

US Forms AI Brain Trust

May 11, 2018

Amid calls for a U.S. strategy for promoting AI development, the Trump administration is forming a senior-level panel to help coordinate government and industry research efforts. The Select Committee on Artificial Intelligence was announced Thursday (May 10) during a White House summit organized by the Office of Science and Technology Policy (OSTP). Read more…

By George Leopold

Emerging Advanced Scale Tech Trends Focus of Annual Tabor Conference

May 9, 2018

At Tabor Communications' annual Advanced Scale Forum (ASF) held this week in Austin, the focus was on enterprise adoption of HPC-class technologies and high performance data analytics (HPDA). It’s a confab that brings together end users (CIOs, IT planners, department heads) and vendors and encourages... Read more…

By the Editorial Team

MLPerf – Will New Machine Learning Benchmark Help Propel AI Forward?

May 2, 2018

Let the AI benchmarking wars begin. Today, a diverse group from academia and industry – Google, Baidu, Intel, AMD, Harvard, and Stanford among them – releas Read more…

By John Russell

How the Cloud Is Falling Short for HPC

March 15, 2018

The last couple of years have seen cloud computing gradually build some legitimacy within the HPC world, but still the HPC industry lies far behind enterprise I Read more…

By Chris Downing

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 19, 2018

Petaflop per second deep learning training performance on the NERSC (National Energy Research Scientific Computing Center) Cori supercomputer has given climate Read more…

By Rob Farber

AI Cloud Competition Heats Up: Google’s TPUs, Amazon Building AI Chip

February 12, 2018

Competition in the white hot AI (and public cloud) market pits Google against Amazon this week, with Google offering AI hardware on its cloud platform intended Read more…

By Doug Black

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Lenovo Unveils Warm Water Cooled ThinkSystem SD650 in Rampup to LRZ Install

February 22, 2018

This week Lenovo took the wraps off the ThinkSystem SD650 high-density server with third-generation direct water cooling technology developed in tandem with par Read more…

By Tiffany Trader

Leading Solution Providers

HPC and AI – Two Communities Same Future

January 25, 2018

According to Al Gara (Intel Fellow, Data Center Group), high performance computing and artificial intelligence will increasingly intertwine as we transition to Read more…

By Rob Farber

Researchers Measure Impact of ‘Meltdown’ and ‘Spectre’ Patches on HPC Workloads

January 17, 2018

Computer scientists from the Center for Computational Research, State University of New York (SUNY), University at Buffalo have examined the effect of Meltdown Read more…

By Tiffany Trader

Google Chases Quantum Supremacy with 72-Qubit Processor

March 7, 2018

Google pulled ahead of the pack this week in the race toward "quantum supremacy," with the introduction of a new 72-qubit quantum processor called Bristlecone. Read more…

By Tiffany Trader

HPE Wins $57 Million DoD Supercomputing Contract

February 20, 2018

Hewlett Packard Enterprise (HPE) today revealed details of its massive $57 million HPC contract with the U.S. Department of Defense (DoD). The deal calls for HP Read more…

By Tiffany Trader

CFO Steps down in Executive Shuffle at Supermicro

January 31, 2018

Supermicro yesterday announced senior management shuffling including prominent departures, the completion of an audit linked to its delayed Nasdaq filings, and Read more…

By John Russell

Deep Learning Portends ‘Sea Change’ for Oil and Gas Sector

February 1, 2018

The billowing compute and data demands that spurred the oil and gas industry to be the largest commercial users of high-performance computing are now propelling Read more…

By Tiffany Trader

Nvidia Ups Hardware Game with 16-GPU DGX-2 Server and 18-Port NVSwitch

March 27, 2018

Nvidia unveiled a raft of new products from its annual technology conference in San Jose today, and despite not offering up a new chip architecture, there were still a few surprises in store for HPC hardware aficionados. Read more…

By Tiffany Trader

Hennessy & Patterson: A New Golden Age for Computer Architecture

April 17, 2018

On Monday June 4, 2018, 2017 A.M. Turing Award Winners John L. Hennessy and David A. Patterson will deliver the Turing Lecture at the 45th International Sympo Read more…

By Staff

  • arrow
  • Click Here for More Headlines
  • arrow
Share This