Utility Supercomputing Heats Up

By Tiffany Trader

February 28, 2013

The HPC in the cloud space continues to evolve and one of the companies leading that charge is Cycle Computing. The utility supercomputing vendor recently reported a record-breaking 2012, punctuated by several impressive big science endeavors. One of Cycle’s most significant projects was the creation of a 50,000-core utility supercomputer inside the Amazon Elastic Compute Cloud.

Built for pharmaceutical companies Schrödinger and Nimbus Discovery, the virtual mega-cluster was able to analyze 21 million drug compounds in just 3 hours for less than $4,900 per hour. The accomplishment caught the attention of IDC analysts Chirag Dekate and Steve Conway, who elected to honor Cycle with their firm’s HPC Innovation Excellence Award.

Research Manager of IDC’s High-Performance Systems Chirag Dekate explained the award recognizes those who have best applied HPC in the ecosystem to solve critical problems. More specifically, IDC is looking for scientific achievement, ROI, and a combination of these two elements.

HPCwire spoke with Cycle CEO Jason Stowe shortly after the award was announced about the growth in HPC cloud and his company. Stowe really sees 2012 as the turning point – both for the space and for Cycle Computing. “We’ve basically hit the hockey stick growth period where there’s more rapid adoption of the technology,” he says. “Relative to utility supercomputing and HPC cloud in general we are definitely seeing a lot of interest in the space.”

During the Amazon Web Services re:Invent show in November, some big-name customers, including Novartis, Johnson & Johnson, Life Technologies, along with Hartford Insurance Group and Pacific Life Insurance, came forward to discuss their use of Cycle’s cluster-building software. The companies highlighted many of their biggest use cases and described how HPC cloud helps move the needle for Fortune500.

“Utility supercomputing applies to a large variety of companies regardless of their industry,” says Stowe, “because it supports business analytics, it supports various forms of engineering simulations and helps get the science done.”

Cycle’s customer base is well-represented across disciplines. “The majority of the top 20 big pharma companies use our software; three of the five largest variable annuity businesses use our software internally and externally or in combination,” says the CEO. The vendor also counts several leading life science companies among its customer base, including Schrödinger, who in addition to their initial 50k core run, continues to use the Cycle-EC2 cluster for ongoing workloads. Manufacturing and energy companies are also plugging into the Cycle cloud.

There are still technical and cultural barriers to cloud adoption, however. Stowe concedes the point, but only half-jokingly he adds that Cycle has solved most of the technical challenges. At this juncture, he believes the lag is more on cultural side, but there are signs of progress.

“We have these traditional companies like Johnson & Johnson and Hartford Life transitioning to a cloud model. That’s a huge cultural indicator, and definitely a C-change from four-to-five years ago,” he says.

Next >> the Business Model

The Business Model

What about the long-term profit potential for a business that relies on data parallel workloads? The question is met with a three-part answer. First off, Stowe says that Cycle has always been profitable. As a bootstrapped company, they have no investors. They’ve built a business off of a real cash-flow stream. Second, he insists that the vast amount of growth in computation is in the area of data-parallel applications.

He considers business analytics, the entirety of big data and a majority of even traditional simulation codes to be strong candidates for the cloud or utility supercomputing model.

“Sure, people still use MPI, they still use fast interconnect – but we have cases (and we hope to publish soon) where folks are running Monte Carlo simulations as a data-parallel problem. There’s a small MPI cluster that’s running the simulation, but the overall structure of the computation is parallel,” says Stowe.

Stowe expects these kinds of data-parallel or high-throughput applications to make up the bulk of new commercial workloads. The activity is coming from a range of verticals: genomics, computational chemistry, even finite element analysis.

Stowe’s final point in the context of MPI applications might be surprising to some. Cycle has seen at least two examples of real-world MPI applications that ran as much as 40 percent better on the Amazon EC2 cloud than on an internal kit that used QDR InfiniBand.

“The only real test of whether or not cloud is right for you is to actually bench it in comparison to the kit you are using in-house,” he advises.

Stowe’s team was not particularly surprised. “A lot of MPI applications under the hood are essentially doing low-interconnect, master-worker kind of workloads,” he adds.

Stowe readily admits there are applications that require the fastest interconnects and highly-tuned systems – “like weather simulations, nuclear bomb testing, the stuff at Oak Ridge or Sandia” – but he contends that some of the newer applications, especially those written in-house or by a domain scientist as opposed to a computer scientist, often run faster on cloud.

“It’s so cheap to do a bench, so why not just verify it. I’m an engineer at heart, so I’m very practical. We can talk about the theory, but it’s hard to argue with results,” he adds.

Next >> Another Tool in the Toolbox

Another Tool in the Toolbox

So much of the discussion around HPC cloud focuses on the so-called I/O problem – the bandwidth and latency challenges associated with a general public cloud like Amazon. “What about performance?” critics will ask.

Stowe feels that questions like this point to cloud necessarily replacing large capability machines, but that’s not how he sees it.

“I think of it as a radically different kind of capability machine,” says Stowe. “The old kind of capability machine required millions of dollars and tons of planning and special environments to be created, heating/cooling/power, expert staff, and so on. These systems are used very heavily for a certain kind of application, and that’s the right thing to do.”

Stowe looks at utility supercomputing as another tool in the toolbox. It doesn’t need to replace traditional capability machines, which will still be needed for certain kinds of applications. In fact, he says you can think of the Cycle-AWS cloud as another kind of capability machine with an attractive set of benefits (on-demand, pay for what you use, scalable, elastic, lower overhead).

It’s a different branch of the same tree, he says.

IDC’s Dekate takes pretty much the same position. He sees HPC in the cloud and dedicated HPC clusters as complementary.

“The HPC ecosystem is diverse and there’s a class of applications that makes sense for utility supercomputing,” says Dekate. “Solving the diverse needs of the user community requires different kinds of technological capabilities, including dedicated hardware infrastructure and HPC cloud frameworks. Our argument is that one does not have to replace the other. It’s more important to find the right kind of matches for applications that work well in either or both of these cases.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combined peak computing capacity, the new systems will extend the a Read more…

By Tiffany Trader

ASC18: Final Results Revealed & Wrapped Up

May 17, 2018

It was an exciting week at ASC18 in Nanyang, China. The student teams braved extreme heat, extremely difficult applications, and extreme competition in order to cross the cluster competition finish line. The gala awards ceremony took place on Wednesday. The auditorium was packed with student teams, various dignitaries, the media, and other interested parties. So what happened? Read more…

By Dan Olds

ASC18: Tough Applications & Tough Luck

May 17, 2018

The applications at the ASC18 Student Cluster Competition were tough. Tougher than the $3.99 steak special at your local greasy spoon restaurant. The apps are so tough that even Chuck Norris backs away from them slowly. Read more…

By Dan Olds

HPE Extreme Performance Solutions

HPC and AI Convergence is Accelerating New Levels of Intelligence

Data analytics is the most valuable tool in the digital marketplace – so much so that organizations are employing high performance computing (HPC) capabilities to rapidly collect, share, and analyze endless streams of data. Read more…

IBM Accelerated Insights

Mastering the Big Data Challenge in Cognitive Healthcare

Patrick Chain, genomics researcher at Los Alamos National Laboratory, posed a question in a recent blog: What if a nurse could swipe a patient’s saliva and run a quick genetic test to determine if the patient’s sore throat was caused by a cold virus or a bacterial infection? Read more…

Spring Meetings Underscore Quantum Computing’s Rise

May 17, 2018

The month of April 2018 saw four very important and interesting meetings to discuss the state of quantum computing technologies, their potential impacts, and the technology challenges ahead. These discussions happened in Read more…

By Alex R. Larzelere

Japan Meteorological Agency Takes Delivery of Pair of Crays

May 21, 2018

Cray has supplied two identical Cray XC50 supercomputers to the Japan Meteorological Agency (JMA) in northwestern Tokyo. Boasting more than 18 petaflops combine Read more…

By Tiffany Trader

ASC18: Final Results Revealed & Wrapped Up

May 17, 2018

It was an exciting week at ASC18 in Nanyang, China. The student teams braved extreme heat, extremely difficult applications, and extreme competition in order to cross the cluster competition finish line. The gala awards ceremony took place on Wednesday. The auditorium was packed with student teams, various dignitaries, the media, and other interested parties. So what happened? Read more…

By Dan Olds

Spring Meetings Underscore Quantum Computing’s Rise

May 17, 2018

The month of April 2018 saw four very important and interesting meetings to discuss the state of quantum computing technologies, their potential impacts, and th Read more…

By Alex R. Larzelere

Quantum Network Hub Opens in Japan

May 17, 2018

Following on the launch of its Q Commercial quantum network last December with 12 industrial and academic partners, the official Japanese hub at Keio University is now open to facilitate the exploration of quantum applications important to science and business. The news comes a week after IBM announced that North Carolina State University was the first U.S. university to join its Q Network. Read more…

By Tiffany Trader

Democratizing HPC: OSC Releases Version 1.3 of OnDemand

May 16, 2018

Making HPC resources readily available and easier to use for scientists who may have less HPC expertise is an ongoing challenge. Open OnDemand is a project by t Read more…

By John Russell

PRACE 2017 Annual Report: Exascale Aspirations; Industry Collaboration; HPC Training

May 15, 2018

The Partnership for Advanced Computing in Europe (PRACE) today released its annual report showcasing 2017 activities and providing a glimpse into thinking about Read more…

By John Russell

US Forms AI Brain Trust

May 11, 2018

Amid calls for a U.S. strategy for promoting AI development, the Trump administration is forming a senior-level panel to help coordinate government and industry research efforts. The Select Committee on Artificial Intelligence was announced Thursday (May 10) during a White House summit organized by the Office of Science and Technology Policy (OSTP). Read more…

By George Leopold

Emerging Advanced Scale Tech Trends Focus of Annual Tabor Conference

May 9, 2018

At Tabor Communications' annual Advanced Scale Forum (ASF) held this week in Austin, the focus was on enterprise adoption of HPC-class technologies and high performance data analytics (HPDA). It’s a confab that brings together end users (CIOs, IT planners, department heads) and vendors and encourages... Read more…

By the Editorial Team

MLPerf – Will New Machine Learning Benchmark Help Propel AI Forward?

May 2, 2018

Let the AI benchmarking wars begin. Today, a diverse group from academia and industry – Google, Baidu, Intel, AMD, Harvard, and Stanford among them – releas Read more…

By John Russell

How the Cloud Is Falling Short for HPC

March 15, 2018

The last couple of years have seen cloud computing gradually build some legitimacy within the HPC world, but still the HPC industry lies far behind enterprise I Read more…

By Chris Downing

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Deep Learning at 15 PFlops Enables Training for Extreme Weather Identification at Scale

March 19, 2018

Petaflop per second deep learning training performance on the NERSC (National Energy Research Scientific Computing Center) Cori supercomputer has given climate Read more…

By Rob Farber

AI Cloud Competition Heats Up: Google’s TPUs, Amazon Building AI Chip

February 12, 2018

Competition in the white hot AI (and public cloud) market pits Google against Amazon this week, with Google offering AI hardware on its cloud platform intended Read more…

By Doug Black

US Plans $1.8 Billion Spend on DOE Exascale Supercomputing

April 11, 2018

On Monday, the United States Department of Energy announced its intention to procure up to three exascale supercomputers at a cost of up to $1.8 billion with th Read more…

By Tiffany Trader

Lenovo Unveils Warm Water Cooled ThinkSystem SD650 in Rampup to LRZ Install

February 22, 2018

This week Lenovo took the wraps off the ThinkSystem SD650 high-density server with third-generation direct water cooling technology developed in tandem with par Read more…

By Tiffany Trader

Leading Solution Providers

HPC and AI – Two Communities Same Future

January 25, 2018

According to Al Gara (Intel Fellow, Data Center Group), high performance computing and artificial intelligence will increasingly intertwine as we transition to Read more…

By Rob Farber

Researchers Measure Impact of ‘Meltdown’ and ‘Spectre’ Patches on HPC Workloads

January 17, 2018

Computer scientists from the Center for Computational Research, State University of New York (SUNY), University at Buffalo have examined the effect of Meltdown Read more…

By Tiffany Trader

Google Chases Quantum Supremacy with 72-Qubit Processor

March 7, 2018

Google pulled ahead of the pack this week in the race toward "quantum supremacy," with the introduction of a new 72-qubit quantum processor called Bristlecone. Read more…

By Tiffany Trader

HPE Wins $57 Million DoD Supercomputing Contract

February 20, 2018

Hewlett Packard Enterprise (HPE) today revealed details of its massive $57 million HPC contract with the U.S. Department of Defense (DoD). The deal calls for HP Read more…

By Tiffany Trader

CFO Steps down in Executive Shuffle at Supermicro

January 31, 2018

Supermicro yesterday announced senior management shuffling including prominent departures, the completion of an audit linked to its delayed Nasdaq filings, and Read more…

By John Russell

Deep Learning Portends ‘Sea Change’ for Oil and Gas Sector

February 1, 2018

The billowing compute and data demands that spurred the oil and gas industry to be the largest commercial users of high-performance computing are now propelling Read more…

By Tiffany Trader

Nvidia Ups Hardware Game with 16-GPU DGX-2 Server and 18-Port NVSwitch

March 27, 2018

Nvidia unveiled a raft of new products from its annual technology conference in San Jose today, and despite not offering up a new chip architecture, there were still a few surprises in store for HPC hardware aficionados. Read more…

By Tiffany Trader

Hennessy & Patterson: A New Golden Age for Computer Architecture

April 17, 2018

On Monday June 4, 2018, 2017 A.M. Turing Award Winners John L. Hennessy and David A. Patterson will deliver the Turing Lecture at the 45th International Sympo Read more…

By Staff

  • arrow
  • Click Here for More Headlines
  • arrow
Share This