Oil Company Drills into HPC Cloud Issues

By Nicole Hemsoth

April 7, 2014

Danish energy company, Maersk Oil, has seen a number of technology trends evolve since its founding in 1962, particularly in terms of its ability to stay ahead of the unending race to new sources of hydrocarbons. The company’s research and technology team hasn’t been immune to the promises of cloud computing in its evaluations of potential options for powering critical simulations.

In a recent experiment in conjunction with the software company behind one of their core codes, Eclipse, which is a product from Schlumberger and Amazon Web Services, Maersk put the cloud to the test for its reservoir simulation runs.

While this experiment didn’t yield a great deal in terms of surprises performance-wise, it did raise some important questions about critical side topics in cloud for high performance computing codes. License models, overall cost, and the potential for “cloud gateway” services as delivered by companies like Panzura made this test use of public cloud resources noteworthy, even if the goal was not to come up with a world class-level infrastructure alternative.

In a recent conversation with Maersk’s head of server infrastructure, Thomas O’Reilly, he told us the intention of what the team did “was not to move our production HPC stuff into the cloud.” Rather, he says, “the point was to open a conversation about the possibilities of cloud in general and use this as an example of what might be possible.” The company provided a diagram to highlight their effort, emphasizing the differences between their bare metal environment, the cloud with multiple options in terms of configuration. He noted that the limitations to their test were Schlumberger’s requirements for running the code on certain CPUs, network cards, etc. as well as running before the newest generation of Amazon Ivy Bridge-based servers were available.

flow

O’Reilly stressed that their collaborators, Schlumberger, whom allowed them to use limited code for the test runs, learned some important lessons about license modeling. The benefit here, says O’Reilly, is that can understand firsthand the possible value of a model that goes beyond per-core pricing.

In addition to being one of the most popular reservoir simulation codes, Eclipse is also one of the most expensive, with license costs in the millions if its deployed across a large cluster given its per core pricing structure and additional required modules. Ultimately, the test case proved the well-known concept that, for smaller companies, it’s possible to secure lower startup and maintenance costs. It also proved that for these users, especially with the newest Generation 3 upgrades with Xeon E5s and enhanced networking, reasonably good performance can be extracted for applications like Eclipse at a cost model that is palatable—as long as they’re not running 24/7. The problem is, most major oil companies are cycling through Eclipse or their code of choice endlessly.

cost

With cost, performance, and licensing in mind, one might wonder why anyone would bother running a code like this on a large system if it doesn’t scale to high CPU counts. O’Reilly says these simulations can take probabilistic approaches where it runs every scenario to look for trends or deterministic simulations, which tap the knowledge of an expert reservoir engineer to point them to the best possible location for solid results to run a more focused simulation. In either case, they’re running “batches” against very large datasets and meshing the results together–divide and conquer style. This means two things for running these jobs in a cloud environment—first, these smaller runs can be nicely handled on cloud-based resources and held together within appliances like their Panzura boxes for access and sharing.

The team tried to keep the cloud configuration and use as close to real production as possible. On site they’re running the same Eclipse code around the clock with little to no downtime. Interestingly, they only have a three HP blades in their Infiniband-connected cluster with around 40 CPUs, says O’Reilly. This is because the code doesn’t scale well past that—something he says Schlumberger has an alternative to meet, called Intersect, but that they haven’t adopted.

This limitation in scalability frees them from maintaining massive clusters devoted to the code (and makes accelerators less appealing), there is no way around the need for ultra high bandwidth and low latency. That also sounds counter-intuitive to choosing cloud, but as we’ll get to in a moment, the team’s experiences were positive on the performance front, albeit at the small scale and without the real production demands for compute that a large house like Maersk requires.

All of this also means that the recent advances cloud providers like Amazon have put forth to beef up their network story could mean more customers since codes like this require Inifiniband connections (although O’Reilly pointed to some details that in the right situation, 10GbE could be effective).

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Geospatial Data Research Leverages GPUs

August 17, 2017

MapD Technologies, the GPU-accelerated database specialist, said it is working with university researchers on leveraging graphics processors to advance geospatial analytics. The San Francisco-based company is collabor Read more…

By George Leopold

Intel, NERSC and University Partners Launch New Big Data Center

August 17, 2017

A collaboration between the Department of Energy’s National Energy Research Scientific Computing Center (NERSC), Intel and five Intel Parallel Computing Centers (IPCCs) has resulted in a new Big Data Center (BDC) that Read more…

By Linda Barney

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last week the cloud giant released deeplearn.js as part of that in Read more…

By John Russell

HPE Extreme Performance Solutions

Leveraging Deep Learning for Fraud Detection

Advancements in computing technologies and the expanding use of e-commerce platforms have dramatically increased the risk of fraud for financial services companies and their customers. Read more…

Spoiler Alert: Glimpse Next Week’s Solar Eclipse Via Simulation from TACC, SDSC, and NASA

August 17, 2017

Can’t wait to see next week’s solar eclipse? You can at least catch glimpses of what scientists expect it will look like. A team from Predictive Science Inc. (PSI), based in San Diego, working with Stampede2 at the Read more…

By John Russell

Microsoft Bolsters Azure With Cloud HPC Deal

August 15, 2017

Microsoft has acquired cloud computing software vendor Cycle Computing in a move designed to bring orchestration tools along with high-end computing access capabilities to the cloud. Terms of the acquisition were not disclosed. Read more…

By George Leopold

HPE Ships Supercomputer to Space Station, Final Destination Mars

August 14, 2017

With a manned mission to Mars on the horizon, the demand for space-based supercomputing is at hand. Today HPE and NASA sent the first off-the-shelf HPC system i Read more…

By Tiffany Trader

AMD EPYC Video Takes Aim at Intel’s Broadwell

August 14, 2017

Let the benchmarking begin. Last week, AMD posted a YouTube video in which one of its EPYC-based systems outperformed a ‘comparable’ Intel Broadwell-based s Read more…

By John Russell

Deep Learning Thrives in Cancer Moonshot

August 8, 2017

The U.S. War on Cancer, certainly a worthy cause, is a collection of programs stretching back more than 40 years and abiding under many banners. The latest is t Read more…

By John Russell

IBM Raises the Bar for Distributed Deep Learning

August 8, 2017

IBM is announcing today an enhancement to its PowerAI software platform aimed at facilitating the practical scaling of AI models on today’s fastest GPUs. Scal Read more…

By Tiffany Trader

IBM Storage Breakthrough Paves Way for 330TB Tape Cartridges

August 3, 2017

IBM announced yesterday a new record for magnetic tape storage that it says will keep tape storage density on a Moore's law-like path far into the next decade. Read more…

By Tiffany Trader

AMD Stuffs a Petaflops of Machine Intelligence into 20-Node Rack

August 1, 2017

With its Radeon “Vega” Instinct datacenter GPUs and EPYC “Naples” server chips entering the market this summer, AMD has positioned itself for a two-head Read more…

By Tiffany Trader

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

Nvidia’s Mammoth Volta GPU Aims High for AI, HPC

May 10, 2017

At Nvidia's GPU Technology Conference (GTC17) in San Jose, Calif., this morning, CEO Jensen Huang announced the company's much-anticipated Volta architecture a Read more…

By Tiffany Trader

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

March 21, 2017

For a technology that’s usually characterized as far off and in a distant galaxy, quantum computing has been steadily picking up steam. Just how close real-wo Read more…

By John Russell

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

HPC Compiler Company PathScale Seeks Life Raft

March 23, 2017

HPCwire has learned that HPC compiler company PathScale has fallen on difficult times and is asking the community for help or actively seeking a buyer for its a Read more…

By Tiffany Trader

Trump Budget Targets NIH, DOE, and EPA; No Mention of NSF

March 16, 2017

President Trump’s proposed U.S. fiscal 2018 budget issued today sharply cuts science spending while bolstering military spending as he promised during the cam Read more…

By John Russell

Leading Solution Providers

Groq This: New AI Chips to Give GPUs a Run for Deep Learning Money

April 24, 2017

CPUs and GPUs, move over. Thanks to recent revelations surrounding Google’s new Tensor Processing Unit (TPU), the computing world appears to be on the cusp of Read more…

By Alex Woodie

CPU-based Visualization Positions for Exascale Supercomputing

March 16, 2017

In this contributed perspective piece, Intel’s Jim Jeffers makes the case that CPU-based visualization is now widely adopted and as such is no longer a contrarian view, but is rather an exascale requirement. Read more…

By Jim Jeffers, Principal Engineer and Engineering Leader, Intel

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

MIT Mathematician Spins Up 220,000-Core Google Compute Cluster

April 21, 2017

On Thursday, Google announced that MIT math professor and computational number theorist Andrew V. Sutherland had set a record for the largest Google Compute Engine (GCE) job. Sutherland ran the massive mathematics workload on 220,000 GCE cores using preemptible virtual machine instances. Read more…

By Tiffany Trader

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

IBM Clears Path to 5nm with Silicon Nanosheets

June 5, 2017

Two years since announcing the industry’s first 7nm node test chip, IBM and its research alliance partners GlobalFoundries and Samsung have developed a proces Read more…

By Tiffany Trader

Messina Update: The US Path to Exascale in 16 Slides

April 26, 2017

Paul Messina, director of the U.S. Exascale Computing Project, provided a wide-ranging review of ECP’s evolving plans last week at the HPC User Forum. Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This