Acceleware Offers GPU-Flavored HPC

By John E. West

July 10, 2008

Market analysts are predicting big things for HPC. As growth in the rest of the server industry flattens out, they see HPC continuing to expand. Most of this growth is at the bottom, and the vision is one of many new customers stepping up to buy clusters of 16, 32, or 64 nodes costing less than $50,000.

It’s a believable vision, but one with serious obstacles to overcome. HPC is hard — inexcusably so. In a blog post several weeks ago, Tabor Research analyst Chris Willard passed on one of the many definitions of the term “supercomputer” he has heard over the years: “Any computer where the compute time is worth more than the programmer time.” This is a pervasive definition that has worked its way deep into our thinking about HPC. It is also totally wrong-headed.

Another obstacle to broad entry-level adoption of HPC is change: the technologies of HPC evolve faster than most of the vendors of the software tools that are driving low-end HPC adoption can keep up with. Cost is another problem; although costs of hardware are falling and have fallen dramatically, they are still out of the reach of much of the small enterprises that could make use of HPC, and the entry path often isn’t a smooth slope. One is either using a $5,000 workstation or a $100,000 cluster. Until recently, there hasn’t been much in the middle, and the offerings are still pretty thin.

Finally, the very people who need to buy the computers to fulfill the visions of dramatic HPC growth do not identify with HPC, or supercomputing, as solutions to their problems. They are a “dark market.” So HPC has to find them, in small corners of very diverse markets, and find a way to sell to them that will cause the vast new market at the bottom of the HPC customer pyramid to finally take shape and start making actual purchases.

Acceleware’s strategy is to to bring HPC to a diverse set of customers, one vertical at a time. I recently spoke with Ryan Schneider, Acceleware’s CTO, and Robert Miller, the vice president of marketing and product management, about the company, its technology, and its approach to the market.

Acceleware’s technology is software deployed on top of NVIDIA’s GPUs and HP’s servers. Although the company doesn’t manufacture the hardware, it does intense integration with the result that it owns the entire solution delivered to a customer, from silicon to application software. The company works with ISVs in a vertical to modify the key software tools in that domain in order to take advantage of GPU acceleration using Acceleware’s middleware, the “Acceleware Technology Platform.” Right now Acceleware has offerings in virtual prototyping, imaging, and the oil and gas industries.

There may be a lot of code work done under the hood, but the users continue to work with their favorite software using their familiar UIs and workflows without disruption. The only change users will see is something akin to a checkbox asking whether they want to run their simulation with acceleration, or not. That’s it.

What’s the payoff? For users it can be tremendous. They buy hardware in a familiar workstation form factor or as one of the smaller C30 clusters, running the same OS and applications they’ve already run. The unit comes ready to plug in, totally configured with their domain application software licensed and ready to run. And they get performance advantages that can be significant. For example, cell phone manufacturers using Acceleware’s kit have seen turnaround times decrease from 10 hours per simulation run — in an analysis suite that typically requires 400 runs — to 15 minutes. Medical providers using Acceleware’s imaging kit for reconstruction of medical image data have seen their run times shrink from a month to a few days, and a large drug company saw their imaging workflow shrink from a week to a few hours.

What’s the motivation for ISVs to work with Acceleware? Ryan Schneider, Acceleware’s CTO, described it as a build-versus-buy value proposition for these companies. For the most part, the software companies Acceleware works with have limited resources that are focused very tightly in their application domain writing the software their users want. Porting to Acceleware’s middleware allows companies to insulate themselves from technology change risk. As new technologies replace or augment GPU acceleration as viable alternatives in technical computing, Acceleware simply adds support for the new hardware underneath the middleware layer.

The relationship with the software vendors gives Acceleware direct access to what would otherwise be a dark market at the low end. Acceleware sells directly into the verticals they are targeting, augmenting channel sales driven by the ISVs.

What’s the motivation for customers? That one is easy, and I’ve already touched on it. For a nominal expense — anywhere from $10,000 for a single acceleration card to $250,000 for a four node, 16 GPU cluster, including hardware and application licensing costs — users get access to what is, in some cases, orders of magnitude better performance without having to change their workflow. This puts Acceleware’s products square in the middle of production design and operational computing for some very serious companies. Firms like Boston Scientific, Philips, Nokia, Samsung, Boeing, Eli Lilly, and Hitachi are using Acceleware’s equipment.

Speaking of equipment, what does Acceleware offer? All of the company’s hardware supports the same performance middleware, so any application ported to one of the company’s products runs on them all. At the very low end, users can buy a single GPU card, the A30, with the application and licensing for around $10,000. The D30 is a dual card configuration installed in a workstation with software for roughly $25,000, and there is a quad card configuration with workstation for roughly $65,000.

The company also recently announced a clustered solution. The C30 series of hardware clusters 4, 8, 12, or 16 nodes together to form a GPU cluster. Each node has two dual-core Intel processors connected to four GPUs; the nodes are interconnected with InfiniBand. MPI is used for communication between the nodes on the coarse-grained problem decomposition, and each node then parcels out work to its GPUs as appropriate for the problem. The company claims that the C30-16, with 64 GPUs, is capable of hitting 32 TFLOPS peak single-precision performance. I asked about the single-precision part; evidently in the verticals the company is currently operating in, that’s A-OK. Seismic acquisition data, for example, starts life as less than 32-bit precision anyway. The C30-4 node configuration is priced around $250,000.

Inside the Acceleware’s middelware platform are a host of common and industry-specific algorithms that have been tuned for optimal performance on NVIDIA’s GPUs. In the library you’ll find optimized FDTD algorithms, Feldkamp image reconstruction, Kirchhoff algorithms for the oil and gas industry, and more general purpose matrix math libraries.

Of course there is competition. Companies like RapidMind are building software akin to Acceleware’s middleware for abstracting the hardware and allowing software to be retargeted at FPGAs, Cell processors, GPUS, and so on, and this will be a good solution for many users outside the verticals that Acceleware is operating in. But for sheer ease of use, Acceleware seems to offer a product that is uniquely focused on making the transition from ordinary computing to supercomputing a seamless one.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Mira Supercomputer Enables Cancer Research Breakthrough

November 11, 2019

Dynamic partial-wave spectroscopic (PWS) microscopy allows researchers to observe intracellular structures as small as 20 nanometers – smaller than those visible by optical microscopes – in three dimensions at a mill Read more…

By Staff report

IBM Adds Support for Ion Trap Quantum Technology to Qiskit

November 11, 2019

After years of percolating in the shadow of quantum computing research based on superconducting semiconductors – think IBM, Rigetti, Google, and D-Wave (quantum annealing) – ion trap technology is edging into the QC Read more…

By John Russell

Tackling HPC’s Memory and I/O Bottlenecks with On-Node, Non-Volatile RAM

November 8, 2019

On-node, non-volatile memory (NVRAM) is a game-changing technology that can remove many I/O and memory bottlenecks and provide a key enabler for exascale. That’s the conclusion drawn by the scientists and researcher Read more…

By Jan Rowell

What’s New in HPC Research: Cosmic Magnetism, Cryptanalysis, Car Navigation & More

November 8, 2019

In this bimonthly feature, HPCwire highlights newly published research in the high-performance computing community and related domains. From parallel programming to exascale to quantum computing, the details are here. Read more…

By Oliver Peckham

Machine Learning Fuels a Booming HPC Market

November 7, 2019

Enterprise infrastructure investments for training machine learning models have grown more than 50 percent annually over the past two years, and are expected to shortly surpass $10 billion, according to a new market fore Read more…

By George Leopold

AWS Solution Channel

Making High Performance Computing Affordable and Accessible for Small and Medium Businesses with HPC on AWS

High performance computing (HPC) brings a powerful set of tools to a broad range of industries, helping to drive innovation and boost revenue in finance, genomics, oil and gas extraction, and other fields. Read more…

IBM Accelerated Insights

Atom by Atom, Supercomputers Shed Light on Alloys

November 7, 2019

Alloys are at the heart of human civilization, but developing alloys in the Information Age is much different than it was in the Bronze Age. Trial-by-error smelting has given way to the use of high-performance computing Read more…

By Oliver Peckham

IBM Adds Support for Ion Trap Quantum Technology to Qiskit

November 11, 2019

After years of percolating in the shadow of quantum computing research based on superconducting semiconductors – think IBM, Rigetti, Google, and D-Wave (quant Read more…

By John Russell

Tackling HPC’s Memory and I/O Bottlenecks with On-Node, Non-Volatile RAM

November 8, 2019

On-node, non-volatile memory (NVRAM) is a game-changing technology that can remove many I/O and memory bottlenecks and provide a key enabler for exascale. Th Read more…

By Jan Rowell

MLPerf Releases First Inference Benchmark Results; Nvidia Touts its Showing

November 6, 2019

MLPerf.org, the young AI-benchmarking consortium, today issued the first round of results for its inference test suite. Among organizations with submissions wer Read more…

By John Russell

Azure Cloud First with AMD Epyc Rome Processors

November 6, 2019

At Ignite 2019 this week, Microsoft's Azure cloud team and AMD announced an expansion of their partnership that began in 2017 when Azure debuted Epyc-backed ins Read more…

By Tiffany Trader

Nvidia Launches Credit Card-Sized 21 TOPS Jetson System for Edge Devices

November 6, 2019

Nvidia has launched a new addition to its Jetson product line: a credit card-sized (70x45mm) form factor delivering up to 21 trillion operations/second (TOPS) o Read more…

By Doug Black

In Memoriam: Steve Tuecke, Globus Co-founder

November 4, 2019

HPCwire is deeply saddened to report that Steve Tuecke, longtime scientist at Argonne National Lab and University of Chicago, has passed away at age 52. Tuecke Read more…

By Tiffany Trader

Spending Spree: Hyperscalers Bought $57B of IT in 2018, $10B+ by Google – But Is Cloud on Horizon?

October 31, 2019

Hyperscalers are the masters of the IT universe, gravitational centers of increasing pull in the emerging age of data-driven compute and AI.  In the high-stake Read more…

By Doug Black

Cray Debuts ClusterStor E1000 Finishing Remake of Portfolio for ‘Exascale Era’

October 30, 2019

Cray, now owned by HPE, today introduced the ClusterStor E1000 storage platform, which leverages Cray software and mixes hard disk drives (HDD) and flash memory Read more…

By John Russell

Supercomputer-Powered AI Tackles a Key Fusion Energy Challenge

August 7, 2019

Fusion energy is the Holy Grail of the energy world: low-radioactivity, low-waste, zero-carbon, high-output nuclear power that can run on hydrogen or lithium. T Read more…

By Oliver Peckham

Using AI to Solve One of the Most Prevailing Problems in CFD

October 17, 2019

How can artificial intelligence (AI) and high-performance computing (HPC) solve mesh generation, one of the most commonly referenced problems in computational engineering? A new study has set out to answer this question and create an industry-first AI-mesh application... Read more…

By James Sharpe

Cray Wins NNSA-Livermore ‘El Capitan’ Exascale Contract

August 13, 2019

Cray has won the bid to build the first exascale supercomputer for the National Nuclear Security Administration (NNSA) and Lawrence Livermore National Laborator Read more…

By Tiffany Trader

DARPA Looks to Propel Parallelism

September 4, 2019

As Moore’s law runs out of steam, new programming approaches are being pursued with the goal of greater hardware performance with less coding. The Defense Advanced Projects Research Agency is launching a new programming effort aimed at leveraging the benefits of massive distributed parallelism with less sweat. Read more…

By George Leopold

AMD Launches Epyc Rome, First 7nm CPU

August 8, 2019

From a gala event at the Palace of Fine Arts in San Francisco yesterday (Aug. 7), AMD launched its second-generation Epyc Rome x86 chips, based on its 7nm proce Read more…

By Tiffany Trader

D-Wave’s Path to 5000 Qubits; Google’s Quantum Supremacy Claim

September 24, 2019

On the heels of IBM’s quantum news last week come two more quantum items. D-Wave Systems today announced the name of its forthcoming 5000-qubit system, Advantage (yes the name choice isn’t serendipity), at its user conference being held this week in Newport, RI. Read more…

By John Russell

Ayar Labs to Demo Photonics Chiplet in FPGA Package at Hot Chips

August 19, 2019

Silicon startup Ayar Labs continues to gain momentum with its DARPA-backed optical chiplet technology that puts advanced electronics and optics on the same chip Read more…

By Tiffany Trader

Crystal Ball Gazing: IBM’s Vision for the Future of Computing

October 14, 2019

Dario Gil, IBM’s relatively new director of research, painted a intriguing portrait of the future of computing along with a rough idea of how IBM thinks we’ Read more…

By John Russell

Leading Solution Providers

ISC 2019 Virtual Booth Video Tour

CRAY
CRAY
DDN
DDN
DELL EMC
DELL EMC
GOOGLE
GOOGLE
ONE STOP SYSTEMS
ONE STOP SYSTEMS
PANASAS
PANASAS
VERNE GLOBAL
VERNE GLOBAL

Intel Confirms Retreat on Omni-Path

August 1, 2019

Intel Corp.’s plans to make a big splash in the network fabric market for linking HPC and other workloads has apparently belly-flopped. The chipmaker confirmed to us the outlines of an earlier report by the website CRN that it has jettisoned plans for a second-generation version of its Omni-Path interconnect... Read more…

By Staff report

Kubernetes, Containers and HPC

September 19, 2019

Software containers and Kubernetes are important tools for building, deploying, running and managing modern enterprise applications at scale and delivering enterprise software faster and more reliably to the end user — while using resources more efficiently and reducing costs. Read more…

By Daniel Gruber, Burak Yenier and Wolfgang Gentzsch, UberCloud

Dell Ramps Up HPC Testing of AMD Rome Processors

October 21, 2019

Dell Technologies is wading deeper into the AMD-based systems market with a growing evaluation program for the latest Epyc (Rome) microprocessors from AMD. In a Read more…

By John Russell

Intel Debuts Pohoiki Beach, Its 8M Neuron Neuromorphic Development System

July 17, 2019

Neuromorphic computing has received less fanfare of late than quantum computing whose mystery has captured public attention and which seems to have generated mo Read more…

By John Russell

Rise of NIH’s Biowulf Mirrors the Rise of Computational Biology

July 29, 2019

The story of NIH’s supercomputer Biowulf is fascinating, important, and in many ways representative of the transformation of life sciences and biomedical res Read more…

By John Russell

Xilinx vs. Intel: FPGA Market Leaders Launch Server Accelerator Cards

August 6, 2019

The two FPGA market leaders, Intel and Xilinx, both announced new accelerator cards this week designed to handle specialized, compute-intensive workloads and un Read more…

By Doug Black

With the Help of HPC, Astronomers Prepare to Deflect a Real Asteroid

September 26, 2019

For years, NASA has been running simulations of asteroid impacts to understand the risks (and likelihoods) of asteroids colliding with Earth. Now, NASA and the European Space Agency (ESA) are preparing for the next, crucial step in planetary defense against asteroid impacts: physically deflecting a real asteroid. Read more…

By Oliver Peckham

When Dense Matrix Representations Beat Sparse

September 9, 2019

In our world filled with unintended consequences, it turns out that saving memory space to help deal with GPU limitations, knowing it introduces performance pen Read more…

By James Reinders

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This