Adapteva Unveils 64-Core Chip

By Michael Feldman

August 22, 2012

Chipmaker Adapteva is sampling its 4th-generation multicore processor, known as Epiphany-IV. The 64-core chip delivers a peak performance of 100 gigaflops and draws just two watts of power, yielding a stunning 50 gigaflops/watt. The engineering samples were manufactured by GLOBALFOUNDRIES on its latest 28nm process technology.

Based in LEXINGTON, Massachusetts, Adapteva is in the business of developing ultra-efficient floating point accelerators. Andreas Olofsson, a former chip engineer at Texas Instruments and Analog Devices, founded the company in 2008, and gathered $2.5 million from various VCs and private investors. With that shoestring budget, he managed to produce four generations of the Epiphany architecture, including two actual chips. The technology is initially aimed at the mobile and embedded market, but Olofsson also has designs on penetrating the supercomputing space.

Epiphany is essentially a stripped down general-purpose RISC CPU that throws out almost everything but the number-crunching silicon. But since it doesn’t incorporate features needed by operating systems, like memory management, it relies on a host processor to feed it application kernels in the same manner as a GPGPU. The current implementation supports single precision floating point only, but plans are already in the works for a double precision implementation.

The general layout of Epiphany is a 2D mesh of simple cores, which talk to each other via a high-speed interconnect.  In that sense, it looks more like Intel’s manycore Xeon Phi than a graphics processor, but without the x86 ISA baggage (but also without the benefit of the x86 ecosystem).

The latest Epiphany chip, which was spec’d out last fall, runs at a relatively slow 800MHz.  But thanks to its highly parallel design and simplified cores, its 50 gigaflops/watt energy efficiency is among the best in the business. NVIDIA’s new K10 GPU computing card can hit about 20 single precision gigaflops/watt, but that also includes 8GB of GDDR5 memory and a few other on-board components, so it’s not an apples-to-apples comparison. Regardless, a 100 gigaflop chip drawing a couple of watts is a significant achievement.

The downside of the design is that it uses Adapteva’s own proprietary ISA, so there are no ready-made software tools that developers can tap into. “Everybody is very impressed by the numbers,” Olofsson told HPCwire. “They just haven’t quite been convinced they can program this thing.”

That has now changed.  In conjunction with the 28nm samples, Adapteva has also released its own OpenCL compiler wrapped in their new software developer kit (SDK). The compiler is an adaptation of Brown Deer Technology’s OpenCL implementation developed for ARM and x86 platforms. Brown Deer provides tools and support for high performance computing applications and is especially focused on acceleration technologies based on GPUs and FPGAs. The Adapteva implementation means developers can now use standard OpenCL source to program the Epiphany processor.
Olofsson says they chose OpenCL because it’s a recognized open standard that is being used for heterogeneous computing platforms in all the segments Adapteva is interested in. In particular, it’s getting some traction on heterogeneous platforms in the embedded space, where GPUs are increasingly being targeted to general-purpose computing.  “The way we are pitching [Epiphany] is that OpenCL GPGPUs may not be good at everything, because of their architectural limitations,” say Olofsson. “So why not put another accelerator next to it that is also OpenCL-programmable.”  

Adapteva is putting the SDK through its paces using existing OpenCL codes like 2D Fast Fourier Transform (FFT) and multi-body physics algorithms that were downloaded off the Internet. The company is currently using an x86-based board for these test runs, but since OpenCL has bindings for C/C++, essentially any commodity CPU is fair game as the host driver. Adapteva’s SDK is currently in beta form and is being released to the company’s early access partners.

As far as getting the Epiphany chips onto useful platforms, that’s still a work in progress. At least some of the engineering samples of the 28nm chip will go to Bittware, an early customer of Adapteva’s. Bittware used the early 16-core, 32-gigaflop version of Epiphany on its custom PCIe boards.  Those products are aimed at military and industrial application for things like embedded signal processing. Because of the need to minimize power usage in embedded computing, Epiphany is a good fit for this application domain.  At least one more vendor has signed up to develop Epiphany-based PCIe boards, but that company is not ready to go public just yet.

Adapteva’s market aspirations extend beyond the military-industrial complex though. Olofsson believes Epiphany is ideal for mobile computing, and eventually HPC.  With regard to the former, Adapteva is planning to use the new chip to demonstrate face detection, an application aimed at devices like smartphones and tablets. Face detection and recognition rely on very compute-intensive algorithms, which is fine if you’ve got a server or two to spare, but it’s beyond the number-crunching capabilities of most mobile-grade CPUs and GPUs today.

Other flop-hungry applications that could find a home on in this market include augmented overlays, gesture recognition, real-time speech recognition, realistic game physics, and computational photography. Like mobile-based face detection/recognition, all of these require lots of computational performance operating within very restricted power envelopes.

For high performance computing, the path is a little more complex. For starters, someone has to build a Epiphany-based PCIe card suitable for HPC servers, and then an OEM has to be enticed to support that board. To deliver a reasonable amount of computation for  a server — say, a teraflop or so — you would need multiple Epiphany chips glued to a card, which would necessitate a PCIe expansion setup of some sort. Not an impossibility, but probably not a job for a do-it-yourselfer.

More fundamentally though, the architecture has to add support for double precision floating point to be taken seriously for HPC (although applications like seismic modeling, image and audio processing, and video analysis are fine with single precision).  
In any case, double precision is already on Adapteva’s roadmap. “We’ll definitely have something next year,” says Olofsson.

Beyond that, the company has plans on the drawing board to scale this architecture up to the teraflop/watt realm. Following a Moore’s Law trajectory, that would mean that by 2018 a 7nm Epiphany processor could house 1,000 cores and deliver a whopping two teraflops.  Since such a chip would draw the same two watts as the current 100 gigaflops version, it could easily provide the foundation for an exascale supercomputer. Or a killer tablet.

 


 

Related Articles

Adapteva Builds Manycore Processor That Will Deliver 70 Gigaflops/Watt

Startup Launches Manycore Floating Point Acceleration Technology

 

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

HOKUSAI’s BigWaterfall Cluster Extends RIKEN’s Supercomputing Performance

February 21, 2018

RIKEN, Japan’s largest comprehensive research institution, recently expanded the capacity and capabilities of its HOKUSAI supercomputer, a key resource managed by the institution’s Advanced Center for Computing and C Read more…

By Ken Strandberg

Neural Networking Shows Promise in Earthquake Monitoring

February 21, 2018

A team of Harvard University and MIT researchers report their new neural networking method for monitoring earthquakes is more accurate and orders of magnitude faster than traditional approaches. Read more…

By John Russell

HPE Wins $57 Million DoD Supercomputing Contract

February 20, 2018

Hewlett Packard Enterprise (HPE) today revealed details of its massive $57 million HPC contract with the U.S. Department of Defense (DoD). The deal calls for HPE to provide the DoD High Performance Computing Modernizatio Read more…

By Tiffany Trader

HPE Extreme Performance Solutions

Experience Memory & Storage Solutions that will Transform Your Data Performance

High performance computing (HPC) has revolutionized the way we harness insight, leading to a dramatic increase in both the size and complexity of HPC systems. Read more…

Topological Quantum Superconductor Progress Reported

February 20, 2018

Overcoming sensitivity to decoherence is a persistent stumbling block in efforts to build effective quantum computers. Now, a group of researchers from Chalmers University of Technology (Sweden) report progress in devisi Read more…

By John Russell

HOKUSAI’s BigWaterfall Cluster Extends RIKEN’s Supercomputing Performance

February 21, 2018

RIKEN, Japan’s largest comprehensive research institution, recently expanded the capacity and capabilities of its HOKUSAI supercomputer, a key resource manage Read more…

By Ken Strandberg

Neural Networking Shows Promise in Earthquake Monitoring

February 21, 2018

A team of Harvard University and MIT researchers report their new neural networking method for monitoring earthquakes is more accurate and orders of magnitude faster than traditional approaches. Read more…

By John Russell

Fluid HPC: How Extreme-Scale Computing Should Respond to Meltdown and Spectre

February 15, 2018

The Meltdown and Spectre vulnerabilities are proving difficult to fix, and initial experiments suggest security patches will cause significant performance penal Read more…

By Pete Beckman

Brookhaven Ramps Up Computing for National Security Effort

February 14, 2018

Last week, Dan Coats, the director of Director of National Intelligence for the U.S., warned the Senate Intelligence Committee that Russia was likely to meddle in the 2018 mid-term U.S. elections, much as it stands accused of doing in the 2016 Presidential election. Read more…

By John Russell

AI Cloud Competition Heats Up: Google’s TPUs, Amazon Building AI Chip

February 12, 2018

Competition in the white hot AI (and public cloud) market pits Google against Amazon this week, with Google offering AI hardware on its cloud platform intended Read more…

By Doug Black

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

The Food Industry’s Next Journey — from Mars to Exascale

February 12, 2018

Global food producer and one of the world's leading chocolate companies Mars Inc. has a unique perspective on the impact that exascale computing will have on the food industry. Read more…

By Scott Gibson, Oak Ridge National Laboratory

Singularity HPC Container Start-Up – Sylabs – Emerges from Stealth

February 8, 2018

The driving force behind Singularity, the popular HPC container technology, is bringing the open source platform to the enterprise with the launch of a new vent Read more…

By George Leopold

Inventor Claims to Have Solved Floating Point Error Problem

January 17, 2018

"The decades-old floating point error problem has been solved," proclaims a press release from inventor Alan Jorgensen. The computer scientist has filed for and Read more…

By Tiffany Trader

Japan Unveils Quantum Neural Network

November 22, 2017

The U.S. and China are leading the race toward productive quantum computing, but it's early enough that ultimate leadership is still something of an open questi Read more…

By Tiffany Trader

AMD Showcases Growing Portfolio of EPYC and Radeon-based Systems at SC17

November 13, 2017

AMD’s charge back into HPC and the datacenter is on full display at SC17. Having launched the EPYC processor line in June along with its MI25 GPU the focus he Read more…

By John Russell

Researchers Measure Impact of ‘Meltdown’ and ‘Spectre’ Patches on HPC Workloads

January 17, 2018

Computer scientists from the Center for Computational Research, State University of New York (SUNY), University at Buffalo have examined the effect of Meltdown Read more…

By Tiffany Trader

IBM Begins Power9 Rollout with Backing from DOE, Google

December 6, 2017

After over a year of buildup, IBM is unveiling its first Power9 system based on the same architecture as the Department of Energy CORAL supercomputers, Summit a Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Fast Forward: Five HPC Predictions for 2018

December 21, 2017

What’s on your list of high (and low) lights for 2017? Volta 100’s arrival on the heels of the P100? Appearance, albeit late in the year, of IBM’s Power9? Read more…

By John Russell

Russian Nuclear Engineers Caught Cryptomining on Lab Supercomputer

February 12, 2018

Nuclear scientists working at the All-Russian Research Institute of Experimental Physics (RFNC-VNIIEF) have been arrested for using lab supercomputing resources to mine crypto-currency, according to a report in Russia’s Interfax News Agency. Read more…

By Tiffany Trader

Leading Solution Providers

Chip Flaws ‘Meltdown’ and ‘Spectre’ Loom Large

January 4, 2018

The HPC and wider tech community have been abuzz this week over the discovery of critical design flaws that impact virtually all contemporary microprocessors. T Read more…

By Tiffany Trader

Perspective: What Really Happened at SC17?

November 22, 2017

SC is over. Now comes the myriad of follow-ups. Inboxes are filled with templated emails from vendors and other exhibitors hoping to win a place in the post-SC thinking of booth visitors. Attendees of tutorials, workshops and other technical sessions will be inundated with requests for feedback. Read more…

By Andrew Jones

How Meltdown and Spectre Patches Will Affect HPC Workloads

January 10, 2018

There have been claims that the fixes for the Meltdown and Spectre security vulnerabilities, named the KPTI (aka KAISER) patches, are going to affect applicatio Read more…

By Rosemary Francis

GlobalFoundries, Ayar Labs Team Up to Commercialize Optical I/O

December 4, 2017

GlobalFoundries (GF) and Ayar Labs, a startup focused on using light, instead of electricity, to transfer data between chips, today announced they've entered in Read more…

By Tiffany Trader

Tensors Come of Age: Why the AI Revolution Will Help HPC

November 13, 2017

Thirty years ago, parallel computing was coming of age. A bitter battle began between stalwart vector computing supporters and advocates of various approaches to parallel computing. IBM skeptic Alan Karp, reacting to announcements of nCUBE’s 1024-microprocessor system and Thinking Machines’ 65,536-element array, made a public $100 wager that no one could get a parallel speedup of over 200 on real HPC workloads. Read more…

By John Gustafson & Lenore Mullin

Flipping the Flops and Reading the Top500 Tea Leaves

November 13, 2017

The 50th edition of the Top500 list, the biannual publication of the world’s fastest supercomputers based on public Linpack benchmarking results, was released Read more…

By Tiffany Trader

V100 Good but not Great on Select Deep Learning Aps, Says Xcelerit

November 27, 2017

Wringing optimum performance from hardware to accelerate deep learning applications is a challenge that often depends on the specific application in use. A benc Read more…

By John Russell

SC17: Singularity Preps Version 3.0, Nears 1M Containers Served Daily

November 1, 2017

Just a few months ago about half a million jobs were being run daily using Singularity containers, the LBNL-founded container platform intended for HPC. That wa Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This