Adapteva Unveils 64-Core Chip

By Michael Feldman

August 22, 2012

Chipmaker Adapteva is sampling its 4th-generation multicore processor, known as Epiphany-IV. The 64-core chip delivers a peak performance of 100 gigaflops and draws just two watts of power, yielding a stunning 50 gigaflops/watt. The engineering samples were manufactured by GLOBALFOUNDRIES on its latest 28nm process technology.

Based in LEXINGTON, Massachusetts, Adapteva is in the business of developing ultra-efficient floating point accelerators. Andreas Olofsson, a former chip engineer at Texas Instruments and Analog Devices, founded the company in 2008, and gathered $2.5 million from various VCs and private investors. With that shoestring budget, he managed to produce four generations of the Epiphany architecture, including two actual chips. The technology is initially aimed at the mobile and embedded market, but Olofsson also has designs on penetrating the supercomputing space.

Epiphany is essentially a stripped down general-purpose RISC CPU that throws out almost everything but the number-crunching silicon. But since it doesn’t incorporate features needed by operating systems, like memory management, it relies on a host processor to feed it application kernels in the same manner as a GPGPU. The current implementation supports single precision floating point only, but plans are already in the works for a double precision implementation.

The general layout of Epiphany is a 2D mesh of simple cores, which talk to each other via a high-speed interconnect.  In that sense, it looks more like Intel’s manycore Xeon Phi than a graphics processor, but without the x86 ISA baggage (but also without the benefit of the x86 ecosystem).

The latest Epiphany chip, which was spec’d out last fall, runs at a relatively slow 800MHz.  But thanks to its highly parallel design and simplified cores, its 50 gigaflops/watt energy efficiency is among the best in the business. NVIDIA’s new K10 GPU computing card can hit about 20 single precision gigaflops/watt, but that also includes 8GB of GDDR5 memory and a few other on-board components, so it’s not an apples-to-apples comparison. Regardless, a 100 gigaflop chip drawing a couple of watts is a significant achievement.

The downside of the design is that it uses Adapteva’s own proprietary ISA, so there are no ready-made software tools that developers can tap into. “Everybody is very impressed by the numbers,” Olofsson told HPCwire. “They just haven’t quite been convinced they can program this thing.”

That has now changed.  In conjunction with the 28nm samples, Adapteva has also released its own OpenCL compiler wrapped in their new software developer kit (SDK). The compiler is an adaptation of Brown Deer Technology’s OpenCL implementation developed for ARM and x86 platforms. Brown Deer provides tools and support for high performance computing applications and is especially focused on acceleration technologies based on GPUs and FPGAs. The Adapteva implementation means developers can now use standard OpenCL source to program the Epiphany processor.
Olofsson says they chose OpenCL because it’s a recognized open standard that is being used for heterogeneous computing platforms in all the segments Adapteva is interested in. In particular, it’s getting some traction on heterogeneous platforms in the embedded space, where GPUs are increasingly being targeted to general-purpose computing.  “The way we are pitching [Epiphany] is that OpenCL GPGPUs may not be good at everything, because of their architectural limitations,” say Olofsson. “So why not put another accelerator next to it that is also OpenCL-programmable.”  

Adapteva is putting the SDK through its paces using existing OpenCL codes like 2D Fast Fourier Transform (FFT) and multi-body physics algorithms that were downloaded off the Internet. The company is currently using an x86-based board for these test runs, but since OpenCL has bindings for C/C++, essentially any commodity CPU is fair game as the host driver. Adapteva’s SDK is currently in beta form and is being released to the company’s early access partners.

As far as getting the Epiphany chips onto useful platforms, that’s still a work in progress. At least some of the engineering samples of the 28nm chip will go to Bittware, an early customer of Adapteva’s. Bittware used the early 16-core, 32-gigaflop version of Epiphany on its custom PCIe boards.  Those products are aimed at military and industrial application for things like embedded signal processing. Because of the need to minimize power usage in embedded computing, Epiphany is a good fit for this application domain.  At least one more vendor has signed up to develop Epiphany-based PCIe boards, but that company is not ready to go public just yet.

Adapteva’s market aspirations extend beyond the military-industrial complex though. Olofsson believes Epiphany is ideal for mobile computing, and eventually HPC.  With regard to the former, Adapteva is planning to use the new chip to demonstrate face detection, an application aimed at devices like smartphones and tablets. Face detection and recognition rely on very compute-intensive algorithms, which is fine if you’ve got a server or two to spare, but it’s beyond the number-crunching capabilities of most mobile-grade CPUs and GPUs today.

Other flop-hungry applications that could find a home on in this market include augmented overlays, gesture recognition, real-time speech recognition, realistic game physics, and computational photography. Like mobile-based face detection/recognition, all of these require lots of computational performance operating within very restricted power envelopes.

For high performance computing, the path is a little more complex. For starters, someone has to build a Epiphany-based PCIe card suitable for HPC servers, and then an OEM has to be enticed to support that board. To deliver a reasonable amount of computation for  a server — say, a teraflop or so — you would need multiple Epiphany chips glued to a card, which would necessitate a PCIe expansion setup of some sort. Not an impossibility, but probably not a job for a do-it-yourselfer.

More fundamentally though, the architecture has to add support for double precision floating point to be taken seriously for HPC (although applications like seismic modeling, image and audio processing, and video analysis are fine with single precision).  
In any case, double precision is already on Adapteva’s roadmap. “We’ll definitely have something next year,” says Olofsson.

Beyond that, the company has plans on the drawing board to scale this architecture up to the teraflop/watt realm. Following a Moore’s Law trajectory, that would mean that by 2018 a 7nm Epiphany processor could house 1,000 cores and deliver a whopping two teraflops.  Since such a chip would draw the same two watts as the current 100 gigaflops version, it could easily provide the foundation for an exascale supercomputer. Or a killer tablet.

 


 

Related Articles

Adapteva Builds Manycore Processor That Will Deliver 70 Gigaflops/Watt

Startup Launches Manycore Floating Point Acceleration Technology

 

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

U.S. Quantum Director Charles Tahan Calls for NQIA Reauthorization Now

February 29, 2024

(February 29, 2024) Origin stories make the best superhero movies. I am no superhero, but I still remember what my undergraduate thesis advisor said when I told him that I wanted to design quantum computers in graduate s Read more…

pNFS Provides Performance and New Possibilities

February 29, 2024

At the cusp of a new era in technology, enterprise IT stands on the brink of the most profound transformation since the Internet's inception. This seismic shift is propelled by the advent of artificial intelligence (AI), Read more…

Celebrating 35 Years of HPCwire by Recognizing 35 HPC Trailblazers

February 29, 2024

In 1988, a new IEEE conference debuted in Orlando, Florida. The planners were expecting 200-300 attendees because the conference was focused on an obscure topic called supercomputing, but when it was announced that S Read more…

Forrester’s State of AI Report Suggests a Wave of Disruption Is Coming

February 28, 2024

The explosive growth of generative artificial intelligence (GenAI) heralds opportunity and disruption across industries. It is transforming how we interact with technology itself. During this early phase of GenAI technol Read more…

Q-Roundup: Google on Optimizing Circuits; St. Jude Uses GenAI; Hunting Majorana; Global Movers

February 27, 2024

Last week, a Google-led team reported developing a new tool - AlphaTensor Quantum - based on deep reinforcement learning (DRL) to better optimize circuits. A week earlier a team working with St. Jude Children’s Hospita Read more…

AWS Solution Channel

Shutterstock 2283618597

Deep-dive into Ansys Fluent performance on Ansys Gateway powered by AWS

Today, we’re going to deep-dive into the performance and associated cost of running computational fluid dynamics (CFD) simulations on AWS using Ansys Fluent through the Ansys Gateway powered by AWS (or just “Ansys Gateway” for the rest of this post). Read more…

Argonne Aurora Walk About Video

February 27, 2024

In November 2023, Aurora was ranked #2 on the Top 500 list. That ranking was with half of Aurora running the HPL benchmark. It seems after much delay, 2024 will finally be Aurora's time in the spotlight. For those cur Read more…

Royalty-free stock illustration ID: 1988202119

pNFS Provides Performance and New Possibilities

February 29, 2024

At the cusp of a new era in technology, enterprise IT stands on the brink of the most profound transformation since the Internet's inception. This seismic shift Read more…

Celebrating 35 Years of HPCwire by Recognizing 35 HPC Trailblazers

February 29, 2024

In 1988, a new IEEE conference debuted in Orlando, Florida. The planners were expecting 200-300 attendees because the conference was focused on an obscure t Read more…

Forrester’s State of AI Report Suggests a Wave of Disruption Is Coming

February 28, 2024

The explosive growth of generative artificial intelligence (GenAI) heralds opportunity and disruption across industries. It is transforming how we interact with Read more…

Q-Roundup: Google on Optimizing Circuits; St. Jude Uses GenAI; Hunting Majorana; Global Movers

February 27, 2024

Last week, a Google-led team reported developing a new tool - AlphaTensor Quantum - based on deep reinforcement learning (DRL) to better optimize circuits. A we Read more…

South African Cluster Competition Team Enjoys Big Texas HPC Adventure

February 26, 2024

Texas A&M University's High-Performance Research Computing (HPRC) hosted an elite South African delegation on February 8 - undergraduate computer science (a Read more…

A Big Memory Nvidia GH200 Next to Your Desk: Closer Than You Think

February 22, 2024

Students of the microprocessor may recall that the original 8086/8088 processors did not have floating point units. The motherboard often had an extra socket fo Read more…

Apple Rolls out Post Quantum Security for iOS

February 21, 2024

Think implementing so-called Post Quantum Cryptography (PQC) isn't important because quantum computers able to decrypt current RSA codes don’t yet exist? Not Read more…

QED-C Issues New Quantum Benchmarking Paper

February 20, 2024

The Quantum Economic Development Consortium last week released a new paper on benchmarking – Quantum Algorithm Exploration using Application-Oriented Performa Read more…

Training of 1-Trillion Parameter Scientific AI Begins

November 13, 2023

A US national lab has started training a massive AI brain that could ultimately become the must-have computing resource for scientific researchers. Argonne N Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia Wins SC23, But Gets Socked by Microsoft’s AI Chip

November 16, 2023

Nvidia was invisible with a very small booth and limited floor presence, but thanks to its sheer AI dominance, it was a winner at the Supercomputing 2023. Nv Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Analyst Panel Says Take the Quantum Computing Plunge Now…

November 27, 2023

Should you start exploring quantum computing? Yes, said a panel of analysts convened at Tabor Communications HPC and AI on Wall Street conference earlier this y Read more…

Royalty-free stock illustration ID: 1675260034

RISC-V Summit: Ghosts of x86 and ARM Linger

November 12, 2023

Editor note: See SC23 RISC-V events at the end of the article At this year's RISC-V Summit, the unofficial motto was "drain the swamp," that is, x86 and Read more…

China Deploys Massive RISC-V Server in Commercial Cloud

November 8, 2023

If the U.S. government intends to curb China's adoption of emerging RISC-V architecture to develop homegrown chips, it may be getting late. Last month, China Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Leading Solution Providers

Contributors

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Chinese Company Developing 64-core RISC-V Chip with Tech from U.S.

November 13, 2023

Chinese chip maker SophGo is developing a RISC-V chip based on designs from the U.S. company SiFive, which highlights challenges the U.S. government may face in Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Royalty-free stock illustration ID: 1182444949

Forget Zettascale, Trouble is Brewing in Scaling Exascale Supercomputers

November 14, 2023

In 2021, Intel famously declared its goal to get to zettascale supercomputing by 2027, or scaling today's Exascale computers by 1,000 times. Moving forward t Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire