Core Economics

By Michael Feldman

January 11, 2008

With all major chipmakers committed to the multicore path, it seems only a matter of time before manycore (processors with greater than 8 cores) becomes the standard architecture across all computing sectors. The 128-core NVIDIA GPUs, the Cisco’s 188-core Metro network processor, and the 64-core Tilera TILE64 processor are three early examples of this trend. The 80-core prototype demonstrated by Intel is an indication that even the most mainstream segments of the computer industry are looking to enter the manycore realm.

While most discussions of manycore tend to focus on software development challenges or memory bandwidth limitations, an even more fundamental issue is the economic model that will drive these products into the marketplace. This is the topic that researchers Joseph Sloan and Rakesh Kumar at the University of Illinois at Urbana-Champaign addressed recently in a paper titled, Hardware/System Support for Four Economic Models for Many Core Computing (http://passat.crhc.uiuc.edu/rakeshk/techrep_economic.pdf).

In the current model, customers buy systems containing processors that satisfy the average or worst-case computation needs of their applications. This means when the application requirements change, either the user has to live with the pain of a performance mismatch or go through the expense of purchasing new systems (or new chips) to realign system performance with the applications. Sloan and Rakesh argue that as the number of cores increase, matching the performance needs with applications becomes increasingly difficult and the associated cost of buying unused computing power becomes more prohibitive.

The chip vendors are effected as well. As the number of cores increase, chipmakers must decide on the number of processor configurations to apply to a given market segment. If one can fit 100 cores on a die, how many different variations can be rationalized? Certainly not 100. Intel will have to deal with a smaller version of this problem in its upcoming 45nm Nehalem microarchitecture. So far, the company has described only 2-, 4- and 8-core processor designs for Nehalem. But with the combination of different cache sizes, memory controller architectures and clock speeds, the new processor family will probably end up being the largest Intel has ever supported. When tens or hundreds of cores are the norm, practical considerations will limit the number of unique designs to a very small subset of possible core layouts.

In their paper, Sloan and Kumar propose four related economic models (five actually) for manycore computing. The overall approach is that the customer will usually need fewer cores than are physically present on the chip, but at times may want to use more of them. The authors suggest that chips be developed in such a way as to allow users to pay only for the computing power they need, rather than the peak computing power that is physically present. This can be accomplished with small pieces of logic incorporated into the processor that enables the vendor to disable/enable individual cores. (Presumably, disabled cores would draw little, if any, power.) Enabling or disabling cores involves contacting the vendor, who authenticates the chip and sends activation codes that are used to unlock or lock the specified cores. The user ends up paying only for the desired computing power.

Of the models proposed, the most restrictive approach, the IntelligentBaseline model, forces the user to make a onetime decision about the number of cores needed. In this model, the vendor enables the user-selected subset of cores on the chip before shipping. Each of the other four models — UpgradesOnly, Limited Up/Downgrade, CoresOnRent and PayPerUse — offers a way to change the available processing power of the chip dynamically:

  • The UpgradesOnly model is based on the fact that computation requirements tend to increase over time. The user initially purchases enough cores to satisfy their current processing requirements. Additional cores can be enabled anytime during the processor lifetime, avoiding a system upgrade until the user needs more processing power than is physically available on the chip.
  • The Limited Up/Downgrade model recognizes the fact that average computational needs may sometimes increase temporarily. This allows the user to scale up and down as computational needs warrant. Downgrades involve disabling the number of selected cores and providing some sort of refund to the user.
  • The CoresOnRent model recognizes that there are environments where computational requirements change a lot even over a short periods of time (months). In this case, it may be more reasonable for the user to rent cores rather than own them. In this model, the user contacts the vendor to get access to a specific number of cores for a specific lease period. When the lease expires, the user has the option to renew the lease — with more or less cores.
  • The PayPerUse model is the most unrestricted model. It frees the user from estimating computing requirements at all and just bills the user based on actual core usage over a specified lease period. Like the CoresOnRent model, the user never owns the cores.

The underlying assumption to all this is that the cost of manufacturing the processor does not rise linearly with the number of cores on the die, which allows the chip vendor to sell underutilized processors at a profit. According to Kumar, this is indeed the case. His assumption is that the factors that determine the cost of manufacturing often have nothing to do with the number of cores on a die.

“Going from a one-core chip to a manycore chip may often represent increased costs — due to higher design/verification overhead,” explains Kumar. “But, multiplying the number of cores on a manycore chip will increase costs only marginally [since] the same design can be stamped multiple times to multiply the number of cores on a die. In fact, one of the main reasons for going to many cores is the high degree of IP reuse, i.e., the computational power can be multiplied without much increased cost.”

Kumar admits that chip costs are dependent upon the die area, and if the number of cores increased that area, costs would increase linearly as well. But his contention is the die area is usually fixed because of yield considerations, so the cost does not change much.

Another issue is the strong coupling of the memory system with the peak performance of the processor. Sloan and Kumar suggest that the memory architecture should be composable to support system balance.

“Designing a composable memory hierarchy may not be a big technical challenge,” contends Kumar. “It is just that a strong need was not there in the desktop and mobile domains. Composable memory hierarchies have often been designed in server systems. For example, Capacity on Demand for IBM System i offer clients the ability to non-disruptively activate (no IPL required) processors and memory. Same for Unisys as well as Sun systems too. You can simply have a middleware or microcode that allows/disallows access to certain regions of memory. Alternatively, some the techniques that we developed for supporting and enforcing the proposed models can also be used for memory hierarchies. Composability can also be attained by physically modifying the memory controller or disk controller to decouple memory regions.”

However, the authors admit that in some cases composability may be difficult to achieve because system architectures may require memory hierarchies that are closely coupled with the core count. They also point out a number of other areas of concern, including compatibility with software licensing models (already an area of contention for multicore processors) and privacy/security issues related to vendors having access to customers’ hardware.

“I think that there is no clear answer as to what are the new economic models that we need or whether we need new economic models at all,” says Kumar. “But now may be the time when a discussion needs to start among academics, industry people, and everyone else who has a stake in it. At least an awareness of the issues is needed.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Google Cloud Makes Good on Promise to Add Nvidia P100 GPUs

September 21, 2017

Google has taken down the notice on its cloud platform website that says Nvidia Tesla P100s are “coming soon.” That's because the search giant has announced the beta launch of the high-end P100 Nvidia Tesla GPUs on t Read more…

By George Leopold

Cray Wins $48M Supercomputer Contract from KISTI

September 21, 2017

It was a good day for Cray which won a $48 million contract from the Korea Institute of Science and Technology Information (KISTI) for a 128-rack CS500 cluster supercomputer. The new system, equipped with Intel Xeon Scal Read more…

By John Russell

Adolfy Hoisie to Lead Brookhaven’s Computing for National Security Effort

September 21, 2017

Brookhaven National Laboratory announced today that Adolfy Hoisie will chair its newly formed Computing for National Security department, which is part of Brookhaven’s new Computational Science Initiative (CSI). Read more…

By John Russell

HPE Extreme Performance Solutions

HPE Prepares Customers for Success with the HPC Software Portfolio

High performance computing (HPC) software is key to harnessing the full power of HPC environments. Development and management tools enable IT departments to streamline installation and maintenance of their systems as well as create, optimize, and run their HPC applications. Read more…

PNNL’s Center for Advanced Tech Evaluation Seeks Wider HPC Community Ties

September 21, 2017

Two years ago the Department of Energy established the Center for Advanced Technology Evaluation (CENATE) at Pacific Northwest National Laboratory (PNNL). CENATE’s ambitious mission was to be a proving ground for near- Read more…

By John Russell

Stanford University and UberCloud Achieve Breakthrough in Living Heart Simulations

September 21, 2017

Cardiac arrhythmia can be an undesirable and potentially lethal side effect of drugs. During this condition, the electrical activity of the heart turns chaotic, Read more…

By Wolfgang Gentzsch, UberCloud, and Francisco Sahli, Stanford University

PNNL’s Center for Advanced Tech Evaluation Seeks Wider HPC Community Ties

September 21, 2017

Two years ago the Department of Energy established the Center for Advanced Technology Evaluation (CENATE) at Pacific Northwest National Laboratory (PNNL). CENAT Read more…

By John Russell

Exascale Computing Project Names Doug Kothe as Director

September 20, 2017

The Department of Energy’s Exascale Computing Project (ECP) has named Doug Kothe as its new director effective October 1. He replaces Paul Messina, who is stepping down after two years to return to Argonne National Laboratory. Kothe is a 32-year veteran of DOE’s National Laboratory System. Read more…

Takeaways from the Milwaukee HPC User Forum

September 19, 2017

Milwaukee’s elegant Pfister Hotel hosted approximately 100 attendees for the 66th HPC User Forum (September 5-7, 2017). In the original home city of Pabst Blu Read more…

By Merle Giles

Kathy Yelick Charts the Promise and Progress of Exascale Science

September 15, 2017

On Friday, Sept. 8, Kathy Yelick of Lawrence Berkeley National Laboratory and the University of California, Berkeley, delivered the keynote address on “Breakthrough Science at the Exascale” at the ACM Europe Conference in Barcelona. In conjunction with her presentation, Yelick agreed to a short Q&A discussion with HPCwire. Read more…

By Tiffany Trader

DARPA Pledges Another $300 Million for Post-Moore’s Readiness

September 14, 2017

The Defense Advanced Research Projects Agency (DARPA) launched a giant funding effort to ensure the United States can sustain the pace of electronic innovation vital to both a flourishing economy and a secure military. Under the banner of the Electronics Resurgence Initiative (ERI), some $500-$800 million will be invested in post-Moore’s Law technologies. Read more…

By Tiffany Trader

IBM Breaks Ground for Complex Quantum Chemistry

September 14, 2017

IBM has reported the use of a novel algorithm to simulate BeH2 (beryllium-hydride) on a quantum computer. This is the largest molecule so far simulated on a quantum computer. The technique, which used six qubits of a seven-qubit system, is an important step forward and may suggest an approach to simulating ever larger molecules. Read more…

By John Russell

Cubes, Culture, and a New Challenge: Trish Damkroger Talks about Life at Intel—and Why HPC Matters More Than Ever

September 13, 2017

Trish Damkroger wasn’t looking to change jobs when she attended SC15 in Austin, Texas. Capping a 15-year career within Department of Energy (DOE) laboratories, she was acting Associate Director for Computation at Lawrence Livermore National Laboratory (LLNL). Her mission was to equip the lab’s scientists and research partners with resources that would advance their cutting-edge work... Read more…

By Jan Rowell

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

NERSC Scales Scientific Deep Learning to 15 Petaflops

August 28, 2017

A collaborative effort between Intel, NERSC and Stanford has delivered the first 15-petaflops deep learning software running on HPC platforms and is, according Read more…

By Rob Farber

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Oracle Layoffs Reportedly Hit SPARC and Solaris Hard

September 7, 2017

Oracle’s latest layoffs have many wondering if this is the end of the line for the SPARC processor and Solaris OS development. As reported by multiple sources Read more…

By John Russell

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Leading Solution Providers

IBM Clears Path to 5nm with Silicon Nanosheets

June 5, 2017

Two years since announcing the industry’s first 7nm node test chip, IBM and its research alliance partners GlobalFoundries and Samsung have developed a proces Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Graphcore Readies Launch of 16nm Colossus-IPU Chip

July 20, 2017

A second $30 million funding round for U.K. AI chip developer Graphcore sets up the company to go to market with its “intelligent processing unit” (IPU) in Read more…

By Tiffany Trader

Google Releases Deeplearn.js to Further Democratize Machine Learning

August 17, 2017

Spreading the use of machine learning tools is one of the goals of Google’s PAIR (People + AI Research) initiative, which was introduced in early July. Last w Read more…

By John Russell

EU Funds 20 Million Euro ARM+FPGA Exascale Project

September 7, 2017

At the Barcelona Supercomputer Centre on Wednesday (Sept. 6), 16 partners gathered to launch the EuroEXA project, which invests €20 million over three-and-a-half years into exascale-focused research and development. Led by the Horizon 2020 program, EuroEXA picks up the banner of a triad of partner projects — ExaNeSt, EcoScale and ExaNoDe — building on their work... Read more…

By Tiffany Trader

Amazon Debuts New AMD-based GPU Instances for Graphics Acceleration

September 12, 2017

Last week Amazon Web Services (AWS) streaming service, AppStream 2.0, introduced a new GPU instance called Graphics Design intended to accelerate graphics. The Read more…

By John Russell

Cray Moves to Acquire the Seagate ClusterStor Line

July 28, 2017

This week Cray announced that it is picking up Seagate's ClusterStor HPC storage array business for an undisclosed sum. "In short we're effectively transitioning the bulk of the ClusterStor product line to Cray," said CEO Peter Ungaro. Read more…

By Tiffany Trader

IBM Advances Web-based Quantum Programming

September 5, 2017

IBM Research is pairing its Jupyter-based Data Science Experience notebook environment with its cloud-based quantum computer, IBM Q, in hopes of encouraging a new class of entrepreneurial user to solve intractable problems that even exceed the capabilities of the best AI systems. Read more…

By Alex Woodie

  • arrow
  • Click Here for More Headlines
  • arrow
Share This