NSF Seeks ‘Breakthroughs’ for Energy-Efficient Computing

By John Russell

March 7, 2016

Breakthroughs, by their nature, are rarely generated on demand. That said, the energy problem in computing today is so acute that the National Science Foundation and Semiconductor Research Corporation are joining forces to stimulate research into developing ‘breakthroughs’ in energy-efficient computing. The new NSF-NRC grants program, Energy-Efficient Computing: from Devices to Architectures (E2CDA), has a budget of up to $4 million per year and is currently seeking proposals due in late March.

“Truly disruptive breakthroughs are now required, and not just from any one segment of the technology stack. Rather, due to the complexity of the challenges, revolutionary new approaches are needed at each level in the hierarchy. Furthermore, simultaneous co-optimization across all levels is essential for the creation of new, sustainable computing platforms,” is the rather dramatic call in the NSF-SRC solicitation.

The new effort, according to NSF, will specifically support new research to minimize the energy impacts of processing, storing, and moving data within future computing systems, and will be synergistic with other research activities that address other aspects of this overarching energy-constrained computing performance challenge.

“Through this joint solicitation, NSF and SRC aim to support game-changing research that can set the stage for the next paradigm of computing – from mobile devices to data centers – by minimizing the energy impact of future computing systems,” wrote Sankar Basu, one of the E2CDA managers, and NSF program Director for Computing and Communication Foundations, in a blogpost last week.

Clearly the HPC community has been bumping up against power issues for some time and resorted to a wide variety of energy management techniques. (Power dissipation alone has prompted a variety of cooling solutions – but they don’t help other power-related computing challenges.) One near-term power challenge that encompasses the full computing infrastructure is DOE’s stated goal to build an exascale computer that operates with a 25MW budget.

The urgent need now, says NSF-SRC, is research and innovation around devices with switching mechanisms that are fundamentally different from that of the conventional FET (field effect transistor), and architectures than are fundamentally different from the von Neumann architecture.

Shown below is a slide taken from a January webinar on E2CDA and its very familiar contents reflect slide presentations that have taken place throughout the computer industry in recent times. (Here’s a link to the slides from the webinar: https://www.nsf.gov/attachments/137395/public/E2CDA_webinar_slides_2016-01-21.pdf)

Screen Shot 2016-03-07 at 5.02.57 PM

The NFS-SRC grants take aim at two broad types (terms of the grant) of approaches to energy reduction and management in computing:

  1. Type One. Disruptive system architectures, circuit microarchitectures, and attendant device and interconnect technology aimed at achieving the highest level of computational energy efficiency for general purpose computing systems.
  1. Type Two. Revolutionary device concepts and associated circuits and architectures that will greatly extend the practical engineering limits of energy-efficient computation.

NSF-SRC has set ambitious goals. For type one (architecture and connectivity), proposals must target at least a 100X reduction or more in energy per delivered operation as compared to projected high-performance computing (HPC) systems utilizing conventional CMOS architectures and deeply scaled technology at the end of the roadmap. “As just one example of a metric goal, demonstrations that achieve system-level performance of > 1 Giga-MAC/s/nW could be targeted (MAC = multiply and accumulate operations),” spells out the RFP.

Investigation of new alternative connectivity technologies such as plasmonic, photonic, terahertz or “any others that can enable a dramatic lowering of overall system energy dissipation” are encouraged; so are interconnect technologies that enable functionality (such as embedded ‘intelligent’ routing, etc.) beyond point-to-point connectivity and the architectures that implement them are “also within the scope of interest.”

In addition, proposals are strongly encouraged to include an approach for merging heuristic learning and predictive functionality on the same physical platform as a programmable algorithmic capability.

Type two proposals also have a high bar; they must demonstrate new device concepts with the potential to reduce the energy dissipation involved in processing, storing, and moving information by two or more orders of magnitude.

“Any new switch is likely to have characteristics very different from those of a conventional field effect transistor. The interplay between device characteristics and optimum circuit architectures therefore means that circuit architectures must be reconsidered – this includes digital circuits, but also analog, memory, communication, and/or other more specialized functions. Devices combining digital/analog/memory functions may lend themselves particularly well to unconventional information processing architectures,” notes the RFP.

Proposed architectures should enable a broad range of useful functions, rather than being dedicated to one function or a few particular functions. More details for the terms of the E2CDA grants are show here (below).

NSF_SRC Energy Grant_2016-03-07 PM

Deadline for proposals is March 28. The E2CDA solicitation was posted on NSF on December 29, 2015, with a webinar held on January 21. No awards have been made as of this writing. That Basu wrote the guest post on the The CCC Blog (Computing Community Consortium) last week possibly suggests that responses to the solicitation may have been low so far.

The solicitation also offers guidance on what not to propose. Here are a few examples for both types:

Type one (proposals not sought):

  • Evolutionary extensions of existing general purpose computing platform architectures.
  • Systems that preclude substantially expanding the functionality and performance capabilities of general purpose computing, even if they are also aimed at significantly improving the overall level of energy efficiency.
  • System architectures that cannot be demonstrated to economically support the levels of reliability and physical dimension requirements projected for the future applications being targeted.
    System architectures that are not sufficiently scalable to support a broad base of applications.

Type two (proposals not sought):

  • Materials or device concepts that incrementally extend the capabilities of commercially established devices for logic and memory.
  • CMOS-based approaches to energy-conserving circuits and architectures;
  • Device concepts already the focus of research within established projects and centers, unless the proposed research is a substantive step beyond the currently-funded research.
  • Highly-specialized circuit architectures (“accelerators”) suited to a particular function or a limited set of functions, unless these circuits can be envisioned as economically integrated in a hybrid system capable of more generalized functions.
  • Devices and architectures for quantum computing – although proposals that explore the semi-classical regime (perhaps instantiating state variables with small ensembles of quantum states) or proposals that embrace some attributes of quantum computing achievable in the classical limit (such as energy-conserving circuits) are welcome.

Here is a link for information on the E2CDA grants: http://www.nsf.gov/pubs/2016/nsf16526/nsf16526.htm

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire