DOE Exascale Plan Gets Support with Caveats

By John Russell

July 28, 2015

The DOE and the National Nuclear Security Administration (NNSA) plan to develop and deploy exascale technology by 2023 received strong backing yesterday from an Advanced Scientific Computing Advisory Committee (ASCAC) sub-committee but with caveats in the form of seven recommendations for strengthening management of the Exascale Computing Initiative (ECI).

Subcommittee chair, Dan Reed, vice president for research and economic development at the University of Iowa, presented the report and findings to an ASCAC meeting. “Like any ambitious undertaking, DOE’s proposed exascale computing initiative (ECI) involves some risks. Despite the risks, the benefits of the initiative to scientific discovery, national security and U.S. economic competitiveness are clear and compelling,” he said.

The subcommittee draft report was approved with a final full ASCAC version expected in August. Reed called the ECI well-crafted and noted DOE’s demonstrated ability to manage complicated, multi-stakeholder projects. Perhaps surprisingly, technology challenges were a subordinate part of the report. Instead, the report focused on project management.

Reed said, “We chose and we think appropriately to focus primarily on the organization and management issues because the technical issues and application issues have been reviewed so extensively for so many years (the by-now-familiar 10 technical challenges defined by DOE are listed further below).” Technology challenges clearly remain, he agreed.

The sub-committee’s detailed recommendations include:

  • Develop a detailed management and execution plan that defines clear responsibilities and decision-making authority to manage resources, risks, and dependencies appropriately across vendors, DOE laboratories, and other participants.
  • As part of the execution plan, clearly distinguish essential system attributes (e.g., sustained performance levels) from aspirational ones (e.g., specific energy consumption goals) and focus effort accordingly.
  • Given the scope, complexity, and potential impact of the ECI, conduct periodic external reviews by a carefully constituted advisory board.
  • Mitigate software risks by developing evolutionary alternatives to more innovative, but risky alternatives.
  • Unlike other elements of the hardware/software ecosystem, application performance and stability are mission critical, necessitating continued focus on hardware/software co-design to meet application needs.
  • Remain cognizant of the need for the ECI to support for data intensive and computation intensive workloads.
  • Where appropriate, work with other federal research agencies and international partners on workforce development and long-term research needs, while not creating dependences that could delay or imperil the execution plan.

Reed emphasized the need to be realistic in approaching the project and cautioned when setting expectations, particularly since the project is receiving wider attention in Congress.

“There’s a lot of uncertainty about the enabling technology still because this is a multi-year R&D plan. Innovation is still required. One of the things we want to ensure is that people don’t focus on the subsidiary metrics at the risk of those becoming part of the public perception of what success criteria should be. People latch onto figures of merit, sometimes rightly and sometimes wrongly. This is as much a political guidance as a technical one,” said Reed.

ASCAC Exascale Report Apps
Co-design, productive use of applications (legacy and new), and focusing on DOE and NNSA goals to advance science, enhance national competitiveness and assure nuclear stockpile stewardship are all emphasized in the report. Extending the benefits of extreme scale computing beyond these rather exclusive communities was also a theme.

“The whole point [of ECI] is to do a revolutionary leap forward. But it’s also important as part of that to the extent possible that we build broad ecosystems because the economic pull from a broad ecosystem will bring in more applications developers, it will lead to not just exascale laboratory systems, but also petascale research lab systems [used by] a much broader user base and shift the economics as well,” said Reed.

The sub-committee also recognized the growth of data-intensive computing to near equal footing with compute-intensive. Reed emphasized the DOE should, “Keep in mind in that data intensive and computationally intensive workflows both matter and in fact most of the time they are the same thing. They are intertwined pretty deeply [and] draw on the same ecosystems of hardware and software. Both matter. That drives as a corollary some focus on a new generation of analysis tools and libraries that will be needed to interpret that data.”

ASCAC had been charged by DOE and NNSA to review the “conceptual design for the Exascale Computing Initiative” and to deliver a report by September. Sub-committee members included: Reed; Martin Berzins, University of Utah; Bob Lucas, Livermore Software Technology Corporation; Satoshi Matsuoka, Tokyo Institute of Technology; Rob Pennington, University of Illinois, retired; Vivek Sarkar, Rice University; and Valerie Taylor, Texas A&M University.

The ECI’s goal is to deploy by 2023, capable exascale computing systems. This is defined as a hundred-fold increase in sustained performance over today’s computing capabilities, enabling applications to address next-generation science, engineering, and data problems to advance Department of Energy (DOE) Office of Science and National Nuclear Security Administration (NNSA) missions.

The plan includes three distinct components: Exascale Research, Development and Deployment (ExaRD); Exascale Application Development (ExaAD) to take full advantage of the emerging exascale hardware and software technologies from ExaRD; and Exascale Platform Deployment (ExaPD) to prepare for and acquire two or more exascale computers.

ASCAC Exascale Report Goals
Given the many technical issues remaining, ECI mission adjustments are inevitable said Reed. Establishing an external advisory board – coordinated by a single individual or a group – and leveraging other collaborations to help monitor and advise the project was strongly recommended. Reed also said, “On interagency and international collaboration, seek collaborations that don’t imperil the execution plan. This is not an open ended research project; it’s an outcome driven project.”

Included in the report was a restatement of the top ten exascale challenges as identified by DOE, shown here:

  • Energy efficiency: Creating more energy-efficient circuit, power, and cooling technologies.
  • Interconnect technology: Increasing the performance and energy efficiency of data movement.
  • Memory technology: Integrating advanced memory technologies to improve both capacity and bandwidth.
  • Scalable system software: Developing scalable system software that is power- and resilience-aware.
  • Programming systems: Inventing new programming environments that express massive parallelism, data locality, and resilience
  • Data management: Creating data management software that can handle the volume, velocity and diversity of data that is anticipated.
  • Exascale algorithms: Reformulating science problems and redesigning, or reinventing, their solution algorithms for exascale systems.
  • Algorithms for discovery, design, and decision: Facilitating mathematical optimization and uncertainty quantification for exascale discovery, design, and decision making.
  • Resilience and correctness: Ensuring correct scientific computation in face of faults, reproducibility, and algorithm verification challenges.
  • Scientific productivity: Increasing the productivity of computational scientists with new software engineering tools and environment

The full ASCAC report is expected to be completed in August. Here is a link to the slides presented by Reed: http://science.energy.gov/~/media/ascr/ascac/pdf/meetings/20150727/Exascale_Computing_Initiative_Review.pdf

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Trinity Supercomputer’s Haswell and KNL Partitions Are Merged

July 19, 2017

Trinity supercomputer’s two partitions – one based on Intel Xeon Haswell processors and the other on Xeon Phi Knights Landing – have been fully integrated are now available for use on classified work in the Nationa Read more…

By HPCwire Staff

Fujitsu Continues HPC, AI Push

July 19, 2017

Summer is well under way, but the so-called summertime slowdown, linked with hot temperatures and longer vacations, does not seem to have impacted Fujitsu's output. The Japanese multinational has made a raft of HPC and A Read more…

By Tiffany Trader

Researchers Use DNA to Store and Retrieve Digital Movie

July 18, 2017

From abacus to pencil and paper to semiconductor chips, the technology of computing has always been an ever-changing target. The human brain is probably the computer we use most (hopefully) and understand least. This mon Read more…

By John Russell

The Exascale FY18 Budget – The Next Step

July 17, 2017

On July 12, 2017, the U.S. federal budget for its Exascale Computing Initiative (ECI) took its next step forward. On that day, the full Appropriations Committee of the House of Representatives voted to accept the recomme Read more…

By Alex R. Larzelere

HPE Extreme Performance Solutions

HPE Servers Deliver High Performance Remote Visualization

Whether generating seismic simulations, locating new productive oil reservoirs, or constructing complex models of the earth’s subsurface, energy, oil, and gas (EO&G) is a highly data-driven industry. Read more…

Summer Reading: IEEE Spectrum’s Chip Hall of Fame

July 17, 2017

Take a trip down memory lane – the Mostek MK4096 4-kilobit DRAM, for instance. Perhaps processors are more to your liking. Remember the Sh-Boom processor (1988), created by Russell Fish and Chuck Moore, and named after Read more…

By John Russell

Women in HPC Luncheon Shines Light on Female-Friendly Hiring Practices

July 13, 2017

The second annual Women in HPC luncheon was held on June 20, 2017, during the International Supercomputing Conference in Frankfurt, Germany. The luncheon provides participants the opportunity to network with industry lea Read more…

By Tiffany Trader

Satellite Advances, NSF Computation Power Rapid Mapping of Earth’s Surface

July 13, 2017

New satellite technologies have completely changed the game in mapping and geographical data gathering, reducing costs and placing a new emphasis on time series and timeliness in general, according to Paul Morin, directo Read more…

By Ken Chiacchia and Tiffany Jolley

Intel Skylake: Xeon Goes from Chip to Platform

July 13, 2017

With yesterday’s New York unveiling of the new “Skylake” Xeon Scalable processors, Intel made multiple runs at multiple competitive threats and strategic markets. Skylake will carry Intel's flag in the fight for le Read more…

By Doug Black

Fujitsu Continues HPC, AI Push

July 19, 2017

Summer is well under way, but the so-called summertime slowdown, linked with hot temperatures and longer vacations, does not seem to have impacted Fujitsu's out Read more…

By Tiffany Trader

Researchers Use DNA to Store and Retrieve Digital Movie

July 18, 2017

From abacus to pencil and paper to semiconductor chips, the technology of computing has always been an ever-changing target. The human brain is probably the com Read more…

By John Russell

The Exascale FY18 Budget – The Next Step

July 17, 2017

On July 12, 2017, the U.S. federal budget for its Exascale Computing Initiative (ECI) took its next step forward. On that day, the full Appropriations Committee Read more…

By Alex R. Larzelere

Women in HPC Luncheon Shines Light on Female-Friendly Hiring Practices

July 13, 2017

The second annual Women in HPC luncheon was held on June 20, 2017, during the International Supercomputing Conference in Frankfurt, Germany. The luncheon provid Read more…

By Tiffany Trader

Satellite Advances, NSF Computation Power Rapid Mapping of Earth’s Surface

July 13, 2017

New satellite technologies have completely changed the game in mapping and geographical data gathering, reducing costs and placing a new emphasis on time series Read more…

By Ken Chiacchia and Tiffany Jolley

Intel Skylake: Xeon Goes from Chip to Platform

July 13, 2017

With yesterday’s New York unveiling of the new “Skylake” Xeon Scalable processors, Intel made multiple runs at multiple competitive threats and strategic Read more…

By Doug Black

Perverse Incentives? How Economics (Mis-)shaped Academic Science

July 12, 2017

The unintended consequences of how we fund academic research—in the U.S. and elsewhere—are strangling innovation, putting universities into debt and creatin Read more…

By Ken Chiacchia, Senior Science Writer, Pittsburgh Supercomputing Center

Why Tech is Failing at Diversity and How It Can Succeed

July 11, 2017

The sectors that are supposed to be all about innovation and the future continue to fail spectacularly at gender equity and diversity. UK, US and Canada still haven’t managed to break the average 20 percent threshold for gender equity across STEM academic disciplines. Read more…

By Kelly Nolan

HPC Compiler Company PathScale Seeks Life Raft

March 23, 2017

HPCwire has learned that HPC compiler company PathScale has fallen on difficult times and is asking the community for help or actively seeking a buyer for its a Read more…

By Tiffany Trader

Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

March 21, 2017

For a technology that’s usually characterized as far off and in a distant galaxy, quantum computing has been steadily picking up steam. Just how close real-wo Read more…

By John Russell

Google Pulls Back the Covers on Its First Machine Learning Chip

April 6, 2017

This week Google released a report detailing the design and performance characteristics of the Tensor Processing Unit (TPU), its custom ASIC for the inference Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Trump Budget Targets NIH, DOE, and EPA; No Mention of NSF

March 16, 2017

President Trump’s proposed U.S. fiscal 2018 budget issued today sharply cuts science spending while bolstering military spending as he promised during the cam Read more…

By John Russell

CPU-based Visualization Positions for Exascale Supercomputing

March 16, 2017

In this contributed perspective piece, Intel’s Jim Jeffers makes the case that CPU-based visualization is now widely adopted and as such is no longer a contrarian view, but is rather an exascale requirement. Read more…

By Jim Jeffers, Principal Engineer and Engineering Leader, Intel

Nvidia’s Mammoth Volta GPU Aims High for AI, HPC

May 10, 2017

At Nvidia's GPU Technology Conference (GTC17) in San Jose, Calif., this morning, CEO Jensen Huang announced the company's much-anticipated Volta architecture a Read more…

By Tiffany Trader

Facebook Open Sources Caffe2; Nvidia, Intel Rush to Optimize

April 18, 2017

From its F8 developer conference in San Jose, Calif., today, Facebook announced Caffe2, a new open-source, cross-platform framework for deep learning. Caffe2 is the successor to Caffe, the deep learning framework developed by Berkeley AI Research and community contributors. Read more…

By Tiffany Trader

Leading Solution Providers

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

MIT Mathematician Spins Up 220,000-Core Google Compute Cluster

April 21, 2017

On Thursday, Google announced that MIT math professor and computational number theorist Andrew V. Sutherland had set a record for the largest Google Compute Engine (GCE) job. Sutherland ran the massive mathematics workload on 220,000 GCE cores using preemptible virtual machine instances. Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Groq This: New AI Chips to Give GPUs a Run for Deep Learning Money

April 24, 2017

CPUs and GPUs, move over. Thanks to recent revelations surrounding Google’s new Tensor Processing Unit (TPU), the computing world appears to be on the cusp of Read more…

By Alex Woodie

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This