PNNL’s Center for Advanced Tech Evaluation Seeks Wider HPC Community Ties

By John Russell

September 21, 2017

Two years ago the Department of Energy established the Center for Advanced Technology Evaluation (CENATE) at Pacific Northwest National Laboratory (PNNL). CENATE’s ambitious mission was to be a proving ground for near-term and long-term technologies that could impact DoE workloads and HPC broadly. This month the leadership baton was passed from founding director Adolfy Hoisie to Kevin Barker, a veteran PNNL researcher and member of the CENATE project since its start. Hoisie has moved to Brookhaven National Lab to lead another new initiative as chair of the just-formed Computing for National Security Department.

In its short lifespan, CENATE has made steady strides. It has assembled an impressive infrastructure of test and measurement capabilities to explore computer technology. It has tackled several specific projects, ranging from the study of novel architecture from Data Vortex and Nvidia’s DGX-1 to longer horizon efforts around neuromorphic technology. The change in leadership, emphasizes Barker, won’t alter CENATE’s ambitious plans, but it will enable refinement of several processes, notably an effort to forge tighter links to the HPC community writ large, including DoE researchers, academia, and commercial technology partners.

Today there are about a dozen CENATE staff at PNNL, says Barker. One of the biggest changes will be standing up a more inclusive, more activist steering committee to guide CENATE.

Kevin Barker, P.I., Center for Advanced Technology Evaluation (CENATE), PNNL

Recently, HPCwire talked with Barker about the developing plans for CENATE and its priorities. Barker is certainly no stranger to HPC. He joined PNNL in 2010 as a senior HPC research scientist rising to team lead for modeling and simulation in 2014. Before joining PNNL, Barker spent nearly six years at Los Alamos National Laboratory as an HPC research scientist.

HPCwire: Your prior CENATE experience will no doubt be helpful. Perhaps you could describe your role and provide a sense of what CENATE has accomplished to date.

Barker: Really, I’ve been with CENATE since it began. It had a couple of focus areas in terms of how it was organized internally. I was in charge of one of those areas around performance predictions. The idea was that CENATE would be a program that encompassed not only of testbed and performance measurements but also would take those performance measurements we could get from physical systems, or prototype systems at small scale, and use performance prediction techniques to explore what those performance impacts would be at large scale. That was my role. Now, I am the PI.

In the first two years of the project, CENATE has deployed testbeds incorporating emerging technologies in the areas of processing, memory, and networking, and has engaged with the research community to assess the applicability of these technologies to workloads of interest to the Department of Energy. Specifically, CENATE has explored high-throughput architectures applied to large-scale Machine Learning frameworks; non-volatile memories; reconfigurable optical networking technology; and self-routing, dynamic, congestion-free networks applied to graph analytics. Through a broad community engagement, CENATE has ensured that its findings are fed back through workshops and deep collaborations with leading researchers.

HPCwire: That’s an extensive list. What’s going to change?

Barker: This change in leadership isn’t a dramatic change in terms of the technical capabilities or what we hope to accomplish technically. Now we want to ensure CENATE is more tightly integrated with the outside community, the HPC community in the DoE, and the vendor space. We also want to make sure the work we are doing at CENATE has an impact back at the DoE.

We’re working on getting a good plan in place to accomplish that – engaging with the vendor community, engaging with application developers, systems software developers, with the DoE complex making CENATE resources available to those people so that we can have a collaborative research environment. They can bring their problems to CENATE and we could provide access to some of these novel and emerging technologies that CENATE is tasked with assessing.

HPCwire: Maybe we should step back and review CENATE’s mission. How do you see it evolving, particularly in light of establishing a more inclusive and activist steering committee?

Barker: Again, it hasn’t really changed. When the steering committee stands up we envision CENATE taking on two kinds of tracks in terms of the research and the resources that we look at in each track. In the first track, we envision a shorter time scale where we are looking at technologies that are very near to market that we can get close to either prototype hardware, early release hardware, or engineering sample hardware. For the second track, in terms of timescale, we want CENATE to have an impact on more novel or high risk architectural approaches. So we might look at such things as beyond Moore’s Law computing technologies.

We envision the steering committee having a big impact because we want to have some indication from the community regarding what are the technologies that we should be most interested in from a community perspective. [Tentatively] we envision a fixed six-month schedule steering committee meetings in particular to discuss what architectures should be look at in the next six months window and feedback from the previous six month windows. We haven’t decided yet whether those will take the form of a meeting or workshop where we have more community involvement from outside the just the steering committee. Those are some of the thing still under discussion.

HPCwire: Given the importance of the steering committee, how big will it be and who will be on it?

Barker: It could be 15 or so organizations, maybe a person from each organization. We would like to have participation from other labs in the DoE community, and potentially academic partners. For example Indiana University is a major user of the Data Vortex architecture so it makes sense for them to participate. The third group being the commercial vendor space. We want to have this settled (and up on the web) before supercomputing (SC17, November 12-17).

HPCwire: There are so many new technologies bubbling up at various stages of development. Adolfy had mentioned interest in developing neuromorphic chips. Is that work ongoing? What’s the thinking on longer term technologies?

Barker: We are definitely interested in these longer term technologies and think that CENATE can have a big impact in the community, presenting that to the funding sources and say hey we want to have CENATE really to be positioned to have an impact beyond the next thing that you can buy from your hardware vendor. To explore the next gen technologies that aren’t necessarily tied to commercial products at this point but may still have real impact, particularly in the generation of beyond exascale.

For example, the exascale systems are pretty well covered with the Exascale Computing Project. We’re very familiar now with what those systems are going to look like. People are very focused to get their applications to run on those architectures. That’s not really where we see CENATE having a play. In looking beyond that, what are the technologies that are going to shape high performance computing beyond exascale timeframe? We really want CENATE to be positioned to have an impact in those areas. This is what I mean by saying a refinement of the CENATE direction. Up until now CENATE has looked at a number of near-to-market or new to market technologies. And they have had a big impact. The DGX-1 is great example. We stood up a DGX-1 and immediately we had users from around the lab complex and academia clamoring to get on the machine to explore how their applications are going to perform to develop system software and things such as that.

But we want CENATE also to look beyond that, things like extreme heterogeneity, software reconfigurable computing. So this is really why we are placing and emphasis on the connection to the research community so that we can get as much as is possible an accurate prediction of why these are the technologies that we think are going to make an impact. How can CENATE position itself to help assess those technologies in the near-term and that might involve a much deeper dive into specific technologies? CENATE doesn’t have an unlimited amount of resources (time personnel dollars) so it’s very important we target those things as effectively as we can.

HPCwire: Funding is always an issue on advanced research and DoE is a big player. What about tapping into programs such as DARPA’s Electronics Resurgence Initiative (ERI) which is focused on post-Moore’s law technology and just received additional funding?

Barker: There are definitely some programs that worth [looking at]. We are not working with ERI in particular although that could be something we explore together with program management from DoE headquarters. But there are some opportunities exactly along those lines that we are looking into but nothing firm at this point.

HPCwire: Given the various changes, and the enhanced role of the steering committee, is it premature to identify the top five technologies we are going to tackle in the next year?

Barker: Exactly. Unfortunately the end of the year is kind of a busy time in the HPC world with SC (supercomputing conference) and everything else going on. We hope those kinds of things pinned down with at least some degree of certainty within the next few months.

HPCwire: One of the distinguishing aspects of CENATE is the diversity and sophistication of the test and measurement equipment and capabilities at PNNL. What’s happening on that front?

Barker: We have equipment for testing power and energy and well as for thermal measurement capability. That is still all in place. We’re expanding the evaluation test suite that we have been using up until this point, the benchmark codes. CENATE itself has an evaluation test suite in addition to reaching out to collaborators who are interested in the equipment and who bring their own software test suite. We’re interested in looking at these machines in the context of numerical simulation, high performance computing codes, as well as graph analytics codes, machine learning codes, so we are expanding that set of benchmark codes, but the measurement capabilities we have in place are still in place.

HPCwire: It sounds like, among other things, you are adapting your capabilities to be able to handle emerging, nontraditional ‘HPC’ needs such as deep learning and data analytics?

Barker: Right. One of the important things when we are looking at these architectures, and the DGX-1 is good example, is we want to evaluate those technologies in the mode they are designed to operate in. The DGX-1 really is designed as a deep learning/ machine learning architecture. Exploring simply traditional HPC simulation codes on it might not be the most appropriate thing to do. We want to paint it [DGX-1 performance and potential] in the light it was designed for. Our evaluation suite of kernels and benchmarks needs to encompass those application areas that these architectures are targeting. And things like machine learning and deep learning are becoming such a part of the DoE workload that for CENATE to remain relevant to the DoE we need to have that capability. The DoE HPC landscape is much more than tightly couple code.

HPCwire: In the past there had been talk of CENATE workshops and other outreach efforts to diffuse CENATE learnings into the community but I don’t think much has happened along those lines yet. How do you share results and what are the plans going forward?

Barker: This is one area where we have decided that some refinement is necessary. Currently the mechanism that we use to present some of these results back to the community is through publications. It’s a pretty typical route. We’ve had some success there and, for example, we have papers on our work with DGX-1 in the submission process right now. We want to expand how we do this and are still developing the plans.

Hosting user group meetings is another way. Just two weeks ago, we hosted the first Data Vortex user group meeting at PNNL and CENATE was a player in that it brought together a couple of other programs that were looking at the data vortex architecture. That was a really successful workshop. Researchers from DoE, other government agencies, academic researchers came here to PNNL specifically about the data vortex architecture which is a big architecture in CENATE. We actually have two Data Vortex machines. That’s an example we can point to where we can say CENATE is making an impact in the community.

The NDA issues are sometimes very tricky but we have some experience with other projects where similar issues have arisen so we do have some strategies to deal with NDA issues.

HPCwire: How will you reach potential collaborators. There’s the technical steering committee but given its relatively small size, who will you reach beyond its immediate interests and attract other collaborators?

Barker: We are standing up a new CENATE web site that we hope to have up very soon, which will solicit this kind of input and have a mechanism where we can say if you’re a commercial or vendor partner and you want to participate in the CENATE program, here’s how you can get in touch with us. We definitely don’t want to be an exclusive club. We want to cast a wide net in terms of the types of technologies that are represented in the steering committee. Some of this is still in progress

One of the things we are exploring [for the web site] is a way to have potential interested external parties propose what they would like to do and the equipment they would be interested in evaluating. Again, this where the technical steering committee again comes to evaluate these proposals. It might be a model where – and this is what we are moving towards – where we essentially put out a call [for proposal]. That sounds a bit formal. CENATE is not a funding organization and won’t fund external collaborators. But it will be a way for submitters to say what interesting problems are they interested in solving that CENATE could then participate in and possibly provide access to technology. So if you are professor with some graduate students you might say, ‘Here’s an application that we want to develop and we want to explore how it might work on architecture x but we don’t have the means to get architecture x, can CENATE help?’

HPCwire: Thank you for your time.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

HPC in Life Sciences Part 1: CPU Choices, Rise of Data Lakes, Networking Challenges, and More

February 21, 2019

For the past few years HPCwire and leaders of BioTeam, a research computing consultancy specializing in life sciences, have convened to examine the state of HPC (and now AI) use in life sciences. Without HPC writ lar Read more…

By John Russell

Arm Unveils Neoverse N1 Platform with up to 128-Cores

February 20, 2019

Following on its Neoverse roadmap announcement last October, Arm today revealed its next-gen Neoverse microarchitecture with compute and throughput-optimized silicon designs catered toward general-purpose cloud computing Read more…

By Tiffany Trader

The Internet of Criminal Things—Trust in the Gods but Verify!

February 20, 2019

“Are we under attack?” asked Professor Elmarie Biermann of the Cyber Security Institute during the recent South African Centre for High Performance Computing’s (CHPC) National Conference in Cape Town. A quick show Read more…

By Elizabeth Leake, STEM-Trek

HPE Extreme Performance Solutions

HPE and Intel® Omni-Path Architecture: How to Power a Cloud

Learn how HPE and Intel® Omni-Path Architecture provide critical infrastructure for leading Nordic HPC provider’s HPCFLOW cloud service.

powercloud_blog.jpgFor decades, HPE has been at the forefront of high-performance computing, and we’ve powered some of the fastest and most robust supercomputers in the world. Read more…

IBM Accelerated Insights

The Perils of Becoming Trapped in the Cloud

Terms like ‘open systems’ have been bandied about for decades. While modern computer systems are relatively open compared to their predecessors, there are still plenty of opportunities to become locked into proprietary interfaces. Read more…

Machine Learning Takes Heat for Science’s Reproducibility Crisis

February 19, 2019

Scientists are raising red flags about the accuracy and reproducibility of conclusions drawn by machine learning frameworks. Among the remedies are developing new ML systems that can question their own predictions, show Read more…

By George Leopold

HPC in Life Sciences Part 1: CPU Choices, Rise of Data Lakes, Networking Challenges, and More

February 21, 2019

For the past few years HPCwire and leaders of BioTeam, a research computing consultancy specializing in life sciences, have convened to examine the state of HP Read more…

By John Russell

Arm Unveils Neoverse N1 Platform with up to 128-Cores

February 20, 2019

Following on its Neoverse roadmap announcement last October, Arm today revealed its next-gen Neoverse microarchitecture with compute and throughput-optimized si Read more…

By Tiffany Trader

Insights from Optimized Codes on Cineca’s Marconi

February 15, 2019

What can you do with 381,392 CPU cores? For Cineca, it means enabling computational scientists to expand a large part of the world’s body of knowledge from th Read more…

By Ken Strandberg

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

UC Berkeley Paper Heralds Rise of Serverless Computing in the Cloud – Do You Agree?

February 13, 2019

Almost exactly ten years to the day from publishing of their widely-read, seminal paper on cloud computing, UC Berkeley researchers have issued another ambitious examination of cloud computing - Cloud Programming Simplified: A Berkeley View on Serverless Computing. The new work heralds the rise of ‘serverless computing’ as the next dominant phase of cloud computing. Read more…

By John Russell

Iowa ‘Grows Its Own’ to Fill the HPC Workforce Pipeline

February 13, 2019

The global workforce that supports advanced computing, scientific software and high-speed research networks is relatively small when you stop to consider the magnitude of the transformative discoveries it empowers. Technical conferences provide a forum where specialists convene to learn about the latest innovations and schedule face-time with colleagues from other institutions. Read more…

By Elizabeth Leake, STEM-Trek

Trump Signs Executive Order Launching U.S. AI Initiative

February 11, 2019

U.S. President Donald Trump issued an Executive Order (EO) today launching a U.S Artificial Intelligence Initiative. The new initiative - Maintaining American L Read more…

By John Russell

Celebrating Women in Science: Meet Four Women Leading the Way in HPC

February 11, 2019

One only needs to look around at virtually any CS/tech conference to realize that women are underrepresented, and that holds true of HPC. SC hosts over 13,000 H Read more…

By AJ Lauer

Quantum Computing Will Never Work

November 27, 2018

Amid the gush of money and enthusiastic predictions being thrown at quantum computing comes a proposed cold shower in the form of an essay by physicist Mikhail Read more…

By John Russell

Cray Unveils Shasta, Lands NERSC-9 Contract

October 30, 2018

Cray revealed today the details of its next-gen supercomputing architecture, Shasta, selected to be the next flagship system at NERSC. We've known of the code-name "Shasta" since the Argonne slice of the CORAL project was announced in 2015 and although the details of that plan have changed considerably, Cray didn't slow down its timeline for Shasta. Read more…

By Tiffany Trader

The Case Against ‘The Case Against Quantum Computing’

January 9, 2019

It’s not easy to be a physicist. Richard Feynman (basically the Jimi Hendrix of physicists) once said: “The first principle is that you must not fool yourse Read more…

By Ben Criger

AMD Sets Up for Epyc Epoch

November 16, 2018

It’s been a good two weeks, AMD’s Gary Silcott and Andy Parma told me on the last day of SC18 in Dallas at the restaurant where we met to discuss their show news and recent successes. Heck, it’s been a good year. Read more…

By Tiffany Trader

Intel Reportedly in $6B Bid for Mellanox

January 30, 2019

The latest rumors and reports around an acquisition of Mellanox focus on Intel, which has reportedly offered a $6 billion bid for the high performance interconn Read more…

By Doug Black

ClusterVision in Bankruptcy, Fate Uncertain

February 13, 2019

ClusterVision, European HPC specialists that have built and installed over 20 Top500-ranked systems in their nearly 17-year history, appear to be in the midst o Read more…

By Tiffany Trader

US Leads Supercomputing with #1, #2 Systems & Petascale Arm

November 12, 2018

The 31st Supercomputing Conference (SC) - commemorating 30 years since the first Supercomputing in 1988 - kicked off in Dallas yesterday, taking over the Kay Ba Read more…

By Tiffany Trader

Looking for Light Reading? NSF-backed ‘Comic Books’ Tackle Quantum Computing

January 28, 2019

Still baffled by quantum computing? How about turning to comic books (graphic novels for the well-read among you) for some clarity and a little humor on QC. The Read more…

By John Russell

Leading Solution Providers

SC 18 Virtual Booth Video Tour

Advania @ SC18 AMD @ SC18
ASRock Rack @ SC18
DDN Storage @ SC18
HPE @ SC18
IBM @ SC18
Lenovo @ SC18 Mellanox Technologies @ SC18
NVIDIA @ SC18
One Stop Systems @ SC18
Oracle @ SC18 Panasas @ SC18
Supermicro @ SC18 SUSE @ SC18 TYAN @ SC18
Verne Global @ SC18

Contract Signed for New Finnish Supercomputer

December 13, 2018

After the official contract signing yesterday, configuration details were made public for the new BullSequana system that the Finnish IT Center for Science (CSC Read more…

By Tiffany Trader

Deep500: ETH Researchers Introduce New Deep Learning Benchmark for HPC

February 5, 2019

ETH researchers have developed a new deep learning benchmarking environment – Deep500 – they say is “the first distributed and reproducible benchmarking s Read more…

By John Russell

IBM Quantum Update: Q System One Launch, New Collaborators, and QC Center Plans

January 10, 2019

IBM made three significant quantum computing announcements at CES this week. One was introduction of IBM Q System One; it’s really the integration of IBM’s Read more…

By John Russell

HPC Reflections and (Mostly Hopeful) Predictions

December 19, 2018

So much ‘spaghetti’ gets tossed on walls by the technology community (vendors and researchers) to see what sticks that it is often difficult to peer through Read more…

By John Russell

IBM Bets $2B Seeking 1000X AI Hardware Performance Boost

February 7, 2019

For now, AI systems are mostly machine learning-based and “narrow” – powerful as they are by today's standards, they're limited to performing a few, narro Read more…

By Doug Black

Nvidia’s Jensen Huang Delivers Vision for the New HPC

November 14, 2018

For nearly two hours on Monday at SC18, Jensen Huang, CEO of Nvidia, presented his expansive view of the future of HPC (and computing in general) as only he can do. Animated. Backstopped by a stream of data charts, product photos, and even a beautiful image of supernovae... Read more…

By John Russell

The Deep500 – Researchers Tackle an HPC Benchmark for Deep Learning

January 7, 2019

How do you know if an HPC system, particularly a larger-scale system, is well-suited for deep learning workloads? Today, that’s not an easy question to answer Read more…

By John Russell

Intel Confirms 48-Core Cascade Lake-AP for 2019

November 4, 2018

As part of the run-up to SC18, taking place in Dallas next week (Nov. 11-16), Intel is doling out info on its next-gen Cascade Lake family of Xeon processors, specifically the “Advanced Processor” version (Cascade Lake-AP), architected for high-performance computing, artificial intelligence and infrastructure-as-a-service workloads. Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This