Enlisting Deep Learning in the War on Cancer

By John Russell

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. The pilots, supported in part by DOE exascale funding, not only seek to do good by advancing cancer research and therapy but also to advance deep learning capabilities and infrastructure with an eye towards eventual use on exascale machines.

By any standard, the U.S. War on Cancer and the Precision Medicine Initiative’s (PMI) are ambitious. Past Wars on Cancer haven’t necessarily fared well, which is not to say much hasn’t been accomplished. Today’s timing seems more promising. Progress in biomedical science and the ramp-up of next gen leadership computers (en route to exascale) are powerful enablers. Stir in the rapid emergence of deep learning to exploit data-driven science and many see greater cause for optimism. Not by chance was the opening plenary panel at SC16 on precision medicine and the role of HPC.

The three JDACS4C pilots span molecular to population scale efforts in support of the CANcer Distributed Learning Environment (CANDLE) project: they are intended to “provide insight into scalable machine learning tools; deep learning, simulation and analytics to reduce time to solution; and inform design of future computing solutions.” The hope is also to establish “a new paradigm for cancer research for years to come by making effective use of the ever-growing volumes and diversity of cancer-related data to build predictive models, provide better understanding of the disease and, ultimately, provide guidance and support decisions on anticipated outcomes of treatment for individual patients.”

Rick Stevens, ANL

These are ambitious goals. Sorting out JDACS4C’s precise lineage is a little challenging – it falls broadly under the Precision Medicine Initiative, NCI Cancer Moonshot, and has been also lumped under NSCI. Stevens noted the early discussion to create the effort started a couple of years ago with the first funding issued in the August time frame. Here’s a snapshot of the three pilots:

  • RAS Molecular Project. This project (Molecular Level Pilot for RAS Structure and Dynamics in Cellular Membranes) is intended to develop new computational approaches supporting research already being done under the RAS Initiative. Ultimately the hope is to refine our understanding of the role of the RAS (gene family) and its associated signaling pathway in cancer and to identify new therapeutic targets uniquely present in RAS protein membrane signaling complexes.
  • Pre-Clinical Screening. This project (Cellular Level Pilot for Predictive Modeling for Pre-clinical Screening) will develop “machine learning, large-scale data and predictive models based on experimental biological data derived from patient-derived xenografts.” The idea is to create a feedback loop, where the experimental models inform the design of the computational models. These predictive models may point to new targets in cancer and help identify new treatments.

Not surprisingly, there are many organizational pieces required. NCI components include the Center for Biomedical Informatics and Information Technology (CBIIT), the Division of Cancer Treatment and Diagnosis (DCTD), the Division of Cancer Control and Population Science (DCCPS), and the Frederick National Laboratory for Cancer Research. There are also four DOE National Laboratories formally designated on the project – Argonne National Laboratory, Oak Ridge National Laboratory, Lawrence Livermore National Laboratory, and Los Alamos National Laboratory.

As the projects came together, “We realized each had a need for deep learning and different uses of it. So the idea is that we would all work together on building both the software environment and network topologies and everything we would need for the three projects so we wouldn’t have duplication,” said Stevens. The researchers defined key benchmarks that “are tractable kinds of deep learning problems that are aligned with what we have to solve for the different cancer sub problems.”

An early first step was attracting vendor participation – something that turned out to be easy said Stevens because virtually all the major HPC vendors are aggressively ramping up DL roadmaps. Most see the JDACS4C pilots as opportunities to learn and refine. Stevens said JDASC4C has collaborations with Intel, Cray, NVIDIA, IBM, among others.

“All of the labs have DGX-1s and NVIDIA has optimized most of the common frameworks for the different GPUs, Pascal, etc. The DGX-1 is like an appliance so anything we build that runs on the DGX-1 can be easily distributed. Intel has it own extensive plans and not all is public yet. I can say that we are collaborating with all the right parts of Intel,” said Stevens, an ANL researcher and leader of the pre-clinical screening project.

Indeed Intel has been busy, buying Nervana (a complete platform for DL) and recently rolling out expanded plans. “They are talking about versions of Knights X series that are optimized for machine learning. Knights Mill is the first version of that part of their roadmap,” said Stevens. The chip giant also introduced a DL inference accelerator card at SC16; it’s a field-programmable gate array (FPGA)-based hardware and software solution for neural network acceleration. Stevens suggests Intel, like NVIDIA, is developing an appliance strategy.

“Intel is very much trying to define a strategy that differentiates some level between the platform for training and for inferencing. Most deep learning systems now do inferencing on the ‘quasi’ client side – on smaller platforms than used for training. Intel wants to ensure “future IA architectures are good at inferencing,” he said.

Not surprisingly  there’s a fair amount of effort assessing the many DL frameworks coming out of the Google, Microsoft, Facebook et al. “We are evaluating which frameworks work best for our problems and we are working with vendors to optimize them on the hardware. We’re also working with Livermore which has an internal project to build a scalable artificial neural network framework call LBANN,” said Stevens.

The plan is to develop “our models in a way that is independent of the frameworks so we can swap out the frameworks as those evolve without having to recode our models. This is a very common approach with deep learning where you have a scripting layer that captures your model representation – the meta algorithms for training and management data, etc. – and we are working with both the academic community and the NVIDIA on the workflow engine at the top. So we have kind of a stacked architecture and it involves collaborating with all of the different groups around the DL landscape.”

“What’s interesting,” said Stevens, “is the vendors for the next-gen platforms are strongly embracing the architectural ideas and features needed for accelerated machine learning in addition to traditional kind of physics-driven simulation.” He noted that market pressures and the fast growth of DL compared to the traditional HPC are pushing them in this direction. “It’s also giving us insight into DOE applications that are going to start looking like this – where there will be traditional physics-driven simulation, but where often we can find a way to leverage machine learning [too].”

Sharing the learning is an important component of the pilot projects. “We are abstracting model problems for the machine learning community to work on that are kind of sanitized versions of the seven candle benchmarks we’re working on,” said Stevens. That will include distributable data, code, all to be available at GitHub. The first of those elements are expected in Q2.

Individual pilot teams are also mounting their own outreach activities with the academic community. In terms of compute power for the pilots, “We are targeting platforms, particularly the CORAL platforms, new machines at Argonne, Oak Ridge and Livermore, and [eventually] exascale. Everything is sort of ecumenical so its not GPU specific or manycore specific.”

It’s interesting to look at the different ways in which the three projects plan to use deep learning.

The RAS project, at the molecular scale, is the smallest dimensional scale of all of the projects. RAS, you may know, is a well-known family of oncogenes that code for signaling proteins embedded in the cell membrane. These proteins control signaling pathways that extend into the cell and drive very many diverse cellular processes. RAS is currently implicated in about 30 percent of cancers including some of the toughest such as pancreatic cancer. The pilot project will combine simulation and wet lab screening data in an effort to elaborate the details of the RAS-related signaling cascades and hopefully identify key places to intervene and new drugs to use.

Even a relatively small tumor may have “thousands of mutations, both driver mutations and many passenger mutations,” said Stevens. These genetic miscues can alter the important details of signaling networks. For many years RAS itself as well as its associated signaling networks have been drug targets but as Stevens pointed out, “the behavior of that signaling network is very non-intuitive. Sometimes if you hit one of the downstream components, it actually creates negative feedback, which actually increases the effect you are trying to inhibit.”

In the RAS project, the simulation is basically a molecular dynamics exercise conducted at various granularities extending all the way down to atomistic behavior including quantum effects. The computational power required, not surprisingly, depends on the level of granularity being simulated and can be substantial.

“Machine learning is being used to track the state space that the simulation is going through and to make decisions – do we zoom in here, do we zoom out, do we change the parameters that we are looking in a different part to the ensemble space. It’s basically acting like a smart supervisor of this simulation to more effectively use it.

“In some sense it’s like the network is watching a movie and saying, “OK, I’ve seen this part of the movie before, let’s fast forward, or wow this is really interesting I’ve never seen this before, let’s use slow motion and zoom in.” That’s sort of what the machine learning is doing in the simulation. It’s able to fast forward and skip around in some sense,” said Stevens.

The pre-clinical screening project, led by Stevens, is an ambitious effort to sift through basically as much cancer  preclinical and clinical data as it can lay hold of and combine that with new data generated from mouse models to build predictive models of drug-tumor interactions. It’s an in silico and experimental feedback approach. Ultimately, given a specific tumor whose molecular attributes (gene expression, SNPs, proteomics, etc) have been characterized, it should be possible to plug that data into a model to determine the best therapeutic regime.

The subtlety here, said Stevens, is there has been a lot of machine learning work in this done at kind of the small scale, that is on single classes of tumors or relatively small classes of drugs. “What we are trying to do with the deep learning is to integrate all of this information across thousands of cell lines, tens of thousands of compounds that have been screened against smaller number of cell lines, and then be able to project that into a mouse. You grow a colony of mice derived from that human tumor, and these mice become proxies for human clinical trials. So I can try different compounds on the colony of tumor mice to provide information about how my tumor might respond to them if given as a drug.”

A huge challenge is being able to make sense of all the historical data, much of which is unstructured and often subjective (e.g. pathology reports). “One of the first things that we have done is to build classifiers to tell us what type the tumor is or where the body site is [based on diverse data],” he said. Not infrequently the data may be suspect. “If it’s a new dataset we run it through our classifiers and they may say, really, this is not from the liver, it’s from some other place.”

As a rule, the preclinical data is outcome based; it doesn’t explain how the result was achieved.

“Right now we can build machine learning models that are pretty accurate at say predicting a drug response or tumor type or outcome but they can’t tell us very effectively why. They are not explanatory, not mechanistic,” said Stevens, “What we want to do is bring in mechanistic models or mechanistic data in some way and hybridize that with machine learning models so that we get two things. We get the ability to have a highly accurate predictable model but also a model that can explain why that prediction. So the idea of this hybrid approach is a wide open space and we think that this will generalize into many fields.” Obtaining large and high quality data for training models remains challenging, he said.

The third project strives to develop models able to make population scale forecasts, what Stevens call “patient trajectories.” It’s basically mining surveillance data across the country. Although somewhat dispersed, there is a great deal of patient data held by NCI, NIH, FDA, pharma, and payor organizations (pathology reports, treatments, outcomes, lifestyle, demographics, etc.). Unfortunately, like a lot of biomedical data, it’s largely unstructured. “We can’t really compute on it in the way we want to so we are using machine learning to translate the unstructured data into structured data we can compute on,” said Stevens

“So for example we want to read all the pathology reports with a machine and pull out, say the biomarkers, the mutational state, or the drugs and so on such that we can then build profiles that are consistent. Think of it as a population-based model. In preclinical screening pilot let’s say we uncover some treatments and strategies that are very effective on a certain type of cancer. We want to take that information and feed it into the population model and say “If this became a common therapy, how much would it change the statistics globally or nationally” or something like that.”

It’s also a way to link all of the pilots, said Steven. Insight from the RAS project might be later used to look at subclasses of cancers where the new treatment might work; that in turn put it into the population model to understand what the impact of that might be.

It’s still early days for the JDACS4C pilot projects, but hopes are high. Stevens noted both NCI and DOE are getting access to things they don’t readily have. “NCI does not have a lot of mathematicians and computer scientists, which DOE has. They also don’t have access to leadership machines. What we (DOE) are getting is access to all of this great experimental data, experimental facilities, public databases.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

NIST/Xanadu Researchers Report Photonic Quantum Computing Advance

March 3, 2021

Researchers from the National Institute of Standards and Technology (NIST) and Xanadu, a young Canada-based quantum computing company, have reported developing a full-stack, photonic quantum computer able to carry out th Read more…

By John Russell

Can Deep Learning Replace Numerical Weather Prediction?

March 3, 2021

Numerical weather prediction (NWP) is a mainstay of supercomputing. Some of the first applications of the first supercomputers dealt with climate modeling, and even to this day, the largest climate models are heavily con Read more…

By Oliver Peckham

HPE Names Justin Hotard New HPC Chief as Pete Ungaro Departs

March 2, 2021

HPE CEO Antonio Neri announced today (March 2, 2020) the appointment of Justin Hotard as general manager of HPC, mission critical solutions and labs, effective immediately. Hotard replaces long-time Cray exec Pete Ungaro Read more…

By Tiffany Trader

ORNL’s Jeffrey Vetter on How IRIS Runtime will Help Deal with Extreme Heterogeneity

March 2, 2021

Jeffery Vetter is a familiar figure in HPC. Last year he became one of the new section heads in a reorganization at Oak Ridge National Laboratory. He had been founding director of ORNL's Future Technologies Group which i Read more…

By John Russell

HPC Career Notes: March 2021 Edition

March 1, 2021

In this monthly feature, we’ll keep you up-to-date on the latest career developments for individuals in the high-performance computing community. Whether it’s a promotion, new company hire, or even an accolade, we’ Read more…

By Mariana Iriarte

AWS Solution Channel

Introducing AWS HPC Tech Shorts

Amazon Web Services (AWS) is excited to announce a new videos series focused on running HPC workloads on AWS. This new video series will cover HPC workloads from genomics, computational chemistry, to computational fluid dynamics (CFD) and more. Read more…

Supercomputers Enable First Holistic Model of SARS-CoV-2, Showing Spike Proteins Move in Tandem

February 28, 2021

Most models of SARS-CoV-2, the coronavirus that causes COVID-19, hone in on key features of the virus: for instance, the spike protein. Some of this is attributable to the relative importance of those features, but most Read more…

By Oliver Peckham

Can Deep Learning Replace Numerical Weather Prediction?

March 3, 2021

Numerical weather prediction (NWP) is a mainstay of supercomputing. Some of the first applications of the first supercomputers dealt with climate modeling, and Read more…

By Oliver Peckham

HPE Names Justin Hotard New HPC Chief as Pete Ungaro Departs

March 2, 2021

HPE CEO Antonio Neri announced today (March 2, 2020) the appointment of Justin Hotard as general manager of HPC, mission critical solutions and labs, effective Read more…

By Tiffany Trader

ORNL’s Jeffrey Vetter on How IRIS Runtime will Help Deal with Extreme Heterogeneity

March 2, 2021

Jeffery Vetter is a familiar figure in HPC. Last year he became one of the new section heads in a reorganization at Oak Ridge National Laboratory. He had been f Read more…

By John Russell

HPC Career Notes: March 2021 Edition

March 1, 2021

In this monthly feature, we’ll keep you up-to-date on the latest career developments for individuals in the high-performance computing community. Whether it Read more…

By Mariana Iriarte

African Supercomputing Center Inaugurates ‘Toubkal,’ Most Powerful Supercomputer on the Continent

February 25, 2021

Historically, Africa hasn’t exactly been synonymous with supercomputing. There are only a handful of supercomputers on the continent, with few ranking on the Read more…

By Oliver Peckham

Japan to Debut Integrated Fujitsu HPC/AI Supercomputer This Spring

February 25, 2021

The integrated Fujitsu HPC/AI Supercomputer, Wisteria, is coming to Japan this spring. The University of Tokyo is preparing to deploy a heterogeneous computing Read more…

By Tiffany Trader

Xilinx Launches Alveo SN1000 SmartNIC

February 24, 2021

FPGA vendor Xilinx has debuted its latest SmartNIC model, the Alveo SN1000, with integrated “composability” features that allow enterprise users to add their own custom networking functions to supplement its built-in networking. By providing deep flexibility... Read more…

By Todd R. Weiss

ASF Keynotes Showcase How HPC and Big Data Have Pervaded the Pandemic

February 24, 2021

Last Thursday, a range of experts joined the Advanced Scale Forum (ASF) in a rapid-fire roundtable to discuss how advanced technologies have transformed the way humanity responded to the COVID-19 pandemic in indelible ways. The roundtable, held near the one-year mark of the first... Read more…

By Oliver Peckham

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

By John Russell

Esperanto Unveils ML Chip with Nearly 1,100 RISC-V Cores

December 8, 2020

At the RISC-V Summit today, Art Swift, CEO of Esperanto Technologies, announced a new, RISC-V based chip aimed at machine learning and containing nearly 1,100 low-power cores based on the open-source RISC-V architecture. Esperanto Technologies, headquartered in... Read more…

By Oliver Peckham

Azure Scaled to Record 86,400 Cores for Molecular Dynamics

November 20, 2020

A new record for HPC scaling on the public cloud has been achieved on Microsoft Azure. Led by Dr. Jer-Ming Chia, the cloud provider partnered with the Beckman I Read more…

By Oliver Peckham

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

By Tracey Bryant

NICS Unleashes ‘Kraken’ Supercomputer

April 4, 2008

A Cray XT4 supercomputer, dubbed Kraken, is scheduled to come online in mid-summer at the National Institute for Computational Sciences (NICS). The soon-to-be petascale system, and the resulting NICS organization, are the result of an NSF Track II award of $65 million to the University of Tennessee and its partners to provide next-generation supercomputing for the nation's science community. Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

Top500: Fugaku Keeps Crown, Nvidia’s Selene Climbs to #5

November 16, 2020

With the publication of the 56th Top500 list today from SC20's virtual proceedings, Japan's Fugaku supercomputer – now fully deployed – notches another win, Read more…

By Tiffany Trader

Gordon Bell Special Prize Goes to Massive SARS-CoV-2 Simulations

November 19, 2020

2020 has proven a harrowing year – but it has produced remarkable heroes. To that end, this year, the Association for Computing Machinery (ACM) introduced the Read more…

By Oliver Peckham

Leading Solution Providers

Contributors

Texas A&M Announces Flagship ‘Grace’ Supercomputer

November 9, 2020

Texas A&M University has announced its next flagship system: Grace. The new supercomputer, named for legendary programming pioneer Grace Hopper, is replacing the Ada system (itself named for mathematician Ada Lovelace) as the primary workhorse for Texas A&M’s High Performance Research Computing (HPRC). Read more…

By Oliver Peckham

Saudi Aramco Unveils Dammam 7, Its New Top Ten Supercomputer

January 21, 2021

By revenue, oil and gas giant Saudi Aramco is one of the largest companies in the world, and it has historically employed commensurate amounts of supercomputing Read more…

By Oliver Peckham

Intel Xe-HP GPU Deployed for Aurora Exascale Development

November 17, 2020

At SC20, Intel announced that it is making its Xe-HP high performance discrete GPUs available to early access developers. Notably, the new chips have been deplo Read more…

By Tiffany Trader

Intel Teases Ice Lake-SP, Shows Competitive Benchmarking

November 17, 2020

At SC20 this week, Intel teased its forthcoming third-generation Xeon "Ice Lake-SP" server processor, claiming competitive benchmarking results against AMD's second-generation Epyc "Rome" processor. Ice Lake-SP, Intel's first server processor with 10nm technology... Read more…

By Tiffany Trader

New Deep Learning Algorithm Solves Rubik’s Cube

July 25, 2018

Solving (and attempting to solve) Rubik’s Cube has delighted millions of puzzle lovers since 1974 when the cube was invented by Hungarian sculptor and archite Read more…

By John Russell

Livermore’s El Capitan Supercomputer to Debut HPE ‘Rabbit’ Near Node Local Storage

February 18, 2021

A near node local storage innovation called Rabbit factored heavily into Lawrence Livermore National Laboratory’s decision to select Cray’s proposal for its CORAL-2 machine, the lab’s first exascale-class supercomputer, El Capitan. Details of this new storage technology were revealed... Read more…

By Tiffany Trader

It’s Fugaku vs. COVID-19: How the World’s Top Supercomputer Is Shaping Our New Normal

November 9, 2020

Fugaku is currently the most powerful publicly ranked supercomputer in the world – but we weren’t supposed to have it yet. The supercomputer, situated at Japan’s Riken scientific research institute, was scheduled to come online in 2021. When the pandemic struck... Read more…

By Oliver Peckham

African Supercomputing Center Inaugurates ‘Toubkal,’ Most Powerful Supercomputer on the Continent

February 25, 2021

Historically, Africa hasn’t exactly been synonymous with supercomputing. There are only a handful of supercomputers on the continent, with few ranking on the Read more…

By Oliver Peckham

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire