Enlisting Deep Learning in the War on Cancer

By John Russell

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. The pilots, supported in part by DOE exascale funding, not only seek to do good by advancing cancer research and therapy but also to advance deep learning capabilities and infrastructure with an eye towards eventual use on exascale machines.

By any standard, the U.S. War on Cancer and the Precision Medicine Initiative’s (PMI) are ambitious. Past Wars on Cancer haven’t necessarily fared well, which is not to say much hasn’t been accomplished. Today’s timing seems more promising. Progress in biomedical science and the ramp-up of next gen leadership computers (en route to exascale) are powerful enablers. Stir in the rapid emergence of deep learning to exploit data-driven science and many see greater cause for optimism. Not by chance was the opening plenary panel at SC16 on precision medicine and the role of HPC.

The three JDACS4C pilots span molecular to population scale efforts in support of the CANcer Distributed Learning Environment (CANDLE) project: they are intended to “provide insight into scalable machine learning tools; deep learning, simulation and analytics to reduce time to solution; and inform design of future computing solutions.” The hope is also to establish “a new paradigm for cancer research for years to come by making effective use of the ever-growing volumes and diversity of cancer-related data to build predictive models, provide better understanding of the disease and, ultimately, provide guidance and support decisions on anticipated outcomes of treatment for individual patients.”

Rick Stevens, ANL

These are ambitious goals. Sorting out JDACS4C’s precise lineage is a little challenging – it falls broadly under the Precision Medicine Initiative, NCI Cancer Moonshot, and has been also lumped under NSCI. Stevens noted the early discussion to create the effort started a couple of years ago with the first funding issued in the August time frame. Here’s a snapshot of the three pilots:

  • RAS Molecular Project. This project (Molecular Level Pilot for RAS Structure and Dynamics in Cellular Membranes) is intended to develop new computational approaches supporting research already being done under the RAS Initiative. Ultimately the hope is to refine our understanding of the role of the RAS (gene family) and its associated signaling pathway in cancer and to identify new therapeutic targets uniquely present in RAS protein membrane signaling complexes.
  • Pre-Clinical Screening. This project (Cellular Level Pilot for Predictive Modeling for Pre-clinical Screening) will develop “machine learning, large-scale data and predictive models based on experimental biological data derived from patient-derived xenografts.” The idea is to create a feedback loop, where the experimental models inform the design of the computational models. These predictive models may point to new targets in cancer and help identify new treatments.

Not surprisingly, there are many organizational pieces required. NCI components include the Center for Biomedical Informatics and Information Technology (CBIIT), the Division of Cancer Treatment and Diagnosis (DCTD), the Division of Cancer Control and Population Science (DCCPS), and the Frederick National Laboratory for Cancer Research. There are also four DOE National Laboratories formally designated on the project – Argonne National Laboratory, Oak Ridge National Laboratory, Lawrence Livermore National Laboratory, and Los Alamos National Laboratory.

As the projects came together, “We realized each had a need for deep learning and different uses of it. So the idea is that we would all work together on building both the software environment and network topologies and everything we would need for the three projects so we wouldn’t have duplication,” said Stevens. The researchers defined key benchmarks that “are tractable kinds of deep learning problems that are aligned with what we have to solve for the different cancer sub problems.”

An early first step was attracting vendor participation – something that turned out to be easy said Stevens because virtually all the major HPC vendors are aggressively ramping up DL roadmaps. Most see the JDACS4C pilots as opportunities to learn and refine. Stevens said JDASC4C has collaborations with Intel, Cray, NVIDIA, IBM, among others.

“All of the labs have DGX-1s and NVIDIA has optimized most of the common frameworks for the different GPUs, Pascal, etc. The DGX-1 is like an appliance so anything we build that runs on the DGX-1 can be easily distributed. Intel has it own extensive plans and not all is public yet. I can say that we are collaborating with all the right parts of Intel,” said Stevens, an ANL researcher and leader of the pre-clinical screening project.

Indeed Intel has been busy, buying Nervana (a complete platform for DL) and recently rolling out expanded plans. “They are talking about versions of Knights X series that are optimized for machine learning. Knights Mill is the first version of that part of their roadmap,” said Stevens. The chip giant also introduced a DL inference accelerator card at SC16; it’s a field-programmable gate array (FPGA)-based hardware and software solution for neural network acceleration. Stevens suggests Intel, like NVIDIA, is developing an appliance strategy.

“Intel is very much trying to define a strategy that differentiates some level between the platform for training and for inferencing. Most deep learning systems now do inferencing on the ‘quasi’ client side – on smaller platforms than used for training. Intel wants to ensure “future IA architectures are good at inferencing,” he said.

Not surprisingly  there’s a fair amount of effort assessing the many DL frameworks coming out of the Google, Microsoft, Facebook et al. “We are evaluating which frameworks work best for our problems and we are working with vendors to optimize them on the hardware. We’re also working with Livermore which has an internal project to build a scalable artificial neural network framework call LBANN,” said Stevens.

The plan is to develop “our models in a way that is independent of the frameworks so we can swap out the frameworks as those evolve without having to recode our models. This is a very common approach with deep learning where you have a scripting layer that captures your model representation – the meta algorithms for training and management data, etc. – and we are working with both the academic community and the NVIDIA on the workflow engine at the top. So we have kind of a stacked architecture and it involves collaborating with all of the different groups around the DL landscape.”

“What’s interesting,” said Stevens, “is the vendors for the next-gen platforms are strongly embracing the architectural ideas and features needed for accelerated machine learning in addition to traditional kind of physics-driven simulation.” He noted that market pressures and the fast growth of DL compared to the traditional HPC are pushing them in this direction. “It’s also giving us insight into DOE applications that are going to start looking like this – where there will be traditional physics-driven simulation, but where often we can find a way to leverage machine learning [too].”

Sharing the learning is an important component of the pilot projects. “We are abstracting model problems for the machine learning community to work on that are kind of sanitized versions of the seven candle benchmarks we’re working on,” said Stevens. That will include distributable data, code, all to be available at GitHub. The first of those elements are expected in Q2.

Individual pilot teams are also mounting their own outreach activities with the academic community. In terms of compute power for the pilots, “We are targeting platforms, particularly the CORAL platforms, new machines at Argonne, Oak Ridge and Livermore, and [eventually] exascale. Everything is sort of ecumenical so its not GPU specific or manycore specific.”

It’s interesting to look at the different ways in which the three projects plan to use deep learning.

The RAS project, at the molecular scale, is the smallest dimensional scale of all of the projects. RAS, you may know, is a well-known family of oncogenes that code for signaling proteins embedded in the cell membrane. These proteins control signaling pathways that extend into the cell and drive very many diverse cellular processes. RAS is currently implicated in about 30 percent of cancers including some of the toughest such as pancreatic cancer. The pilot project will combine simulation and wet lab screening data in an effort to elaborate the details of the RAS-related signaling cascades and hopefully identify key places to intervene and new drugs to use.

Even a relatively small tumor may have “thousands of mutations, both driver mutations and many passenger mutations,” said Stevens. These genetic miscues can alter the important details of signaling networks. For many years RAS itself as well as its associated signaling networks have been drug targets but as Stevens pointed out, “the behavior of that signaling network is very non-intuitive. Sometimes if you hit one of the downstream components, it actually creates negative feedback, which actually increases the effect you are trying to inhibit.”

In the RAS project, the simulation is basically a molecular dynamics exercise conducted at various granularities extending all the way down to atomistic behavior including quantum effects. The computational power required, not surprisingly, depends on the level of granularity being simulated and can be substantial.

“Machine learning is being used to track the state space that the simulation is going through and to make decisions – do we zoom in here, do we zoom out, do we change the parameters that we are looking in a different part to the ensemble space. It’s basically acting like a smart supervisor of this simulation to more effectively use it.

“In some sense it’s like the network is watching a movie and saying, “OK, I’ve seen this part of the movie before, let’s fast forward, or wow this is really interesting I’ve never seen this before, let’s use slow motion and zoom in.” That’s sort of what the machine learning is doing in the simulation. It’s able to fast forward and skip around in some sense,” said Stevens.

The pre-clinical screening project, led by Stevens, is an ambitious effort to sift through basically as much cancer  preclinical and clinical data as it can lay hold of and combine that with new data generated from mouse models to build predictive models of drug-tumor interactions. It’s an in silico and experimental feedback approach. Ultimately, given a specific tumor whose molecular attributes (gene expression, SNPs, proteomics, etc) have been characterized, it should be possible to plug that data into a model to determine the best therapeutic regime.

The subtlety here, said Stevens, is there has been a lot of machine learning work in this done at kind of the small scale, that is on single classes of tumors or relatively small classes of drugs. “What we are trying to do with the deep learning is to integrate all of this information across thousands of cell lines, tens of thousands of compounds that have been screened against smaller number of cell lines, and then be able to project that into a mouse. You grow a colony of mice derived from that human tumor, and these mice become proxies for human clinical trials. So I can try different compounds on the colony of tumor mice to provide information about how my tumor might respond to them if given as a drug.”

A huge challenge is being able to make sense of all the historical data, much of which is unstructured and often subjective (e.g. pathology reports). “One of the first things that we have done is to build classifiers to tell us what type the tumor is or where the body site is [based on diverse data],” he said. Not infrequently the data may be suspect. “If it’s a new dataset we run it through our classifiers and they may say, really, this is not from the liver, it’s from some other place.”

As a rule, the preclinical data is outcome based; it doesn’t explain how the result was achieved.

“Right now we can build machine learning models that are pretty accurate at say predicting a drug response or tumor type or outcome but they can’t tell us very effectively why. They are not explanatory, not mechanistic,” said Stevens, “What we want to do is bring in mechanistic models or mechanistic data in some way and hybridize that with machine learning models so that we get two things. We get the ability to have a highly accurate predictable model but also a model that can explain why that prediction. So the idea of this hybrid approach is a wide open space and we think that this will generalize into many fields.” Obtaining large and high quality data for training models remains challenging, he said.

The third project strives to develop models able to make population scale forecasts, what Stevens call “patient trajectories.” It’s basically mining surveillance data across the country. Although somewhat dispersed, there is a great deal of patient data held by NCI, NIH, FDA, pharma, and payor organizations (pathology reports, treatments, outcomes, lifestyle, demographics, etc.). Unfortunately, like a lot of biomedical data, it’s largely unstructured. “We can’t really compute on it in the way we want to so we are using machine learning to translate the unstructured data into structured data we can compute on,” said Stevens

“So for example we want to read all the pathology reports with a machine and pull out, say the biomarkers, the mutational state, or the drugs and so on such that we can then build profiles that are consistent. Think of it as a population-based model. In preclinical screening pilot let’s say we uncover some treatments and strategies that are very effective on a certain type of cancer. We want to take that information and feed it into the population model and say “If this became a common therapy, how much would it change the statistics globally or nationally” or something like that.”

It’s also a way to link all of the pilots, said Steven. Insight from the RAS project might be later used to look at subclasses of cancers where the new treatment might work; that in turn put it into the population model to understand what the impact of that might be.

It’s still early days for the JDACS4C pilot projects, but hopes are high. Stevens noted both NCI and DOE are getting access to things they don’t readily have. “NCI does not have a lot of mathematicians and computer scientists, which DOE has. They also don’t have access to leadership machines. What we (DOE) are getting is access to all of this great experimental data, experimental facilities, public databases.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Graphcore Introduces Next-Gen Intelligence Processing Unit for AI Workloads

July 15, 2020

British hardware designer Graphcore, which emerged from stealth in 2016 to launch its first-generation Intelligence Processing Unit (IPU), has announced its next-generation IPU platform: the IPU-Machine M2000. With the n Read more…

By Oliver Peckham

heFFTe: Scaling FFT for Exascale

July 15, 2020

Exascale computing aspires to provide breakthrough solutions addressing today’s most critical challenges in scientific discovery, energy assurance, economic competitiveness, and national security. This has been the mai Read more…

By Jack Dongarra and Stanimire Tomov

There’s No Storage Like ATGC: Breakthrough Helps to Store ‘The Wizard of Oz’ in DNA

July 15, 2020

Even as storage density reaches new heights, many researchers have their eyes set on a paradigm shift in high-density information storage: storing data in the four nucleotides (A, T, G and C) that constitute DNA, a metho Read more…

By Oliver Peckham

Get a Grip: Intel Neuromorphic Chip Used to Give Robotics Arm a Sense of Touch

July 15, 2020

Moving neuromorphic technology from the laboratory into practice has proven slow-going. This week, National University of Singapore researchers moved the needle forward demonstrating an event-driven, visual-tactile perce Read more…

By John Russell

What’s New in HPC Research: Volcanoes, Mobile Games, Proteins & More

July 14, 2020

In this bimonthly feature, HPCwire highlights newly published research in the high-performance computing community and related domains. From parallel programming to exascale to quantum computing, the details are here. Read more…

By Oliver Peckham

AWS Solution Channel

INEOS TEAM UK Accelerates Boat Design for America’s Cup Using HPC on AWS

The America’s Cup Dream

The 36th America’s Cup race will be decided in Auckland, New Zealand in 2021. Like all the teams, INEOS TEAM UK will compete in a boat whose design will have followed guidelines set by race organizers to ensure the crew’s sailing skills are fully tested. Read more…

Intel® HPC + AI Pavilion

Supercomputing the Pandemic: Scientific Community Tackles COVID-19 from Multiple Perspectives

Since their inception, supercomputers have taken on the biggest, most complex, and most data-intensive computing challenges—from confirming Einstein’s theories about gravitational waves to predicting the impacts of climate change. Read more…

Joliot-Curie Supercomputer Used to Build First Full, High-Fidelity Aircraft Engine Simulation

July 14, 2020

When industrial designers plan the design of a new element of a vehicle’s propulsion or exterior, they typically use fluid dynamics to optimize airflow and increase the vehicle’s speed and efficiency. These fluid dyn Read more…

By Oliver Peckham

Graphcore Introduces Next-Gen Intelligence Processing Unit for AI Workloads

July 15, 2020

British hardware designer Graphcore, which emerged from stealth in 2016 to launch its first-generation Intelligence Processing Unit (IPU), has announced its nex Read more…

By Oliver Peckham

heFFTe: Scaling FFT for Exascale

July 15, 2020

Exascale computing aspires to provide breakthrough solutions addressing today’s most critical challenges in scientific discovery, energy assurance, economic c Read more…

By Jack Dongarra and Stanimire Tomov

Get a Grip: Intel Neuromorphic Chip Used to Give Robotics Arm a Sense of Touch

July 15, 2020

Moving neuromorphic technology from the laboratory into practice has proven slow-going. This week, National University of Singapore researchers moved the needle Read more…

By John Russell

Max Planck Society Begins Installation of Liquid-Cooled Supercomputer from Lenovo

July 9, 2020

Lenovo announced today that it is supplying a new high performance computer to the Max Planck Society, one of Germany's premier research organizations. Comprise Read more…

By Tiffany Trader

President’s Council Targets AI, Quantum, STEM; Recommends Spending Growth

July 9, 2020

Last week the President Council of Advisors on Science and Technology (PCAST) met (webinar) to review policy recommendations around three sub-committee reports: Read more…

By John Russell

Google Cloud Debuts 16-GPU Ampere A100 Instances

July 7, 2020

On the heels of the Nvidia’s Ampere A100 GPU launch in May, Google Cloud is announcing alpha availability of the A100 “Accelerator Optimized” VM A2 instance family on Google Compute Engine. The instances are powered by the HGX A100 16-GPU platform, which combines two HGX A100 8-GPU baseboards using... Read more…

By Tiffany Trader

Q&A: HLRS’s Bastian Koller Tackles HPC and Industry in Germany and Europe

July 6, 2020

In this exclusive interview for HPCwire – sadly not face to face – Steve Conway, senior advisor for Hyperion Research, talks with Dr.-Ing Bastian Koller about the state of HPC and its collaboration with Industry in Europe. Koller is a familiar figure in HPC. He is the managing director at High Performance Computing Center Stuttgart (HLRS) and also serves... Read more…

By Steve Conway, Hyperion

OpenPOWER Reboot – New Director, New Silicon Partners, Leveraging Linux Foundation Connections

July 2, 2020

Earlier this week the OpenPOWER Foundation announced the contribution of IBM’s A21 Power processor core design to the open source community. Roughly this time Read more…

By John Russell

Supercomputer Modeling Tests How COVID-19 Spreads in Grocery Stores

April 8, 2020

In the COVID-19 era, many people are treating simple activities like getting gas or groceries with caution as they try to heed social distancing mandates and protect their own health. Still, significant uncertainty surrounds the relative risk of different activities, and conflicting information is prevalent. A team of Finnish researchers set out to address some of these uncertainties by... Read more…

By Oliver Peckham

[email protected] Turns Its Massive Crowdsourced Computer Network Against COVID-19

March 16, 2020

For gamers, fighting against a global crisis is usually pure fantasy – but now, it’s looking more like a reality. As supercomputers around the world spin up Read more…

By Oliver Peckham

[email protected] Rallies a Legion of Computers Against the Coronavirus

March 24, 2020

Last week, we highlighted [email protected], a massive, crowdsourced computer network that has turned its resources against the coronavirus pandemic sweeping the globe – but [email protected] isn’t the only game in town. The internet is buzzing with crowdsourced computing... Read more…

By Oliver Peckham

Supercomputer Simulations Reveal the Fate of the Neanderthals

May 25, 2020

For hundreds of thousands of years, neanderthals roamed the planet, eventually (almost 50,000 years ago) giving way to homo sapiens, which quickly became the do Read more…

By Oliver Peckham

DoE Expands on Role of COVID-19 Supercomputing Consortium

March 25, 2020

After announcing the launch of the COVID-19 High Performance Computing Consortium on Sunday, the Department of Energy yesterday provided more details on its sco Read more…

By John Russell

Neocortex Will Be First-of-Its-Kind 800,000-Core AI Supercomputer

June 9, 2020

Pittsburgh Supercomputing Center (PSC - a joint research organization of Carnegie Mellon University and the University of Pittsburgh) has won a $5 million award Read more…

By Tiffany Trader

Honeywell’s Big Bet on Trapped Ion Quantum Computing

April 7, 2020

Honeywell doesn’t spring to mind when thinking of quantum computing pioneers, but a decade ago the high-tech conglomerate better known for its control systems waded deliberately into the then calmer quantum computing (QC) waters. Fast forward to March when Honeywell announced plans to introduce an ion trap-based quantum computer whose ‘performance’ would... Read more…

By John Russell

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

By Doug Black

Leading Solution Providers

Contributors

Nvidia’s Ampere A100 GPU: Up to 2.5X the HPC, 20X the AI

May 14, 2020

Nvidia's first Ampere-based graphics card, the A100 GPU, packs a whopping 54 billion transistors on 826mm2 of silicon, making it the world's largest seven-nanom Read more…

By Tiffany Trader

‘Billion Molecules Against COVID-19’ Challenge to Launch with Massive Supercomputing Support

April 22, 2020

Around the world, supercomputing centers have spun up and opened their doors for COVID-19 research in what may be the most unified supercomputing effort in hist Read more…

By Oliver Peckham

Australian Researchers Break All-Time Internet Speed Record

May 26, 2020

If you’ve been stuck at home for the last few months, you’ve probably become more attuned to the quality (or lack thereof) of your internet connection. Even Read more…

By Oliver Peckham

15 Slides on Programming Aurora and Exascale Systems

May 7, 2020

Sometime in 2021, Aurora, the first planned U.S. exascale system, is scheduled to be fired up at Argonne National Laboratory. Cray (now HPE) and Intel are the k Read more…

By John Russell

Summit Supercomputer is Already Making its Mark on Science

September 20, 2018

Summit, now the fastest supercomputer in the world, is quickly making its mark in science – five of the six finalists just announced for the prestigious 2018 Read more…

By John Russell

TACC Supercomputers Run Simulations Illuminating COVID-19, DNA Replication

March 19, 2020

As supercomputers around the world spin up to combat the coronavirus, the Texas Advanced Computing Center (TACC) is announcing results that may help to illumina Read more…

By Staff report

$100B Plan Submitted for Massive Remake and Expansion of NSF

May 27, 2020

Legislation to reshape, expand - and rename - the National Science Foundation has been submitted in both the U.S. House and Senate. The proposal, which seems to Read more…

By John Russell

John Martinis Reportedly Leaves Google Quantum Effort

April 21, 2020

John Martinis, who led Google’s quantum computing effort since establishing its quantum hardware group in 2014, has left Google after being moved into an advi Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This