Preparing for the Next Pandemic: Developing an Open Science Platform for Better Decision-Making in Public Health

By Doug Eadline

August 16, 2024

Ask any Epidemiologist about another pandemic, and they will reply, “It is not a question of if, but of when.” Many hard lessons were learned during the height of the recent COVID-19 pandemic (which continues into the fall of 2024). One of these lessons was the large amounts of heroic, overlapping, and redundant work that was taking place. In this interview, HPCwire talks with Dr. Jonathan Ozik and Dr. Valerie Hayot-Sasson about algorithm driven HPC workflows, improved data ingestion, curation, and management capabilities, and a shared development environment for rapid response and collaboration. One of the key goals is developing an open science data-flow platform, using tools like those developed by the Globus Project, to aid in better and timely public health decisions.


HPCwire: Hello and welcome. I’m Doug Eadline, managing editor of HPCwire. Today, we’re going to be talking about how the lessons from COVID-19 have helped shape better ways to manage resources and distribute critical lifesaving information. COVID-19 had an unprecedented impact on scientific collaboration, the pandemic and its broad response from the scientific community has forged new relationships among public health practitioners, mathematical modelers and scientific computing specialists, while revealing critical gaps in exploiting advanced computing systems to support urgent decision making.

Today, we’re speaking with Dr. Jonathan Ozik and Dr. Valerie Hayot-Sasson by way of introduction, Dr. Valerie Hayot-Sasson is a postdoctoral scholar at the University of Chicago and holds a joint appointment at Argonne National Laboratory. Her interests focus on improving accessibility and scientific research by developing software solutions that speed up processing of workflows or improve ease of use. Doctor Jonathan Ozik is a principal computational scientist at Argonne National Laboratory, senior scientist at the consortium for Advanced Science and Engineering with Public Health Science affiliation at the University of Chicago and Senior Institute Fellow, Northwestern Oregon Institute of Science and Engineering, Northwestern University. Doctor Ozik leads multiple US National Science Foundation and National Institute of Health projects, some of which we’re going to talk about today.

So first, I want to thank both of you for being here. And I’d like to start off with Jonathan and ask the basic, you know, COVID pandemic brought together valiant efforts from many sectors, yet there were many challenges that seemed to occur, and trying to support critical public health decision making and data modeling seemed a little strained while this was going on, given that there will be more pandemics. What lessons did you learn or have we learned from COVID-19? Thank you.

Dr. Jonathan Ozik: Doug, and thank you for having us on. So, let me first provide some background on our experiences. So, during COVID, our group dropped what we were doing. And as part of DOE’s multi lab and NVBL effort, we were asked to support the nation’s effort in the COVID response. For us at Argonne, we were one of four academic and national lab groups that made up Illinois Governor’s COVID-19 Task Force, which had us meeting twice weekly to develop and provide analyses to the Illinois and Chicago departments of Public Health and the governor’s and the Chicago mayor’s offices. We brought to bear our previous experiences with HPC workflows, machine learning, and large-scale agent-based epidemiological modeling. But unlike research efforts, we were responding to rapidly evolving policy questions as our understanding of Covid epidemiology changed as well.

So this experience in working directly with public health stakeholders and supporting their decision making was, on one hand very gratifying and on the other extremely difficult. So based on these experiences, we identified ways in which we thought we could improve our ability to enhance evidence based decision making through better use of compute, automation and analytics pipelines. And since then, our goal has been to develop and provide these capabilities to the public health modeling and analysis community as an open science platform. What we experienced with COVID is very likely not a one off event. The better question isn’t if something like that will happen again, but when.

HPCwire: During this your initial efforts into this, what were some of the gaps you found that slowed things down or were real challenges in trying to deliver this information?

Dr. Jonathan Ozik: That’s a really good question. So during COVID, individual research groups across the world really were independently using HPC data management, machine learning and AI and attempting to use automation methods to develop, calibrate, modify, verify and validate their epidemiological models, which involved really large amounts of heroic and overlapping work and also unfortunately, lacked generalizability. So we identified three broad areas where we thought we could push the field forward. So the first is we determined that there was a need for integrated algorithm driven HPC workflows.

This integration is critical in bringing three important areas of computational science, which are simulation, large scale workflow coordination, and then the algorithms that strategically and efficiently guide the simulation and machine learning based analysis. These workflows need to coordinate across distributed and heterogeneous resources, and this is because epidemiological modeling includes a range of different types of computational tasks. They also need to be fault tolerant. They need to be secure, and they also need to facilitate automated access to these heterogeneous distributed resources. And also scalability is important to handle the varying demands of epidemiological workflows, as is really fast time to solution approaches that can provide actionable insights quickly.

Dr. Jonathan Ozik: So that was the first area. The second area is really a need for improved data ingestion, curation, and management capabilities. With all the different data sources that kept changing over time, there was a need to access, move, and track the diverse data sets from their origin to their use within our computational analyses. And there was also an overwhelming need to automate data curation, through data analysis pipelines for data de-biasing, integration, uncertainty quantification, and also metadata and provenance tracking. And the third broad area was a real need for a shared development environment for rapid response and collaboration.

So, this includes the ability to quickly and efficiently share models and create portable workflows that can run across federated HPC systems and systems to house models include how to reproduce, extend, or scale them by others. Based on these requirements, we’re developing the Open Science Platform for Robust Epidemic analysis, or Ospreys, we call it, and we’re taking a decentralized approach that leverages cloud services to connect compute and storage across national laboratory and university resources.

HPCwire: That’s a lot. Can you touch on the decentralized approach and the tools you’re using to connect all these services?

Dr. Jonathan Ozik: Sure thing. I’m happy to start. And then Valerie can expand. So one aspect of decentralization is that different researchers will have different compute and storage resources that they can access. And we want to be able to allow for any combination of these to support their epidemiological analysis needs. The second aspect of decentralization has to do with the data that, some of which might be, you know, proprietary or sensitive, which requires us to really move away from a central monolithic storage or compute requirement.

Dr. Valerie Hayot-Sasson: Yeah. So because we needed to find a solution that can work with all the resources available to researchers, whether that be workstations, high performance computing or cloud, we needed to look at solutions which could seamlessly use all of these services. We decided to use Globus services, which have been designed with the scientific community and diverse and distributed resources in mind. While the services themselves are cloud-hosted, they operate on a hybrid model. These services are also widely used by the scientific community and are already deployed in most research. Cyber infrastructure, making adoption of the tools easier.

HPCwire: So here’s a question I have. I’m always thinking of like you have all this infrastructure and everything. What does it look like to the end user or what? What are you planning to make it look like to the end user?

Dr. Jonathan Ozik: So this is our vision for the end user as an analyst on the bottom left that defines a policy or a timer that reaches out to data portals. And then data from these portals are ingested automatically then, verified, validated, transformed as needed, and then stored in an accessible location, and then subsequently analyses are kicked off with this newly updated data, where we employ HPC workflows to generate, for example, estimates of epidemiological parameters and forecasts of future trends and public health stakeholders are able to access these analysis products and in turn, they provide additional data into the data portals. The ultimate vision is that all of this happens in the background and individuals to access automatically generated analyses, similarly to how we experience, let’s say, weather forecasting. Now Valerie will introduce the automated events based research orchestration or aero tool that we have been building together.

Dr. Valerie Hayot-Sasson: Okay, so this is kind of what it looks like behind the scenes. In this figure, we can see that there are two different types of flows, each of which will be Globus flows. So on one side you have the on the left hand side you have the data ingestion, verification and curation. And on the right hand side you have the data analysis flow. Now for the data ingestion, verification, and curation. This flow is typically used to process data that are not created by an arrow flow. So for example this is for consolidating data made available by different sources. We can think of open data portals hospitals etc. and preparing it for future analysis. So preparing it for the right hand side of or like a right hand side flow.

These flows are expected to be periodical in nature, only being updated at the same cadence in which the data portals are updated. And now if we look at the right hand side, we have our analysis flow. These flows typically operate on data already available in the system. So unlike the data ingestion flows and, produce outputs which can then be used as inputs within other analysis flows, and in this case users can define two types of policies, either timer based or on a database update policies to relaunch their analyses.

HPCwire: Yeah, thanks. That’s interesting. And what struck me was the “weather forecast” ability to I could just imagine during the COVID pandemic if, if the ability to put up heat maps, as it were, of COVID and so forth, just like with the nightly weather forecast, how that may have helped many people. Guessing rather than guessing about what was going on or getting misinformation from who knows where. Is there any place where this is being used right now and is it ready to run, or where do you see this going at this point?

Dr. Jonathan Ozik: So we are working with the Chicago Department of Public Health to integrate these capabilities into their processes. Now, they have been both very supportive of our work and really keen to incorporate what we develop into evidence-based decision-making as we develop the infrastructure. We’re also co-developing test use cases, for example, being able to automatically ingest wastewater data and kick off automated analyses as data is updated. And this is of particular interest. As it turns out, due to the passive nature of this type of data source, especially, for example, as reporting requirements for COVID have largely gone away. So, as a result, previously available testing or hospitalization are not as readily available, if at all. However, I do want to say that our goal is to go beyond Department of Public Health partners, but to have researchers be able to build on what we build to further advance the role of computation in supporting decision making in public health.

Dr. Valerie Hayot-Sasson: And what’s great about Arrow is that it’s very generalizable. So at its base, it’s fundamentally a data sharing platform. And the data types and storage it uses can be at the discretion of the users. The same can be said for computation, which can be executed on any infrastructure that you, the user, has and enables users to describe their analysis using software they’re more most familiar with. So this kind of results in a type of bring-your-own infrastructure model.

HPCwire: So, Valerie that intrigues me. What is a bring-your-own infrastructure model?

Dr. Valerie Hayot-Sasson: So the bring your own infrastructure model means that both compute and storage resources are actually provided by the users. This results in users maintaining full ownership of their data, as the data never gets transmitted to the error servers, and the users get to reuse their configured infrastructure for their services, not having to specify to Aero how to install their code. All the users need to do when specifying a flow is to provide their Globus compute endpoint information, so that arrow knows where the automated flow needs to be executed, and a Globus collection URL that provides information to error on where to store and retrieve the data from.

HPCwire: So this is all very interesting to me. The fact that globally, locally bringing data together and making it usable, actionable data in times of, you know, in this case pandemic. So, what really excites you about taking this beyond where we are now with it today?

Dr. Valerie Hayot-Sasson: So the world is filled with problems that are worthy of investigation. And researchers progress is often hindered by these repetitive tasks that could easily be automated. So the goal of these tools is really to enhance the day to day lives of researchers, such that they can focus on their energy, on the problems that they set out to solve, rather than spend significant amounts of time on these repetitive tasks. And so what I hope when we build these kinds of tools is that it can ultimately lead to faster discoveries. And yeah.

HPCwire: I think being able to manage data and bring it to bear is really one of the challenges we have at this point in, in HPC and in many other areas as well. You don’t hear a lot anymore about big data, although it’s still there and being able to manipulate, move and calculate with big data, I think is really important. And to, really provide these tools I think is a great effort on your part. Jonathan, any closing comments?

Dr. Jonathan Ozik: Yes, Doug. Thank you. Really glad that we had this opportunity to sit down and present to you all the work that we’re doing now and then that we’re hoping to do to really expand the ways in which computational science can really push this field forward.

HPCwire: I want to thank you as well, both you and Valerie. It’s been a great interview hearing about this, that how this can apply to many different areas and in particularly with, with pandemics, which is very clear; the lessons learned from COVID that we had some work to do to get things right and hope to hear more in the future. And we’ll be talking to you soon.


This material is based upon work supported by the National Science Foundation under Grant 2200234, the U.S. Department of Energy, Office of Science, under contract number DE-AC02-06CH11357 and the Bio-preparedness Research Virtual Environment (BRaVE) initiative. This research was completed with resources provided by the Research Computing Center at the University of Chicago, the Laboratory Computing Resource Center at Argonne National Laboratory, and the Argonne Leadership Computing Facility, which is a DOE Office of Science User Facility.

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks.  These benchmarks have focused on mathematical ML operations and accelerators (e.g., Nvidia GPUs). Recently, MLCommons introduced the results of i Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago today emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whatever physical processor they want, without making code changes, the Read more…

IBM Quantum Summit Evolves into Developer Conference

October 2, 2024

Instead of its usual quantum summit this year, IBM will hold its first IBM Quantum Developer Conference which the company is calling, “an exclusive, first-of-its-kind.” It’s planned as an in-person conference at th Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed that the company will release Falcon Shores as a GPU. The com Read more…

Texas A&M HPRC at PEARC24: Building the National CI Workforce

October 1, 2024

Texas A&M High-Performance Research Computing (HPRC) significantly contributed to the PEARC24 (Practice & Experience in Advanced Research Computing 2024) conference. Eleven HPRC and ACES’ (Accelerating Computin Read more…

A Q&A with Quantum Systems Accelerator Director Bert de Jong

September 30, 2024

Quantum technologies may still be in development, but these systems are evolving rapidly and existing prototypes are already making a big impact on science and industry. One of the major hubs of quantum R&D is the Q Read more…

The New MLPerf Storage Benchmark Runs Without ML Accelerators

October 3, 2024

MLCommons is known for its independent Machine Learning (ML) benchmarks.  These benchmarks have focused on mathematical ML operations and accelerators (e.g., N Read more…

DataPelago Unveils Universal Engine to Unite Big Data, Advanced Analytics, HPC, and AI Workloads

October 3, 2024

DataPelago today emerged from stealth with a new virtualization layer that it says will allow users to move AI, data analytics, and ETL workloads to whatever ph Read more…

Stayin’ Alive: Intel’s Falcon Shores GPU Will Survive Restructuring

October 2, 2024

Intel's upcoming Falcon Shores GPU will survive the brutal cost-cutting measures as part of its "next phase of transformation." An Intel spokeswoman confirmed t Read more…

How GenAI Will Impact Jobs In the Real World

September 30, 2024

There’s been a lot of fear, uncertainty, and doubt (FUD) about the potential for generative AI to take people’s jobs. The capability of large language model Read more…

IBM and NASA Launch Open-Source AI Model for Advanced Climate and Weather Research

September 25, 2024

IBM and NASA have developed a new AI foundation model for a wide range of climate and weather applications, with contributions from the Department of Energy’s Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

Building the Quantum Economy — Chicago Style

September 24, 2024

Will there be regional winner in the global quantum economy sweepstakes? With visions of Silicon Valley’s iconic success in electronics and Boston/Cambridge� Read more…

How GPUs Are Embedded in the HPC Landscape

September 23, 2024

Grasping the basics of Graphics Processing Unit (GPU) architecture is crucial for understanding how these powerful processors function, particularly in high-per Read more…

Shutterstock_2176157037

Intel’s Falcon Shores Future Looks Bleak as It Concedes AI Training to GPU Rivals

September 17, 2024

Intel's Falcon Shores future looks bleak as it concedes AI training to GPU rivals On Monday, Intel sent a letter to employees detailing its comeback plan after Read more…

Nvidia Shipped 3.76 Million Data-center GPUs in 2023, According to Study

June 10, 2024

Nvidia had an explosive 2023 in data-center GPU shipments, which totaled roughly 3.76 million units, according to a study conducted by semiconductor analyst fir Read more…

AMD Clears Up Messy GPU Roadmap, Upgrades Chips Annually

June 3, 2024

In the world of AI, there's a desperate search for an alternative to Nvidia's GPUs, and AMD is stepping up to the plate. AMD detailed its updated GPU roadmap, w Read more…

Granite Rapids HPC Benchmarks: I’m Thinking Intel Is Back (Updated)

September 25, 2024

Waiting is the hardest part. In the fall of 2023, HPCwire wrote about the new diverging Xeon processor strategy from Intel. Instead of a on-size-fits all approa Read more…

Ansys Fluent® Adds AMD Instinct™ MI200 and MI300 Acceleration to Power CFD Simulations

September 23, 2024

Ansys Fluent® is well-known in the commercial computational fluid dynamics (CFD) space and is praised for its versatility as a general-purpose solver. Its impr Read more…

Shutterstock_1687123447

Nvidia Economics: Make $5-$7 for Every $1 Spent on GPUs

June 30, 2024

Nvidia is saying that companies could make $5 to $7 for every $1 invested in GPUs over a four-year period. Customers are investing billions in new Nvidia hardwa Read more…

Shutterstock 1024337068

Researchers Benchmark Nvidia’s GH200 Supercomputing Chips

September 4, 2024

Nvidia is putting its GH200 chips in European supercomputers, and researchers are getting their hands on those systems and releasing research papers with perfor Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Leading Solution Providers

Contributors

Everyone Except Nvidia Forms Ultra Accelerator Link (UALink) Consortium

May 30, 2024

Consider the GPU. An island of SIMD greatness that makes light work of matrix math. Originally designed to rapidly paint dots on a computer monitor, it was then Read more…

Quantum and AI: Navigating the Resource Challenge

September 18, 2024

Rapid advancements in quantum computing are bringing a new era of technological possibilities. However, as quantum technology progresses, there are growing conc Read more…

IBM Develops New Quantum Benchmarking Tool — Benchpress

September 26, 2024

Benchmarking is an important topic in quantum computing. There’s consensus it’s needed but opinions vary widely on how to go about it. Last week, IBM introd Read more…

Google’s DataGemma Tackles AI Hallucination

September 18, 2024

The rapid evolution of large language models (LLMs) has fueled significant advancement in AI, enabling these systems to analyze text, generate summaries, sugges Read more…

Microsoft, Quantinuum Use Hybrid Workflow to Simulate Catalyst

September 13, 2024

Microsoft and Quantinuum reported the ability to create 12 logical qubits on Quantinuum's H2 trapped ion system this week and also reported using two logical qu Read more…

IonQ Plots Path to Commercial (Quantum) Advantage

July 2, 2024

IonQ, the trapped ion quantum computing specialist, delivered a progress report last week firming up 2024/25 product goals and reviewing its technology roadmap. Read more…

Intel Customizing Granite Rapids Server Chips for Nvidia GPUs

September 25, 2024

Intel is now customizing its latest Xeon 6 server chips for use with Nvidia's GPUs that dominate the AI landscape. The chipmaker's new Xeon 6 chips, also called Read more…

US Implements Controls on Quantum Computing and other Technologies

September 27, 2024

Yesterday the Commerce Department announced  export controls on quantum computing technologies as well as new controls for advanced semiconductors and additiv Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire