SC13 Talk Pushes HPC in New Educational Directions

By Nicole Hemsoth

November 19, 2013

According to Dr. Thom Dunning from University of Illinois at Urbana-Champaign, science and engineering research has been revolutionized by computation but, to date computing has largely been used to organize, prepare for and disseminate courses. Dunning says that “The potential of using these technologies to teach students the fundamental principles of a subject through authentic computational simulation is largely unexplored.”

During a presentation today (Tuesday) as part of the SC13 Invited Speakers program, Dunning will discuss how “computational tools and simulations has been used to help teach high school and college chemistry. The simulations enable students to gain a deep, rich appreciation for the basic principles of chemistry. Further, the use of computational tools is enthusiastically embraced by teachers and students, results in improved performance of both, and leads to increased student interest in chemistry.”

Dunning’s talk is at 1:30 today at SC13 in Denver…a full interview is below that hits on the main topics he is set to discuss, plus a broader view of what tools and opportunities HPC pushes to other domains.

HPCwire: Can you give us a sense of what the various roles are for HPC in education—where does it fit in with scientific and broader educational goals?

Dunning: The overarching thesis I have is, as computational simulation has revolutionized scientific and engineering research, it can also revolutionize scientific and engineering education.

It’s very clear that computing and particular simulations had a major impact on most fields of science. It adds a new way for us to understand how the world operates, whether it’s the chemical world, the physical world, the biological world or some combination of all of them, which is of course what happens with global climate simulation. And we really have to think about how it we effectively use those kinds of simulations in education.

I’m a chemist by background, and all chemical phenomena are determined by molecules. But students never see molecules. It might at times almost sound to them like a fiction. But in fact they’re very real, and what you need to do for the student is to make them real. And that’s something we can do with computational simulation.

And in that case it could be as simple as allowing them to draw a molecule on a computer screen and manipulate it—turn it around, look at various sides of it, zoom in and out. Or it could be as sophisticated as taking a molecule and computing its vibrational spectra and then tying that in with global warming.

A teacher might ask why it is that the CO2 molecule and the buildup of CO2 in the atmosphere leads to global warming. To help students answer the question, the teacher could use computational simulation to show the students how the structure and the way the molecule moves is tied to the global warming phenomenon.

We have a project that has now been completed at NCSA for six years, called the Institute for Chemical Literacy through Computational Science. Although the name says “computational science,” computational simulation was really key. There we were looking at tools that would allow the teachers to work with students to visualize these kinds of molecular phenomena, and we saw a significant increase in not only teacher knowledge, but we also saw a significant increase in student knowledge as a result of this intervention.

HPCwire: So it sounds as though making science real and tangible then is one of the core components of what you’ll be talking about.

Dunning: Yes, and if you think about it, this is one of the issues that we have in science: many of the fundamental, underlying concepts from continental drift to the inner workings of the universe aren’t easily visualized.

But by having these simulations one can really make all of this very much more real for students and I think that’s the major benefit of computational simulation in the classroom. It really makes whatever the student is studying more real because they can interact with it and watch how it responds and thereby learn the behavior they’re supposed to be studying.

HPCwire: You’re a chemist by training, so is this the way you found your way into the high performance computing space?

Dunning: Yes, it certainly was a factor when I was an undergraduate. I went to a university that, when I was a sophomore, had just gotten a new computer: an IBM 1620. I was taking a course in quantum mechanics and suddenly it dawned on me that quantum mechanics was actually the basis for all of the chemistry that I had been taking. Then I started looking at what the equations looked like and I quickly realized I couldn’t solve them, and it took me a little while longer to realize that nobody could solve some of them. But with computers you could.

So that really got me interested in the use of computing to study various kinds of chemical research problems. And even for me it made it a lot more real—I was actually doing something. And I was manipulating chemistry at its most fundamental level: the molecular level.

When you go into the laboratory and you pour two solutions together, the outcome is being determined at that fundamental level, but that’s not what you see and that’s not what you experience. But when you actually do the computational simulations you can get a direct experience.

HPCwire: Speaking of that new generation, you’ve seen a number of trends come and go when it comes to computer scientists and students interested in HPC. Have you seen a dip or rise or stagnation around high performance computing as a career path? What ways are people coming into it? What trends have you seen over time?

Dunning: Well I guess what we’re seeing really is an increasing number of scientific communities that really need HPC. Either the models they use to describe the phenomena they’re interested in have gotten sufficiently complicated and sophisticated that now they need HPC to solve them, or on the data side communities now get to the point where the amount of data that they’re generating also requires HPC to be able to manage the data and to analyze it. So I think this is an area that is going to continue to grow over the next decade. And of course in chemistry it’s grown tremendously since I’ve been in the field.

I like to say that when I first started in the 1960s, if you had a conference that had a couple hundred theoretical and computational chemists in it, that was the world’s supply. Now there are thousands of theoretical and computational chemists—even tens of thousands. But even more important than that, even experimental chemists now routinely use computational simulation to gain insights into the experiments that they do. So it’s really expanded far beyond the folks who have traditionally done that type of work, to the point that I would say half of all chemists use computational simulations in one way or another.

HPCwire: So a lot of the interest in HPC then is being fed by the actual end-users of some of these applications? What about on the other side of it—people who are actively pursuing a career in high performance systems to help empower these systems. Are there some trends that you’ve noticed there?

Dunning: I don’t think I can be definitive there, but I can at least give you an impression. My impression is that when I started back in the 60s, most of the people that I ran into who were computer scientists were focused on scientific computing and what would today be called high performance computing, although it was CDC 6600s and things like that back then.

Now computing has gotten to be such a large, broad field that the number of computer scientists who are actually focused on high performance computing is now a pretty small fraction of the total field of computer science. And I would say that that area has certainly not seen the growth that computer science as a whole has seen because of exactly that.

HPCwire: Could you explain in more detail what some of those reasons may be?

Dunning: I think that for people who have an interest in computing, back in the 60s and 70s, scientific computing was the hot thing to do. In fact, it was the only thing you could do. Now if you’re interested in computing and you want to be a computer scientist, there’s a very broad field that you could be involved in: everything from the Web to real-time processing or embedded processing, and high performance computing is only a very small fraction of computer science at this point.

HPCwire: I see a lot of job postings and kind of “hipness” around the big data jobs and the label of “data scientist,” which is a way to call advanced computing something else completely. Maybe you just answered this question by what you just said, but do you want to comment on that transition into this culture of big data? There’s a piece of it in HPC, and a lot of these guys say, “Well, HPC has always been big data, it’s just data, and yes it’s gotten bigger.” There’s a little bit of push and pull between those two cultures, but there’s really nothing new.

Dunning: I think there’s a measure of truth to both of the comments. I think it’s correct for people in HPC to say, “We’ve been dealing with big data basically from the beginning. It’s been growing year by year.” And that’s true. On the other hand there are new sources of data now. There are sensor networks that are being put in place. All of the telescopes that are designed now basically are digital instruments—they are no longer analog instruments.

So there’s a huge explosion that’s occurring now both inside and outside of HPC. And I think that’s the big change that’s happened over the course of the past few years.

When I first came to NCSA in 2004 we were talking about the rapid increase in the growth of data. It took NCSA the first 19 years of its existence to archive a petabyte of data. By the twentieth year we had a second petabyte of data. And we’ve been adding substantial amounts of data into the archive ever since then.

But now we’re talking about instruments like the LSST, the Large Synoptic Survey Telescope. Once it’s in operation it will, by itself, generate something on the order of 100 to 200 petabytes of data. So in some ways we’re not quite there yet, but it’s coming at us pretty fast at this point. For these other sources of data are going to become as dominant if not more dominant than the amount of data that’s produced in high performance computing.

And I think that one of the things that’s going to happen here is that it’s not only going to revolutionize science and engineering, (the telescopes being a good example of how it revolutionizes science,) but it’s also going to revolutionize the behavioral, social and economic sciences also. Because as they get to the point that they start collecting the data, which is now so much easier to do because so much data is born digitally to begin with, they’re going to get new insights into the problems they’re studying that wouldn’t be possible to get otherwise.

HPCwire: It sounds like that brings us full circle back to the work that you’ll eventually be doing with the sustainable cities project, where you’re gathering sociological data plus sensor data plus GIS and other types of data, right?

Dunning: That’s correct. Looking at cities is putting all of that information together and really understanding what that data is telling you about how you can make the city operate more efficiently both energy-wise and people-wise.

HPCwire: Does this intersect at all with a project in the Middle East that Thomas Zachariah was doing regarding connected cities. Are these similar in nature?

Dunning: I’m not sure exactly what it is that Thomas is doing, but at NYU there’s a center called the Center for Urban Science and Policy and that’s one of its goal also: to use the city of New York as a testbed for understanding how you can improve the efficiency of cities and improve the quality of life in cities.

And a lot of this is drive by the fact that right now. I believe, in the U.S. about 50 percent of the population lives in cities and by the time we get to 2050 it’s going to be more like 70 percent. And this is a trend that’s happening worldwide, not just in the U.S So if you’re really going to focus on a major problem where you need to enhance energy efficiency and sustainability, cities are a natural testbed for that and one of the problems you really want to settle.

HPCwire: Is there any other takeaway from the talks you’re expected to give at SC that you want to share here and now?

Dunning: I guess my goal is to make this a call to arms—to really get the folks in HPC interested in talking to their colleagues in science and engineering.

Right now most of that is done in research, and I think it’s time for us to start that discussion on the teaching side. And I know here at Illinois I’ve had conversations with my colleagues on the faculty here and there is a lot of pent-up desire to introduce computational simulation into the undergraduate classroom because they know it could have a big impact. But it really takes a partnership between the folks in HPC and the faculty to be able to bring that about in each of the universities.

So I’m kind of hoping that people come away from my talk saying “You know, I need to go over and have a talk with Joe in Physics because I know he’s interested in computational simulation. I wonder what his thoughts are on using computational simulation in the classroom?”

I think it would just be terrific if you had a class in meteorology and the students each got to run their weather forecasts. In fact you change the initial conditions slightly so that all of a sudden when you say, “There’s a 30 percent chance of rain,” the students would really have a real understanding of what that meant because they could look at all the different simulations that all the other students ran to see in 30 percent of them there was rain in this area. It would make it really concrete as to what terms like that mean, which now people sort of having a feeling for, but not an in-depth understanding.

Same kind of thing with climate simulation. Having students run climate simulations and understand where some of the issues might be in some of the simulations and where the simulations are very stable. They’re all predicting very similar kinds of trends, which would give you much more confidence in that particular number. Then seeing where they aren’t producing things that are as consistent, so maybe that’s an area where the models need to be improved.

HPCwire: By the way, what are you up to now? We recently published news that you’ve retired as NCSA director…what next?

Dunning: Well, that doesn’t actually happen until December 31st, so I’m still officially the NCSA director. After that I will be joining Pacific Northwest National Laboratory and the University of Washington to help them set up a joint institute for advanced computing between the two institutions.

They’re really just formulating it at this point. This is an entirely new institute, so one of the things that we’ll be doing over the next couple of years is looking at areas of common interest between the institutions and looking for opportunities for them to collaborate on various projects in advanced computing.

And it will be both what I refer to as “scientific computing” and data analysis, which is the fundamental base for the activity, but it will also look at various kinds of applications. For example: exascale science and engineering, which involves looking at the science of cities to try to understand how they operate and how one can make them more sustainable and energy efficient.

But those specific areas have not yet been decided on—we’ll be holding workshops over the next couple of years to define what some of those major thrusts will be. And then there will be lots of individual collaborations between staff at Pacific Northwest National Laboratory and faculty at the University of Washington.

HPCwire: That doesn’t sound much like retirement.

Dunning: I really was planning on retiring. Although this is the second time I’ve tried to retire and as soon as I do people come knocking on my door, saying, “Hey, I’ve got something great for you to do.”

This one actually holds a special place for me in that when I was at Pacific Northwest National Laboratory one of my goals was, in fact, to get the two institutions, who are the two major research institutions in the state of Washington, working more closely together. So this is something that’s close to my heart.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Nvidia Aims Clara Healthcare at Drug Discovery, Imaging via DGX

April 12, 2021

Nvidia Corp. continues to expand its Clara healthcare platform with the addition of computational drug discovery and medical imaging tools based on its DGX A100 platform, related InfiniBand networking and its AGX develop Read more…

Nvidia Serves Up Its First Arm Datacenter CPU ‘Grace’ During Kitchen Keynote

April 12, 2021

Today at Nvidia’s annual spring GPU technology conference, held virtually once more due to the ongoing pandemic, the company announced its first ever Arm-based CPU, called Grace in honor of the famous American programmer Grace Hopper. Read more…

Nvidia Debuts BlueField-3 – Its Next DPU with Big Plans for an Expanded Role

April 12, 2021

Nvidia today announced its next generation data processing unit (DPU) – BlueField-3 – adding more substance to its evolving concept of the DPU as a full-fledged partner to CPUs and GPUs in delivering advanced computi Read more…

Nvidia’s Newly DPU-Enabled SuperPOD Is a Multi-Tenant, Cloud-Native Supercomputer

April 12, 2021

At GTC 2021, Nvidia has announced an upgraded iteration of its DGX SuperPods, calling the new offering “the first cloud-native, multi-tenant supercomputer.” The newly announced SuperPods come just two years after the Read more…

Tune in to Watch Nvidia’s GTC21 Keynote with Jensen Huang – Recording Now Available

April 12, 2021

Join HPCwire right here on Monday, April 12, at 8:30 am PT to see the Nvidia GTC21 keynote from Nvidia’s CEO, Jensen Huang, livestreamed in its entirety. Hosted by HPCwire, you can click to join the Huang keynote on our livestream to hear Nvidia’s expected news and... Read more…

AWS Solution Channel

Volkswagen Passenger Cars Uses NICE DCV for High-Performance 3D Remote Visualization

 

Volkswagen Passenger Cars has been one of the world’s largest car manufacturers for over 70 years. The company delivers more than 6 million automobiles to global customers every year, from 50 production locations on five continents. Read more…

The US Places Seven Additional Chinese Supercomputing Entities on Blacklist

April 8, 2021

As tensions between the U.S. and China continue to simmer, the U.S. government today added seven Chinese supercomputing entities to an economic blacklist. The U.S. Entity List bars U.S. firms from supplying key technolog Read more…

Nvidia Serves Up Its First Arm Datacenter CPU ‘Grace’ During Kitchen Keynote

April 12, 2021

Today at Nvidia’s annual spring GPU technology conference, held virtually once more due to the ongoing pandemic, the company announced its first ever Arm-based CPU, called Grace in honor of the famous American programmer Grace Hopper. Read more…

Nvidia Debuts BlueField-3 – Its Next DPU with Big Plans for an Expanded Role

April 12, 2021

Nvidia today announced its next generation data processing unit (DPU) – BlueField-3 – adding more substance to its evolving concept of the DPU as a full-fle Read more…

Nvidia’s Newly DPU-Enabled SuperPOD Is a Multi-Tenant, Cloud-Native Supercomputer

April 12, 2021

At GTC 2021, Nvidia has announced an upgraded iteration of its DGX SuperPods, calling the new offering “the first cloud-native, multi-tenant supercomputer.” Read more…

Tune in to Watch Nvidia’s GTC21 Keynote with Jensen Huang – Recording Now Available

April 12, 2021

Join HPCwire right here on Monday, April 12, at 8:30 am PT to see the Nvidia GTC21 keynote from Nvidia’s CEO, Jensen Huang, livestreamed in its entirety. Hosted by HPCwire, you can click to join the Huang keynote on our livestream to hear Nvidia’s expected news and... Read more…

The US Places Seven Additional Chinese Supercomputing Entities on Blacklist

April 8, 2021

As tensions between the U.S. and China continue to simmer, the U.S. government today added seven Chinese supercomputing entities to an economic blacklist. The U Read more…

Habana’s AI Silicon Comes to San Diego Supercomputer Center

April 8, 2021

Habana Labs, an Intel-owned AI company, has partnered with server maker Supermicro to provide high-performance, high-efficiency AI computing in the form of new Read more…

Intel Partners Debut Latest Servers Based on the New Intel Gen 3 ‘Ice Lake’ Xeons

April 7, 2021

Fresh from Intel’s launch of the company’s latest third-generation Xeon Scalable “Ice Lake” processors on April 6 (Tuesday), Intel server partners Cisco, Dell EMC, HPE and Lenovo simultaneously unveiled their first server models built around the latest chips. And though arch-rival AMD may... Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

Julia Update: Adoption Keeps Climbing; Is It a Python Challenger?

January 13, 2021

The rapid adoption of Julia, the open source, high level programing language with roots at MIT, shows no sign of slowing according to data from Julialang.org. I Read more…

Intel Launches 10nm ‘Ice Lake’ Datacenter CPU with Up to 40 Cores

April 6, 2021

The wait is over. Today Intel officially launched its 10nm datacenter CPU, the third-generation Intel Xeon Scalable processor, codenamed Ice Lake. With up to 40 Read more…

CERN Is Betting Big on Exascale

April 1, 2021

The European Organization for Nuclear Research (CERN) involves 23 countries, 15,000 researchers, billions of dollars a year, and the biggest machine in the worl Read more…

Programming the Soon-to-Be World’s Fastest Supercomputer, Frontier

January 5, 2021

What’s it like designing an app for the world’s fastest supercomputer, set to come online in the United States in 2021? The University of Delaware’s Sunita Chandrasekaran is leading an elite international team in just that task. Chandrasekaran, assistant professor of computer and information sciences, recently was named... Read more…

HPE Launches Storage Line Loaded with IBM’s Spectrum Scale File System

April 6, 2021

HPE today launched a new family of storage solutions bundled with IBM’s Spectrum Scale Erasure Code Edition parallel file system (description below) and featu Read more…

10nm, 7nm, 5nm…. Should the Chip Nanometer Metric Be Replaced?

June 1, 2020

The biggest cool factor in server chips is the nanometer. AMD beating Intel to a CPU built on a 7nm process node* – with 5nm and 3nm on the way – has been i Read more…

Saudi Aramco Unveils Dammam 7, Its New Top Ten Supercomputer

January 21, 2021

By revenue, oil and gas giant Saudi Aramco is one of the largest companies in the world, and it has historically employed commensurate amounts of supercomputing Read more…

Quantum Computer Start-up IonQ Plans IPO via SPAC

March 8, 2021

IonQ, a Maryland-based quantum computing start-up working with ion trap technology, plans to go public via a Special Purpose Acquisition Company (SPAC) merger a Read more…

Leading Solution Providers

Contributors

Can Deep Learning Replace Numerical Weather Prediction?

March 3, 2021

Numerical weather prediction (NWP) is a mainstay of supercomputing. Some of the first applications of the first supercomputers dealt with climate modeling, and Read more…

Livermore’s El Capitan Supercomputer to Debut HPE ‘Rabbit’ Near Node Local Storage

February 18, 2021

A near node local storage innovation called Rabbit factored heavily into Lawrence Livermore National Laboratory’s decision to select Cray’s proposal for its CORAL-2 machine, the lab’s first exascale-class supercomputer, El Capitan. Details of this new storage technology were revealed... Read more…

New Deep Learning Algorithm Solves Rubik’s Cube

July 25, 2018

Solving (and attempting to solve) Rubik’s Cube has delighted millions of puzzle lovers since 1974 when the cube was invented by Hungarian sculptor and archite Read more…

African Supercomputing Center Inaugurates ‘Toubkal,’ Most Powerful Supercomputer on the Continent

February 25, 2021

Historically, Africa hasn’t exactly been synonymous with supercomputing. There are only a handful of supercomputers on the continent, with few ranking on the Read more…

The History of Supercomputing vs. COVID-19

March 9, 2021

The COVID-19 pandemic poses a greater challenge to the high-performance computing community than any before. HPCwire's coverage of the supercomputing response t Read more…

HPE Names Justin Hotard New HPC Chief as Pete Ungaro Departs

March 2, 2021

HPE CEO Antonio Neri announced today (March 2, 2021) the appointment of Justin Hotard as general manager of HPC, mission critical solutions and labs, effective Read more…

AMD Launches Epyc ‘Milan’ with 19 SKUs for HPC, Enterprise and Hyperscale

March 15, 2021

At a virtual launch event held today (Monday), AMD revealed its third-generation Epyc “Milan” CPU lineup: a set of 19 SKUs -- including the flagship 64-core, 280-watt 7763 part --  aimed at HPC, enterprise and cloud workloads. Notably, the third-gen Epyc Milan chips achieve 19 percent... Read more…

Microsoft, HPE Bringing AI, Edge, Cloud to Earth Orbit in Preparation for Mars Missions

February 12, 2021

The International Space Station will soon get a delivery of powerful AI, edge and cloud computing tools from HPE and Microsoft Azure to expand technology experi Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire