The Top Supercomputing Led Discoveries of 2013

By Nicole Hemsoth

January 2, 2014

2013 has been an incredible year for the entire ecosystem around supercomputing; from vendors pushing new technologies to boost performance, capacity, and programmability to researchers turning over new insights with fresh techniques. While exascale has taken more of a backseat than we might have predicted at the year’s end of 2010, there are plenty of signs that production HPC environments are blazing plenty of new trails.

As the calendar flips into 2014, we wanted to cast a backward glance at select discoveries and opportunities made possible by the fastest systems in the world and the people who run them—all pulled from our news archives of the past year along some important thematic lines.

We’ve pulled over 30 examples of how supercomputers are set to change the world in 2014 and beyond and while this list is anything but exhaustive, it does show how key segments in research and industry are evolving with HPC.

In a Galaxy Far, Far Away…

galaxyOne of the most famous “showcase” areas where HPC finds a mainstream shine is when news breaks of startling answers emerge to questions as big as “where do we come from” and “what is the universe made of.” As one might expect, 2013 was a banner year for discoveries that reached well beyond earth.

This year, Kraken at the National Institute for Computational Sciences (NICS) at the University of Tennessee Knoxville, addressed some large-scale, stubborn classical physics problems with revolutionary protoplanetary disk research while another massive system, the Opteron-powered “Hopper” Cray XE6 system that is part of the National Energy Research Scientific Computing Center (NERSC) at Lawrence Berkeley National Lab lit up the comos.

Back in October, a team of scientists from ETH Zurich and the University of Leeds solved a 300-year-old riddle about the nature of the Earth’s rotation. Using the Cray XE6 supercomputer “Monte Rosa” installed at CSCS, the researchers uncovered the reason for the gradual westward movement of the Earth’s magnetic field.

This past November, astrophysics researchers at UC San Diego advanced their understanding of star formation with the help of some major computational resources from San Diego Supercomputer Center (SDSC) at the UC San Diego and the National Institute for Computational Science (NICS) at Oak Ridge National Laboratory.

Researchers at the Universities of Leeds and Chicago harnessed supercomputing power to uncover an important mechanism behind the generation of astrophysical magnetic fields such as that of the sun. And researchers at the Institute for Computational Cosmology (ICC), are using HPC to model phenomena ranging from solar flares to the formation of galaxies. Others, including an NSF-supported team from the University of Iowa used 2013 and plenty of supercomputing might to measure space turbulence directly for the first time in the laboratory, allowing the world to finally see the dynamics behind it.

Some spent some of their year bolstering current resources to aid in new discoveries. For instance, a new 60 teraflops supercomputer and 1 petabyte high speed storage system recently installed on the campus of the University of California at Santa Cruz, which will give astrophysicists at the college the computational and storage headroom they need to model the heavens like never before.

Medical Discovery and Research

CONNECTOME3Ever-growing datasets fed by information from a widening pool of sources pushed medical research even further into supercomputing territory this year. The modeling and simulation requirements of viruses, genomic data, organ and

Supercomputing’s reach into the human brain was one of the most widely-cited research items in medical circles in 2013. This year the Human Brain Project was carried forward by a host of supporting institutions and systems and the topic was the subject of several lectures and keynotes that are worthy of reviewing before a fresh year begins.

Cancer research is another important field that is increasingly reliant on powerful systems. For instance, this year, researchers at Emory University reported a significant improvement in their ability to analyze and understand changes of cancer tumors over time thanks to HPC work done on a Keeneland Project supercomputer. Analysis of high resolution cancer tumor images that used to take weeks can now be completed in a matter of minutes on the hybrid GPU-CPU system.

Complex viruses, including HIV/AIDS were also the subject of a great deal of supercomputer-powered research this year. Researchers at the University of Illinois Urbana-Champaign have successfully modeled the interior of the HIV-1 virus using the Blue Waters system, opening the door to new antiretroviral drugs that target HIV-1, the virus that causes AIDS.

Other viruses, including malaria, were the target of additional innovative research. Pittsburgh Supercomputing Center (PSC) and the University of Notre Dame received up to $1.6 million in funding from the Bill & Melinda Gates Foundation to develop a system of computers and software for the Vector Ecology and Control Network (VECNet), an international consortium to eradicate malaria. The new VECNet Cyber-Infrastructure Project (CI) will support VECNet’s effort to unite research, industrial and public policy efforts to attack one of the worst diseases in the developing world in more effective, economical ways.

Armed with vaccines, however, viruses can be stopped in their tracks, assuming the delivery of such life-saving measures is done effectively. A supercomputer simulation of the West African nation of Niger showed that improving transportation as well could improve vaccine availability among children and mothers from roughly 50 percent to more than 90 percent.

On another front, researchers came a step closer to understanding strokes this year. A team from UC Berkeley and the University of California San Diego (UCSD) used the supercomputing resources of the National Energy Research Scientific Computing Center (NERSC) to model the efficacy of microbubbles and high intensity focused ultrasound (HIFU) for breaking up stroke-causing clots.

Other noteworthy advances powered by world-class systems emerged this year in medical areas as diverse as autism research and pushing new boundaries in medical physics. As ever-more computing capacity comes online in the coming year, we expect the diverse medical field to produce stunning stories and discoveries in 2014.

Climate Change and Earth Science

earfThe volumes of scientific support are growing in support of climate change, a process which has been powered by massive simulations, including those that put the changes in the context of global shifts over vast lengths of time.

For example, HPCwire’s Tiffany Trader wrote back in September on the “Climate Time Machine”, which relies on a database of extreme global weather events from 1871 to the present day, culled from newspaper weather reports, measurements on land and sea for the first decades along with more modern data. The team of top climate scientists fed the data into powerful supercomputers, including those at NERSC and the Oak Ridge Leadership Computing Facility in Tennessee, to create a virtual climate time machine. A sizable portion (12 percent) of the supercomputing resources at NERSC is allocated to global climate change research. That’s nearly 150 million processor-hours of highly-tuned computational might focused on an issue that is critical to humanity’s future.

New systems emerged to tackle climate change data. For instance, Berkeley Lab’s Green Flash is a specialized supercomputer designed to showcase a way to perform more detailed climate modeling. The system uses customized Tensilica-based processors, similar to those found in iPhones, and communication-minimizing algorithms that cut down on the movement of data, to model the movement of clouds around the earth at a higher resolution than was previously possible, without consuming huge amounts of electricity.

This year large-scale systems, like Blue Waters at NCSA, were used by a research team including Penn State engineers to enhance scientists’ understanding of global precipitation. The team used Blue Waters to tackle the problem of large gaps in precipitation data for large parts of the world. The goal is to help scientists and researchers move toward an integrated global water cycle observatory.

Other advances using new and enhanced data sources, including GIS, advanced satellite and emerging sensor technologies were made to aid research into other aspects of climate change. From an NCAR-led project to predict air pollution to others, the global climate change picture is filling in rapidly.

Manufacturing and Heavy Industry

planetcityManufacturing has been a notable target of investments on both the vendor and research fronts in 2013 as political rhetoric, changes in traditional manufacturing jobs, the need for STEM-based education to support a new manufacturing future and new technologies have all stepped up.

At the heart of industrial progress is a constant march toward more automation, efficiency and data-driven progress. As one might imagine, this offers significant opportunities for HPC modeling and simulation—not to mention for supercomputer-fed innovations in materials science, manufacturing processes and other areas.

Several facilities, including the Ohio Supercomputer Center, have lent helping hands to bring HPC to industry in 2013–and fresh efforts are springing up, including at Lawrence Livermore National Laboratory (LLNL). For instance, this year select industrial users had a crack at Vulcan, a 5 petafopper with 390,000 cores. With this, and a new host of commercial applications to tweak, LLNL is providing a much-needed slew of software and scaling support. The lab spent 2013 lining up participants to step to the high-core line to see how more compute horsepower can push modeling and simulation limits while solving specific scalability issues.

In July, companies interested in testing the latest in low-cost carbon fiber had a new opportunity to partner with the Department of Energy’s Carbon Fiber Technology Facility. The CFTF, operated by Oak Ridge National Laboratory as part of the Department’s Clean Energy Manufacturing Initiative, opened earlier this year to find ways to reduce carbon fiber production costs and to work with the private sector to stimulate widespread use of the strong, lightweight material.

Research was made into reality in a few interesting projects this year. For instance, a team of scientists and mathematicians at the DOE’s Lawrence Berkeley National Laboratory used their powerful number crunchers together with sophisticated algorithms to create cleaner combustion technologies to reduce the footprint of vehicles and machines. In another addition to cleaner manufacturing futures, scientists turned a lowly crustacean’s habits into a potentially beneficially process. The “Gribble” creature landed on the biofuel industry’s radar for its unique ability to digest wood in salty conditions. Now, researchers in the US and the UK are putting the University of Tennessee’s Kraken supercomputer to work modeling an enzyme in the Gribble’s gut, which could unlock the key to developing better industrial enzymes in the future.

Another notable story related to industry came from Oak Ridge National Lab, where researchers noted the importance of big rig trucks—a backbone to industry supply chains and product delivery. Most trucks only get about 6 miles to the gallon and altogether they emit about 423 million pounds of CO2 into the atmosphere each year. South Carolina-based BMI Corp. partnered with researchers at Oak Ridge National Laboratory (ORNL) to develop the SmartTruck UnderTray System, “a set of integrated aerodynamic fairings that improve the aerodynamics of 18-wheeler (Class 8) long-haul trucks.” After installation, the typical big rig can expect to achieve a fuel savings of between 7 and 12 percent, amounting to $5,000 annual savings in fuel costs.

CONTINUE with more of 2013’s progress… >>>

 

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment group International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

Weekly Twitter Roundup (Jan. 19, 2017)

January 19, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

France’s CEA and Japan’s RIKEN to Partner on ARM and Exascale

January 19, 2017

France’s CEA and Japan’s RIKEN institute announced a multi-faceted five-year collaboration to advance HPC generally and prepare for exascale computing. Among the particulars are efforts to: build out the ARM ecosystem; work on code development and code sharing on the existing and future platforms; share expertise in specific application areas (material and seismic sciences for example); improve techniques for using numerical simulation with big data; and expand HPC workforce training. It seems to be a very full agenda. Read more…

By Nishi Katsuya and John Russell

ARM Waving: Attention, Deployments, and Development

January 18, 2017

It’s been a heady two weeks for the ARM HPC advocacy camp. At this week’s Mont-Blanc Project meeting held at the Barcelona Supercomputer Center, Cray announced plans to build an ARM-based supercomputer in the U.K. while Mont-Blanc selected Cavium’s ThunderX2 ARM chip for its third phase of development. Last week, France’s CEA and Japan’s Riken announced a deep collaboration aimed largely at fostering the ARM ecosystem. This activity follows a busy 2016 when SoftBank acquired ARM, OpenHPC announced ARM support, ARM released its SVE spec, Fujistu chose ARM for the post K machine, and ARM acquired HPC tool provider Allinea in December. Read more…

By John Russell

HPE Extreme Performance Solutions

Remote Visualization: An Integral Technology for Upstream Oil & Gas

As the exploration and production (E&P) of natural resources evolves into an even more complex and vital task, visualization technology has become integral for the upstream oil and gas industry. Read more…

Women Coders from Russia, Italy, and Poland Top Study

January 17, 2017

According to a study posted on HackerRank today the best women coders as judged by performance on HackerRank challenges come from Russia, Italy, and Poland. Read more…

By John Russell

Spurred by Global Ambitions, Inspur in Joint HPC Deal with DDN

January 17, 2017

Inspur, the fast-growth cloud computing and server vendor from China that has several systems on the current Top500 list, and DDN, a leader in high-end storage, have announced a joint sales and marketing agreement to produce solutions based on DDN storage platforms integrated with servers, networking, software and services from Inspur. Read more…

By Doug Black

Weekly Twitter Roundup (Jan. 12, 2017)

January 12, 2017

Here at HPCwire, we aim to keep the HPC community apprised of the most relevant and interesting news items that get tweeted throughout the week. Read more…

By Thomas Ayres

NSF Seeks Input on Cyberinfrastructure Advances Needed

January 12, 2017

In cased you missed it, the National Science Foundation posted a “Dear Colleague Letter” (DCL) late last week seeking input on needs for the next generation of cyberinfrastructure to support science and engineering. Read more…

By John Russell

IDG to Be Bought by Chinese Investors; IDC to Spin Out HPC Group

January 19, 2017

US-based publishing and investment group International Data Group, Inc. (IDG) will be acquired by a pair of Chinese investors, China Oceanwide Holdings Group Co., Ltd. Read more…

By Tiffany Trader

France’s CEA and Japan’s RIKEN to Partner on ARM and Exascale

January 19, 2017

France’s CEA and Japan’s RIKEN institute announced a multi-faceted five-year collaboration to advance HPC generally and prepare for exascale computing. Among the particulars are efforts to: build out the ARM ecosystem; work on code development and code sharing on the existing and future platforms; share expertise in specific application areas (material and seismic sciences for example); improve techniques for using numerical simulation with big data; and expand HPC workforce training. It seems to be a very full agenda. Read more…

By Nishi Katsuya and John Russell

ARM Waving: Attention, Deployments, and Development

January 18, 2017

It’s been a heady two weeks for the ARM HPC advocacy camp. At this week’s Mont-Blanc Project meeting held at the Barcelona Supercomputer Center, Cray announced plans to build an ARM-based supercomputer in the U.K. while Mont-Blanc selected Cavium’s ThunderX2 ARM chip for its third phase of development. Last week, France’s CEA and Japan’s Riken announced a deep collaboration aimed largely at fostering the ARM ecosystem. This activity follows a busy 2016 when SoftBank acquired ARM, OpenHPC announced ARM support, ARM released its SVE spec, Fujistu chose ARM for the post K machine, and ARM acquired HPC tool provider Allinea in December. Read more…

By John Russell

Spurred by Global Ambitions, Inspur in Joint HPC Deal with DDN

January 17, 2017

Inspur, the fast-growth cloud computing and server vendor from China that has several systems on the current Top500 list, and DDN, a leader in high-end storage, have announced a joint sales and marketing agreement to produce solutions based on DDN storage platforms integrated with servers, networking, software and services from Inspur. Read more…

By Doug Black

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

UberCloud Cites Progress in HPC Cloud Computing

January 10, 2017

200 HPC cloud experiments, 80 case studies, and a ton of hands-on experience gained, that’s the harvest of four years of UberCloud HPC Experiments. Read more…

By Wolfgang Gentzsch and Burak Yenier

A Conversation with Women in HPC Director Toni Collis

January 6, 2017

In this SC16 video interview, HPCwire Managing Editor Tiffany Trader sits down with Toni Collis, the director and founder of the Women in HPC (WHPC) network, to discuss the strides made since the organization’s debut in 2014. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

AWS Beats Azure to K80 General Availability

September 30, 2016

Amazon Web Services has seeded its cloud with Nvidia Tesla K80 GPUs to meet the growing demand for accelerated computing across an increasingly-diverse range of workloads. The P2 instance family is a welcome addition for compute- and data-focused users who were growing frustrated with the performance limitations of Amazon's G2 instances, which are backed by three-year-old Nvidia GRID K520 graphics cards. Read more…

By Tiffany Trader

US, China Vie for Supercomputing Supremacy

November 14, 2016

The 48th edition of the TOP500 list is fresh off the presses and while there is no new number one system, as previously teased by China, there are a number of notable entrants from the US and around the world and significant trends to report on. Read more…

By Tiffany Trader

Vectors: How the Old Became New Again in Supercomputing

September 26, 2016

Vector instructions, once a powerful performance innovation of supercomputing in the 1970s and 1980s became an obsolete technology in the 1990s. But like the mythical phoenix bird, vector instructions have arisen from the ashes. Here is the history of a technology that went from new to old then back to new. Read more…

By Lynd Stringer

For IBM/OpenPOWER: Success in 2017 = (Volume) Sales

January 11, 2017

To a large degree IBM and the OpenPOWER Foundation have done what they said they would – assembling a substantial and growing ecosystem and bringing Power-based products to market, all in about three years. Read more…

By John Russell

Container App ‘Singularity’ Eases Scientific Computing

October 20, 2016

HPC container platform Singularity is just six months out from its 1.0 release but already is making inroads across the HPC research landscape. It's in use at Lawrence Berkeley National Laboratory (LBNL), where Singularity founder Gregory Kurtzer has worked in the High Performance Computing Services (HPCS) group for 16 years. Read more…

By Tiffany Trader

Dell EMC Engineers Strategy to Democratize HPC

September 29, 2016

The freshly minted Dell EMC division of Dell Technologies is on a mission to take HPC mainstream with a strategy that hinges on engineered solutions, beginning with a focus on three industry verticals: manufacturing, research and life sciences. "Unlike traditional HPC where everybody bought parts, assembled parts and ran the workloads and did iterative engineering, we want folks to focus on time to innovation and let us worry about the infrastructure," said Jim Ganthier, senior vice president, validated solutions organization at Dell EMC Converged Platforms Solution Division. Read more…

By Tiffany Trader

Lighting up Aurora: Behind the Scenes at the Creation of the DOE’s Upcoming 200 Petaflops Supercomputer

December 1, 2016

In April 2015, U.S. Department of Energy Undersecretary Franklin Orr announced that Intel would be the prime contractor for Aurora: Read more…

By Jan Rowell

Enlisting Deep Learning in the War on Cancer

December 7, 2016

Sometime in Q2 2017 the first ‘results’ of the Joint Design of Advanced Computing Solutions for Cancer (JDACS4C) will become publicly available according to Rick Stevens. He leads one of three JDACS4C pilot projects pressing deep learning (DL) into service in the War on Cancer. Read more…

By John Russell

Leading Solution Providers

D-Wave SC16 Update: What’s Bo Ewald Saying These Days

November 18, 2016

Tucked in a back section of the SC16 exhibit hall, quantum computing pioneer D-Wave has been talking up its new 2000-qubit processor announced in September. Forget for a moment the criticism sometimes aimed at D-Wave. This small Canadian company has sold several machines including, for example, ones to Lockheed and NASA, and has worked with Google on mapping machine learning problems to quantum computing. In July Los Alamos National Laboratory took possession of a 1000-quibit D-Wave 2X system that LANL ordered a year ago around the time of SC15. Read more…

By John Russell

CPU Benchmarking: Haswell Versus POWER8

June 2, 2015

With OpenPOWER activity ramping up and IBM’s prominent role in the upcoming DOE machines Summit and Sierra, it’s a good time to look at how the IBM POWER CPU stacks up against the x86 Xeon Haswell CPU from Intel. Read more…

By Tiffany Trader

Nvidia Sees Bright Future for AI Supercomputing

November 23, 2016

Graphics chipmaker Nvidia made a strong showing at SC16 in Salt Lake City last week. Read more…

By Tiffany Trader

New Genomics Pipeline Combines AWS, Local HPC, and Supercomputing

September 22, 2016

Declining DNA sequencing costs and the rush to do whole genome sequencing (WGS) of large cohort populations – think 5000 subjects now, but many more thousands soon – presents a formidable computational challenge to researchers attempting to make sense of large cohort datasets. Read more…

By John Russell

Beyond von Neumann, Neuromorphic Computing Steadily Advances

March 21, 2016

Neuromorphic computing – brain inspired computing – has long been a tantalizing goal. The human brain does with around 20 watts what supercomputers do with megawatts. And power consumption isn’t the only difference. Fundamentally, brains ‘think differently’ than the von Neumann architecture-based computers. While neuromorphic computing progress has been intriguing, it has still not proven very practical. Read more…

By John Russell

The Exascale Computing Project Awards $39.8M to 22 Projects

September 7, 2016

The Department of Energy’s Exascale Computing Project (ECP) hit an important milestone today with the announcement of its first round of funding, moving the nation closer to its goal of reaching capable exascale computing by 2023. Read more…

By Tiffany Trader

Dell Knights Landing Machine Sets New STAC Records

November 2, 2016

The Securities Technology Analysis Center, commonly known as STAC, has released a new report characterizing the performance of the Knight Landing-based Dell PowerEdge C6320p server on the STAC-A2 benchmarking suite, widely used by the financial services industry to test and evaluate computing platforms. The Dell machine has set new records for both the baseline Greeks benchmark and the large Greeks benchmark. Read more…

By Tiffany Trader

BioTeam’s Berman Charts 2017 HPC Trends in Life Sciences

January 4, 2017

Twenty years ago high performance computing was nearly absent from life sciences. Today it’s used throughout life sciences and biomedical research. Genomics and the data deluge from modern lab instruments are the main drivers, but so is the longer-term desire to perform predictive simulation in support of Precision Medicine (PM). There’s even a specialized life sciences supercomputer, ‘Anton’ from D.E. Shaw Research, and the Pittsburgh Supercomputing Center is standing up its second Anton 2 and actively soliciting project proposals. There’s a lot going on. Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This