ASC16 Kicks Off in Wuhan, China: 16 Teams, $40k in Prizes

By Tiffany Trader

April 18, 2016

The fifth annual Asia Student Supercomputer Challenge (ASC16) got off to an exciting start this morning at the Huazhong University of Science and Technology (HUST) in Wuhan, the capital of Hubei province, China. From an initial pool of 175 teams, representing 148 universities across six continents, 16 teams advanced through the preliminary rounds into the final round, including returning champion and “triple crown” team, Tsinghua University, and the only US team to make it into the competition, the Boston Green Team.

With each iteration of the Asian Student Competition and its “sister” events at SC and ISC, the roster of competitors becomes more formidable. Many of the teams have accumulated a number of wins already, but there are new entrants too, including Hong Kong Baptist University and Dalian University of Technology, which have earned their place in this final based on their high rankings  in the preliminary rounds. There’s also a steady influx of new students, as the competition torch gets passed on.

ASC16 Boston Green Team
Boston Green Team at ASC16

The teams that HPCwire spoke with who have been here before are aiming to put that experience to their advantage. Students from the Boston Green Team (team includes students from Boston University and Northeastern) gave a lot of consideration to the cluster design process. Some of the teams brought in Nvidia Tesla K80 GPUs at their own expense, but the Boston team opted to outfit their Inspur NF5280M4 server with the standard parts that Inspur and Intel supplied, including the Xeon E5-2650 v3 processor and the Intel Xeon Phi-31S1 coprocessor cards.

We also spoke with returning champs from Tsinghua University who had a somewhat different point of view. The team leader said that because the teams will all be using similar hardware with the same CPU, they will really be forced to focus on application testing and development. “There is not that large a difference in architecture, so I think we have to beat them by software and that’s a really big challenge.”

ASC16 Tsinghua University team
Tsinghua University team at ASC16

The purpose of the contest is for each team to design its cluster for the best application performance within the power consumption mandates. Per the contest regulations, power consumption must be kept under 3000W or the result of a given task becomes invalid (except for the e-Prize, which we’ll get to in a moment). While teams can build larger clusters — and indeed some teams on the contest floor have ten nodes, the hard power limit constrains how many of those nodes can be harnessed.

Boston Green Team members Winston Chen and John Smith pointed out that at five standard nodes (with two Xeon CPUs and a Phi coprocessor) that is right at the limit for 3,000 watts. Last year at ASC15, the team had an eight node straight-CPU cluster and it hit 3,000 watts. This year, they have six Phi-accelerated nodes — five to run their workloads and a spare.

In the five years since its inception, ASC has developed into the largest student supercomputing competition and is also one with the highest award levels. During the four days of the competition, the 16 teams at ASC16 will race to conduct real-world benchmarking and science workloads as they vie for a total of six prizes worth nearly $40,000.

The application set includes HPL, HPCG, MASNUM, Abinit, DNN and a mystery application; students also must deliver and are evaluated on a team presentation. The ASC program committee is dedicated to making the program as “life-like” as possible. “ASC provides a learning platform for students to get hands-on experience through the use of high performance computers,” observes Professor Pak-Chung Ching, Chinese University of Hong Kong. “But training supercomputing talents is more than that…they should be trained to understand a complex computational problem, break it down and then use computers to solve different types of problems from a professional perspective. ASC provides a platform for students to participate in practical projects as well as an opportunity to share experience and gain knowledge.”

The Applications

ASC16 Zhejiang University with Tesla K80 800x
Zhejiang University team with Tesla K80 GPU at ASC16

Students will be tasked with running two benchmarking applications. The High Performance LINPACK (HPL) benchmark — currently used to rank the TOP500 computing systems — plus the HPCG High Performance Conjugate Gradients (HPCG), an emerging international standard that exercises computational and data access patterns that are representative of a broad range of modern applications. The authors of the HPCG embrace both HPL and HPCG as bookends that enable application teams to assess the full balance of a system, including processor performance, memory capacity, memory bandwidth, and interconnection capability.

Teams will conduct three science applications: a surface wave numerical model MASNUM; a material simulation software ABINIT; and a mystery application, which will be announced on the day of the final.

e-Prize for Deep Learning Prowess

Even if teams have opted to forego using Phis in their cluster configuration, all teams get the chance to access the Phi MIC (Many Integrated Core) architecture when they compete for this year’s e-Prize, which requires that students set up and deploy a deep neural network (DNN). After developing the DNN on a Phi platform, teams will optimize the DNN algorithm using a remote login to Phi-based Tianhe-2 nodes.

The cluster has eight Phi nodes, each comprising two CPUs (Intel Xeon E5-2692 v2, 12 cores, 2.20GHz) and three Xeon Phi cards (Intel Xeon Phi 31S1P, 57 cores, 1.1GHz, 8GB memory). The Tianhe-2 network employs a custom high speed interconnect system with a bandwidth between nodes of 160 Gbps. There is no power consumption limit for this part of the contest.

The students will apply a data set comprising more than 100,000 pieces of speech data to the DNN with the objective of training the machine to be able to recognize speech as efficiently and effectively as the human brain.

The DNN performance optimization is given the highest weight in the contest, 25 percent of the total score. The ASC committee recognizes DNN as “one of the most important deep learning algorithms in artificial intelligence and the most popular cutting-edge emerging field in high-performance computing.” The organizers cite the recent “Man vs. Machine Battle” between AlphaGo and Lee Sedol for generating widespread interest in deep learning.

After the performance testing is concluded, the teams will present the results of their work to a panel of expert judges. The team presentation is worth ten percent of their final score. The winning teams will be announced during an awards ceremony on Friday, April 22.

HPC Workshop

The event will also be host to the 12th HPC Connection Workshop on April 21, 2016. The workshop features a roster of prominent HPC experts from around the world. Jack Dongarra, University of Tennessee professor and founder of the HPL benchmark, will discuss the current trends and future challenges in the HPC field. Depei Qian, chief leader of the “HPC and Core Software” Project in the China 863 Program, will share the vision of supercomputing mapped in China’s 2016-2020 Five-Year Plan. Yutong Lu, professor at National University of Defense Technology and deputy chief designer on the Tianhe-2 project, will discuss the convergence of big data and HPC on the Tianhe-2 supercomputer. The full listing of invited talks can be viewed here.

ASC16 entrance signage 1000x
ASC16 Welcome Sign

 

ASC16 scoreboard
ASC16 scoreboard awaits results
Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Dell’s AMD-Powered Server Line Targets High-End Jobs

September 17, 2019

Dell Technologies rolled out five new servers this week based on AMD’s latest Epyc processor that are geared toward data-driven workloads running on increasingly popular multi-cloud platforms as well as in the HPC data Read more…

By George Leopold

Cerebras to Supply DOE with Wafer-Scale AI Supercomputing Technology

September 17, 2019

Cerebras Systems, which debuted its wafer-scale AI silicon at Hot Chips last month, has entered into a multi-year partnership with Argonne National Laboratory and Lawrence Livermore National Laboratory as part of a larger collaboration with the U.S. Department of Energy... Read more…

By Tiffany Trader

Better Scientific Software: Turn Your Passion into Cash

September 13, 2019

Do you know your way around scientific software and programming? You think you can contribute to the community by making scientific software better? If so, then the Better Scientific Software (BSSW) organization wants yo Read more…

By Dan Olds

AWS Solution Channel

A Guide to Discovering the Best AWS Instances and Configurations for Your HPC Workload

The flexibility and heterogeneity of HPC cloud services provide a welcome contrast to the constraints of on-premises HPC. Every HPC configuration is potentially accessible to any given workload in a well-resourced cloud HPC deployment, with vast scalability to spin up as much compute as that workload demands in any given moment. Read more…

HPE Extreme Performance Solutions

Intel FPGAs: More Than Just an Accelerator Card

FPGA (Field Programmable Gate Array) acceleration cards are not new, as they’ve been commercially available since 1984. Typically, the emphasis around FPGAs has centered on the fact that they’re programmable accelerators, and that they can truly offer workload specific hardware acceleration solutions without requiring custom silicon. Read more…

IBM Accelerated Insights

Rumors of My Death Are Still Exaggerated: The Mainframe

[Connect with Spectrum users and learn new skills in the IBM Spectrum LSF User Community.]

As of 2017, 92 of the world’s top 100 banks used mainframes. Read more…

Google’s ML Compiler Initiative Advances

September 12, 2019

Machine learning models running on everything from cloud platforms to mobile phones are posing new challenges for developers faced with growing tool complexity. Google’s TensorFlow team unveiled an open-source machine Read more…

By George Leopold

Cerebras to Supply DOE with Wafer-Scale AI Supercomputing Technology

September 17, 2019

Cerebras Systems, which debuted its wafer-scale AI silicon at Hot Chips last month, has entered into a multi-year partnership with Argonne National Laboratory and Lawrence Livermore National Laboratory as part of a larger collaboration with the U.S. Department of Energy... Read more…

By Tiffany Trader

IDAS: ‘Automagic’ HPC With Training Wheels

September 12, 2019

High-performance computing (HPC) for research is notorious for having steep barriers to entry. For this reason, high-tech disciplines were early adopters, have Read more…

By Elizabeth Leake

Univa Brings Cloud Automation to Slurm Users with Navops Launch 2.0

September 11, 2019

Univa, the company behind Grid Engine, announced today its HPC cloud-automation platform NavOps Launch will support the popular open-source workload scheduler Slurm. With the release of NavOps Launch 2.0, “Slurm users will have access to the same cloud automation capabilities... Read more…

By Tiffany Trader

When Dense Matrix Representations Beat Sparse

September 9, 2019

In our world filled with unintended consequences, it turns out that saving memory space to help deal with GPU limitations, knowing it introduces performance pen Read more…

By James Reinders

Eyes on the Prize: TACC’s Frontera Quickly Ramps up Science Agenda

September 9, 2019

Announced a year ago and officially launched a week ago, the Texas Advanced Computing Center’s Frontera – now the fastest academic supercomputer (~25 petefl Read more…

By John Russell

Quantum Roundup: IBM Goes to School, Delft Tackles Networking, Rigetti Updates

September 5, 2019

IBM today announced a new open source quantum ‘textbook’, a series of quantum education videos, and plans to expand its nascent quantum hackathon program. L Read more…

By John Russell

DARPA Looks to Propel Parallelism

September 4, 2019

As Moore’s law runs out of steam, new programming approaches are being pursued with the goal of greater hardware performance with less coding. The Defense Advanced Projects Research Agency is launching a new programming effort aimed at leveraging the benefits of massive distributed parallelism with less sweat. Read more…

By George Leopold

Fastest Academic Supercomputer Enters Full Production at TACC, Just in Time for Hurricane Season

September 3, 2019

Frontera, the NSF supercomputer installed at the Texas Advanced Computing Center (TACC) in June, passed its formal acceptance last week and is now officially la Read more…

By Tiffany Trader

High Performance (Potato) Chips

May 5, 2006

In this article, we focus on how Procter & Gamble is using high performance computing to create some common, everyday supermarket products. Tom Lange, a 27-year veteran of the company, tells us how P&G models products, processes and production systems for the betterment of consumer package goods. Read more…

By Michael Feldman

Supercomputer-Powered AI Tackles a Key Fusion Energy Challenge

August 7, 2019

Fusion energy is the Holy Grail of the energy world: low-radioactivity, low-waste, zero-carbon, high-output nuclear power that can run on hydrogen or lithium. T Read more…

By Oliver Peckham

AMD Verifies Its Largest 7nm Chip Design in Ten Hours

June 5, 2019

AMD announced last week that its engineers had successfully executed the first physical verification of its largest 7nm chip design – in just ten hours. The AMD Radeon Instinct Vega20 – which boasts 13.2 billion transistors – was tested using a TSMC-certified Calibre nmDRC software platform from Mentor. Read more…

By Oliver Peckham

TSMC and Samsung Moving to 5nm; Whither Moore’s Law?

June 12, 2019

With reports that Taiwan Semiconductor Manufacturing Co. (TMSC) and Samsung are moving quickly to 5nm manufacturing, it’s a good time to again ponder whither goes the venerable Moore’s law. Shrinking feature size has of course been the primary hallmark of achieving Moore’s law... Read more…

By John Russell

DARPA Looks to Propel Parallelism

September 4, 2019

As Moore’s law runs out of steam, new programming approaches are being pursued with the goal of greater hardware performance with less coding. The Defense Advanced Projects Research Agency is launching a new programming effort aimed at leveraging the benefits of massive distributed parallelism with less sweat. Read more…

By George Leopold

Cray Wins NNSA-Livermore ‘El Capitan’ Exascale Contract

August 13, 2019

Cray has won the bid to build the first exascale supercomputer for the National Nuclear Security Administration (NNSA) and Lawrence Livermore National Laborator Read more…

By Tiffany Trader

AMD Launches Epyc Rome, First 7nm CPU

August 8, 2019

From a gala event at the Palace of Fine Arts in San Francisco yesterday (Aug. 7), AMD launched its second-generation Epyc Rome x86 chips, based on its 7nm proce Read more…

By Tiffany Trader

Nvidia Embraces Arm, Declares Intent to Accelerate All CPU Architectures

June 17, 2019

As the Top500 list was being announced at ISC in Frankfurt today with an upgraded petascale Arm supercomputer in the top third of the list, Nvidia announced its Read more…

By Tiffany Trader

Leading Solution Providers

ISC 2019 Virtual Booth Video Tour

CRAY
CRAY
DDN
DDN
DELL EMC
DELL EMC
GOOGLE
GOOGLE
ONE STOP SYSTEMS
ONE STOP SYSTEMS
PANASAS
PANASAS
VERNE GLOBAL
VERNE GLOBAL

Ayar Labs to Demo Photonics Chiplet in FPGA Package at Hot Chips

August 19, 2019

Silicon startup Ayar Labs continues to gain momentum with its DARPA-backed optical chiplet technology that puts advanced electronics and optics on the same chip Read more…

By Tiffany Trader

Top500 Purely Petaflops; US Maintains Performance Lead

June 17, 2019

With the kick-off of the International Supercomputing Conference (ISC) in Frankfurt this morning, the 53rd Top500 list made its debut, and this one's for petafl Read more…

By Tiffany Trader

A Behind-the-Scenes Look at the Hardware That Powered the Black Hole Image

June 24, 2019

Two months ago, the first-ever image of a black hole took the internet by storm. A team of scientists took years to produce and verify the striking image – an Read more…

By Oliver Peckham

Cray – and the Cray Brand – to Be Positioned at Tip of HPE’s HPC Spear

May 22, 2019

More so than with most acquisitions of this kind, HPE’s purchase of Cray for $1.3 billion, announced last week, seems to have elements of that overused, often Read more…

By Doug Black and Tiffany Trader

Chinese Company Sugon Placed on US ‘Entity List’ After Strong Showing at International Supercomputing Conference

June 26, 2019

After more than a decade of advancing its supercomputing prowess, operating the world’s most powerful supercomputer from June 2013 to June 2018, China is keep Read more…

By Tiffany Trader

Qualcomm Invests in RISC-V Startup SiFive

June 7, 2019

Investors are zeroing in on the open standard RISC-V instruction set architecture and the processor intellectual property being developed by a batch of high-flying chip startups. Last fall, Esperanto Technologies announced a $58 million funding round. Read more…

By George Leopold

Intel Confirms Retreat on Omni-Path

August 1, 2019

Intel Corp.’s plans to make a big splash in the network fabric market for linking HPC and other workloads has apparently belly-flopped. The chipmaker confirmed to us the outlines of an earlier report by the website CRN that it has jettisoned plans for a second-generation version of its Omni-Path interconnect... Read more…

By Staff report

Intel Debuts Pohoiki Beach, Its 8M Neuron Neuromorphic Development System

July 17, 2019

Neuromorphic computing has received less fanfare of late than quantum computing whose mystery has captured public attention and which seems to have generated mo Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This