UC San Diego Students Win MLPerf Contest at SC23

February 2, 2024

Feb. 2, 2024 — A team of UC San Diego undergraduate students won third place in the annual Student Cluster Competition (SCC), which was held at the recent 2023 Supercomputing Conference in Denver, Colorado. The students, from the San Diego Supercomputer Center (SDSC) and Jacobs School of Engineering at UC San Diego, called their team Triton LLC (Last Level Cache).

This year’s UC San Diego SCC award-winning team. Front row: Mary Thomas (mentor), Francisco (Paco) Guitierrez, Gloria Seo, Rachel Handran, Jeremy Tow, Bryan Chin. Back row: Anish Govind, Khai Vu, Kyle Smith, Zixian Wang, Austin Garcia, Triston Babers. Credit: Jeremy Tow.

They were among a total of 11 in-person teams from around the world selected to compete. This is the second year in a row the student team has garnered awards at the supercomputing competition.

Six UC San Diego undergraduate students— Austin Garcia, Francisco Gutierrez, Khai Vu, Kyle Smith, Song Hee Seo and Zixian Wang— made up the competition team. The students were mentored by Mary Thomas (SDSC) and Bryan Chin, a UC San Diego computer science lecturer–with assistance from SDSC staff members Andreas Goetz, Martin Kandes, Mahidhar Tatineni and Bob Sinkovits as well as several mentors from AMD.

The competition consisted of the team running industry benchmarks such as High Performance Linpack, High Performance Conjugate Gradients (HPCG) and MLPerf Inference Benchmark, a leading AI Inference benchmark suite that measures the speed of a trained neural network when performing inference tasks on new data. Students had to fine-tune their computer cluster to maximize performance while staying under a power limit.

In addition, they were graded on their ability to run two programs simulating atmosphere and 3D magnetohydrodynamics, as well as reproducing results on a technique for Cholesky Decomposition, and were surprised with a Capture the Flag mystery application.

“The competition provides undergraduate students with the opportunity to learn about, and work with, advanced HPC hardware and software which provides them with skills and experience that they can apply in their studies, research and future careers,” said Thomas, SDSC’s HPC training lead.

For the benchmarking award in MLPerf Inference Benchmark, UC San Diego shared the win with Peking University.

“It has been my most valuable experience working and learning with the whole team, spending 30+ hours every week for three months, striving for better performance on benchmarks and applications,” said Wang, a UC San Diego student majoring in computer science. “The competition result was very rewarding and I truly appreciate everyone’s effort and enjoy the time together aiming for the same goal.”

According to Thomas, the win was especially rewarding since Peking used NVIDIA software on the newest H100 accelerator hardware, while the UC San Diego team developed new software solutions to run the benchmark on AMD Instinct™ accelerators. The SCC23 team, led by Wang and Vu and advised by Miro Hodak from AMD, ported the code to the AMD Instinct GPUs and created the necessary system configurations and environment. Together, they were able to achieve some of the first results of the benchmark on the AMD hardware and to publish their results on the MLPerf website.

“AMD Instinct accelerators deliver leadership performance for the most demanding AI and HPC workloads. The AMD ROCm™ open software platform supports a wide range of AI inference and training tools making AMD the premier platform for AI development. It was the seamless support for PyTorch and ONNX runtime that enabled the team to port the code and optimize its performance,” said Hodak, senior member of technical staff for AI and ML Solutions Architecture at AMD, co-chair of MLPerf Inference working group.

In addition to the students who competed in Denver, the following students also contributed their skills to the team’s ultimate success: Triston Babers, Jeremy Tow, Anish Govind who presented a paper at the conference and student volunteer Rachel Handran. As part of the program at UC San Diego, the SCC program includes independent study course credit. To extend the impact of the SCC program and student experiences, the course also included a “home” team of students who co-trained with the competition team: Aarush Mehrotra, Akhilm Ram Shankar, Arnav Talreja, Brandon Szeto, Henry (Keng-Lien) Lin, James Choi, James Han, Pranav Prabu. Many of these students are interested in applying to be on the team next year.

To learn more about Triton LLC (Last Level Cache), their SC23 poster shares details and more information can be found on the SDSC/UCSD SCC23 team website.

The Student Cluster Competition was developed in 2007 to immerse undergraduate and high school students in HPC. SCC teams consist of worldwide participants, in a non-stop, 48-hour challenge to complete a real-world scientific workload, while keeping the cluster up and running, and demonstrating to the judges their HPC skills and knowledge.

Each SCC team consists of six students who design and build a small cluster with support from mentors, as well as hardware and software industry partners. The teams learn designated scientific applications and apply optimization techniques for their chosen architectures.

To learn more about the SCC series, see the Student Cluster Competition website. To learn more about the UC San Diego/SDSC SCC teams, see the HPC Students website.


Source: Kimberly Mann Bruch, SDSC

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Read more…

Core42 Building an 172 Million-core AI Supercomputer in Texas

May 20, 2024

UAE-based Core42 is building an AI supercomputer with 172 million cores which will become operational later this year. The system, Condor Galaxy 3, was announced earlier this year and will have 192 nodes with Cerebras Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's latest weapon in the AI battle with GPU maker Nvidia and clou Read more…

ISC 2024 Student Cluster Competition

May 16, 2024

The 2024 ISC 2024 competition welcomed 19 virtual (remote) and eight in-person teams. The in-person teams participated in the conference venue and, while the virtual teams competed using the Bridges-2 supercomputers at t Read more…

Grace Hopper Gets Busy with Science 

May 16, 2024

Nvidia’s new Grace Hopper Superchip (GH200) processor has landed in nine new worldwide systems. The GH200 is a recently announced chip from Nvidia that eliminates the PCI bus from the CPU/GPU communications pathway.  Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of the last panels at ISC 2024 — the discussion was fascinat Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of Read more…

The Future of AI in Science

May 15, 2024

AI is one of the most transformative and valuable scientific tools ever developed. By harnessing vast amounts of data and computational power, AI systems can un Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

ISC 2024 Keynote: High-precision Computing Will Be a Foundation for AI Models

May 15, 2024

Some scientific computing applications cannot sacrifice accuracy and will always require high-precision computing. Therefore, conventional high-performance c Read more…

Shutterstock 493860193

Linux Foundation Announces the Launch of the High-Performance Software Foundation

May 14, 2024

The Linux Foundation, the nonprofit organization enabling mass innovation through open source, is excited to announce the launch of the High-Performance Softw Read more…

ISC 2024: Hyperion Research Predicts HPC Market Rebound after Flat 2023

May 13, 2024

First, the top line: the overall HPC market was flat in 2023 at roughly $37 billion, bogged down by supply chain issues and slowed acceptance of some larger sys Read more…

Top 500: Aurora Breaks into Exascale, but Can’t Get to the Frontier of HPC

May 13, 2024

The 63rd installment of the TOP500 list is available today in coordination with the kickoff of ISC 2024 in Hamburg, Germany. Once again, the Frontier system at Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

Leading Solution Providers

Contributors

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

How the Chip Industry is Helping a Battery Company

May 8, 2024

Chip companies, once seen as engineering pure plays, are now at the center of geopolitical intrigue. Chip manufacturing firms, especially TSMC and Intel, have b Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire