SC17 US Student Cluster Competition Teams: Defending the Home Turf

By Dan Olds

November 24, 2017

Nine US universities showed up to the SC17 Student Cluster Competition in an attempt to keep the trophies in the United States. Let’s use our video lens to get to know them a bit better….

Georgia Tech is a newbie team composed of budding computer scientists and other tech types. They’ve dubbed themselves “Team Swarm” in a nod to their school mascot, the yellow jacket. They’re mostly juniors and seniors, with the exception of one 5th year senior who took a redshirt year in order to, assumedly, gain some more experience.

The team seems to be adjusting to the competition fairly well, but is facing some of the usual problems. The Yellow Jacket configuration isn’t quite what they need in order to compete in the upper echelon, but they’re moving ahead with determination. Check out the video for more details.

On another note, their team song still sucks….and it got stuck in my head again when they mentioned it.

University of Illinois Urbana-Champaign is a second time competitor at the big SC17 cluster dance. This team has some great experience, one member actually works on Blue Waters, and the others (with one exception) are seniors who are just about ready to graduate into the world of professional clustering.

The team feels pretty good about their performance, although they concede that they won’t have the highest benchmark scores. They’re happy that they put their best foot forward and feel that they got the most out of their box.

They also have been working on training their next team, with nine other members on the way to cluster stardom next year. What’s really gratifying about this team is that they all seem like they’re committed to careers in HPC, which is fantastic. We definitely need new blood in this industry and these kids are supplying it – hmm….that didn’t come out exactly like I hoped, but you get my point, right?

Northeastern University had to sit through two of my inane interviews because I screwed up the first one. Needless to say, they’re a patient bunch of students. Things have been going pretty smoothly with the NEU team, but they did do a last minute hardware change.

They were originally decked out with Radeon MI25 GPUs, which are pretty good at single-precision computation and great at machine learning. However, all of the benchmarks and applications at SC17 require double-precision. I did have an idea for the team, as you’ll see on the video. My idea? Run the apps in single precision but do them twice. The team wasn’t very impressed with my shortcut.

They ended up swapping their Radeon GPUs for four NVIDIA P100s, which was quite the last minute fire drill. Team Boston provided the NVIDIA cards out of their hardware stash. Nice job, Team Boston, way to help out another team.

San Diego State University/SW Oklahoma State University is another combination team that is competing together for the first time. When we catch up with the team in the video, they’re doing ok….but not great.

They’ve having a problem getting MrBays running on their GPUs, but they’re trying various work arounds to see if they can get it to work. In the meantime, they’re running the application on their CPU, although they’re pretty sure they won’t get close to finishing if they don’t get it running on GPUs.

The team had originally figured they’d be able to run a Power 8 based head node, storage node, and a couple of compute nodes. However, there’s no way that amount of hardware will run under the 3,000 watt power cap. So the team had to slim down their cluster considerably, as they discuss in the video.

William Henry Harrison High School is the first high school only team to compete at any major Student Cluster Competition. Who would have figured that we’d get to a place where a high school can compete credibly against university cluster teams? But WHHHS is doing just that, holding their own against older students who have much more experience and expertise.

When we got our cameras to their booth, the team is feeling good about their progress. They had just completed MrBayes and were driving on. The team was currently having a few problems with the Reproducibility task and MPAS, but felt confident that they would overcome the challenges.

One bad note was that the team wasn’t able to submit a HPCG score due to some unnamed problem. That’s not a good thing, but it’s not an unrecoverable error, they still have time and apps to make up for the shortfall on HPCG.

The Chicago Fusion Team is composed of students from the Illinois Institute of Technology, Maine South High School, and Adlai Stevenson High School. We couldn’t find the time to film a video introduction due to both of our busy schedules. However, this is a team that looks like a pretty solid competitor and has had very few problems as near as I can tell. We’ll see how they look when the benchmarks and apps are scored – that’ll tell the tale.

Team Boston is a frequent competitor in the Student Cluster Competitions. As usual, the Big Man, Evan DeNato (or is it DeNado?) is at the helm, leading the team as usual.

When we catch up with Team Chowder, they’re looking at their cloud budget and figuring out how to apportion it for the Born application. Boston is doing well on MrBayes and doesn’t seem to see any problems ahead of them.

In the video, we talk about various ways to present data (my favorite: spider chart). The team is also working on the mystery application (MPAS). In talking to them, we discover their HPL (LINPACK) benchmark would have been a record breaking score just a year ago, but this year? No such luck.

Team Texas is comprised of students from University of Texas, Austin, and Texas State University. The team had just undergone the second of two (planned) power outages and was hard at work bringing up their cluster to complete the competition.

The team feels that their strongest application so far has been MPASS, the mystery application.

One of the students was all dressed up, having attended the job fair earlier in the day. He looks like a great candidate with a great attitude. I promised to post his resume in this video, but forgot. You can find his name in the video, if you have a position open and are looking for a motivated and knowledgeable employee with a great attitude.

Something a student said triggered me, and I went into my IEEE FP rant (it’s the shame of the computer industry) towards the end of the video and almost lose control when a student disagrees with me. But I manage to keep my rage bottled up so we could all part as friends.

Team Utah is looking to build upon their second place finish at SC16 last year. The newbie team burst onto the scene in a home court battle that saw them top all competitors but one. Can they do it again?

When our lenses catch up to them during SC17, the team seems calm and composed. They had a few problems setting up Born, but it’s running smoothly in the cloud at the time of the video. During the video, I discuss some personal GPU-enabling business with their HPL expert; I might be putting him to work at a later date.

The Utes have a couple of new faces, but the rest of the team competed at SC16 and are at least somewhat used to the pressure. I dangle the prospect of dropping out of college to participate in the underground cluster competitions that still exist in out of the way locales. The team wasn’t buying it and figured they’d stay in school.

If you missed any of the other “Meet the Teams” features, you can find the European teams here, and the Asian team here.

Now that we’ve given the teams their fifteen minutes of video fame (more like 5 minutes, but who’s counting?), it’s time to see how they finished. In our next article, we’ll be exhaustively going over the results to see who won and how they won. Stay tuned……

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire