TACC Provides Researchers With Data Analysis Capabilities That Advance Discoveries

By Nicole Hemsoth

February 9, 2011

As the machines in high-performance computing (HPC) centers increase in processing power, they are able to create more realistic simulations to help study and solve problems that affect the lives of many people. Predicting the spread of viruses and the landfall of hurricanes, for instance, are real-world problems that scientists can analyze using statistical algorithms with computers.

When dealing with large datasets, visualization technology becomes an important factor in enabling scientists to analyze their results. As Kelly Gaither, director, data and information analysis at the Texas Advanced Computing Center (TACC) at The University of Texas at Austin, puts it, “Without visualization, we would need to analyze numerically. We would have to make sense of stacks and stacks of zeroes and ones. We have in our brains the most powerful supercomputer that you can get access to, and visualization technology lets us take advantage of that and leverage our automatic real-time pattern matching, feature recognition and visual acuity to make sense of very large datasets.”

Breeding new discoveries

The Texas Advanced Computing Center provides ample opportunities for scientists to crunch numbers and visualize the results in great detail using graphics programs. TACC’s visualization laboratory serves researchers at The University of Texas at Austin and across the nation, and it is increasing its impact on science thanks in part to its partnership with Dell. Two of TACC’s systems, Stallion and Longhorn, are the largest of their kind in the world.

Stallion, a 307-megapixel display of 75 Dell UltraSharp™ 30-inch monitors, provides users with the ability to perform visualizations on a 15’x5’ tiled display.

Powered by Dell XPS 720 tower desktops with Intel Core 2 processors and dual NVIDIA 8800 GTX (G80) graphics cards, the visualization cluster allows users to access over 36 gigabytes of graphics memory, 108 gigabytes of system memory and 100 processing cores. This configuration enables the processing of datasets of a massive scale, and the interactive visualization of substantial geometries. A large, shared file system is available to enable the storage of terascale size datasets.

“The beauty of Stallion is that we’re able to see the imagery at its native resolution,” says Gaither. “And Stallion can be used for a variety of things. Certainly we’re using it to display very large imagery from simulations when we want to see the resolution of the native data, but we also use it for non-scientific purposes. Stallion is the only place in town where an Austin-based photographer is able to see his digital pictures at native resolution.”

For researchers who don’t require such a large display, TACC offers a workstation area dedicated to visualization with Dell Precision 690 workstations connected to large LCD displays, and a collaboration room for video-conferencing and small group meetings.

World-class data analysis

TACC’s most recent venture with Dell is Longhorn, a cluster designed for remote interactive visualization and data analysis. The system consists of 16 Dell PowerEdge R710 nodes with 144Gb of RAM each, and 240 Dell PowerEdge R610 compute nodes, each with two Intel Xeon 5500 series processors and 48 gigabytes of RAM. Storage nodes are Dell PowerEdge R710 servers with Intel Xeon 5500 series processors. Fourteen Dell PowerVault MD1000 direct-attached storage arrays provide a 210 terabyte global file system, managed by the Lustre Parallel File System. A Mellanox InfiniBand quad-data rate (QDR) fabric provides the interconnect.

To accelerate data analysis and make interactive visualization possible, Longhorn uses 128 NVIDIA Quadro Plex S4 units sourced through Dell, each with four NVIDIA FX 5800 graphics processing units (GPUs), 16 gigabytes of graphics memory and two independent graphics busses (one per GPU pair). Compute nodes are each connected to two dedicated NVIDIA FX 5800 GPUs via the Quadro Plex graphics bus.

“Longhorn is the largest graphics accelerated remote interactive visualization cluster in the world,” says Gaither. “We built it with a grant from the National Science Foundation and in cooperation with Dell. Dell helped us design the cluster architecture using a hybrid approach, combining CPU cycles with NVIDIA graphics processing units.”

Traditionally, GPU acceleration in visualization clusters is handled by desktop computers that leverage internal graphics cards. “Dell has a very compact, rack-mountable server footprint in collaboration with NVIDIA that allows us to have rack-mountable nodes while still gaining the benefit of real-time performance for graphics applications,” says Gaither. “You can imagine what the footprint would look like for a visualization cluster this powerful using hundreds of desktop machines—the Dell and NVIDIA solution reduced our space requirements for Longhorn.”

The Intel Xeon 5500 series processors in the Dell PowerEdge servers are helping TACC solve problems faster. “Although visualization and data analysis are often more memory-bound, we’ve run a number of CPU-intensive jobs that have been quite successful,” says Gaither.

Staying on the bleeding edge

Receiving a grant from the National Science Foundation for a visualization resource of this magnitude was a major step for TACC. It was also critically important for visualization and data analysis as scientific fields—a recognition that visualization and data analysis services are a very important part of the scientific process.

TACC was an early adopter of distributed memory HPC clusters, mostly for reasons of scale. Because analyzing larger and larger problems drives up the cost of processing with shared memory machines prohibitively, distributed memory machines are now the preferred alternative for many HPC centers, using cluster configurations of “commodity” machines.

Dell plays an important part in the visualization facilities at TACC by promptly responding to the center’s needs as it provides greater capabilities to scientists at The University of Texas and across the nation. By purchasing servers, storage, visual computing hardware, monitors and networking components through Dell, TACC is able to realize economies of scale, streamline procurement and benefit from Dell’s technology expertise.

“We keep an open collaboration with Dell, as we do with all of our technology partners,” says Gaither. “We communicate with Dell about the kinds of scientific problems that we are interested in, which enables them to respond to our needs.”

Solving real-world problems

Although Longhorn itself does not have a visualization display attached, data can be streamed to Stallion for local viewing or to remote locations to accommodate researchers all over the world. TACC has worked with the National Oceanic and Atmospheric Administration (NOAA) to run simulations to more accurately predict landfall for hurricanes using new techniques with statistical models, and hopes to use Longhorn during hurricane season to perform the visualization in real time.

“Longhorn will allow us to do three-dimensional visualizations as the storm data is being analyzed,” says Gaither.

TACC also met with researchers to help predict how certain diseases might spread geographically. “I think we’ve had an impact on understanding how to prevent the spread of viruses,” says Gaither. Researchers are also using TACC systems to produce 2D and 3D simulations of the impact of BP’s massive Gulf of Mexico oil spill on coastal areas, and plan for the possibility that a hurricane moving through the Gulf might bring some of the oil inland.

The robust information analysis and visualization capabilities the Dell systems provide are drawing more researchers to TACC. “Stallion and Longhorn are the crown jewels of our visualization resources,” Gaither concludes. “Researchers come here to get capabilities that they wouldn’t otherwise have access to, and that in turn breeds new science and new discoveries.”

HPC Visualization, a conversation with TACC’s Dr. Kelly Gaither

Visualizing Science at Scale with Longhorn, with TACC’s Dr. Kelly Gaither

Learn more about Dell’s HPC solutions

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

ISC 2024 Takeaways: Love for Top500, Extending HPC Systems, and Media Bashing

May 23, 2024

The ISC High Performance show is typically about time-to-science, but breakout sessions also focused on Europe's tech sovereignty, server infrastructure, storage, throughput, and new computing technologies, This round Read more…

HPC Pioneer Gordon Bell Passed Away

May 22, 2024

Legendary computer scientist Gordon Bell passed away last Friday at his home in Coronado, CA. He was 89. The New York Times has a nice tribute piece. A long-time pioneer with Digital Equipment Corp, he pushed hard for de Read more…

ISC 2024 — A Few Quantum Gems and Slides from a Packed QC Agenda

May 22, 2024

If you were looking for quantum computing content, ISC 2024 was a good place to be last week — there were around 20 quantum computing related sessions. QC even earned a slide in Kathy Yelick’s opening keynote — Bey Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Core42 Is Building Its 172 Million-core AI Supercomputer in Texas

May 20, 2024

UAE-based Core42 is building an AI supercomputer with 172 million cores which will become operational later this year. The system, Condor Galaxy 3, was announced earlier this year and will have 192 nodes with Cerebras Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's latest weapon in the AI battle with GPU maker Nvidia and clou Read more…

ISC 2024 Takeaways: Love for Top500, Extending HPC Systems, and Media Bashing

May 23, 2024

The ISC High Performance show is typically about time-to-science, but breakout sessions also focused on Europe's tech sovereignty, server infrastructure, storag Read more…

ISC 2024 — A Few Quantum Gems and Slides from a Packed QC Agenda

May 22, 2024

If you were looking for quantum computing content, ISC 2024 was a good place to be last week — there were around 20 quantum computing related sessions. QC eve Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Google Announces Sixth-generation AI Chip, a TPU Called Trillium

May 17, 2024

On Tuesday May 14th, Google announced its sixth-generation TPU (tensor processing unit) called Trillium.  The chip, essentially a TPU v6, is the company's l Read more…

Europe’s Race towards Quantum-HPC Integration and Quantum Advantage

May 16, 2024

What an interesting panel, Quantum Advantage — Where are We and What is Needed? While the panelists looked slightly weary — their’s was, after all, one of Read more…

The Future of AI in Science

May 15, 2024

AI is one of the most transformative and valuable scientific tools ever developed. By harnessing vast amounts of data and computational power, AI systems can un Read more…

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

ISC 2024 Keynote: High-precision Computing Will Be a Foundation for AI Models

May 15, 2024

Some scientific computing applications cannot sacrifice accuracy and will always require high-precision computing. Therefore, conventional high-performance c Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Atos Outlines Plans to Get Acquired, and a Path Forward

May 21, 2024

Atos – via its subsidiary Eviden – is the second major supercomputer maker outside of HPE, while others have largely dropped out. The lack of integrators and Atos' financial turmoil have the HPC market worried. If Atos goes under, HPE will be the only major option for building large-scale systems. Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Leading Solution Providers

Contributors

Some Reasons Why Aurora Didn’t Take First Place in the Top500 List

May 15, 2024

The makers of the Aurora supercomputer, which is housed at the Argonne National Laboratory, gave some reasons why the system didn't make the top spot on the Top Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

Intel Plans Falcon Shores 2 GPU Supercomputing Chip for 2026  

August 8, 2023

Intel is planning to onboard a new version of the Falcon Shores chip in 2026, which is code-named Falcon Shores 2. The new product was announced by CEO Pat Gel Read more…

The NASA Black Hole Plunge

May 7, 2024

We have all thought about it. No one has done it, but now, thanks to HPC, we see what it looks like. Hold on to your feet because NASA has released videos of wh Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

How the Chip Industry is Helping a Battery Company

May 8, 2024

Chip companies, once seen as engineering pure plays, are now at the center of geopolitical intrigue. Chip manufacturing firms, especially TSMC and Intel, have b Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire