SC13 Wrapup: Supercomputing’s Top Themes

By Nicole Hemsoth

November 24, 2013

For those of us who traveled to Denver for SC13, it’s now back to “normal” as the year in high performance computing begins its slow descent into relative silence before a fresh start in 2014.

Sitting down to plow through the plethora of new items to pluck for a top announcements article seemed impossible without first discussing some of the broader trends and themes—they beg to be heard. The hard news breakdown can be found here, but context is everything during a time of flux and each one of our newsy picks embodies at least one of these.

Outside of some of the vendors and organizations who had a great showing last week, there are a few topics and specific machines worth mentioning as topical “best in show” picks.  Forgive the rare “personal pronouning” I’m about to do for once, but with so many great conversations with you all last week, it’s hard to leave those experiences out.

For now, we shall begin this thematic breakdown with the topic that you expected…but with some (possibly) unexpected details about its relative weight during the show…

Exascale

Let me guess…you probably saw this topic at the top of the list and said, “well, of course”… While this might not be a surprise because of its meaning for the HPC community (in terms of research and commercially-driven technology development, funding drive, and competitive appeal), in some ways this topic wasn’t the star of the show.

smnetworkLet’s just be honest here. Ever since China topped the Top 500 charts with what some in the U.S. are calling its “insurmountably” high performance system, the momentum and excitement around the “race” seems to have cooled. It’s hard to get excited about a dash to a finish line when there are thousands of yards between the runners.

But it’s just a matter of timing and technology refreshes, say many. The introduction of some innovative processor, memory and interconnect technologies, especially around 2015, are set to breathe new life into the race, spawning a new set of runners and adding some major ripples into what appears, for now anyway, to be very still waters. In the meantime, it’s slow and steady toward the goal.

This topic of exascale on the U.S. front was not without its own news announcement, however. Early in the week we broke word of a new investment in exascale technologies, this time from the Department of Energy’s Office of Science and the National Nuclear Security Administration (NNSA). The organization awarded $25.4 million in R&D contracts to “accelerate the development of next-generation supercomputers.”

This new funding effort rests under the DoE’s “DesignForward” initiative, which is a follow-on to the wider exascale ambitions put forth by the FastForward project. As one might imagine, it involves a number of the “usual suspects” for this sort of project. AMD, Cray, IBM, Intel’s federal division, and NVIDIA are all going to “work to advance extreme-scale, on the path to exascale, computing technology that is vital to national security, scientific research, energy security and the nation’s economic competitiveness.”

The emphasis of the DesignForward contracts is on the development of interconnect technologies that are architected with energy efficiency, high bandwidth and I/O capabilities. According to project leaders, “Under the new contract, Intel will focus on interconnect architectures and implementation approaches, Cray on open network protocol standards, AMD on interconnect architectures and associated execution models, IBM on energy-efficient interconnect architectures and messaging models and NVIDIA on interconnect architectures for massively threaded processors.” They note that, “The vendors will collaborate with DOE’s Exascale Co-design Centers to determine how changes in the system architectures will affect how well the scientific applications perform.”

Notice the lack of urgency in the language there… “working to advance”… “on the path to exascale”… but after all, it’s the thought (and money) that counts, right? And there are many who are counting. Counting down to the reality, counting up the number of government dollars that have been pushed toward the efforts, and counting on the fact that the investments will be returned to the public following the sustained focus on supercomputing—some are even counting by twos to keep up with the continued push-back on the projected year.

Interestingly, the technical program’s emphasis on exascale shared the stage with a few other topics of more contemporary appeal, most notably Hadoop (more on that in a moment). Still, the challenges on the energy, programming, reliability and other fronts were explored in great detail by a number of key presenters and served as the topical backdrop for many of the larger conversations and innovations.

hadoopelephantHadoop and Big Data

Let’s all agree that these are not the same thing, even if they are generally lumped into the same conversations.

In fact, this week the resounding sentiment I picked up from numerous non-vendor conversations is that HPC has always been about data and yes, that data has always been big.

While many seem to feel that the attention around big data is driven by the vendor and commercial user communities, there’s no doubt that the tooling—both on the systems and software fronts—are definitely worth the attention this community is starting to pay to it. And shouldn’t the big data folks be looking here too, because after all…

If your definition of big data revolves around complex datasets (structured versus unstructured), or of data use that needs to think beyond (or even before) MPI, or if there’s just plain too much of it and a way to manage/store it (off to tape, in memory, in a cloud somewhere) is a challenge, there was likely a lot at SC13 for you. Again, it’s not just about the Hadoopery that so often serves as focal point.  We will hit on a few of the specific announcements around “big data” in the news edition of our SC13 wrapup, but it’s fair to say that every vendor had a story—and often a solid one—about how to manage massive, complex datasets.

With that said, aside from the larger trend of categorizing “big data” as a natural part of HPC (or the reverse, depending on who you ask), Hadoop and MapReduce were at the core of almost as many sessions that emphasized exascale in the session title or description. Further, many vendors saved their key announcements for the supercomputing show, even if the audience was tuned for a wider world of technology users. Intel expanded on its Hadoop distro in detail, Cray and others emphasized the role their boxes play for Hadoop workloads with customized hooks, all the storage vendors danced a strange little dance with the topic (when they weren’t busy spinning Lustre around), and Adaptive Computing and others made announcements around how their tech can play nice with the tech world’s biggest buzzword since “big data” itself hit the show floors. It is dizzying, isn’t it?

Actually, some of the most compelling of those “big data” stories were from those you might not expect (or hear as much from). This is especially true on the “orchestration” and management front. Traditional workload management software, for example, is doing double-duty (and managing to double its reach for the first time outside of “pure” HPC and into the enterprise) by being robust enough to scale to cope with some dramatic data demands. We talked at length with Univa, Adaptive Computing, and even a smaller company from France, SysFera, about what they’re doing at the orchestration level to make management of complex data more practical for both scientific and commercial environments. Again, more on that during our news recap.

locationHPC: It’s Not Just Academia Anymore..

To this you could probably argue that it never was, depending on your perspective and current place of employment on the academia/commercial spectrum. But this year, perhaps (far) more than ever, most of what we were hearing from those who are “traditional” HPC vendors is that there is an ever-increasing demand for their goods and services outside of the expected quarters.

The concept of “productizing HPC” is really taking off and there are a few vendors who seem to manage this split very well while others struggle wrapping their unique technologies around a message that kicks a much broader appeal. But let’s face it—now, more than ever, companies with large-scale infrastructure concerns (and that’s almost anyone whose business success hinges on adept data wrangling) are looking to tried and true technologies that are proven at massive scale. And who are they going to learn this from? HPC.

From the largest systems on earth, the most robust software to manage all that iron and the breed of applications, tools and support ecosystems that have been purpose-built and designed to run at mind-boggling core count (and throw in a dash of acceleration) is finally sounding its wake up call to the rest of the world. The era of broader application of the technologies all of you folks are developing have a home…look around. And let me assure you, this isn’t a shameless plug when I tell you, but HPCwire spun out a new publication this year called EnterpriseTech exactly because of these reasons. HPC is growing up and out—we don’t see a need to divide the community into two pieces (scientific vs. commercial) but the expansion of supercomputing technologies into the types of mainstream large-scale environments is happening fast and deserved a more focused outlet that directs its attention to the wider world of these technologies you folks are developing, refining and leading as they trickle down the enterprise ladder. It’s cool. Plain and simple.

We handed out a couple of Editor’s Choice Awards this year simply due to companies’ unique ability to expand some traditional supercomputing technologies into far wider markets. Notable winners there include Cray (which has captured some compelling enterprise customers and managed to take its messaging as a “supercomputing company” into a bigger plain by listening to the market), SGI (which has managed to fine-tune a message and product line that balances supercomputing/HPC with a much wider commercial appeal), and Univa (which boasts massive commercial growth of a technology based on HPC efforts via GridEngine). We watched as other companies, including Penguin Computing, tweaked its offerings by listening to what’s going on at the hyperscale/large-scale shops that are asking for Open Compute designs that are backed by the perceived reliability of a company that’s built large-scale systems. IBM and NVDIA hooked up in an effort to expand GPU computing to a wider group of potential users. Even tape storage vendors, especially SpectraLogic, have found new life in catering to an expanding array of commercial needs with new tooling. It’s fun to watch, isn’t it?

This is certainly not to say that at SC13 and those ahead scientific computing won’t take the topical cake. But this is to say that these tools are going to see an explosion of interest, adoption and hell, for that matter, press from the wider world of technology. HPC has arrived.

So with so much momentum, potential and exploration possible, this begs another question entirely—one that is its own “top topical pick” from the show…

Where Are All The Startups in HPC?

Seriously.

Each day, the news feed here at HPCwire HQ is flooded with “big data” vendor announcements of x-million dollars in series A funding for your typical, often rather vague and difficult-to-determine competitive angle-based tooling. More database vendors than one can shake a stick at. And why? Because “big data” is sexy. Don’t ask me why, but in a very all-encompassing, hopelessly generalized, technologically fleeting sort of way, it just is.

The real question for you many innovators out there is how do we bring the sexy back? To HPC, that is, because there was a day when this was all very fancy and special and, yes, sexy.

Dazzling scientific simulations? Yep, we have those. Dramatic feats of massive scale? Check. Theoretical technologies being developed in stealth mode. Ab-so-lutely! .. So where is the missing link? We’re going to be exploring that throughout 2014. Every hyper-hyped technology lately got its start because it scaled, because it was big, and because it powered the unfathomable. You, holy halls of scientific computing at the national lab scale, have something to learn from them, they’ll say. But they are well aware that you have a great deal to offer. MPI, Lustre, GPU computing—these are filtering in, trickling down from supercomputing mountain. Look out, world!

As the wider vendor and user world wakes up to the fact that the HPC community has been doing the truly awe-inspiring work before the Hadoop elephant was ever stuffed and it’s always been about “big data” on this side of the fence, we’re going to be here to catch that news and push it out. HPC needs investment. These technologies are the only thing proven at large scale. This is our year—send me your stories, your stealth mode progress, your ideas, your vision—and let’s share HPC with the rest of the world. I have a feeling that none of us have ever been the “cool kids” (sorry if that’s inaccurate, but I know a lot of you—ha!), but this is our chance to take over the technology lunchroom. Know what I mean?

Forward-Looking Processors/Accelerators

If you stay tuned tomorrow for the announcement/news based SC13 wrapup we’ll shed more light on the processor and accelerator new picks front, but suffice to say, there were some great “looking ahead” announcements from some surprise vendors, including Convey Computing and Micron.

We sat down for a close-knit briefing with Intel to discuss some of the specifics of the Knight’s Landing chip, which has the potential to shake up the HPC processing ecosystem, watched NVIDIA roll out more power with its K40, and as noted above, drew in our breaths at some of the neat ideas coming from new processor outliers, including Micron (please do read this) which has done something really interesting with exploiting the inherent parallelism of memory, and Convey, which took a noteworthy dip in the specialty processor pool.

Although it doesn’t necessarily fit neatly into the mix, there was a lot of talk about quantum computing at the show. And of course, wild speculation about whether or not this “thing” from D-Wave can technically be called such given the entanglement questions. Again, this is an issue we’ll explore more in 2014, but suffice to say, the mainstream media has picked up on this idea in a big way, so expect a plethora of (creatively inaccurate and under-researched/informed) material about this topic. We’ll do what we can to stretch our brains in the coming year to deliver some perspective from its primary research leaders at D-Wave, Google, Lockheed Martin and others.

It’s Lustre’s Year to Shine

Lustre marks a great example of an HPC-born technology that is bound for great things in the larger enterprise world. A few of the forward-looking vendors are taking notice of this momentum and adding it to their offerings for reasons that scale past the orders they’re taking from X National Lab or university.

It seemed to make sense to mention it here because it was such an important part of many vendor offerings and more important, conversations with the very few potential end users who were cruising the floor shopping solutions (that’s another topic—where are all the end users at this show and how do we reel them in?). In the news edition of the SC13 wrapup jabber that will come out tomorrow evening the vendor spotlight will be on these announcements.

Denver Has Awesome Beer

That is all.

And Now, Talk Amongst Yourselves…

Please send along your thoughts (for publication or fun) about a few other topics that we noticed, including:

  • The range, depth and scale of the technical sessions is something to behold. For sys admins to center directors, it was hard to find something that wouldn’t appeal to someone. Impossible. Kudos to the SC committee who puts these sorts of things together.
  • How many storage vendors are there exactly? And how to differentiate?
  • Did you notice a difference in the show’s size or “bling” due to government shutdown?
  • Those student cluster kids are outstanding. Will you hire them?
  • Who had the best booth in terms of demonstrations?
  • Did it seem like there were more young people milling about than usual (or I am just so old now that everyone under 35 looks 25?).
  • Denver has awesome beer. New Orleans (SC14) is a better place to drink it.
Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion XL — were added to the benchmark suite as MLPerf continues Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing power it brings to artificial intelligence.  Nvidia's DGX Read more…

Call for Participation in Workshop on Potential NSF CISE Quantum Initiative

March 26, 2024

Editor’s Note: Next month there will be a workshop to discuss what a quantum initiative led by NSF’s Computer, Information Science and Engineering (CISE) directorate could entail. The details are posted below in a Ca Read more…

Waseda U. Researchers Reports New Quantum Algorithm for Speeding Optimization

March 25, 2024

Optimization problems cover a wide range of applications and are often cited as good candidates for quantum computing. However, the execution time for constrained combinatorial optimization applications on quantum device Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at the network layer threatens to make bigger and brawnier pro Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HBM3E memory as well as the the ability to train 1 trillion pa Read more…

MLPerf Inference 4.0 Results Showcase GenAI; Nvidia Still Dominates

March 28, 2024

There were no startling surprises in the latest MLPerf Inference benchmark (4.0) results released yesterday. Two new workloads — Llama 2 and Stable Diffusion Read more…

Q&A with Nvidia’s Chief of DGX Systems on the DGX-GB200 Rack-scale System

March 27, 2024

Pictures of Nvidia's new flagship mega-server, the DGX GB200, on the GTC show floor got favorable reactions on social media for the sheer amount of computing po Read more…

NVLink: Faster Interconnects and Switches to Help Relieve Data Bottlenecks

March 25, 2024

Nvidia’s new Blackwell architecture may have stolen the show this week at the GPU Technology Conference in San Jose, California. But an emerging bottleneck at Read more…

Who is David Blackwell?

March 22, 2024

During GTC24, co-founder and president of NVIDIA Jensen Huang unveiled the Blackwell GPU. This GPU itself is heavily optimized for AI work, boasting 192GB of HB Read more…

Nvidia Looks to Accelerate GenAI Adoption with NIM

March 19, 2024

Today at the GPU Technology Conference, Nvidia launched a new offering aimed at helping customers quickly deploy their generative AI applications in a secure, s Read more…

The Generative AI Future Is Now, Nvidia’s Huang Says

March 19, 2024

We are in the early days of a transformative shift in how business gets done thanks to the advent of generative AI, according to Nvidia CEO and cofounder Jensen Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

Nvidia Showcases Quantum Cloud, Expanding Quantum Portfolio at GTC24

March 18, 2024

Nvidia’s barrage of quantum news at GTC24 this week includes new products, signature collaborations, and a new Nvidia Quantum Cloud for quantum developers. Wh Read more…

Alibaba Shuts Down its Quantum Computing Effort

November 30, 2023

In case you missed it, China’s e-commerce giant Alibaba has shut down its quantum computing research effort. It’s not entirely clear what drove the change. Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Leading Solution Providers

Contributors

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

Google Introduces ‘Hypercomputer’ to Its AI Infrastructure

December 11, 2023

Google ran out of monikers to describe its new AI system released on December 7. Supercomputer perhaps wasn't an apt description, so it settled on Hypercomputer Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Intel Won’t Have a Xeon Max Chip with New Emerald Rapids CPU

December 14, 2023

As expected, Intel officially announced its 5th generation Xeon server chips codenamed Emerald Rapids at an event in New York City, where the focus was really o Read more…

IBM Quantum Summit: Two New QPUs, Upgraded Qiskit, 10-year Roadmap and More

December 4, 2023

IBM kicks off its annual Quantum Summit today and will announce a broad range of advances including its much-anticipated 1121-qubit Condor QPU, a smaller 133-qu Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire