SC13 Wrapup: Supercomputing’s Top Themes

By Nicole Hemsoth

November 24, 2013

For those of us who traveled to Denver for SC13, it’s now back to “normal” as the year in high performance computing begins its slow descent into relative silence before a fresh start in 2014.

Sitting down to plow through the plethora of new items to pluck for a top announcements article seemed impossible without first discussing some of the broader trends and themes—they beg to be heard. The hard news breakdown can be found here, but context is everything during a time of flux and each one of our newsy picks embodies at least one of these.

Outside of some of the vendors and organizations who had a great showing last week, there are a few topics and specific machines worth mentioning as topical “best in show” picks.  Forgive the rare “personal pronouning” I’m about to do for once, but with so many great conversations with you all last week, it’s hard to leave those experiences out.

For now, we shall begin this thematic breakdown with the topic that you expected…but with some (possibly) unexpected details about its relative weight during the show…

Exascale

Let me guess…you probably saw this topic at the top of the list and said, “well, of course”… While this might not be a surprise because of its meaning for the HPC community (in terms of research and commercially-driven technology development, funding drive, and competitive appeal), in some ways this topic wasn’t the star of the show.

smnetworkLet’s just be honest here. Ever since China topped the Top 500 charts with what some in the U.S. are calling its “insurmountably” high performance system, the momentum and excitement around the “race” seems to have cooled. It’s hard to get excited about a dash to a finish line when there are thousands of yards between the runners.

But it’s just a matter of timing and technology refreshes, say many. The introduction of some innovative processor, memory and interconnect technologies, especially around 2015, are set to breathe new life into the race, spawning a new set of runners and adding some major ripples into what appears, for now anyway, to be very still waters. In the meantime, it’s slow and steady toward the goal.

This topic of exascale on the U.S. front was not without its own news announcement, however. Early in the week we broke word of a new investment in exascale technologies, this time from the Department of Energy’s Office of Science and the National Nuclear Security Administration (NNSA). The organization awarded $25.4 million in R&D contracts to “accelerate the development of next-generation supercomputers.”

This new funding effort rests under the DoE’s “DesignForward” initiative, which is a follow-on to the wider exascale ambitions put forth by the FastForward project. As one might imagine, it involves a number of the “usual suspects” for this sort of project. AMD, Cray, IBM, Intel’s federal division, and NVIDIA are all going to “work to advance extreme-scale, on the path to exascale, computing technology that is vital to national security, scientific research, energy security and the nation’s economic competitiveness.”

The emphasis of the DesignForward contracts is on the development of interconnect technologies that are architected with energy efficiency, high bandwidth and I/O capabilities. According to project leaders, “Under the new contract, Intel will focus on interconnect architectures and implementation approaches, Cray on open network protocol standards, AMD on interconnect architectures and associated execution models, IBM on energy-efficient interconnect architectures and messaging models and NVIDIA on interconnect architectures for massively threaded processors.” They note that, “The vendors will collaborate with DOE’s Exascale Co-design Centers to determine how changes in the system architectures will affect how well the scientific applications perform.”

Notice the lack of urgency in the language there… “working to advance”… “on the path to exascale”… but after all, it’s the thought (and money) that counts, right? And there are many who are counting. Counting down to the reality, counting up the number of government dollars that have been pushed toward the efforts, and counting on the fact that the investments will be returned to the public following the sustained focus on supercomputing—some are even counting by twos to keep up with the continued push-back on the projected year.

Interestingly, the technical program’s emphasis on exascale shared the stage with a few other topics of more contemporary appeal, most notably Hadoop (more on that in a moment). Still, the challenges on the energy, programming, reliability and other fronts were explored in great detail by a number of key presenters and served as the topical backdrop for many of the larger conversations and innovations.

hadoopelephantHadoop and Big Data

Let’s all agree that these are not the same thing, even if they are generally lumped into the same conversations.

In fact, this week the resounding sentiment I picked up from numerous non-vendor conversations is that HPC has always been about data and yes, that data has always been big.

While many seem to feel that the attention around big data is driven by the vendor and commercial user communities, there’s no doubt that the tooling—both on the systems and software fronts—are definitely worth the attention this community is starting to pay to it. And shouldn’t the big data folks be looking here too, because after all…

If your definition of big data revolves around complex datasets (structured versus unstructured), or of data use that needs to think beyond (or even before) MPI, or if there’s just plain too much of it and a way to manage/store it (off to tape, in memory, in a cloud somewhere) is a challenge, there was likely a lot at SC13 for you. Again, it’s not just about the Hadoopery that so often serves as focal point.  We will hit on a few of the specific announcements around “big data” in the news edition of our SC13 wrapup, but it’s fair to say that every vendor had a story—and often a solid one—about how to manage massive, complex datasets.

With that said, aside from the larger trend of categorizing “big data” as a natural part of HPC (or the reverse, depending on who you ask), Hadoop and MapReduce were at the core of almost as many sessions that emphasized exascale in the session title or description. Further, many vendors saved their key announcements for the supercomputing show, even if the audience was tuned for a wider world of technology users. Intel expanded on its Hadoop distro in detail, Cray and others emphasized the role their boxes play for Hadoop workloads with customized hooks, all the storage vendors danced a strange little dance with the topic (when they weren’t busy spinning Lustre around), and Adaptive Computing and others made announcements around how their tech can play nice with the tech world’s biggest buzzword since “big data” itself hit the show floors. It is dizzying, isn’t it?

Actually, some of the most compelling of those “big data” stories were from those you might not expect (or hear as much from). This is especially true on the “orchestration” and management front. Traditional workload management software, for example, is doing double-duty (and managing to double its reach for the first time outside of “pure” HPC and into the enterprise) by being robust enough to scale to cope with some dramatic data demands. We talked at length with Univa, Adaptive Computing, and even a smaller company from France, SysFera, about what they’re doing at the orchestration level to make management of complex data more practical for both scientific and commercial environments. Again, more on that during our news recap.

locationHPC: It’s Not Just Academia Anymore..

To this you could probably argue that it never was, depending on your perspective and current place of employment on the academia/commercial spectrum. But this year, perhaps (far) more than ever, most of what we were hearing from those who are “traditional” HPC vendors is that there is an ever-increasing demand for their goods and services outside of the expected quarters.

The concept of “productizing HPC” is really taking off and there are a few vendors who seem to manage this split very well while others struggle wrapping their unique technologies around a message that kicks a much broader appeal. But let’s face it—now, more than ever, companies with large-scale infrastructure concerns (and that’s almost anyone whose business success hinges on adept data wrangling) are looking to tried and true technologies that are proven at massive scale. And who are they going to learn this from? HPC.

From the largest systems on earth, the most robust software to manage all that iron and the breed of applications, tools and support ecosystems that have been purpose-built and designed to run at mind-boggling core count (and throw in a dash of acceleration) is finally sounding its wake up call to the rest of the world. The era of broader application of the technologies all of you folks are developing have a home…look around. And let me assure you, this isn’t a shameless plug when I tell you, but HPCwire spun out a new publication this year called EnterpriseTech exactly because of these reasons. HPC is growing up and out—we don’t see a need to divide the community into two pieces (scientific vs. commercial) but the expansion of supercomputing technologies into the types of mainstream large-scale environments is happening fast and deserved a more focused outlet that directs its attention to the wider world of these technologies you folks are developing, refining and leading as they trickle down the enterprise ladder. It’s cool. Plain and simple.

We handed out a couple of Editor’s Choice Awards this year simply due to companies’ unique ability to expand some traditional supercomputing technologies into far wider markets. Notable winners there include Cray (which has captured some compelling enterprise customers and managed to take its messaging as a “supercomputing company” into a bigger plain by listening to the market), SGI (which has managed to fine-tune a message and product line that balances supercomputing/HPC with a much wider commercial appeal), and Univa (which boasts massive commercial growth of a technology based on HPC efforts via GridEngine). We watched as other companies, including Penguin Computing, tweaked its offerings by listening to what’s going on at the hyperscale/large-scale shops that are asking for Open Compute designs that are backed by the perceived reliability of a company that’s built large-scale systems. IBM and NVDIA hooked up in an effort to expand GPU computing to a wider group of potential users. Even tape storage vendors, especially SpectraLogic, have found new life in catering to an expanding array of commercial needs with new tooling. It’s fun to watch, isn’t it?

This is certainly not to say that at SC13 and those ahead scientific computing won’t take the topical cake. But this is to say that these tools are going to see an explosion of interest, adoption and hell, for that matter, press from the wider world of technology. HPC has arrived.

So with so much momentum, potential and exploration possible, this begs another question entirely—one that is its own “top topical pick” from the show…

Where Are All The Startups in HPC?

Seriously.

Each day, the news feed here at HPCwire HQ is flooded with “big data” vendor announcements of x-million dollars in series A funding for your typical, often rather vague and difficult-to-determine competitive angle-based tooling. More database vendors than one can shake a stick at. And why? Because “big data” is sexy. Don’t ask me why, but in a very all-encompassing, hopelessly generalized, technologically fleeting sort of way, it just is.

The real question for you many innovators out there is how do we bring the sexy back? To HPC, that is, because there was a day when this was all very fancy and special and, yes, sexy.

Dazzling scientific simulations? Yep, we have those. Dramatic feats of massive scale? Check. Theoretical technologies being developed in stealth mode. Ab-so-lutely! .. So where is the missing link? We’re going to be exploring that throughout 2014. Every hyper-hyped technology lately got its start because it scaled, because it was big, and because it powered the unfathomable. You, holy halls of scientific computing at the national lab scale, have something to learn from them, they’ll say. But they are well aware that you have a great deal to offer. MPI, Lustre, GPU computing—these are filtering in, trickling down from supercomputing mountain. Look out, world!

As the wider vendor and user world wakes up to the fact that the HPC community has been doing the truly awe-inspiring work before the Hadoop elephant was ever stuffed and it’s always been about “big data” on this side of the fence, we’re going to be here to catch that news and push it out. HPC needs investment. These technologies are the only thing proven at large scale. This is our year—send me your stories, your stealth mode progress, your ideas, your vision—and let’s share HPC with the rest of the world. I have a feeling that none of us have ever been the “cool kids” (sorry if that’s inaccurate, but I know a lot of you—ha!), but this is our chance to take over the technology lunchroom. Know what I mean?

Forward-Looking Processors/Accelerators

If you stay tuned tomorrow for the announcement/news based SC13 wrapup we’ll shed more light on the processor and accelerator new picks front, but suffice to say, there were some great “looking ahead” announcements from some surprise vendors, including Convey Computing and Micron.

We sat down for a close-knit briefing with Intel to discuss some of the specifics of the Knight’s Landing chip, which has the potential to shake up the HPC processing ecosystem, watched NVIDIA roll out more power with its K40, and as noted above, drew in our breaths at some of the neat ideas coming from new processor outliers, including Micron (please do read this) which has done something really interesting with exploiting the inherent parallelism of memory, and Convey, which took a noteworthy dip in the specialty processor pool.

Although it doesn’t necessarily fit neatly into the mix, there was a lot of talk about quantum computing at the show. And of course, wild speculation about whether or not this “thing” from D-Wave can technically be called such given the entanglement questions. Again, this is an issue we’ll explore more in 2014, but suffice to say, the mainstream media has picked up on this idea in a big way, so expect a plethora of (creatively inaccurate and under-researched/informed) material about this topic. We’ll do what we can to stretch our brains in the coming year to deliver some perspective from its primary research leaders at D-Wave, Google, Lockheed Martin and others.

It’s Lustre’s Year to Shine

Lustre marks a great example of an HPC-born technology that is bound for great things in the larger enterprise world. A few of the forward-looking vendors are taking notice of this momentum and adding it to their offerings for reasons that scale past the orders they’re taking from X National Lab or university.

It seemed to make sense to mention it here because it was such an important part of many vendor offerings and more important, conversations with the very few potential end users who were cruising the floor shopping solutions (that’s another topic—where are all the end users at this show and how do we reel them in?). In the news edition of the SC13 wrapup jabber that will come out tomorrow evening the vendor spotlight will be on these announcements.

Denver Has Awesome Beer

That is all.

And Now, Talk Amongst Yourselves…

Please send along your thoughts (for publication or fun) about a few other topics that we noticed, including:

  • The range, depth and scale of the technical sessions is something to behold. For sys admins to center directors, it was hard to find something that wouldn’t appeal to someone. Impossible. Kudos to the SC committee who puts these sorts of things together.
  • How many storage vendors are there exactly? And how to differentiate?
  • Did you notice a difference in the show’s size or “bling” due to government shutdown?
  • Those student cluster kids are outstanding. Will you hire them?
  • Who had the best booth in terms of demonstrations?
  • Did it seem like there were more young people milling about than usual (or I am just so old now that everyone under 35 looks 25?).
  • Denver has awesome beer. New Orleans (SC14) is a better place to drink it.
Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

At Long Last, Supercomputing Helps to Map the Poles

August 22, 2019

“For years,” Paul Morin wrote, “those of us that made maps of the Poles apologized. We apologized for the blank spaces on maps, we apologized for mountains being in the wrong place and out-of-date information.” Read more…

By Oliver Peckham

Xilinx Says Its New FPGA is World’s Largest

August 21, 2019

In this age of exploding “technology disaggregation” – in which the Big Bang emanating from the Intel x86 CPU has produced significant advances in CPU chips and a raft of alternative, accelerated architectures... Read more…

By Doug Black

Supercomputers Generate Universes to Illuminate Galactic Formation

August 20, 2019

With advanced imaging and satellite technologies, it’s easier than ever to see a galaxy – but understanding how they form (a process that can take billions of years) is a different story. Now, a team of researchers f Read more…

By Oliver Peckham

AWS Solution Channel

Efficiency and Cost-Optimization for HPC Workloads – AWS Batch and Amazon EC2 Spot Instances

High Performance Computing on AWS leverages the power of cloud computing and the extreme scale it offers to achieve optimal HPC price/performance. With AWS you can right size your services to meet exactly the capacity requirements you need without having to overprovision or compromise capacity. Read more…

HPE Extreme Performance Solutions

Bring the combined power of HPC and AI to your business transformation

FPGA (Field Programmable Gate Array) acceleration cards are not new, as they’ve been commercially available since 1984. Typically, the emphasis around FPGAs has centered on the fact that they’re programmable accelerators, and that they can truly offer workload specific hardware acceleration solutions without requiring custom silicon. Read more…

IBM Accelerated Insights

Keys to Attracting the Newest HPC Talent – Post-Millennials

[Connect with HPC users and learn new skills in the IBM Spectrum LSF User Community.]

For engineers and scientists growing up in the 80s, the current state of HPC makes perfect sense. Read more…

Singularity Moves Up the Container Value Chain

August 20, 2019

The enterprise version of the Singularity HPC container platform released this week by Sylabs is designed to allow users to create, secure and share the high-end containers in self-hosted production deployments. The e Read more…

By George Leopold

At Long Last, Supercomputing Helps to Map the Poles

August 22, 2019

“For years,” Paul Morin wrote, “those of us that made maps of the Poles apologized. We apologized for the blank spaces on maps, we apologized for mountains being in the wrong place and out-of-date information.” Read more…

By Oliver Peckham

IBM Deepens Plunge into Open Source; OpenPOWER to Join Linux Foundation

August 20, 2019

IBM today announced it was contributing the instruction set (ISA) for its Power microprocessor and the designs for the Open Coherent Accelerator Processor Inter Read more…

By John Russell

Ayar Labs to Demo Photonics Chiplet in FPGA Package at Hot Chips

August 19, 2019

Silicon startup Ayar Labs continues to gain momentum with its DARPA-backed optical chiplet technology that puts advanced electronics and optics on the same chip Read more…

By Tiffany Trader

Scientists to Tap Exascale Computing to Unlock the Mystery of our Accelerating Universe

August 14, 2019

The universe and everything in it roared to life with the Big Bang approximately 13.8 billion years ago. It has continued expanding ever since. While we have a Read more…

By Rob Johnson

AI is the Next Exascale – Rick Stevens on What that Means and Why It’s Important

August 13, 2019

Twelve years ago the Department of Energy (DOE) was just beginning to explore what an exascale computing program might look like and what it might accomplish. Today, DOE is repeating that process for AI, once again starting with science community town halls to gather input and stimulate conversation. The town hall program... Read more…

By Tiffany Trader and John Russell

Cray Wins NNSA-Livermore ‘El Capitan’ Exascale Contract

August 13, 2019

Cray has won the bid to build the first exascale supercomputer for the National Nuclear Security Administration (NNSA) and Lawrence Livermore National Laborator Read more…

By Tiffany Trader

AMD Launches Epyc Rome, First 7nm CPU

August 8, 2019

From a gala event at the Palace of Fine Arts in San Francisco yesterday (Aug. 7), AMD launched its second-generation Epyc Rome x86 chips, based on its 7nm proce Read more…

By Tiffany Trader

Lenovo Drives Single-Socket Servers with AMD Epyc Rome CPUs

August 7, 2019

No summer doldrums here. As part of the AMD Epyc Rome launch event in San Francisco today, Lenovo announced two new single-socket servers, the ThinkSystem SR635 Read more…

By Doug Black

High Performance (Potato) Chips

May 5, 2006

In this article, we focus on how Procter & Gamble is using high performance computing to create some common, everyday supermarket products. Tom Lange, a 27-year veteran of the company, tells us how P&G models products, processes and production systems for the betterment of consumer package goods. Read more…

By Michael Feldman

Supercomputer-Powered AI Tackles a Key Fusion Energy Challenge

August 7, 2019

Fusion energy is the Holy Grail of the energy world: low-radioactivity, low-waste, zero-carbon, high-output nuclear power that can run on hydrogen or lithium. T Read more…

By Oliver Peckham

Cray, AMD to Extend DOE’s Exascale Frontier

May 7, 2019

Cray and AMD are coming back to Oak Ridge National Laboratory to partner on the world’s largest and most expensive supercomputer. The Department of Energy’s Read more…

By Tiffany Trader

Graphene Surprises Again, This Time for Quantum Computing

May 8, 2019

Graphene is fascinating stuff with promise for use in a seeming endless number of applications. This month researchers from the University of Vienna and Institu Read more…

By John Russell

AMD Verifies Its Largest 7nm Chip Design in Ten Hours

June 5, 2019

AMD announced last week that its engineers had successfully executed the first physical verification of its largest 7nm chip design – in just ten hours. The AMD Radeon Instinct Vega20 – which boasts 13.2 billion transistors – was tested using a TSMC-certified Calibre nmDRC software platform from Mentor. Read more…

By Oliver Peckham

TSMC and Samsung Moving to 5nm; Whither Moore’s Law?

June 12, 2019

With reports that Taiwan Semiconductor Manufacturing Co. (TMSC) and Samsung are moving quickly to 5nm manufacturing, it’s a good time to again ponder whither goes the venerable Moore’s law. Shrinking feature size has of course been the primary hallmark of achieving Moore’s law... Read more…

By John Russell

Cray Wins NNSA-Livermore ‘El Capitan’ Exascale Contract

August 13, 2019

Cray has won the bid to build the first exascale supercomputer for the National Nuclear Security Administration (NNSA) and Lawrence Livermore National Laborator Read more…

By Tiffany Trader

Deep Learning Competitors Stalk Nvidia

May 14, 2019

There is no shortage of processing architectures emerging to accelerate deep learning workloads, with two more options emerging this week to challenge GPU leader Nvidia. First, Intel researchers claimed a new deep learning record for image classification on the ResNet-50 convolutional neural network. Separately, Israeli AI chip startup Hailo.ai... Read more…

By George Leopold

Leading Solution Providers

ISC 2019 Virtual Booth Video Tour

CRAY
CRAY
DDN
DDN
DELL EMC
DELL EMC
GOOGLE
GOOGLE
ONE STOP SYSTEMS
ONE STOP SYSTEMS
PANASAS
PANASAS
VERNE GLOBAL
VERNE GLOBAL

Nvidia Embraces Arm, Declares Intent to Accelerate All CPU Architectures

June 17, 2019

As the Top500 list was being announced at ISC in Frankfurt today with an upgraded petascale Arm supercomputer in the top third of the list, Nvidia announced its Read more…

By Tiffany Trader

Top500 Purely Petaflops; US Maintains Performance Lead

June 17, 2019

With the kick-off of the International Supercomputing Conference (ISC) in Frankfurt this morning, the 53rd Top500 list made its debut, and this one's for petafl Read more…

By Tiffany Trader

AMD Launches Epyc Rome, First 7nm CPU

August 8, 2019

From a gala event at the Palace of Fine Arts in San Francisco yesterday (Aug. 7), AMD launched its second-generation Epyc Rome x86 chips, based on its 7nm proce Read more…

By Tiffany Trader

A Behind-the-Scenes Look at the Hardware That Powered the Black Hole Image

June 24, 2019

Two months ago, the first-ever image of a black hole took the internet by storm. A team of scientists took years to produce and verify the striking image – an Read more…

By Oliver Peckham

Cray – and the Cray Brand – to Be Positioned at Tip of HPE’s HPC Spear

May 22, 2019

More so than with most acquisitions of this kind, HPE’s purchase of Cray for $1.3 billion, announced last week, seems to have elements of that overused, often Read more…

By Doug Black and Tiffany Trader

Chinese Company Sugon Placed on US ‘Entity List’ After Strong Showing at International Supercomputing Conference

June 26, 2019

After more than a decade of advancing its supercomputing prowess, operating the world’s most powerful supercomputer from June 2013 to June 2018, China is keep Read more…

By Tiffany Trader

Qualcomm Invests in RISC-V Startup SiFive

June 7, 2019

Investors are zeroing in on the open standard RISC-V instruction set architecture and the processor intellectual property being developed by a batch of high-flying chip startups. Last fall, Esperanto Technologies announced a $58 million funding round. Read more…

By George Leopold

Intel 7nm GPU on Roadmap for 2021, OneAPI Coming This Year

May 8, 2019

At Intel's investor meeting today in Santa Clara, Calif., the company filled in details of its roadmap and product launch plans and sought to allay concerns about delays of its 10nm chips. In laying out its 10nm and 7nm timelines, Intel revealed that its first 7nm product would be... Read more…

By Tiffany Trader

  • arrow
  • Click Here for More Headlines
  • arrow
Do NOT follow this link or you will be banned from the site!
Share This