Today's Top Feature

Advancing Modular Supercomputing with DEEP and DEEP-ER Architectures

Knowing that the jump to exascale will require novel architectural approaches capable of delivering dramatic

By Sean Thielen

Center Stage

HPC Technique Propels Deep Learning at Scale

Researchers from Baidu’s Silicon Valley AI Lab (SVAIL) have adapted a well-known HPC communication technique

By Tiffany Trader

IDC: Will the Real Exascale Race Please Stand Up?

So the exascale race is on. And lots of organizations are in the pack. Government

By Bob Sorensen

TSUBAME3.0 Points to Future HPE Pascal-NVLink-OPA Server

Since our initial coverage of the TSUBAME3.0 supercomputer yesterday, more details have come to light on this innovative project. Of particular interest is a new board design for NVLink-equipped Pascal P100 GPUs that will create another entrant to the space currently occupied by Nvidia's DGX-1 system, IBM's "Minsky" platform and the Supermicro SuperServer (1028GQ-TXR).

By Tiffany Trader

People to Watch 2017

With 2017 underway, we’re looking to the future of high performance computing and the milestones that are growing ever closer.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

HPE Server Shows Low Latency on STAC-N1 Test

February 22, 2017

The performance of trade and match servers can be a critical differentiator for financial trading houses. Read more…

By John Russell

Rethinking HPC Platforms for ‘Second Gen’ Applications

February 22, 2017

Just what constitutes HPC and how best to support it is a keen topic currently. Read more…

By John Russell

ExxonMobil, NCSA, Cray Scale Reservoir Simulation to 700,000+ Processors

February 17, 2017

In a scaling breakthrough for oil and gas discovery, ExxonMobil geoscientists report they have harnessed the power of 717,000 processors – the equivalent of 22,000 32-processor computers – to run complex oil and gas reservoir simulation models. Read more…

By Doug Black

Tokyo Tech’s TSUBAME3.0 Will Be First HPE-SGI Super

February 16, 2017

In a press event Friday afternoon local time in Japan, Tokyo Institute of Technology (Tokyo Tech) announced its plans for the TSUBAME3.0 supercomputer, which will be Japan’s “fastest AI supercomputer,” Read more…

By Tiffany Trader

Drug Developers Use Google Cloud HPC in the Fight Against ALS

February 16, 2017

Within the haystack of a lethal disease such as ALS (amyotrophic lateral sclerosis / Lou Gehrig’s Disease) there exists, somewhere, the needle that will pierce this therapy-resistant affliction. Read more…

By Doug Black

HPC Cloud Startup Launches ‘App Store’ for HPC Workflows

February 9, 2017

“Civilization advances by extending the number of important operations which we can perform without thinking about them,” Read more…

By Tiffany Trader

UMass Rolls Out New GPU Cluster for Deep Learning

February 8, 2017

UMass today rolled out its new GPU cluster – Gypsum – aimed at deep learning. Like many institutions, UMass is seeking to attract Ph.D. students drawn to deep learning and artificial intelligence. At 400 GPUs, Gypsum is on the large side for academic GPU clusters according the university. Read more…

By John Russell

Van Andel Research Optimizes HPC Pipeline with DDN

February 7, 2017

For more than a decade the swelling output from life sciences experimental instruments has been overwhelming research computing infrastructures intended to support them. Read more…

By John Russell

Leading Solution Providers

  • arrow
  • Click Here for More Headlines
  • arrow

Whitepaper:

Sorting Fact from Fiction: HPC-enabled Engineering Simulations, On-premises or in the Cloud

HPC may once have been the sole province for huge corporations and national labs, but with hardware and cloud resources becoming more affordable even small and mid-sized companies are taking advantage.

Download this Report

Sponsored by ANSYS

Webinar:

Enabling Open Source High Performance Workloads with Red Hat

High performance workloads, big data, and analytics are increasingly important in finding real value in today's applications and data. Before we deploy applications and mine data for mission and business insights, we need a high-performance, rapidly scalable, resilient infrastructure foundation that can accurately, securely, and quickly access data from all relevant sources. Red Hat has technology that allows high performance workloads with a scale-out foundation that integrates multiple data sources and can transition workloads across on-premise and cloud boundaries.

Register to attend this LIVE webinar

Sponsored by Red Hat

Virtual Booth Tours

Not able to attend SC16?
See what you missed

SC16

Advanced Scale Career Development & Workforce Enhancement Center

Featured Advanced Scale Jobs:

Receive the Monthly
Advanced Computing Job Bank Resource:

HPCwire Product Showcase

Subscribe to the Monthly
Technology Product Showcase:

Subscribe