Intel Launches Silicon Photonics Chip, Previews Next-Gen Phi for AI

By Tiffany Trader

August 18, 2016

At the Intel Developer Forum, held in San Francisco this week, Intel Senior Vice President and General Manager Diane Bryant announced the launch of Intel’s Silicon Photonics product line and teased a brand-new Phi product, codenamed “Knights Mill,” aimed at machine learning workloads.

With the introduction of Silicon Photonics, Intel is debuting two new 100G optical transceivers. Sixteen years in the making, the small form-factor design fuses optical components with silicon integrated circuits to provide 100 gigabits per second over a distance of two kilometers. Initial target applications include connectivity for cloud and enterprise datacenters as well as Ethernet switch, router, and client-side telecom interfaces. Microsoft is adopting the technology for its scale-loving Azure datacenters.

“Electrons running over network cables won’t cut it,” said Bryant in her keynote address, “Intel is the only one to build the laser on silicon and therefore we are the first to light up silicon. We integrate the laser light emitting material, which is indium phosphide onto the silicon, and we use silicon lithography to align the laser with precision. This gives us a cost advantage because it is automatically aligned versus manually aligned as with traditional silicon photonics.”

The two QSFP28 optical transceivers, now shipping in volume, are based on industry standards at 100G for switch, router, and server use, notes Intel. The 100G PSM4 (Parallel Single Mode fiber 4-lane) optical transceiver features up to 2 kilometer reach on parallel single-mode fiber and the 100G CWDM4 (Coarse Wavelength Division Multiplexing 4-lane) optical transceiver offers up to 2 kilometer reach on duplex single-mode fiber.

The first Intel Silicon Photonics products will fulfill the need for faster connections from rack to rack and across the datacenter, said Bryant. “As the server network bandwidth increases from 10 Gig to 25 Gig to 50 Gig, optics will be required down into the server as well. We see a future where silicon photonics, optical I/O is everywhere in the datacenter and then integrated into the switch and the controller silicon. Our ability to run optics on silicon gives the end user a compelling benefit.”

Kushagra Vaid, general manager for Micrsoft Azure Cloud hardware engineering, emphasized the need to keep up with continued growth in its datacenter, especially relating to cloud networking. “Back in 2009 the server bandwidth used to be around a GB/sec, and if you fast forward to later this year into early next year, we anticipate it to be around 50 GB/sec, so that’s a growth of 50 times on bandwidth to the server. As the server data rates increase, from 1 to 10 to 25 Gbps, when we start getting to 100 Gbps to the server, you will hit a brick wall. There is no way copper can scale beyond 100 Gbps. It is already getting difficult to scale copper at 25 Gbps over 3 meters, so we do need some new technologies that are going to be used for this scaling. That’s why Silicon Photonics is very interesting to us.”

Microsoft will initially be deploying Intel’s Silicon Photonics technology for switch-to-switch interconnectivity at 100 Gbps in its Azure datacenter. “We found it’s a great cost-effective way to do these deployments,” said Vaid. “It’s optimized versus what we are doing today and I think the best part is it gives us a mechanism to scale to even higher bandwidth — up to 400 Gbps in the near future.”

Intel Puts AI-focused ‘Knights Mill’ on Phi Roadmap

Bryant also revealed that the next-generation Xeon Phi product would not be the 10nm “Knights Hill” that we’d been expecting but rather a brand-new Phi entry, codenamed “Knights Mill” and optimized for deep learning. The surprise Phi product will feature AI-targeted design elements such as enhanced variable precision compute and high capacity memory.

Like its second-gen cousin “Knights Landing,” the third-generation Phi is also a bootable host CPU. “It maintains that onload model,” said Bryant, “but we’ve included new instructions into the Intel instruction set – enhancements for variable precision floating point so the result is you will get even higher efficiency for deep learning models and training of those models complex neural data sets.”

IDF16 Phi Knights Mill slide 850x

Intel’s move to optimize for single-precision (and likely half-precision) follows the same path that NVIDIA started when it launched the highly FP32-optimized Titan X at its 2015 GTC event. Pascal, debuted at GTC16, is the company’s first high-end GPU to feature mixed-precision floating point capability, meaning the architecture will be able to process FP16 operations twice as quickly as FP32 operations. While double-precision FLOPS are standard fare in HPC, machine learning typically does quite well with single or half-precision compute.

There is still a lot we don’t know about Knights Mill, such as what manufacturing process it will use and whether it replaces Knights Hill, the chip that is supposed to power Argonne Lab’s CORAL installation in the 2018 timeframe. Bryant didn’t indicate if or how the new chip would affect previous disclosures, but emphasized Intel’s commitment to “a very long roadmap of optimized solutions for artificial intelligence.”

The War for AI Dominance

With the launch of both Nvidia Pascal GPUs and the Intel Knights Landing Phi this year, there’s a battle brewing between the reigning GPU champ and Chipzilla for AI supremacy with the most recent shot being fired by NVIDIA this week in the form of a blog post contesting performance claims made by Intel. Intel said it stands by its numbers.

During Bryant’s keynote, representatives from Chinese cloud giant Baidu and machine learning startup Indico took to the stage to sing the praises of Xeon and Xeon Phi for machine learning workloads. In one exchange Indico founder Slater Victoroff noted that “the issue with that is once you move to thousands of models, GPUs don’t make sense anymore.” “I certainly like the idea of GPUs not making sense,” Bryant quipped back.

Baidu provided an even heftier endorsement. The company, which has relied heavily on NVIDIA GPUs to run its deep learning models, announced that it will be using Xeon Phi chips to train and run Deep Speech, its speech recognition service.

“We are always trying to find ways to train neural networks faster,” said Baidu’s Jing Wang. “A big part of our approach is our use of techniques normally reserved for high-performance computing and that has helped us achieve a 7X speedup over our previous system. When it comes to AI, Intel Xeon Phi processors are a great fit in terms of running our machine learning networks. The increased memory size that Intel Phi provides makes it easier for us to train our models efficiently compared to other solutions. We find Xeon Phi very promising and consider performance across a wide range of kernel shapes and sizes relevant to the state-of-art along short-term memory models.”

Baidu also announced a new HPC cloud service, featuring Xeon Phis. “The Xeon Phi-based public cloud solutions will help bring HPC to a much broader audience,” said Wang. “We think it will mean not only lower cost but greater velocity of HPC and AI innovations.”

Bryant observed that machine learning is also a prime workload at government and academic high-performance computing centers. Increasingly, researchers are applying machine learning to what are traditional data-intensive science problems. At NERSC, the DOE computing facility where the Knights Landing-based Cori machine is currently being installed, Intel is partnering with researchers to advance machine learning at scale. Together, she said, they’ll tackle “previously unsolved problems that require the entire Cori supercomputer for challenges such as creating a catalogue of all objects in the universe.”

The final AI note hit by Bryant was Intel’s planned acquisition of Nervana Systems, announced last week. “Their IP as well as their deep expertise in accelerating deep learning algorithms will directly apply to our advancements in artificial intelligence,” said Bryant. “They have solutions at the silicon levels, at the libraries and at the framework level.”

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industry updates delivered to you every week!

Empowering High-Performance Computing for Artificial Intelligence

April 19, 2024

Artificial intelligence (AI) presents some of the most challenging demands in information technology, especially concerning computing power and data movement. As a result of these challenges, high-performance computing Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that have occurred about once a decade. With this in mind, the ISC Read more…

2024 Winter Classic: Texas Two Step

April 18, 2024

Texas Tech University. Their middle name is ‘tech’, so it’s no surprise that they’ve been fielding not one, but two teams in the last three Winter Classic cluster competitions. Their teams, dubbed Matador and Red Read more…

2024 Winter Classic: The Return of Team Fayetteville

April 18, 2024

Hailing from Fayetteville, NC, Fayetteville State University stayed under the radar in their first Winter Classic competition in 2022. Solid students for sure, but not a lot of HPC experience. All good. They didn’t Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use of Rigetti’s Novera 9-qubit QPU. The approach by a quantum Read more…

2024 Winter Classic: Meet Team Morehouse

April 17, 2024

Morehouse College? The university is well-known for their long list of illustrious graduates, the rigor of their academics, and the quality of the instruction. They were one of the first schools to sign up for the Winter Read more…

Kathy Yelick on Post-Exascale Challenges

April 18, 2024

With the exascale era underway, the HPC community is already turning its attention to zettascale computing, the next of the 1,000-fold performance leaps that ha Read more…

Software Specialist Horizon Quantum to Build First-of-a-Kind Hardware Testbed

April 18, 2024

Horizon Quantum Computing, a Singapore-based quantum software start-up, announced today it would build its own testbed of quantum computers, starting with use o Read more…

MLCommons Launches New AI Safety Benchmark Initiative

April 16, 2024

MLCommons, organizer of the popular MLPerf benchmarking exercises (training and inference), is starting a new effort to benchmark AI Safety, one of the most pre Read more…

Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

April 15, 2024

As the AI revolution marches on, it is vital to continually reassess how this technology is reshaping our world. To that end, researchers at Stanford’s Instit Read more…

Intel’s Vision Advantage: Chips Are Available Off-the-Shelf

April 11, 2024

The chip market is facing a crisis: chip development is now concentrated in the hands of the few. A confluence of events this week reminded us how few chips Read more…

The VC View: Quantonation’s Deep Dive into Funding Quantum Start-ups

April 11, 2024

Yesterday Quantonation — which promotes itself as a one-of-a-kind venture capital (VC) company specializing in quantum science and deep physics  — announce Read more…

Nvidia’s GTC Is the New Intel IDF

April 9, 2024

After many years, Nvidia's GPU Technology Conference (GTC) was back in person and has become the conference for those who care about semiconductors and AI. I Read more…

Google Announces Homegrown ARM-based CPUs 

April 9, 2024

Google sprang a surprise at the ongoing Google Next Cloud conference by introducing its own ARM-based CPU called Axion, which will be offered to customers in it Read more…

Nvidia H100: Are 550,000 GPUs Enough for This Year?

August 17, 2023

The GPU Squeeze continues to place a premium on Nvidia H100 GPUs. In a recent Financial Times article, Nvidia reports that it expects to ship 550,000 of its lat Read more…

Synopsys Eats Ansys: Does HPC Get Indigestion?

February 8, 2024

Recently, it was announced that Synopsys is buying HPC tool developer Ansys. Started in Pittsburgh, Pa., in 1970 as Swanson Analysis Systems, Inc. (SASI) by John Swanson (and eventually renamed), Ansys serves the CAE (Computer Aided Engineering)/multiphysics engineering simulation market. Read more…

Intel’s Server and PC Chip Development Will Blur After 2025

January 15, 2024

Intel's dealing with much more than chip rivals breathing down its neck; it is simultaneously integrating a bevy of new technologies such as chiplets, artificia Read more…

Choosing the Right GPU for LLM Inference and Training

December 11, 2023

Accelerating the training and inference processes of deep learning models is crucial for unleashing their true potential and NVIDIA GPUs have emerged as a game- Read more…

Baidu Exits Quantum, Closely Following Alibaba’s Earlier Move

January 5, 2024

Reuters reported this week that Baidu, China’s giant e-commerce and services provider, is exiting the quantum computing development arena. Reuters reported � Read more…

Comparing NVIDIA A100 and NVIDIA L40S: Which GPU is Ideal for AI and Graphics-Intensive Workloads?

October 30, 2023

With long lead times for the NVIDIA H100 and A100 GPUs, many organizations are looking at the new NVIDIA L40S GPU, which it’s a new GPU optimized for AI and g Read more…

Shutterstock 1179408610

Google Addresses the Mysteries of Its Hypercomputer 

December 28, 2023

When Google launched its Hypercomputer earlier this month (December 2023), the first reaction was, "Say what?" It turns out that the Hypercomputer is Google's t Read more…

AMD MI3000A

How AMD May Get Across the CUDA Moat

October 5, 2023

When discussing GenAI, the term "GPU" almost always enters the conversation and the topic often moves toward performance and access. Interestingly, the word "GPU" is assumed to mean "Nvidia" products. (As an aside, the popular Nvidia hardware used in GenAI are not technically... Read more…

Leading Solution Providers

Contributors

Shutterstock 1606064203

Meta’s Zuckerberg Puts Its AI Future in the Hands of 600,000 GPUs

January 25, 2024

In under two minutes, Meta's CEO, Mark Zuckerberg, laid out the company's AI plans, which included a plan to build an artificial intelligence system with the eq Read more…

China Is All In on a RISC-V Future

January 8, 2024

The state of RISC-V in China was discussed in a recent report released by the Jamestown Foundation, a Washington, D.C.-based think tank. The report, entitled "E Read more…

Shutterstock 1285747942

AMD’s Horsepower-packed MI300X GPU Beats Nvidia’s Upcoming H200

December 7, 2023

AMD and Nvidia are locked in an AI performance battle – much like the gaming GPU performance clash the companies have waged for decades. AMD has claimed it Read more…

Nvidia’s New Blackwell GPU Can Train AI Models with Trillions of Parameters

March 18, 2024

Nvidia's latest and fastest GPU, codenamed Blackwell, is here and will underpin the company's AI plans this year. The chip offers performance improvements from Read more…

DoD Takes a Long View of Quantum Computing

December 19, 2023

Given the large sums tied to expensive weapon systems – think $100-million-plus per F-35 fighter – it’s easy to forget the U.S. Department of Defense is a Read more…

Eyes on the Quantum Prize – D-Wave Says its Time is Now

January 30, 2024

Early quantum computing pioneer D-Wave again asserted – that at least for D-Wave – the commercial quantum era has begun. Speaking at its first in-person Ana Read more…

GenAI Having Major Impact on Data Culture, Survey Says

February 21, 2024

While 2023 was the year of GenAI, the adoption rates for GenAI did not match expectations. Most organizations are continuing to invest in GenAI but are yet to Read more…

The GenAI Datacenter Squeeze Is Here

February 1, 2024

The immediate effect of the GenAI GPU Squeeze was to reduce availability, either direct purchase or cloud access, increase cost, and push demand through the roof. A secondary issue has been developing over the last several years. Even though your organization secured several racks... Read more…

  • arrow
  • Click Here for More Headlines
  • arrow
HPCwire