Integrated Photonics Coming of Age

By Nicole Hemsoth

November 17, 2011

Thanks to Moore’s Law and advances in silicon photonic fabrication, over the past 10 years more and more photonic components are being integrated onto chips. Such integration will present an opportunity for hardware engineers to reconsider basic computer designs.

That topic is the theme of a Disruptive Technology session at SC11 on Thursday, conducted by Keren Bergman of Columbia University and Nadya Bliss of MIT Lincoln Laboratory. In conjunction with photonic research at their respective organizations, Bergen and Bliss are working on the DARPA POEM Program (Photonically Optimized Embedded Microprocessor) led by Dr. Jagdeep Shah.

MIT Lincoln Laboratory’s role in the effort is to consider potential impact of silicon photonics on both applications and architectures. They are considering this problem both top down (i.e., what are the key application drivers that would benefit from or need performance advantage that photonic interconnects could provide) and bottom up (i.e., what possible architectural changes can be motivated by availability of photonic interconnects both on-chip and to memory).

At Columbia they have been developing an extensive suite of design tools for creating optically interconnected networks-on-chip that are physical layer accurate. This is the basis for their architectural design exploration and validation that the photonic technologies will function as intended.

Prior to their SC11 session, HPCwire asked Bergen and Bliss to discuss the technology issues surrounding integrated photonics and how it could impact computer systems, including HPC machines.

HPCwire: Where are photonic technologies deployed today and what are the main impacts it has had on computing to date?

Keren Bergman: Optical interconnects have historically been used in the longer distance connectivity of HPC systems to storage area networks. With increased bandwidth requirements optical links have been used for inter-rack cluster communications significantly reducing cabling congestion.

Current HPC systems with vastly increased parallelism have accelerated the need for communications bandwidth and driven optical interconnects further into backplane, even placing photonic technologies within the router package in the most advanced systems. Active cables have come into widespread use between racks. At modern 5 to 10 gigabits per second data rates, electrical cables don’t have adequate reach for this application in an HPC, and are far too bulky.

HPCwire: What is the current status of on-chip silicon photonics technology? What are some of the different approaches being explored?

Bergman: There has been significant progress in creating the key silicon photonics device components, however large scale monolithic integration with electrical circuitry in CMOS compatible processes remains a major challenge. Several schemes are currently under development for the integration of silicon photonic components with microelectronic components — both the transmitter and receiver circuits, and the microprocessor or memory components that will utilize the optical links.

The major approaches explored include front-end of CMOS line (FEOL) integration for building modulator and photodetector photonic circuitry and low-temperature processing where optical devices can be monolithically integrated with the metallization levels of the chip as a back-end of line (BEOL) fabrication step. The first use of silicon photonics utilizing some of these approaches appears to be in a module-based technology for the active optical cable market.

VCSEL-based optical modules are perhaps the most widely used technology today. Recent efforts have focused on approaches of directly integrating VCSELs onto the chip package.

HPCwire: How much better is on-chip silicon photonics compared to today’s copper interconnects?

Bergman: On-chip silicon photonic is a potentially disruptive communications platform for building high performance computing systems. Immense bandwidth densities are enabled by the low-loss single mode silicon photonic interconnect which can propagate numerous high bitrate — 40 to 100 gigabit per second — signals in dense WDM, corresponding to terabits per second in a single waveguide.

Furthermore, unlike electronic routing circuitry which requires individual switching elements for each data channel, a single broadband photonic switching element can route multiple high-bandwidth optical channels for the equivalent power of switching a single channel. The optical interconnect can therefore enable extremely high-bandwidth, low-latency end-to-end data transmission from on-chip to off-chip and potentially across the system without the need for power consuming repeaters, buffers, and regenerators.

Combining the power advantage with the bandwidth advantage could yield approximately an order of magnitude communication performance improvement over today’s interconnects — say 10 to 20x better.

HPCwire: How do you see on-chip silicon photonics changing processor, memory, and motherboard designs?

Nadya Bliss: This is exactly the right question to ask. While alone the bandwidth and energy efficiency improvements are significant, the true power of integrated silicon photonics can be demonstrated by considering new architectures and instruction sets. The physical and performance characteristics of silicon photonics enable consideration of new network architectures, new memory hierarchies such as flatter, fewer levels of cache, and pushing parallelism into the hardware.

All of the current computer architecture trends point to multicore systems, with increasing number of cores on chip. Photonic interconnects have the potential to both balance the compute capability of emerging multicore and simplify the programming model by enabling balanced communication.

Realizing these potential benefits in future commercial systems will require significant advancements in high density, low cost optical packaging technology that can meet the reliability challenges.

HPCwire: What will be the impact on system software: the operating system, compilers, communication libraries, etc.?

Bliss: While any new technology has the potential to require new compilers, libraries, etc, the current multicore platforms and the on-chip and to-memory communication challenges are requiring a re-evaluation of programming models. New programming models are emerging that improve programmability, reducing the burden on the programmer, while also allowing the users to increase parallel efficiency of computation. Better communication/computation balance and pushing parallelism closer to the hardware has the potential of simplifying programming models and therefore associated compilers and libraries.

HPCwire: Will applications have to change to take advantage of on-chip photonics?

Bliss: I don’t think applications have to change per se, but new capabilities can be enabled, for example: more complex algorithms can be implemented in smaller form factors.

HPCwire: When do you think we can expect to see on-chip photonics makes its way into commercial silicon?

Bliss: Given the existing research efforts and pending successful demonstrations of both technologies and application capabilities, it is possible to imagine that this would happen over the next 5 to 10 years. To be honest, if it doesn’t, the programmability and performance challenges will continue to get worse and we will see decreased performance scaling in the near future.

Subscribe to HPCwire's Weekly Update!

Be the most informed person in the room! Stay ahead of the tech trends with industy updates delivered to you every week!

Trinity Supercomputer’s Haswell and KNL Partitions Are Merged

July 19, 2017

Trinity supercomputer’s two partitions – one based on Intel Xeon Haswell processors and the other on Xeon Phi Knights Landing – have been fully integrated are now available for use on classified work in the Nationa Read more…

By HPCwire Staff

Fujitsu Continues HPC, AI Push

July 19, 2017

Summer is well under way, but the so-called summertime slowdown, linked with hot temperatures and longer vacations, does not seem to have impacted Fujitsu's output. The Japanese multinational has made a raft of HPC and A Read more…

By Tiffany Trader

Researchers Use DNA to Store and Retrieve Digital Movie

July 18, 2017

From abacus to pencil and paper to semiconductor chips, the technology of computing has always been an ever-changing target. The human brain is probably the computer we use most (hopefully) and understand least. This mon Read more…

By John Russell

The Exascale FY18 Budget – The Next Step

July 17, 2017

On July 12, 2017, the U.S. federal budget for its Exascale Computing Initiative (ECI) took its next step forward. On that day, the full Appropriations Committee of the House of Representatives voted to accept the recomme Read more…

By Alex R. Larzelere

HPE Extreme Performance Solutions

HPE Servers Deliver High Performance Remote Visualization

Whether generating seismic simulations, locating new productive oil reservoirs, or constructing complex models of the earth’s subsurface, energy, oil, and gas (EO&G) is a highly data-driven industry. Read more…

Summer Reading: IEEE Spectrum’s Chip Hall of Fame

July 17, 2017

Take a trip down memory lane – the Mostek MK4096 4-kilobit DRAM, for instance. Perhaps processors are more to your liking. Remember the Sh-Boom processor (1988), created by Russell Fish and Chuck Moore, and named after Read more…

By John Russell

Women in HPC Luncheon Shines Light on Female-Friendly Hiring Practices

July 13, 2017

The second annual Women in HPC luncheon was held on June 20, 2017, during the International Supercomputing Conference in Frankfurt, Germany. The luncheon provides participants the opportunity to network with industry lea Read more…

By Tiffany Trader

Satellite Advances, NSF Computation Power Rapid Mapping of Earth’s Surface

July 13, 2017

New satellite technologies have completely changed the game in mapping and geographical data gathering, reducing costs and placing a new emphasis on time series and timeliness in general, according to Paul Morin, directo Read more…

By Ken Chiacchia and Tiffany Jolley

Intel Skylake: Xeon Goes from Chip to Platform

July 13, 2017

With yesterday’s New York unveiling of the new “Skylake” Xeon Scalable processors, Intel made multiple runs at multiple competitive threats and strategic markets. Skylake will carry Intel's flag in the fight for le Read more…

By Doug Black

Fujitsu Continues HPC, AI Push

July 19, 2017

Summer is well under way, but the so-called summertime slowdown, linked with hot temperatures and longer vacations, does not seem to have impacted Fujitsu's out Read more…

By Tiffany Trader

Researchers Use DNA to Store and Retrieve Digital Movie

July 18, 2017

From abacus to pencil and paper to semiconductor chips, the technology of computing has always been an ever-changing target. The human brain is probably the com Read more…

By John Russell

The Exascale FY18 Budget – The Next Step

July 17, 2017

On July 12, 2017, the U.S. federal budget for its Exascale Computing Initiative (ECI) took its next step forward. On that day, the full Appropriations Committee Read more…

By Alex R. Larzelere

Women in HPC Luncheon Shines Light on Female-Friendly Hiring Practices

July 13, 2017

The second annual Women in HPC luncheon was held on June 20, 2017, during the International Supercomputing Conference in Frankfurt, Germany. The luncheon provid Read more…

By Tiffany Trader

Satellite Advances, NSF Computation Power Rapid Mapping of Earth’s Surface

July 13, 2017

New satellite technologies have completely changed the game in mapping and geographical data gathering, reducing costs and placing a new emphasis on time series Read more…

By Ken Chiacchia and Tiffany Jolley

Intel Skylake: Xeon Goes from Chip to Platform

July 13, 2017

With yesterday’s New York unveiling of the new “Skylake” Xeon Scalable processors, Intel made multiple runs at multiple competitive threats and strategic Read more…

By Doug Black

Perverse Incentives? How Economics (Mis-)shaped Academic Science

July 12, 2017

The unintended consequences of how we fund academic research—in the U.S. and elsewhere—are strangling innovation, putting universities into debt and creatin Read more…

By Ken Chiacchia, Senior Science Writer, Pittsburgh Supercomputing Center

Why Tech is Failing at Diversity and How It Can Succeed

July 11, 2017

The sectors that are supposed to be all about innovation and the future continue to fail spectacularly at gender equity and diversity. UK, US and Canada still haven’t managed to break the average 20 percent threshold for gender equity across STEM academic disciplines. Read more…

By Kelly Nolan

HPC Compiler Company PathScale Seeks Life Raft

March 23, 2017

HPCwire has learned that HPC compiler company PathScale has fallen on difficult times and is asking the community for help or actively seeking a buyer for its a Read more…

By Tiffany Trader

Quantum Bits: D-Wave and VW; Google Quantum Lab; IBM Expands Access

March 21, 2017

For a technology that’s usually characterized as far off and in a distant galaxy, quantum computing has been steadily picking up steam. Just how close real-wo Read more…

By John Russell

Google Pulls Back the Covers on Its First Machine Learning Chip

April 6, 2017

This week Google released a report detailing the design and performance characteristics of the Tensor Processing Unit (TPU), its custom ASIC for the inference Read more…

By Tiffany Trader

Nvidia Responds to Google TPU Benchmarking

April 10, 2017

Nvidia highlights strengths of its newest GPU silicon in response to Google's report on the performance and energy advantages of its custom tensor processor. Read more…

By Tiffany Trader

Trump Budget Targets NIH, DOE, and EPA; No Mention of NSF

March 16, 2017

President Trump’s proposed U.S. fiscal 2018 budget issued today sharply cuts science spending while bolstering military spending as he promised during the cam Read more…

By John Russell

CPU-based Visualization Positions for Exascale Supercomputing

March 16, 2017

In this contributed perspective piece, Intel’s Jim Jeffers makes the case that CPU-based visualization is now widely adopted and as such is no longer a contrarian view, but is rather an exascale requirement. Read more…

By Jim Jeffers, Principal Engineer and Engineering Leader, Intel

Nvidia’s Mammoth Volta GPU Aims High for AI, HPC

May 10, 2017

At Nvidia's GPU Technology Conference (GTC17) in San Jose, Calif., this morning, CEO Jensen Huang announced the company's much-anticipated Volta architecture a Read more…

By Tiffany Trader

Facebook Open Sources Caffe2; Nvidia, Intel Rush to Optimize

April 18, 2017

From its F8 developer conference in San Jose, Calif., today, Facebook announced Caffe2, a new open-source, cross-platform framework for deep learning. Caffe2 is the successor to Caffe, the deep learning framework developed by Berkeley AI Research and community contributors. Read more…

By Tiffany Trader

Leading Solution Providers

How ‘Knights Mill’ Gets Its Deep Learning Flops

June 22, 2017

Intel, the subject of much speculation regarding the delayed, rewritten or potentially canceled “Aurora” contract (the Argonne Lab part of the CORAL “ Read more…

By Tiffany Trader

MIT Mathematician Spins Up 220,000-Core Google Compute Cluster

April 21, 2017

On Thursday, Google announced that MIT math professor and computational number theorist Andrew V. Sutherland had set a record for the largest Google Compute Engine (GCE) job. Sutherland ran the massive mathematics workload on 220,000 GCE cores using preemptible virtual machine instances. Read more…

By Tiffany Trader

Reinders: “AVX-512 May Be a Hidden Gem” in Intel Xeon Scalable Processors

June 29, 2017

Imagine if we could use vector processing on something other than just floating point problems.  Today, GPUs and CPUs work tirelessly to accelerate algorithms Read more…

By James Reinders

Google Debuts TPU v2 and will Add to Google Cloud

May 25, 2017

Not long after stirring attention in the deep learning/AI community by revealing the details of its Tensor Processing Unit (TPU), Google last week announced the Read more…

By John Russell

Russian Researchers Claim First Quantum-Safe Blockchain

May 25, 2017

The Russian Quantum Center today announced it has overcome the threat of quantum cryptography by creating the first quantum-safe blockchain, securing cryptocurrencies like Bitcoin, along with classified government communications and other sensitive digital transfers. Read more…

By Doug Black

Groq This: New AI Chips to Give GPUs a Run for Deep Learning Money

April 24, 2017

CPUs and GPUs, move over. Thanks to recent revelations surrounding Google’s new Tensor Processing Unit (TPU), the computing world appears to be on the cusp of Read more…

By Alex Woodie

Top500 Results: Latest List Trends and What’s in Store

June 19, 2017

Greetings from Frankfurt and the 2017 International Supercomputing Conference where the latest Top500 list has just been revealed. Although there were no major Read more…

By Tiffany Trader

Six Exascale PathForward Vendors Selected; DoE Providing $258M

June 15, 2017

The much-anticipated PathForward awards for hardware R&D in support of the Exascale Computing Project were announced today with six vendors selected – AMD Read more…

By John Russell

  • arrow
  • Click Here for More Headlines
  • arrow
Share This