Silicon Based Packaging for 400/800/1600 Gb/s Optical Interconnects

Similar documents
Scaling the Compute and High Speed Networking Needs of the Data Center with Silicon Photonics ECOC 2017

VCSEL-based solderable optical modules

Driving the future of datacenters

High-bandwidth CX4 optical connector

CDFP MSA Delivers 400 Gb/s Today

40 Gbit/s Small-Form Low-Power Optical Transceiver

Kotura Analysis: WDM PICs improve cost over LR4

Optical Trends in the Data Center. Doug Coleman Manager, Technology & Standards Distinguished Associate Corning Optical Communications

Reflex Photonics Inc. The Light on Board Company. Document #: LA Rev 3.1 June 2009 Slide 1

Optical Interconnects: Trend and Applications

Intel: Driving the Future of IT Technologies. Kevin C. Kahn Senior Fellow, Intel Labs Intel Corporation

Photonics Integration in Si P Platform May 27 th Fiber to the Chip

DC Network Connectivity

Using On-Board Optics for Networking Technology Innovation

Singlemode vs Multimode Optical Fibre

WHITE PAPER. Photonic Integration

MARKET PERSPECTIVE: SEMICONDUCTOR TREND OF 2.5D/3D IC WITH OPTICAL INTERFACES PHILIPPE ABSIL, IMEC

400G PAM4 The Wave of the Future. Michael G. Furlong - Senior Director, Product Marketing

Wideband Multimode Fiber What is it and why does it make sense?

Arista 400G Transceivers and Cables: Q&A

Active Optical Cables. Dr. Stan Swirhun VP & GM, Optical Communications April 2008

SOI at the heart of the silicon photonics design. Arnaud Rigny, Business Development Manager Semicon Europa, TechArena

Future Datacenter Interfaces Based on Existing and Emerging Technologies

Organics in Photonics: Opportunities & Challenges. Louay Eldada DuPont Photonics Technologies

Fiber Optic Cabling Systems for High Performance Applications

AIM Photonics: Manufacturing Challenges for Photonic Integrated Circuits

Development of Optical Wiring Technology for Optical Interconnects

2000 Technology Roadmap Optoelectronics. John Stafford, Motorola January 17, 2001

WHITE PAPER. Photonic Integration

1x40 Gbit/s and 4x25 Gbit/s Transmission at 850 nm on Multimode Fiber

VCSEL Technology and Digital

PSMC Roadmap For Integrated Photonics Manufacturing

Cisco 100GBASE QSFP-100G Modules

Integrated Optical Devices

Roadmap to 400 Gigabit Ethernet over Multimode Fiber WHITEPAPER

100G DWDM QSFP Datasheet

Optical Engine? What s That?

Interconnect Challenges in a Many Core Compute Environment. Jerry Bautista, PhD Gen Mgr, New Business Initiatives Intel, Tech and Manuf Grp

A 3-stage CLOS architecture for high-throughput optical packet switching

1 of 5 12/31/2008 8:41 AM

Intel Silicon Photonics: from Research to Product

Packaging for parallel optical interconnects with on-chip optical access

Name of Course : E1-E2 CFA. Chapter 15. Topic : DWDM

Intra Optical Data Center Interconnection Session 2: Debating Intra-DC solutions and Photonic Integration approaches

Leverage Data Center Innovation for Greater Efficiency and Profitability

Imaging, BiCMOS ASIC and Silicon Photonics. Eric Aussedat Executive Vice President General Manager, Imaging, Bi-CMOS ASIC and Silicon Photonics Group

From Majorca with love

Cisco 100GBASE QSFP-100G Modules

Facebook: CWDM4-OCP. 100G Optical Transceiver Specification

Pivotal Issues for 400 Gb/s Ethernet

OIF CEI-56G Project Activity

Next Generation Transceivers: The Roadmap Component Driver Contributions from Roadmap team. Dominic O Brien Mike Schabel

100 Gbps/λ vs. 50 Gbps/λ : Cost and Power Comparisons in Silicon Photonics. Brian Welch

Silicon Photonics and the Future of Optical Connectivity in the Data Center

SILICON PHOTONICS WAVEGUIDE AND ITS FIBER INTERCONNECT TECHNOLOGY. Jeong Hwan Song

ETHERNET OPTICS TODAY: 25G NRZ

The Impact of Emerging Data Rates on Layer One Fiber Cabling Infrastructures. Rick Dallmann Senior Data Center Infrastructure Architect CABLExpress

Non-destructive, High-resolution Fault Imaging for Package Failure Analysis. with 3D X-ray Microscopy. Application Note

Packaging and Integration Technologies for Silicon Photonics. Dr. Peter O Brien, Tyndall National Institute, Ireland.

Insuring 40/100G Performance Multimode Fiber Bandwidth Testing & Array Solutions. Kevin Paschal Sr. Product Manager Enterprise Fiber Solutions

Silicon Photonics: Failing to Deliver on WDM Promises for the Datacenter

40 Gigabit QSFP CWDM 1310 nm LR4 for 10 km

ECOC Market Focus State of the Optical Transport Market

The MIT Communications Technology Roadmap Program IPI TWG Report

FUJITSU COMPONENT LIMITED

PLANAR LIGHTWAVE CIRCUITS FOR USE IN ADVANCED OPTICAL INSTRUMENTATION

Trickle Up: Photonics and the Future of Computing Justin Rattner Chief Technology Officer Intel Corporation

Signature Core Fiber Optic Cabling System

Photon-to-Photon CMOS Imager: Opto-Electronic 3D Integration

Coarse and Dense Wavelength Division Multiplexing

PIC design across platforms. Ronald Broeke Bright Photonics

Fiber Backbone Cabling in Buildings

Trends in Data Centre

Beyond 100 Gb/s High-speed Optical (and Copper) Interconnects

Heterogeneous Integration and the Photonics Packaging Roadmap

How Adolite Breaks the Optical Interconnect Supply Chain Bottleneck

Consensus building discussion on next-generation MMF PMDs. Presented by Robert Lingle, Jr. IEEE Interim Huntington Beach, CA January 9 th, 2017

AMP NETCONNECT XG FIBER SYSTEM

Outline. August 18-19, 1998 NEC

Conflicts in Data Center Fiber Structured Cabling Standards

CHAPTER 1 INTRODUCTION

ExtremeSwitching TM. 100Gb Ethernet QSFP28 Transceivers and Cables. Product Overview. Data Sheet. Highlights. 100Gb Ethernet QSFP28 SR4 MMF

Cabling Solutions for Nexus 9000 Switches With Dual-Rate 40/100G BiDi Transceiver and Panduit Connectivity

HIGH DENSITY DATA CENTER CABLING GUIDE. Optics DACs Cables Cable Management Fiber TAPs Media Converters

Product Data Sheet. 3M Active Optical Cable (AOC) Assemblies for CX4 and QSFP+ Applications

New Data Center and Transport Interconnect Technology

Active Optical Cable. Optical Interconnection Design Innovator 25G SFP28 AOC 10G SFP+ AOC. Shenzhen Gigalight Technology Co.,Ltd.

OPERATIONAL UP TO. 300 c. Microcontrollers Memories Logic

High-Speed Jitter Testing of XFP Transceivers

Silicon Photonics PDK Development

Fiber backbone cabling in buildings

Virtual & Mixed Reality > Near-Eye Displays. Light Propagation through Waveguide with In- & Outcoupling Surface Gratings

Data Center High Speed Migration: Infrastructure issues, trends, drivers and recommendations

A block diagram of the HXFP-XDX1XX XFP optical transceiver is shown below

Cable Network Components Supporting the Infrastructures of the Optical Network

40Gbps QSFP SR4 Optical Transceiver Module

State-of-the-Art and direction for high density optical fiber interconnects

White Paper: Testing Strategies for SYSTIMAX High Speed Migration Solutions

Leaders in the Advancement of Multimode Fiber. Fiber Future and Beyond

PSM4 Technology & Relative Cost Analysis Update

Transcription:

Silicon Based Packaging for 400/800/1600 Gb/s Optical Interconnects The Low Cost Solution for Parallel Optical Interconnects Into the Terabit per Second Age Executive Summary White Paper PhotonX Networks August 2018 Exponential demand for interconnect bandwidth will see data center links running at 400Gb/sec in the near future. Hot on their tails are the anticipated 800 and 1.6Tb/sec links based on either 28 Gbaud-PAM4 or the more future looking 56 Gbaud-PAM4 base rates. Current pluggable module solutions are quickly running out of steam as is evident from the recent annoucement of the COBO specification. How can these complicated and massive interconnect solutions scale to support such massive data loads? The solution is a novel integration platform developed by PhotonX Networks. A wafer scale packaging solution for surface normal Multi-mode light emitters and detectors, which enable a leap forward in bandwidth density at significant lower cost and dramatically simplifies the challenge of scaling to 800,1600 Gb/sec and beyond.

Introduction Data centers have seen exponential growth over the past two decades. Modern data centers host many tens of thousands of servers interconnected with an optical network. While these servers are becoming ever more powerful due to rapid developments in CPU & memory technologies, the main bottleneck in data center performance is becoming the available network bandwidth. Servers are continuously processing data at higher rates necessitating faster transport of data between servers. This puts very stringent demands on the interconnect technology and the design of the switching fabrics. As interconnect technology becomes more and more complex to implement, the price associated with adopting and deploying it grows out of proportion when compared to the cost of compute. Today the bill of materials (BOM) for a high end Ethernet switch in a data center is dominated (>50%) by the cost of the optical interconnect modules. More than that, the power consumption of optical interconnects, at ever increasing bit rates, is also beginning to dominate the total power consumption. At bit rates of 25 Gbaud and beyond, as is proposed for next generation switch fabrics, additional circuitry would be needed to support the connections between switching ASICs and interconnect modules. Current interconnect modules on their own will require >100 Watts. Looking forwards to emerging signaling rates of 25Gbaud (& 50 Gbaud) requiring the inclusion of additional regeneration and clock recovery power consumption is expected to exceed >200 Watts which makes the interconnect consume ~50% of power used by such a switch. It is clear that a new paradigm is needed to solve the interconnect challenges facing data center network builders and operators. PhotonX Networks, a spin-off of the Eindhoven University of Technology, is commercializing a novel and unique solution that will revolutionize the way high capacity parallel optical interconnects are fabricated. By using proprietary technology, already used to demonstrate state-of-the-art 200Gb/sec SR-8 optical engines 1 a low cost and wafer scale packaging solution has been developed which allows co-integration of more than 40 separate optical channels with easy direct cable attached optical interfacing. The critical components are interconnected using carefully designed metallic waveguides to insure optimal signal integrity which has been tested to support 28 Gbaud-PAM4 signaling and possesses a 3dB bandwidth >40GHz. 1 C. Li, O. Raz, F. Kraemer, R. Stabile, Silicon Interposer Based QSFP-DD Transceiver Demonstrator with >10 Gbps/mm2 Bandwidth Density, proc. Of European Conference on Optical Communications 2018, Rome, Italy

Technology breakthrough Optical interconnects, a key component in the DCN (Data Center Network), have been developed and deployed by using different approaches. PhotonX networks, building on the advanced prototyping activities at Eindhoven University of Technology, is pioneering a novel packaging solution for parallel optical interconnects. The solution is based on the use of silicon wafers as a packaging platform for surface normal lasers (vertical cavity surface emitting lasers or VCSELs for short) and detectors which are commonly deployed in short reach data center interconnects today. VCSELs are today the most efficient, lowest cost, and most widely used laser source for interconnects due to their low threshold current, small size, direct modulation and wafer level testing capabilities, as well as their ability to be integrated in multi-component arrays. VCSELs have prevailed in short reach (up to hundreds of meters) interconnects using MM fiber and operating at 850 nm. VCSELs may also find applications for communication over longer distances in data centers if maturity of long wavelength VCSELs and high volume production becomes a reality. High speed VCSELs at 1550 nm have already been shown enabling data transmission at bit rates up to 35 Gb/s in data communication applications. As VCSELs lend themselves easily to production in arrays (1D and 2D), they have found many applications in parallel optical interconnect solutions. The main challenge for parallel optical transceiver design is the packaging of the components. Over the years different substrates have been put forward for the packaging of VCSEL s based interconnects among which: printed and flexible circuit boards, glass and ceramics. More recently there is increased interest in the use of silicon wafers for such packaging solutions. Table 1 summarizes the main attributes of using the different packaging platforms. Will Silicon Photonics solve the interconnect problem? SiPh devices are mostly fabricated on silicon on insulator (SOI) platforms. SOI wafers are composed of a high quality crystalline silicon layer over a buried silicon dioxide (SiO2) layer and a bulk silicon support layer. The top silicon layer can be used to fabricate optical waveguides and devices, because the buried SiO2 enables light propagation on the basis of total internal reflection on the top Si layer. Data center optical interconnects based on SiPh are beginning to emerge with an anticipated sweet spot at 100Gb/sec (QSFP28) for 500-2000 meters transmission distances. Both parallel single mode (PSM) and coarse wavelength division multiplexing (CWDM) have been commercialized. However, because of the lack of silicon light sources a SiPh solution is always dependent on an external light source. This greatly complicates the packaging of SiPh transceivers and leads to higher costs and increased power consumption. Table 1. Comparison of different packaging substrates for parallel optical interconnects Properties PCB FPC GLASS CERAMIC SILICON Cost Very low Low Low Moderate Low Obtainable resolution [micron] >150 >50 >1 >10 >1 Thermal conduction Poor Poor Poor Good Excellent Frequency range [GHz] <10GHz <25GHz >25GHz >25GHz >25GHz

Technology demonstrators In the figure on the left we show an image of a fully assembled 200Gb/sec optical engine based on arrays of 8x25 Gb/sec VCSEL lasers and 8x25 Gb/sec surface normal photo detectors. The module is shown under testing and performance obtained exceeds that of a commercial module. The improved performance is driven by the choice of substrate and the careful design of very short metallic connections between the ICs and the opto-electronic devices. As can be seen in the image below the required metallic connections can be made as short as a few hundreds of microns. Also clearly visible in the image are the gold bumps which are lithographically deposited on the interposer (saving the cost and time of stud bumping the entire wafer) and the optical through holes which are being formed in the process. For devices with an increased number of ports, the same lithographic process can be used to create A 200 Gb/sec prototype optical engine under testing. Whole assembly measures ~6x6 mm leading to a bandwidth density higher than 10 Gb/sec/mm 2 several rows of emitters/detectors as was used for the 200 Gb/sec engine example without loss of performance, thanks to our clever design methodology. Two lay-outs of metallic interconnect highlighting the versatility of the PhotonX packaging concept To complement the electronic part of the solution, a low cost optical interface has been designed and tested based on a collimated MPO connector, modified to directly interface with our optical engine. A side view of the interposer with MPO connector attached in shown below. Obtained insertion loss is ~1dB which is mostly due to collimating loss and not coupling between light emitters/detectors and lenses array. Two lay-outs of metallic interconnect highlighting the versatility of the PhotonX packaging concept

Technology performance The integrated components show excellent performance under testing. Error free operation is easily obtained with performance equivalent of better than commercial devices. The performance of the interconnecting metallic wires has also been characterized beyond the tested performance of the fully integrated module. 28 Gbaud PAM4 signaling has been successfully transmitted without any 28 Gbaud PAM-4 eye patterns and 8x25 Gb/sec NRZ Tx and Rx eye patterns distortion over up to 5 mm of transmission lines on the processed interposer. Measured 8x25 Gb/sec Tx and Rx eye patterns, also show exceptional jitter free operation even when CDR functionality inside TIA CMOS ICs has been switched off! Technology potential The PhotonX technology has been developed in close collaboration with the Technical University of Eindhoven in The Netherlands. The underlying processes and concept collimating in the demonstrators have been patented. As switching ASICs move towards signaling at 50 Gbaud (NRZ and/or PAM-4) the packaging challenges facing traditional techniques are only getting bigger. The PhotonX technology has proven its resilience at ever increasing rates and channel numbers 2. The further increase in the number of channels and single channel speed is actively being pursued. In addition, the potential of combining the technology with silicon photonics is being evaluated Summary The challenges facing manufacturers of optical interconnect modules are putting an increasing strain on the traditional packaging techniques used by the industry. In order to meet the demand for more optical channels running at ever higher speeds, a new packaging technology is needed. PhotonX Networks is introducing a novel packaging solution which has been proven through multiple demonstrators to provide scalability in both number of channel and channel bit rates as well as drastically reduce the cost of packaging. For more info contact us at info@photonxnetworks.com Copyright PhotonX networks 2018 2 C. Li, X. Zhang, T. Li, O. Raz, R. Stabile, 48-Channel Matrix Optical Transmitter on a Single Direct Fiber Connector, IEEE TRANSACTIONS ON ELECTRON DEVICES, 2018