Network Virtualization. Duane de Witt

Similar documents
Cisco ASR 9000 Architecture Overview BRKARC Christian Calixto, IP NGN Consulting Systems Engineer

Configuring the Satellite Network Virtualization (nv) System

Cisco Carrier Packet Transport (CPT) 50 Series

Cisco CPT Packet Transport Module 4x10GE

Cisco ASR 9000v. Product Overview

Configuring the Satellite Network Virtualization (nv) System

ASR Single Chassis Migration to nv Edge System Configuration Example

Configuring Virtual Private LAN Services

Taking MPLS to the Edge. Irit Gillath

XR Release 6.2.x. First Published: Last Modified:

Introduction to Cisco ASR 9000 Series Network Virtualization Technology

Ponoření do architektury ASR9000

Configuring Modular QoS Service Packet Classification

Designing Next Generation Carrier Ethernet Access Network. Benjamin Lee Session ID 20PT Consulting Systems Engineer

Cisco CPT Packet Transport Fabric 256G Fabric Card with 4x10GE

Configuring Traffic Mirroring

Configuring Traffic Mirroring

Cisco Innovations for SP Access

Implementing IP in IP Tunnel

Configuring Link Aggregation

MC-LAG to VPLS Technology and Solution Overview

Junos Fusion Data Center

Configuring Resilient Ethernet Protocol

Carrier Ethernet Services

Configuring Link Aggregation

Pass-Through Technology

Cisco ASR 9000 Series 4-Port and 8-Port 100 Gigabit Ethernet Line Cards

Configuring Interfaces

Configuring Port Channels

Layer 2 Implementation

Cisco EXAM Cisco ADVDESIGN. Buy Full Product.

MX ALS DATACENTER EDGE

NS-090. Carrier Ethernet Based on MPLS-TP SERIES NS: NEW TECHNOLOGIES. PTCL Specifications NS-090 PAKISTAN TELECOMMUNICATION COMPANY LIMITED

Configuring VPLS. VPLS overview. Operation of VPLS. Basic VPLS concepts

ENTERPRISE MPLS. Kireeti Kompella

Configuring Ethernet Virtual Connections on the Cisco ASR 1000 Series Router

ISCOM2948GF-4C Intelligent Ethernet Service Aggregation

Cisco ASR 1000 Series Ethernet Line Cards

Configuring Interfaces

Cisco ASR 1000 Series Ethernet Line Cards

Carrier Ethernet Evolution

Cisco NCS 5011 Router Data Sheet

Transparent SONET or SDH over Packet (TSoP) Protocol

Transparent SONET or SDH over Packet (TSoP) Protocol

Dell Networking OS 9.7 and 6.2

DEPLOYING A STACK OF DELL M-SERIES BLADE SWITCHES IN SIMPLE SWITCH MODE (SSM)

Implementing IEEE 802.1ah Provider Backbone Bridge

Cisco ASR 9001 Router

Overview. Cisco ASR 920 Router Features

Configuration and Management of Networks. Pedro Amaral

L2VPN Protocol-Based CLIs

Configuring Virtual Port Channels

MPLS design. Massimiliano Sbaraglia

High Availability and Redundant Operation

Deep Dive QFX5100 & Virtual Chassis Fabric Washid Lootfun Sr. System Engineer

Arista 7170 series: Q&A

ISCOM RAX 711 (B) Ethernet Demarcation Device

For Ethernet and IP/MPLS networks

VLAN Configuration. Understanding VLANs CHAPTER

Understanding Basic 802.1ah Provider Backbone Bridge

Instant Access - Virtual Switching System Hands on Lab

Cisco Nexus 9508 Switch Power and Performance

Deploying MPLS L2VPN

Junos Fusion. Christoph Plum, Systems Engineer

Configuring StackWise Virtual

For Ethernet and IP/MPLS networks

Understanding the Carrier Packet Transport System

Configuring Virtual Port Channels

Cisco Nexus 9200 Switch Datasheet

Configuring the Fabric Extender

itn201-4gf - Packet Transport Network Device

Configure Virtual LANs in Layer 2 VPNs

Loop-IP6763 TDMoEthernet Aggregator

Data Center InterConnect (DCI) Technologies. Session ID 20PT

Cisco ACI Multi-Pod/Multi-Site Deployment Options Max Ardica Principal Engineer BRKACI-2003

Cisco SD-WAN and DNA-C

Arista 7300X and 7250X Series: Q&A

Cisco SCE 2020 Service Control Engine

Unified RAN Transport Solution: Cisco-SIAE MICROELETTRONICA Interoperability. June, 2012

FibeAir IP-10 Series. Product Guide

Interfaces for Firepower Threat Defense

Overview of Ports and Interfaces

Cisco ME 6524 Ethernet Switch

Configuring VLANs. Understanding VLANs CHAPTER

VPLS configuration commands

Cisco Meraki MS400 Series Cloud-Managed Aggregation Switches

Configuring VLANs. Understanding VLANs CHAPTER

Centec V350 Product Introduction. Centec Networks (Suzhou) Co. Ltd R

ETHERNET Transport Service in Wide Area Network

VeloCloud Cloud-Delivered WAN Fast. Simple. Secure. KUHN CONSULTING GmbH

Ethernet Virtual Connections Configuration

Cisco Evolved Programmable Network Implementation Guide for Large Network with End-to-End Segment Routing, Release 5.0

Arista 7160 series: Q&A

Configuring Port Channels

Configuring Private VLANs

IBM Europe Announcement ZG , dated February 13, 2007

Configuring IEEE 802.1Q Tunneling and Layer 2 Protocol Tunneling

Anticipating the future

Practice exam questions for the Nokia NRS II Composite Exam

Optimised redundancy for Security Gateway deployments

Transcription:

Network Virtualization Duane de Witt

nv Edge System Overview System Deep Dive nv System Overview System Deep Dive

NV EDGE SYSTEM OVERVIEW

SUPERIOR, SIMPLE NETWORK DUAL-HOMING SOLUTION L3 Router dualhoming (L3 ECMP) IP/Service Edge L2 Ethernet Ring (MST/REP-AG, G.8032) IP/MPLS Ethernet spokeand-hub (MC-LAG) nv Edge

EXTEND THE NODES CAPACITY What if I need more line card slots? Here are two solutions: + Solution 1 Add new Router Solution 2 Add new chassis and cluster it with existing Router

Active A Active PW A Active Solution1: MC-LAG + 2-way PW redundancy (Considered one of the best solution in the market) Standby PW LACP S S LACP Standby Standby Solution 2: ASR 9000 nv Edge

Two Routing Adjacency Single Routing Adjacency

SUPER, SIMPLE, FASTER Traditional : Protocol based approach Operationally complex: need multilayer and many different resiliency protocols to address different services and different network topology MC-LAG is active/standby redundancy. Bandwidth is not fully utilized Evolution: nv Edge System Operationally simple: completely replace the dual-homing scenario with singlehoming. No need for those service resiliency protocols Single chassis link bundle support both active/active and active/standby redundancy, bandwidth efficiency and flexibility L2VPN: more nodes need more PWs, more control plane and data plane overhead L3: two separated routing adjacencies. L3 topology changes during link/node failure Flexible deployment: co-locate or Geo-redundant L2VPN: Single node cut the PWs by ¾, control plane and data plane efficiency L3: single routing adjacency. Link/node failure is protected by link bundle, L3 topology is not aware Flexible deployment: co-locate or Georedundant

NV EDGE SYSTEM DEEP DIVE

Control Plane EOBC Extension, two 1G links from each Special external EOBC 1G ports on External EOBC link failure won t cause RP failover as long as it has an alternative EOBC link 0 Active Secondary LC LC LC LC 1 Standby Secondary LC LC LC LC Internal EOBC Inter-chassis data link (L1 connection) 10G or 100 G bundle (up to 32 ports) For best high availability, recommend 4 control plane links, two s on each chassis and a minimum of two data plane links Regular 10G or 100G data ports

Active control plane Standby control plane 0 Active Secondary Standby 1 Standby Active Secondary Standby DSC Chassis LC LC LC LC Non DSC Chassis LC LC LC LC

0 Active Secondary 1 Standby Secondary LC LC LC LC LC LC LC LC Simulated switch fabric

Encapsulation Decapsulation Lookup Load Balance Inter-Chassis Link bundle LOOKUP Encapsulation Decapsulation KEEP EXISTING IOS-XR TWO-STAGE FORWARDING Chassis 0 Chassis 1 Inter- Chassis LC Inter- Chassis LC Data Plane 3 Data Plane 4 Ingress LC P1 P2 P1 P2 Egress LC Data Plane Data Plane P1 P 1 P 2 Inter- Chassis LC Inter- Chassis LC P2 1 2 Data Plane Data Plane 5 3 4 P1 P2 P1 P2

Packets will cross interchassis links ONLY if it s not dual-homing multicast unicast L3 ECMP Active/active link bundle

LOSING ALL CONTROL PLANE LINKS 0 Active Secondary 1 Standby Secondary DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC

LOSING ALL DATA PLANE LINKS 0 Active Secondary 1 Standby Secondary DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC

LOSING ALL CONTROL PLANE AND DATA PLANE LINKS Optional keepalive over any L2 cloud 0 Active Secondary 1 Standby Secondary Management port DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC Any regular data port interface or sub-interface

NV SATELLITE SYSTEM OVERVIEW

UNI L2: VLAN provisioning, S-tag, etc L3: routing, vrf, bfd, etc NNI L2: MC-LAG, REP, STP, etc L3: IGP/LDP/BFD, etc MPLS

Protocol nv fabric links One virtual system Host is plug-n-play, zero configuration for opex savings

Power Feeds Redundant -48vDC Power Feeds Single AC power feed 1 RU ANSI & ETSI Compliant Field Replaceable Fan Tray Redundant Fans ToD/PSS Output Bits Out 44x10/100/1000Mbps Pluggables Full Line Rate Packet Processing and Traffic Management Copper and fiber SFP optics Speed/duplex auto negotiation 4x10G SFP+ Initially used as Fabric Ports ONLY (could be used as access port in the future) Copper and fiber SFP+ optics Max Power 210 Watts Nominal Power 159 Watts Industrial Temp Rated -40C to +65C Operational Temperature -40C to +70C Storage Temperature

Note, not all SW features are supported in satellite mode in 4.3.0. For example, TDM interface is not supported ASR 901-E (Ethernet only) Metro/Carr. Eth Switch w/ 12 GE ports ASR 901 (TDM + Ethernet) Cell Site Router w/ 16 T1/E1 + 12 GE ports Accelerating the migration from 2G/3G to 4G/LTE Cisco ASR 901 Cell Site Router for 2G, 3G & 4G Compact 1RU, ETSI 300mm depth, < 40W Hardened/Extended temp range -40c to +65c Reliable Power Supply: Dual line feed Redundant power supply (optional) Flexible LTE ready Pay-as-you-grow license model Scalable 12 GE ports + 16 T1/E1 ports L2 Switching, L3 Routing capabilities with MPLS, QOS at line rate SyncE, 1588v2 BC/TC, 10Mhz, BITS, 1PPS, ToD, p2p

Note, not all SW features are supported in satellite mode in 4.3.0. For example, TDM interface is not supported Cisco ASR 903 Unified Ethernet Access Breaking the silos of Residential, Business & Mobile Compact 3RU, 6 Interface Slots Fits in 300mm cabinets (9.2 deep) Reliable Redundant PSUs, FANs and s Extended operating temp. range -40 to 65 C Scalable Ethernet : 1x10GE and 8xGE TDM/ATM: 16x T1/E1 and 4x STM1 / 1x STM4 Roadmap CHOC-12

FTTB/CE Access CE Aggregation 5 ASR 9000 nv Systems cluster Plug-n-Play, zero touch remote satellite Single feature set across small and big POP sites Single user CLI and management across small and big POP sites One ASR 9000 nv System Huge GE ports fan-out Simplified accessaggregation network redundancy Single feature set across access and aggregation Single user CLI and management across access and aggregation

Next-gen BNG One ASR 9000 nv System nv cluster Mobile Access CO 9k BNG 9k BNG One ASR 9000 nv System Stateful failover Simplified network redundancy Simplified access switch VLAN provisioning Single user CLI and management across access and aggregation Plug-n-Play, zero touch cell site router 1000s cell site routers are managed by single ASR 9K system, huge operational saving: image management and configuration Simplified features on the cell site satellite

NV SATELLITE SYSTEM DEEP DIVE

CPU MAC-DA MAC-SA Control VID Payload/FCS CPU ASR 9000v Discovery Phase A CDP-like link-level protocol that discovers satellites and maintains a periodic heartbeat Heartbeat sent once every second, used to detect satellite or fabric link failures. BFD based fast failure detection plan for future release ASR 9000 Host Control Phase Used for Inter-Process Communication between Host and Cisco proprietary protocol over TCP socket for the time being. It could move to standard in the future Get/ Set style messages to provision the satellites and also to retrieve notifications from the satellite

MAC-DA MAC-SA VLANs (OPT) Payload MAC-DA MAC-SA VLANs (OPT) Payload MAC-DA MAC-SA nv-tag VLANs (OPT) Payload/FCS ASR 9000v ASR 9000 Host

ONLY does local connect between access and fabric port access port to normal port communication ASR 9000v ASR 9000 Host

All the control plane protocols (L2 and L3) are all running on the ASR9K Host ASR 9000 nv System does the simple local connect, no FIB/MAC required doesn t run any control plane protocol. However, it does run some limited local features like QoS, T1/E1 framing Virtual interface works and feels just as local interface interface GigabitEthernet 100/0/0/1 ipv4 address 2.2.2.2 255.255.255.0 interface GigabitEthernet 100/0/0/2.2 l2 encap dot1q 2 rewrite ingress tag pop 1 sym

44x1GE Access ports 4x10GE Fabric ports 44x1GE Access ports 4x10GE Fabric ports Model 1: Static pinning No fabric port redundancy It can t mix model 1 and 2 on the same satellite Model 2: Fabric bundle Fabric port redundancy

Model 1: Static pinning No fabric port redundancy Model 2: Fabric bundle Fabric port redundancy Mixed model example, NOT supported in 4.2.1 Mixed model example, NOT supported in 4.2.1

ASR 9000 Cluster Single home, static pinning Phase 1 target Single home, fabric link bundle Dual home to cluster, static pinning ASR 9000 Cluster Dual home to cluster with fabric link bundle

Single satellite dual home to two standalone ASR9K Single ring dual-home to two standalone ASR9K cascading to single ASR9K

Virtual L2 Fabric link: single or dual home 2-level spoke-and-hub 2-level cascaded ring Single ring + cascading

SATELLITE VIRTUAL L2 FABRIC LINK MAC-DA MAC-SA T-VLAN nv-tag VLANs (OPT) Payload/FCS Transport VLAN (T-VLAN) is used for packet forwarding in the cloud Native L2 (802.1q) hand off Each L2 sub-interface is mapped to one satellite fabric port

SATELLITE DUAL-HOME TO TWO SEPARATED HOSTS L1 fabric link L2 fabric link For both L1 or L2 fabric links Run in active/standby mode, satellite maps the access port ONLY to the primary Host fabric ports When it detects a fabric link failure to the primary Host, re-maps the port to the backup Host L2 and L3 service works the similar way as MC-LAG

SATELLITE RING TOPOLOGY Each satellite is assigned unique T-VLAN for forwarding in the ring Primary fabric link Phase 1 target If the T-VLAN belongs to the satellite itself, then it does the access-fabric port xconnect, otherwise, it does fabric-fabric xconnect If satellite detects primary fabric link fail, it remaps access port to its backup fabric port No Local switching within the ring Efficient local multicast replication can still be achieved Backup fabric link Physical Ring Topology Logical Spoke-and-Hub Topology with L2 Fabric Links

CE CE MC-LAG on the Host CE CE Regular link bundle on the Cluster

Support either access bundle or fabric port bundle, but can t support both at same time on the same satellite Bundle member ports across satellite access port and local port is not supported CE CE