Network Virtualization Duane de Witt
nv Edge System Overview System Deep Dive nv System Overview System Deep Dive
NV EDGE SYSTEM OVERVIEW
SUPERIOR, SIMPLE NETWORK DUAL-HOMING SOLUTION L3 Router dualhoming (L3 ECMP) IP/Service Edge L2 Ethernet Ring (MST/REP-AG, G.8032) IP/MPLS Ethernet spokeand-hub (MC-LAG) nv Edge
EXTEND THE NODES CAPACITY What if I need more line card slots? Here are two solutions: + Solution 1 Add new Router Solution 2 Add new chassis and cluster it with existing Router
Active A Active PW A Active Solution1: MC-LAG + 2-way PW redundancy (Considered one of the best solution in the market) Standby PW LACP S S LACP Standby Standby Solution 2: ASR 9000 nv Edge
Two Routing Adjacency Single Routing Adjacency
SUPER, SIMPLE, FASTER Traditional : Protocol based approach Operationally complex: need multilayer and many different resiliency protocols to address different services and different network topology MC-LAG is active/standby redundancy. Bandwidth is not fully utilized Evolution: nv Edge System Operationally simple: completely replace the dual-homing scenario with singlehoming. No need for those service resiliency protocols Single chassis link bundle support both active/active and active/standby redundancy, bandwidth efficiency and flexibility L2VPN: more nodes need more PWs, more control plane and data plane overhead L3: two separated routing adjacencies. L3 topology changes during link/node failure Flexible deployment: co-locate or Geo-redundant L2VPN: Single node cut the PWs by ¾, control plane and data plane efficiency L3: single routing adjacency. Link/node failure is protected by link bundle, L3 topology is not aware Flexible deployment: co-locate or Georedundant
NV EDGE SYSTEM DEEP DIVE
Control Plane EOBC Extension, two 1G links from each Special external EOBC 1G ports on External EOBC link failure won t cause RP failover as long as it has an alternative EOBC link 0 Active Secondary LC LC LC LC 1 Standby Secondary LC LC LC LC Internal EOBC Inter-chassis data link (L1 connection) 10G or 100 G bundle (up to 32 ports) For best high availability, recommend 4 control plane links, two s on each chassis and a minimum of two data plane links Regular 10G or 100G data ports
Active control plane Standby control plane 0 Active Secondary Standby 1 Standby Active Secondary Standby DSC Chassis LC LC LC LC Non DSC Chassis LC LC LC LC
0 Active Secondary 1 Standby Secondary LC LC LC LC LC LC LC LC Simulated switch fabric
Encapsulation Decapsulation Lookup Load Balance Inter-Chassis Link bundle LOOKUP Encapsulation Decapsulation KEEP EXISTING IOS-XR TWO-STAGE FORWARDING Chassis 0 Chassis 1 Inter- Chassis LC Inter- Chassis LC Data Plane 3 Data Plane 4 Ingress LC P1 P2 P1 P2 Egress LC Data Plane Data Plane P1 P 1 P 2 Inter- Chassis LC Inter- Chassis LC P2 1 2 Data Plane Data Plane 5 3 4 P1 P2 P1 P2
Packets will cross interchassis links ONLY if it s not dual-homing multicast unicast L3 ECMP Active/active link bundle
LOSING ALL CONTROL PLANE LINKS 0 Active Secondary 1 Standby Secondary DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC
LOSING ALL DATA PLANE LINKS 0 Active Secondary 1 Standby Secondary DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC
LOSING ALL CONTROL PLANE AND DATA PLANE LINKS Optional keepalive over any L2 cloud 0 Active Secondary 1 Standby Secondary Management port DSC Chassis Non DSC Chassis LC LC LC LC LC LC LC LC Any regular data port interface or sub-interface
NV SATELLITE SYSTEM OVERVIEW
UNI L2: VLAN provisioning, S-tag, etc L3: routing, vrf, bfd, etc NNI L2: MC-LAG, REP, STP, etc L3: IGP/LDP/BFD, etc MPLS
Protocol nv fabric links One virtual system Host is plug-n-play, zero configuration for opex savings
Power Feeds Redundant -48vDC Power Feeds Single AC power feed 1 RU ANSI & ETSI Compliant Field Replaceable Fan Tray Redundant Fans ToD/PSS Output Bits Out 44x10/100/1000Mbps Pluggables Full Line Rate Packet Processing and Traffic Management Copper and fiber SFP optics Speed/duplex auto negotiation 4x10G SFP+ Initially used as Fabric Ports ONLY (could be used as access port in the future) Copper and fiber SFP+ optics Max Power 210 Watts Nominal Power 159 Watts Industrial Temp Rated -40C to +65C Operational Temperature -40C to +70C Storage Temperature
Note, not all SW features are supported in satellite mode in 4.3.0. For example, TDM interface is not supported ASR 901-E (Ethernet only) Metro/Carr. Eth Switch w/ 12 GE ports ASR 901 (TDM + Ethernet) Cell Site Router w/ 16 T1/E1 + 12 GE ports Accelerating the migration from 2G/3G to 4G/LTE Cisco ASR 901 Cell Site Router for 2G, 3G & 4G Compact 1RU, ETSI 300mm depth, < 40W Hardened/Extended temp range -40c to +65c Reliable Power Supply: Dual line feed Redundant power supply (optional) Flexible LTE ready Pay-as-you-grow license model Scalable 12 GE ports + 16 T1/E1 ports L2 Switching, L3 Routing capabilities with MPLS, QOS at line rate SyncE, 1588v2 BC/TC, 10Mhz, BITS, 1PPS, ToD, p2p
Note, not all SW features are supported in satellite mode in 4.3.0. For example, TDM interface is not supported Cisco ASR 903 Unified Ethernet Access Breaking the silos of Residential, Business & Mobile Compact 3RU, 6 Interface Slots Fits in 300mm cabinets (9.2 deep) Reliable Redundant PSUs, FANs and s Extended operating temp. range -40 to 65 C Scalable Ethernet : 1x10GE and 8xGE TDM/ATM: 16x T1/E1 and 4x STM1 / 1x STM4 Roadmap CHOC-12
FTTB/CE Access CE Aggregation 5 ASR 9000 nv Systems cluster Plug-n-Play, zero touch remote satellite Single feature set across small and big POP sites Single user CLI and management across small and big POP sites One ASR 9000 nv System Huge GE ports fan-out Simplified accessaggregation network redundancy Single feature set across access and aggregation Single user CLI and management across access and aggregation
Next-gen BNG One ASR 9000 nv System nv cluster Mobile Access CO 9k BNG 9k BNG One ASR 9000 nv System Stateful failover Simplified network redundancy Simplified access switch VLAN provisioning Single user CLI and management across access and aggregation Plug-n-Play, zero touch cell site router 1000s cell site routers are managed by single ASR 9K system, huge operational saving: image management and configuration Simplified features on the cell site satellite
NV SATELLITE SYSTEM DEEP DIVE
CPU MAC-DA MAC-SA Control VID Payload/FCS CPU ASR 9000v Discovery Phase A CDP-like link-level protocol that discovers satellites and maintains a periodic heartbeat Heartbeat sent once every second, used to detect satellite or fabric link failures. BFD based fast failure detection plan for future release ASR 9000 Host Control Phase Used for Inter-Process Communication between Host and Cisco proprietary protocol over TCP socket for the time being. It could move to standard in the future Get/ Set style messages to provision the satellites and also to retrieve notifications from the satellite
MAC-DA MAC-SA VLANs (OPT) Payload MAC-DA MAC-SA VLANs (OPT) Payload MAC-DA MAC-SA nv-tag VLANs (OPT) Payload/FCS ASR 9000v ASR 9000 Host
ONLY does local connect between access and fabric port access port to normal port communication ASR 9000v ASR 9000 Host
All the control plane protocols (L2 and L3) are all running on the ASR9K Host ASR 9000 nv System does the simple local connect, no FIB/MAC required doesn t run any control plane protocol. However, it does run some limited local features like QoS, T1/E1 framing Virtual interface works and feels just as local interface interface GigabitEthernet 100/0/0/1 ipv4 address 2.2.2.2 255.255.255.0 interface GigabitEthernet 100/0/0/2.2 l2 encap dot1q 2 rewrite ingress tag pop 1 sym
44x1GE Access ports 4x10GE Fabric ports 44x1GE Access ports 4x10GE Fabric ports Model 1: Static pinning No fabric port redundancy It can t mix model 1 and 2 on the same satellite Model 2: Fabric bundle Fabric port redundancy
Model 1: Static pinning No fabric port redundancy Model 2: Fabric bundle Fabric port redundancy Mixed model example, NOT supported in 4.2.1 Mixed model example, NOT supported in 4.2.1
ASR 9000 Cluster Single home, static pinning Phase 1 target Single home, fabric link bundle Dual home to cluster, static pinning ASR 9000 Cluster Dual home to cluster with fabric link bundle
Single satellite dual home to two standalone ASR9K Single ring dual-home to two standalone ASR9K cascading to single ASR9K
Virtual L2 Fabric link: single or dual home 2-level spoke-and-hub 2-level cascaded ring Single ring + cascading
SATELLITE VIRTUAL L2 FABRIC LINK MAC-DA MAC-SA T-VLAN nv-tag VLANs (OPT) Payload/FCS Transport VLAN (T-VLAN) is used for packet forwarding in the cloud Native L2 (802.1q) hand off Each L2 sub-interface is mapped to one satellite fabric port
SATELLITE DUAL-HOME TO TWO SEPARATED HOSTS L1 fabric link L2 fabric link For both L1 or L2 fabric links Run in active/standby mode, satellite maps the access port ONLY to the primary Host fabric ports When it detects a fabric link failure to the primary Host, re-maps the port to the backup Host L2 and L3 service works the similar way as MC-LAG
SATELLITE RING TOPOLOGY Each satellite is assigned unique T-VLAN for forwarding in the ring Primary fabric link Phase 1 target If the T-VLAN belongs to the satellite itself, then it does the access-fabric port xconnect, otherwise, it does fabric-fabric xconnect If satellite detects primary fabric link fail, it remaps access port to its backup fabric port No Local switching within the ring Efficient local multicast replication can still be achieved Backup fabric link Physical Ring Topology Logical Spoke-and-Hub Topology with L2 Fabric Links
CE CE MC-LAG on the Host CE CE Regular link bundle on the Cluster
Support either access bundle or fabric port bundle, but can t support both at same time on the same satellite Bundle member ports across satellite access port and local port is not supported CE CE