Global Routing Instabilities Triggered by Code Red II and Nimda Worm Attacks Λ

Similar documents
Global Routing Instabilities

Toward Understanding the Behavior of BGP During Large-Scale Power Outages

BGP Routing: A study at Large Time Scale

Understanding BGP Behavior through a Study of DoD Prefixes Λ

Detecting Behavior Propagation in BGP Trace Data Brian J. Premore Michael Liljenstam David Nicol

Measurement of Highly Active Prefixes in BGP

Happy Packets: Some Initial Results

Dynamics of Hot-Potato Routing in IP Networks

On the Evaluation of AS Relationship Inferences

The missing links in the BGP-based AS connectivity maps

Internet Routing Dynamics

On characterizing BGP routing table growth

Measuring BGP. Geoff Huston. CAIA SEMINAR 31 May

BGP Routing Stability of Popular Destinations

Flooding Attacks by Exploiting Persistent Forwarding Loops

J. A. Drew Hamilton, Jr., Ph.D. Director, Information Assurance Laboratory and Associate Professor Computer Science & Software Engineering

Internet Routing Protocols Lecture 01 & 02

BGP Scaling Techniques

USING SIMULATION TO UNDERSTAND DYNAMIC CONNECTIVITY AT THE CORE OF THE INTERNET

Lecture outline. Internet Routing Security Issues. Previous lecture: Effect of MinRouteAdver Timer. Recap of previous lecture

An Empirical Study of Router Response to Large BGP Routing Table Load

Taming BGP. An incremental approach to improving the dynamic properties of BGP. Geoff Huston. CAIA Seminar 18 August

Visualization of Internet Traffic Features

Internet inter-as routing: BGP

Internet Routing Protocols Lecture 03 Inter-domain Routing

Traffic in Network /8. Background. Initial Experience. Geoff Huston George Michaelson APNIC R&D. April 2010

BGP-4 Protocol Patterns and Their Impact on QoS Behavior

Investigating occurrence of duplicate updates in BGP announcements

Early Detection of BGP Instabilities Resulting from Internet Worm Attacks

Network Protocols. Routing. TDC375 Autumn 03/04 John Kristoff - DePaul University 1

BGP Scaling Techniques

BGP Routing inside an AS

Measurement: Techniques, Strategies, and Pitfalls. David Andersen CMU

On the Impact of Route Processing and MRAI Timers on BGP Convergence Times

internet technologies and standards

Configuring BGP. Cisco s BGP Implementation

Measuring and Characterizing IPv6 Router Availability

BGP for Internet Service Providers

BGP Route Flap Damping Algorithms

On Routing Table Growth

A Measurement Study of BGP Misconfiguration

Implementation of BGP in a Network Simulator

Network Working Group. Category: Experimental June 1996

Network Working Group. J. Scudder Cisco Systems, Inc. February 2001

AS Connectedness Based on Multiple Vantage Points and the Resulting Topologies

BGP. Autonomous system (AS) BGP version 4. Definition (AS Autonomous System)

Border Gateway Protocol Complexity and Stability

Operation Manual BGP. Table of Contents

CSCD 433/533 Network Programming Fall Lecture 14 Global Address Space Autonomous Systems, BGP Protocol Routing

Border Gateway Protocol (an introduction) Karst Koymans. Monday, March 10, 2014

BGP Route Flap Damping Algorithms

Routing, Routing Algorithms & Protocols

Routing in Geoff Huston Chief Scientist, APNIC

Network Forensics Prefix Hijacking Theory Prefix Hijacking Forensics Concluding Remarks. Network Forensics:

SaaS Providers. ThousandEyes for. Summary

RIPE NCC Routing Information Service (RIS)

Routing Basics. What is Routing? Routing Components. Path Determination CHAPTER

ECE 428 Internet Protocols (Network Layer: Layer 3)

Interdomain Routing Reading: Sections K&R EE122: Intro to Communication Networks Fall 2007 (WF 4:00-5:30 in Cory 277)

The Interconnection Structure of. The Internet. EECC694 - Shaaban

Simulating Internet Scale Topologies with Metarouting

Internet Routing : Fundamentals of Computer Networks Bill Nace

An Empirical Study of Behavioral Characteristics of Spammers: Findings and Implications

Chapter 13 Configuring BGP4

Congestion Propagation among Routers in the Internet

CS4450. Computer Networks: Architecture and Protocols. Lecture 15 BGP. Spring 2018 Rachit Agarwal

Parallel Routing Table Computation for Scalable IP Routers

Quantifying Path Exploration in the Internet

Virtual Multi-homing: On the Feasibility of Combining Overlay Routing with BGP Routing

Routing Geoff Huston Chief Scientist, APNIC. #apricot2017

BGP. BGP Overview. Formats of BGP Messages. I. Header

Configuring BGP community 43 Configuring a BGP route reflector 44 Configuring a BGP confederation 44 Configuring BGP GR 45 Enabling Guard route

Quantifying Path Exploration in the Internet

Routing on the Internet. Routing on the Internet. Hierarchical Routing. Computer Networks. Lecture 17: Inter-domain Routing and BGP

BGP Path Exploration Damping (PED)

Network Protocols. Routing. TDC375 Winter 2002 John Kristoff - DePaul University 1

Investigating occurrence of duplicate updates in BGP announcements

Routing the Internet in Geoff Huston APNIC March 2007

BGP Route Flap Damping Algorithms

CSCD 433/533 Advanced Networks Spring 2016

A Measurement Study on the Impact of Routing Events on End-to-End Internet Path Performance

CS4700/CS5700 Fundamentals of Computer Networks

BGP Commands. Network Protocols Command Reference, Part 1 P1R-355

BGP The Movie. Geoff Huston September 2004 APNIC

Resilient IP Backbones. Debanjan Saha Tellium, Inc.

BGP. Autonomous system (AS) BGP version 4

Introduction. Keith Barker, CCIE #6783. YouTube - Keith6783.

CS519: Computer Networks. Lecture 4, Part 5: Mar 1, 2004 Internet Routing:

CS 640: Introduction to Computer Networks. Intra-domain routing. Inter-domain Routing: Hierarchy. Aditya Akella

ITEC310 Computer Networks II

Network Working Group Request for Comments: 2519 Category: Informational Juniper February A Framework for Inter-Domain Route Aggregation

PREVENTING NETWORK INSTABILITY CAUSED BY PROPAGATION OF CONTROL PLANE POISON MESSAGES *

s i g c o m m Public Review for BGP Routing Dynamics Revisited Jun Li, Michael Guidero, Zhen Wu, Eric Purpus, and Toby Ehrenkranz

Request for Comments: 3345 Category: Informational AOL Time Warner, Inc. D. Walton A. Retana Cisco Systems, Inc. August 2002

BGP Issues. Geoff Huston

Multicast Routing Instabilities

A Longitudinal Study of BGP MOAS Prefixes

BGP. Autonomous system (AS) BGP version 4

THE Internet is an interconnection of separately administered

COMP/ELEC 429 Introduction to Computer Networks

Transcription:

Global Routing Instabilities Triggered by Code Red II and Nimda Worm Attacks Λ James Cowie, Andy T. Ogielski, BJ Premore y, and Yougu Yuan y Renesys Corporation Hanover, NH 03750 December 2001 Abstract We analyze the large, long-lasting, widespread instabilities of the global BGP routing system observed during the Code Red II and Nimda worm attacks in July and September 2001, respectively. The identification and characterization of global routing instabilities employs heuristic spatio-temporal correlation analysis of multiple BGP message streams collected from over 150 autonomous systems border routers in the RIPE RIS project, and their correlation with the worm traffic is exposed by the analysis of TCP packet traces collected in several /16 networks during the worm attacks. We analyze router failure modes that can be triggered by such abnormal traffic and lead to destabilization of the BGP routing system. To further illustrate the occurrence of cascading routing failures we also present data on another type of global routing instabilities associated with common router misconfigurations generating malformed BGP update messages. Our results show previously unrecognized global routing failure modes, and suggest new research directions. 1 Introduction The Internet is highly dynamic in two major respects: first, the user traffic exhibits enormous variability that has been extensively measured and studied [31, 36, 35], second, the IP routing topology changes at a relatively high 1 This study arose from work partially supported by the Defense Advanced Research Projects Agency (DARPA), under grant N66001-00- 8065 from the U.S. Department of Defense. Its contents are solely the responsibility of the authors and do not necessarily represent the official views of the Department of Defense. This is the extended version of the preliminary report [12]. 2 BJP and YY are also at Dartmouth College, Hanover, NH. This research was performed during summer internship at Renesys Corp. rate. The global routing dynamics is considerably less understood, and to date there have been few measurementbased analyses [30, 28]. Recent availability of dedicated routers collecting raw BGP message traces from many default-free peers at several well-connected locations [39] offers the opportunity to accurately track the dynamic changes of the global routing topology (union of valid best routes from any location to every reachable prefix), delayed only by seconds to minutes. Such a detailed dynamical view of routing changes cannot be easily obtained from the analysis of periodically collected BGP routing tables, which have been successfully exploited to analyze the gross Internet topology [20, 16, 7] and its growth trends [22, 8, 42] at the autonomous system (AS) level. The focus of this paper is on the emergence of longlived routing instability events observed in the collective dynamic behavior of the streams of BGP route update messages between June and October 2001. The timestamped BGP message streams collected from 162 border routers from 115 autonomous systems and stored in the RIPE RIS database [39] have been analyzed and correlated at multiple resolutions. In July and September 2001 we detected hours-long periods of exponential growth and decay in the route change rates, across all default-free peering sessions and most prefixes, indicating significant widespread degradation in the end-to-end functioning of the global Internet. Contrary to popular lore, these events did not correlate with localized failures in the Internet infrastructure, such as fiber cuts or power outages. Instead, we have documented a compelling connection between global routing instabilities and the propagation phase of Microsoft worms such as Code Red and Nimda: What were believed to be attacks targeting solely Microsoft servers in fact also turned out to generate widespread endto-end routing instabilities. To explain these phenomena, we rely on case reports we received from many network operators and adminis-

trators, reports posted on network security mailing lists, as well as on the router vendor advisories [11, 9, 10], and on router stress-test results [6]. We present an analysis of the mechanisms by which the peculiar characteristics of the worm-generated traffic high rate scanning for potentially vulnerable hosts can destabilize the global routing by triggering BGP session failures. Although these particular worms do not infect the routers themselves, they impact the routers indirectly through router failures due to excessive CPU and memory utilization, excessive BGP message traffic, as well as through more obscure effects such as proactive reconfiguration or disconnection of certain routers by network administrators, and by other means. The volume of worm traffic generated by Code Red II and Nimda, however, did not appear to be high enough to cause significant congestion losses. In order to further explore the propagation of routing instabilities in a simpler context, we also analyze the propagation of common instabilities triggered by misconfigured confederation routers that send easily identifiable malformed route announcements. While the BGP standards [38] require that any BGP router receiving a malformed update terminate the offending session in order to contain the error, this is not the case in practice: certain widely deployed RFC-nonconforming routers will reannounce such updates to their peers, causing the downstream RFC-conforming routers to reset, which in turn causes their neighbors to send a flood of updates. This behavior persists until the error-originating router is reconfigured, which as we have seen may take many days. The plan of the paper is as follows: Section 2 puts our results in the context of related research. Section 3 presents the methodology of our studies and the characteristics of the BGP data sources. Sections 4, 5 and 6 contain the analysis of routing instabilities and reachability failures observed during the worm spread periods while Section 7 discusses possible mechanisms of BGP router failures under worm-generated traffic conditions. Section 8 extends the analysis of routing instability mechanisms by focusing on a common type of routing instabilities associated with router misconfigurations. Preliminary conclusions and suggestions for future research are in Section 9. 2 Related Research Considerable work has been published on studies of BGP and inter-domain routing. Due to our focus on empirical analyses, we will review only that prior research which dealt with measured BGP data. Govindan and Reddy were the first to examine the ASlevel Internet graph topology and route stability, showing that certain properties, such as diameter, remain unchanged despite growth [20]. Faloutsos, Faloutsos and Faloutsos described power-laws related to AS-level topology metrics such as the vertex-degree distribution [16]. Chen, Chang, Govindan, Jamin, Shenker, and Willinger subsequently showed that using additional data, a more complete picture of Internet topology could be obtained which did not match the power-laws as well [8]. All three of these studies relied heavily on BGP routing tables as a source of data. Broido and Claffy report specifically on the process of analyzing these tables, including using multiple merged tables, and further suggest a new method for reducing redundant data [1, 2]. While we also use BGP data extensively, our focus is not on analyzing the general features of the AS-level Internet topology, but on identification and characterization of instances of global routing instability. The value of BGP table statistics has prompted a number of free public resources which periodically publish such information. Notable examples are those of Huston [21], NLANR [17], and APNIC [5]. Typical statistics are forwarding table size (number of prefixes), AS and prefix path length distribution, and prefix aggregation information. While these are valuable statistics, our analysis required a more detailed spatio-temporal look at BGP message arrivals. The first research to look specifically at BGP messages was conducted by Labovitz, Malan and Jahanian [30]. Their data came from core border routers, and they observed that an extremely large fraction of BGP updates were pathological. Their continued study identified several origins of the misbehavior and eventually resulted in major reductions in BGP update volume [28]. While we too have observed continuously large volumes of BGP traffic, we take particular interest in anomalies in the collective dynamic behavior of BGP message traffic. In this report, we restrict ourselves to a search for the causes of these anomalies, which stand out in stark contrast to the typical pulse of inter-domain routing traffic. 3 Methodological Background To set the context, consider that at the end of 2001 the global Internet has been a coalition of over 12,000 autonomous systems (ASes). An AS a corporation, network provider, or another entity is an administratively closed collection of IP networks communicating using a unified internal routing policy, that is globally identified by an assigned autonomous system number (ASN). Worldwide IP communication among the ASes is established in a multitude of business agreements, exclusively using the Border Gateway Protocol (BGP) [38] for distributing the routing information among BGP routers in 2

all ASes. Therefore, BGP failures and instabilities immediately affect the connectivity of the global Internet. When a BGP router s preferred route to a given network prefix has changed, it sends out a BGP update message to each connected peer router. Therefore, by establishing BGP peering connections with a large number of BGP routers from well-connected organizations, analysis of traffic gathered at a single BGP monitoring point can provide a great deal of information about the way those organizations view the Internet, and about the dynamics of how paths change over a wide range of time scales. The RIPE RIS project [39] maintains several such monitoring points across Europe; they peer with many of the so-called global tier-1 providers, plus very many smaller regional European networks. Access to multiple BGP monitoring points provides opportunities to filter the localized infrastructure failures that are close to individual collection points, clearing the way to unambiguously identify and study routing instability features that affect large portions of the Internet simultaneously. The results shown in this article are primarily based on analysis of time-stamped BGP messages collected at the RIPE NCC site (rrc00) and the AMS-IX site (rrc03), both in Amsterdam, the Netherlands. We also analyzed BGP traffic from other Internet exchanges that host RIPE RIS collection sites, including LINX (London), SFINX (Paris), CIXP (Geneva), and VIX (Vienna). Details of the extended analysis are omitted for brevity. The raw BGP message archives at RIPE contain various errors and anomalies that require certain amount of care in their analysis, in particular: occasional timestamp clock shifts, missing data, corrupted MRT headers, truncated BGP messages, and the opening and closing of collecting routers BGP sessions. The RIPE NCC facility is particularly interesting, as it collects BGP routing updates from several large Internet providers via multi-hop links, and thus provides a good and fairly complete dynamic view second by second of the evolving state of global routing. The autonomous systems that had routers peering with the RIPE NCC collecting router in the period covered in this paper are listed in table 1 There are two primary strategies to measure global Internet instability: ffl Reachability changes: measuring the number and distribution of prefixes that appear in the routing tables, and their change in time. ffl Route changes: measuring the number of prefix advertisements and withdrawals in BGP update messages sent out per unit time. The BGP protocol contains route flap dampening AS286 AS513 AS1103 AS2914 AS3257 AS3549 AS3549 AS4608 AS4777 AS7018 AS9177 AS13129 KPNQwest Backbone CERN SURFnet Verio Tiscali Global Global Crossing UK Global Crossing USA Telstra Internet APNIC Tokyo AT&T Internet4 Nextra (Schweiz) Global Access Telecommunications Table 1: Autonomous systems peering with RIPE NCC router. mechanism that prevents a BGP router from sending too many messages about an unstable route [43]; and a timer (the Minimum Route Advertisement Interval Timer) that maintains a minimum separation between consecutive announcements to a given peer, with default value of 30 seconds. Therefore, if we see large increases in the number of BGP update messages, it s an unambiguous sign that the diversity of network prefixes is rising. Furthermore, the duration of these BGP message traffic surges, and the rate of their growth are what distinguish global instabilities from the pervasive background noise, daily rhythms and localized failures. We operationally define a global routing instability in terms of its rate, duration and diversity as follows: Exponential or similarly fast growth of the rate of prefix updates, high update rates lasting hours to days, with almost all prefixes churning, in BGP updates from almost all default-free peers. Very short, high spikes in advertisement rates are very common whenever a peer BGP session undergoes a hard reset, for example, a full table dump will follow. More surprisingly, our examination of the data indicates that localized failures in the core Internet infrastructure (fiber cuts, flooding, power failures, building collapses) tend to generate only short-term increases in the external BGP prefix advertisement rate, which decline in a matter of minutes as the highly redundant peering in the core Internet topology routes around the damage, although specific networks may remain unreachable until the damage is repaired. Of far greater concern are the appearance of sustained rises in aggregate BGP update rates that last for hours. To expose their mechanisms, in this work we look for route rate change correlations among peers, origin ASes, prefixes, prefix lengths, and route lifetimes. 3

Figure 1: Aggregate rate of BGP prefix advertisements, notifications, and opens in 30-second bins at the RIPE NCC collecting router from June through September 2001. The x-axis is time, and the y-axis is the log of the total number of network addresses (prefixes) advertised in consecutive 30-second windows. Each dot in the plot represents a 30-second count of the number of announced prefixes received by rrc00 from all its peers. 4 Global Routing Stability: Summer of 2001 Figure 1 shows the trends in the aggregated rate of BGP prefix advertisements received from all peers of the RIPE NCC collection router from 1 June through 24 September 2001. This time series represents a measure of the coarse-resolution route change activity for the entire Internet. The aggregation exposes several strong patterns and features, for example: ffl One can observe strong weekly and daily trends in the median, an effect which may be due to interactions with either the diurnal patterns of traffic, or the diurnal patterns of activity by network operators performing routine maintenance on BGP routers. The weekly trend ramps up from Monday through Wednesday, and then steadily declines towards the weekend. ffl Very common high and narrow peaks in update rates also tend to follow a weekly trend, thinning over the weekends, and are assumed to be dominated by router reconfigurations, session resets and other network maintenance, although other time-localized causes cannot be excluded at this time. ffl Two strange non-periodic features jump out when the baseline is examined: A nearly tenfold magnitude rise in the baseline on 19 July (see Figure 2) A more rapidly rising and longer-lasting rise in the baseline on 18-22 September (see Figure 5). Figure 2: A zoom-in on the BGP message storm of 19 July. Note that the above aggregated time series does not serve as a measurement of reachability over time this would be achieved for instance by plotting the fraction of prefixes with valid routes, and watching for dips. At this aggregation level, and in the absence of internal BGP updates in the aggregate, there are few interesting features when network infrastructure is broken at localized geographic points. Neither the Baltimore tunnel train wreck that severed multiple fiber links on 18 July nor the attacks of 11 September appear as features in this plot. These events did not destabilize the global Internet. In general, the high levels of routing activity following fiber cuts between tier-1 and other major providers remain localized within the immediately affected autonomous systems, and do not create external BGP message storms that are highly visible worldwide. 4

Figure 3: Left: BGP prefix advertisements. Right: BGP prefix withdrawals. Each row represents the time series from one BGP peer in RIPE NCC aggregated in 60-minute bins. Note the high correlation of the waves of withdrawals and advertisements across all peers on 19 July. Figure 4: Code Red II port 80 probes per hour recorded in two unrelated /16 networks. 5 Code Red II and Nimda Events In particular, we are concerned with the two non-periodic features visible on 19 20 July and 18 22 September. These two storms in BGP update rates correlate with the propagation phases of the Microsoft worms known as Code Red II (in July) and Nimda (in September). On 19 July, we observed an exponentially growing eight-fold increase in the advertisement rate, over a period of about eight hours (all times are in GMT; subtract 4 hours for EDT). This surge faded over the same time scale as it arrived. When one considers the current estimates of BGP convergence times (several minutes [29]), it is more than a little disturbing to see a fundamental quantity like BGP advertisement rate exhibiting exponential growth for eight hours. One initial guess was a delayed effect from the 18 July Baltimore tunnel fiber cuts, and whose impact was highly visible in the discussions on network operators mailing lists such as NANOG [32], as network engineers tweaked routing for the next day or two. But this does not appear to have caused the BGP storm. In order to gain a better understanding of the mechanism driving this BGP storm, we conduct a finer analysis. We began by separating the contributions of individual BGP peering sessions to the total BGP update message traffic, and by separately following the time courses of BGP route advertisement and withdrawal messages. In Figure 3, BGP prefix advertisements and withdrawals are graphed along the z-axis as impulses. The x-axis is time, from 1 July through 31 July. The y-axis 5

(going into the page) separates the contributions of the 13 individual peer ASes at the rrc00 collecting router. On other days, it is common for individual peers to contribute spikes of high prefix count, reflecting BGP sessions closing and opening close to the collection point. On 19 July, however, all peers experience a wave of updates, sustained for many hours. Further analysis of the BGP message traffic has indicated that no specific autonomous system or set of autonomous systems seems to be generating the traffic surge, and that no specific IP prefix or set of prefixes was flapping significantly more than before the onset of the surge (see Section 5). Instead, the net effect was that routes to most of the approximately 110,000 prefixes in the Internet were changing more than normal: The data reflect a broad-based BGP update storm with no apparent single-point cause. Correlation with Code Red II Attack The time course of the 19 July BGP storm suggests that it has been triggered by the sudden spread of a variant of the Microsoft worm known as Code Red II [13]. Our analysis has been materially aided by quantitative data and qualitative reports on the network security mailing lists [23, 26, 33] concerning worm scans rates, sudden connectivity losses, ARP storms, and other worm effects. Ideal data on the worm-generated traffic storm would show the time series of worm activity for a good statistical sample of networks of known size, so that the global activity levels could be inferred by extrapolation. However, only a limited number of such datasets are available. Figure 4 shows the Code Red II propagation data collected independently on two /16 networks (each nominally containing 64k IP addresses) during the entire day of 19 July. The original data were obtained from [19, 15], also summarized in [24, 25]. The two time series show the number of TCP SYN packets received in the two distinct /16 networks hour after hour. Note the virtually identical time course of these attacks as seen from different networks. These plots give a measure of the intensity of worm scanning traffic at all affected networks. We also analyzed the tcpdump trace for the network shown on the left [14], which further revealed very high diversity of source and destination IP addresses in the worm scan packets: during most of the worm attack on this /16 network there were over 3,000 unique attacking hosts in any one-minute interval, at rates exceeding 8,000 probes per minute. On a longer time scale, the corresponding aggregate rates were over 100,000 unique attacking hosts per hour, and about 500,000 probes per hour, that were probing over 90% of the address space in this network. We will return to this point later. Figure 5: A zoom-in on the BGP message storm of 18 22 September. Further infrmation on the Code Red II virus is available from [18]. Nimda Worm Attack On Tuesday, 18 September, simultaneous with the onset of the propagation phase of the Nimda worm, we observed another BGP storm. This one came on faster, rode the trend higher, and then turned itself off, though much more slowly. Over a period of roughly two hours, starting at about 13:00 GMT, the rrc00 aggregate BGP advertisement rates exponentially ramped up by a factor of 25, from 400 per minute to 10,000 per minute, with sustained gusts to more than 200,000 per minute. The advertisement rate then decayed gradually over several days, reaching pre-nimda levels by 24 September. Figure 6: BGP prefix withdrawals in 15-minute periods. Each row is one BGP peer in RIPE NCC. Note the steep onset of a wave of withdrawals on 18 September. Plotted in the back is the Nimda worm scan rate (not to scale) observed in two /16 networks. 6

The analysis of the BGP storm triggered by the Nimda worm follows a similar course as the analysis of Code Red presented above. By separating the contributions of individual BGP peering sessions to the total BGP update message traffic at RIPE NCC collecting router, and by separately following the time courses of BGP route advertisement and withdrawal messages we can demonstrate correlation across all peers. Similar analyses were performed also for the other data collecting routers. The steep exponential growth of the 18 September BGP storm is aligned with the exponential spread of Nimda, the most virulent Microsoft worm seen to that date. The Nimda worm exhibits extremely high scan rates, multiple attack modes generating very heavy traffic, and has been much more damaging than the July Code Red worm [37, 41]. We have also analyzed the tcpdump traces of the Nimda worm attack on a /16 network [14]. The time series in Figure 9 illustrates the trend in the worm scanning rate. Notice a faster onset than with the 19 July Code Red II attack. It is interesting to speculate that narrow peaks in the worm scan rate on the three or four occasions preceding the actual attack may represent trials of the worm. 6 Network Reachability Failures During the Code Red and Nimda Attacks The temporal correlation of the worm attacks and routing instabilities shown above does not tell much about the statistics of affected prefixes and routes, and in particular does not characterize the populations of high churn networks. In this Section we argue that Nimda and Code Red triggered long-term BGP instabilities unlike any localized network failure, in particular that a detailed analysis shows no suspect prefixes most prefixes churn, and no suspect routes most routes churn. Since the analysis is qualitatively similar for both worm attacks, we will not duplicate each example. Figure 7 compares the populations of prefixes churning during a quiet hour (weekend night of July 7, with very low rates of BGP updates) and during the peak hour of the Code Red II attack (July 19). The data represent all prefixes seen in the BGP updates (announcements and withdrawals) at the AMS-IX collecting router (rrc03). It is seen that the primary difference is in the sizes of prefix populations, with no other obvious patterns that might localize the sources of routing instability. We have also analyzed the count of withdrawals and announcements for each prefix shown in Figure 7, with the result that the primary difference is in their number, and again no obvious pattern can be found. Figure 8 shows that all prefix lengths were similarly affected by the effects of the Nimda worm attack on global routing stability. A preliminary comparative analysis of the BGP routing tables collected a few hours before and a few hours after the onset of the Nimda storm shows that the increase in prefix withdrawals during the worm period, while significant, did not result in long-lasting losses of reachability to the edge networks. These were, by and large, transient failures. 7 Mechanisms of BGP Router Failures During the Worm Attacks In general, long lasting, high diversity, high rate route churn can be produced by any mechanism that is causing a large number of BGP sessions to close and reopen repeatedly 1. At this time we are not aware of any direct measurements showing step-by-step how the the worm propagation traffic causes BGP router failures, moreover, it is expected that different router models and software releases will show different failure modes, and different failure recovery. However, there is strong evidence that BGP sessions do fail under several stress conditions (such as excessive memory demands [6]), and due to software bugs. The router vendor advisories [11, 9, 10] list the primary causes of BGP protocol and router failures: ffl router CPU overload ffl out of memory ffl overflows ffl router software bugs Case reports obtained by us from network engineers and administrators who observed and described the effects of the Code Red and Nimda worm attacks, as well as the router vendor advisories, indicate that while these particular worms do not infect the routers themselves, they do impact the routers indirectly via CPU overload (for example, protocol and slow path processing, high number of flows, high interrupt rates), excessive memory demands (for example, large numbers of routes or flows), and cache overflows (for example, ARP storms). Each of these may 1 Another mechanism would require a very large number of non-bgp originated, globally announced networks to keep failing and recovering repeatedly 7

0 prefixes seen in a quiet hour, rrc03 prefixes seen during a busy hour, rrc03 32 32 24 24 prefix length 16 prefix length 16 8 8 0 0.0.0.0 128.0.0.0 192.0.0.0 255.255.255.255 0.0.0.0 128.0.0.0 192.0.0.0 255.255.255.255 IPv4 address space IPv4 address space Figure 7: Left: Prefixes churning during a typical quiet hour of low instability (July 7, AMS-IX). Right: Prefixes churning during the peak Code Red hour (July 19, AMS-IX). The x-axis represents the numerical values of IP addresses from 0.0.0.0 to 255.255.255.255, the y-axis represents the prefix length. The band pattern reflects the allocation of IP addresses to existing networks. Figure 8: Smoothed rate of prefix withdrawals in 30-second intervals for distinct prefix lengths during the Nimda worm attack. cause either a BGP session failure, or the router failure. The following list shows the aspects of the worm traffic, and worm-induced BGP traffic, that may lead to such failure conditions: ffl traffic diversity (number of flows), ffl traffic intensity and congestion losses, ffl high BGP message load, ffl vulnerability of HTTP servers in routers (management interfaces), ffl failures in network gear (such as DSL routers and other components), ffl IGP (Intra-AS) flapping and routing failures, ffl proactive reconfiguration or disconnection of routers by operators. Among the case reports we obtained detailed measurements of the effect of the worm traffic on router processing load. Figure 9 shows a striking correlation between the Nimda worm scan rate in one /16 corporate network in the Midwest, and a router CPU utilization in a tier-1 provider s network on the West Coast. This router s CPU utilization increases four fold from 10% to over 40% at the peak of Nimda attack. So far, data analysis indicates that the most likely common cause of worm-induced router failures is the worm traffic diversity, that is abnormally high number of IP packet source-destination address pairs seen in a short time under conditions of heavy scanning traffic. This peculiar traffic pattern may not be included in standard router test scenarios, leaving the routers highly vulnerable to such traffic conditions. While failure details may vary, one failure mode may involve generation of extremely many short flow records in a very short time, stressing both the CPU and router memory. Another failure mode may be triggered by ARP storms on directly attached networks targeted by the worm probes. Once BGP traffic increases significantly due to multiple router failures, another BGP router failure scenario becomes possible: Extremely high BGP update rates, impacting the router CPU and memory, can either lead to BGP message losses or to malloc failures. Such a scenario is particularly troubling, as it has the potential for triggering cascading failures. Finally, it has to be noted that exponential spread of the worm may cause a large number of network operators from corporations and smaller ISPs at the Internet s edge to independently shut down, or reboot, or attempt to 8

8 Misconfiguration Instabilities According to BGP standards [38], upon receiving a malformed BGP message the BGP speaker must send a notification to the offending peer, and subsequently close this BGP session. The standards do not specify how quickly this session may be re-established. However, not all deployed BGP routers conform to the RFC, creating the possibility of the following scenario and its variants: 1. A misconfigured router starts announcing a private AS number in a (confederation) ASPATH. 2. Certain routers ignore but propagate the malformed route 3. Other, RFC-compliant routers close and reopen the BGP sessions. 4. The combination may propagate wildly depending on the routers in the neighborhood. Figure 9: Top: time course of the Nimda scan rate in a corporate /16 network. Bottom: time course of a large router s CPU utilization in a tier-1 provider s network thousands of miles away, in the overlapping time frame. 5. Instability ends only when the original leak is plugged. reconfigure their border routers; and the total amount of BGP message traffic grows exponentially with the number of edge domains feeling the effects of the worm. However, according to simple bandwidth utilization estimates based on measured worm scan rates it appears that congestion losses of BGP messages are not a likely effect with these particular worm spread scenarios. Although one would expect BGP messages to be very high priority traffic, and thus not subject to congestion-related loss until the situation were dire indeed (such prioritization has been shown to be a good thing [40]), it s not clear that network operators routinely enable this kind of prioritization. In summary, we tentatively conclude that due to large diversity of network hardware and software the worminduced routing instability results from the combination of effects listed above. This does not make it easy to propose a simple conceptual model amenable to theoretical or simulation-based analyses. However, we have also observed a common occurrence of relatively simpler global routing instabilities that may be much easier to analyze. They are briefly described in the following section. Figure 10: October 6-15 routing instabilities associated with the propagation of malformed BGP update messages. The bars in the background show the time and intensity of malformed update leakages from several ASes (not the same scale as the BGP prefix withdrawal time series). RIPE rrc00 location, 15- minute prefix withdrawal counts. We have recorded numerous cases of BGP instabilities that were correlated with malformed ASPATH announcements, for instance in Figures 3 and 6 they can be seen as smaller instability waves correlated across all peers. In Figure 10 we show a recent instance recorded on October 9

8 that has developed into a substantial global instability event. We are currently investigating this in greater detail, to be reported separately. We have already begun building simulation models to understand the spread of such instabilities in greater detail. 9 Conclusions and Future Research Analysis of BGP message streams collected from multiple routers is only beginning, and it is fair to say that we have barely scratched the surface. At this point, several broad observations are in order: First, there is a strong interdependence between the routing topology and the end-user traffic: Route fluctuations and failures necessarily lead to data traffic failures along affected paths, and, vice versa, certain abnormal data traffic conditions may trigger routing failures as we show in this paper. These interactions are addressed to some extent in this article, and should be further investigated in future research. Second, this work demonstrates the importance of correlating network measurements concerning both routing and user traffic. Not doing so may well result in incorrect analyses, as recently demonstrated by a commercial traffic monitoring service [27, 34] who denied any impact of the worm attacks on the Internet performance by relying solely on monitoring test traffic latencies and losses among agents placed in the largest Internet core backbones, and thus entirely missing the action. In the context of large network instabilities, it is of interest to notice that according to recent research on the dynamics of complex engineered systems [4, 3] the Internet is likely to exhibit unexpected catastrophic failure modes. The empirical analysis of BGP routing failures demonstrated in this paper directly leads to a number of new research directions. Among those are: the need for standardizing the router behavior during recovery from failures to prevent same failure recurrency, graceful degradation of routing protocols running on routers faced with excessive demands on their CPU and memory resources, and detection and containment of attacks on the routing infrastructure. 10 Acknowledgments Thanks to Henk Uijterwaal and his group for the the RIPE RIS project which has been the source of the raw BGP message traffic archives explored in this paper. Vicki Irwin, Ken Eichman and Vern Paxson kindly offered us worm traffic traces from several /16 networks. Special thanks are due to very many network engineers and administrators for sending us detailed case histories and observations on router misbehaviors under worm traffic stress conditions. Finally, we acknowledge interesting discussions with Dave Donoho, Tim Griffin, and others at the 2001 Leiden Workshop on Multiresolution Analysis of Global Internet Measurements. References [1] Andre Broido and kc claffy. Analysis of RouteViews BGP data: Policy atoms, May 2001. Network Resource Data Management Workshop. [2] Andre Broido and kc claffy. Internet topology: Connectivity of IP graphs, August 2001. SPIE International Symposium on Convergence of IT and Communication. [3] J. M. Carlson and J. Doyle. Highly optimized tolerance: A mechanism for power laws in designed systems. Phys. Rev. E 60, 1412, 1999. [4] J. M. Carlson and J. Doyle. Power laws, highly optimized tolerance and generalized source coding. Phys. Rev. Lett. 84, 2529, 2000. [5] Asia Pacific Network Information Centre. BGP statistics. http://www.apnic.net/stats/bgp/. [6] Di-Fa Chang, Ramesh Govindan, and John Heidemann. An empirical study of router response to large BGP routing table load. Technical Report ISI-TR-2001-552, USC/Information Sciences Institute, December 2001. [7] Hyunseok Chang, Ramesh Govindan, Sugih Jamin, Scott J. Shenker, and Walter Willinger. On inferring ASlevel connectivity from BGP route tables. ACM Internet Measurements Workshop 2001. [8] Q. Chen, H. Chang, R. Govindan, S. Jamin, S. Shenker, and W. Willinger. The origin of power laws in Internet topologies revisited. In Proceedings of INFOCOM 2002, 2002. [9] Cisco. Troubleshooting high CPU utilization on Cisco routers. http://www.cisco.com/warp/- public/63/highcpu.html. [10] Cisco. Cisco security advisory: ICMP unreachable vulnerability in Cisco 12000 Series Internet router, November 2001. http://www.cisco.com/warp/public/- 707/GSR-unreachables-pub.shtml. [11] Cisco. Dealing with mallocfail and high CPU utilization resulting from the Code Red worm, October 2001. http://www.cisco.com/warp/public/- 63/ts codred worm.shtml. [12] James Cowie, Andy Ogielski, BJ Premore, and Yougu Yuan. Global routing instabilities during Code Red II and Nimda worm propagation: Preliminary report, September 2001. http://www.renesys.com/projects/- bgp instability/. 10

[13] eeye Digital Security..ida Code Red worm. http://- www.eeye.com/html/research/advisories/- AL20010717.html. [14] Ken Eichman. tcpdump traces of the Code Red II and Nimda attacks. [15] Ken Eichman. incidents.org mailing list: Re: Possible CodeRed connection attempts. http://lists.- jammed.com/incidents/2001/07/0159.html. [16] Michalis Faloutsos, Petros Faloutsos, and Christos Faloutsos. On power-law relationsships of the Internet topology. In Proceedings of ACM SIGCOMM 1999, pages 251 262, 1999. [17] National Laboratory for Applied Network Research. Analysis and visualization of BGP statistics using vbns GateD log data. http://calorie.nlanr.net/bgp/. [18] Cooperative Association for Internet Data Analysis. Caida analysis of code-red. http://www.caida.org/- analysis/security/code-red/. [19] Dave Goldsmith. incidents.org mailing list: Possible CodeRed connection attempts. http://lists.- jammed.com/incidents/2001/07/0149.html. [20] Ramesh Govindan and Anoop Reddy. An analysis of Internet inter-domain topology and route stability. In Proceedings of INFOCOM 1997, 1997. [21] Geoff Huston. BGP table data. http://www.- telstra.net/ops/bgp/. [22] Geoff Huston. Analyzing the Internet s BGP routing table. The Internet Protocol Journal, 4(1), March 2001. http://www.cisco.com/warp/public/- 759/ipj 4-1/ipj 4-1 bgp.html. [23] incidents.org by The SANS Institute. http://www.- incidents.org/. [24] Vicki Irwin. incidents.org: Handler s diary 07/20/01. http://www.incidents.org/- archives/intrusions/msg01134.html. [25] Vicki Irwin. incidents.org: Handler s diary 07/21/01. http://www.incidents.org/- archives/intrusions/msg01137.html. [26] jammed.com. http://www.jammed.com/. [27] Keynote. Press announcement: September 20, 2001. http://www.keynote.com/press/html/- nimda 092001.html. [28] C. Labovitz, G.R. Malan, and F. Jahanian. Origins of Internet routing instability. In Proceedings of INFOCOM 1999, pages 218 226, 1999. [29] Craig Labovitz, Adha Ahuja, Abhijit Bose, and Farnam Jahanian. Delayed Internet routing convergence. In Proceedings of SIGCOMM 1999, 1999. [30] Craig Labovitz, G. Robert Malan, and Farnam Jahanian. Internet routing instability. In Proceedings of ACM SIG- COMM 1997. Association for Computing Machinery, Inc., 1997. [31] Will E. Leland, Murad S. Taqqu, Walter Willinger, and Daniel V. Wilson. On the self-similar nature of Ethernet traffic. In Proceedings of SIGCOMM 1993, September 1993. [32] North American Network Operators Group (NANOG). http://www.nanog.org/. [33] Neohapsis. http://www.neohapsis.com/. [34] BBC News. Code Red was never a threat, August 2001. http://news.bbc.co.uk/hi/english/- sci/tech/newsid 1470000/1470246.stm. [35] K. Park, G. Kim, and M. Crovella. On the relationship between file sizes, transport protocols, and self-similar network traffic. In Proceedings of IEEE International Conference on Network Protocols 1996, pages 171 180, 1996. [36] V. Paxson and S. Floyd. Wide-area traffic: The failure of Poisson modeling. IEEE/ACM Transactions on Networking, 3:226 244, June 1995. [37] Kevin Poulsen. Nimda worm hits net. http://www.- securityfocus.com/news/253. [38] Y. Rekhter and T. Li. RFC 1771: A Border Gateway Protocol 4 (BGP-4). Technical report, Internet Engineering Task Force, March 1995. [39] Reséaux IP Européens (RIPE). Routing Information Service (RIS). http://www.ripe.net/ripencc/- pub-services/np/ris-index.html. [40] A. Shaikh, L. Kalampoukas, R. Dube, and A. Varma. Routing stability in congested networks: Experimentation and analysis. In Proceedings of ACM SIGCOMM 2000, pages 163 174, 2000. [41] System Administration, Networking, and Security Institute (SANS). Nimda worm/virus report. http://www.- incidents.org/react/nimda.pdf. [42] Hongsuda Tangmunarunkit, John Doyle, Ramesh Govindan, Sugih Jamin, Scott J. Shenker, and Walter Willinger. Does AS size determine degree in AS topology? ACM Computer Communication Review, October 2001. [43] C. Villamizar, R. Chandra, and R. Govindan. RFC 2439: BGP route flap damping. Technical report, Internet Engineering Task Force, November 1999. 11