Jana Kosecka. Red-Black Trees Graph Algorithms. Many slides here are based on E. Demaine, D. Luebke slides

Similar documents
COT 6405 Introduction to Theory of Algorithms

Graph Algorithms. Chapter 22. CPTR 430 Algorithms Graph Algorithms 1

CS583 Lecture 09. Jana Kosecka. Graph Algorithms Topological Sort Strongly Connected Component Minimum Spanning Tree

Lecture 6: Analysis of Algorithms (CS )

Outlines: Graphs Part-2

Red-Black Trees. Based on materials by Dennis Frey, Yun Peng, Jian Chen, and Daniel Hood

Graph Representation

Trees and Graphs Shabsi Walfish NYU - Fundamental Algorithms Summer 2006

Graphs. Graph G = (V, E) Types of graphs E = O( V 2 ) V = set of vertices E = set of edges (V V)

Algorithm Design and Analysis

Data Structures. Elementary Graph Algorithms BFS, DFS & Topological Sort

Lecture 10. Elementary Graph Algorithm Minimum Spanning Trees

CSI 604 Elementary Graph Algorithms

Final Exam. EECS 2011 Prof. J. Elder - 1 -

Elementary Graph Algorithms. Ref: Chapter 22 of the text by Cormen et al. Representing a graph:

Review: Graph Theory and Representation

Lecture: Analysis of Algorithms (CS )

Data Structures. Dynamic Sets

Elementary Graph Algorithms


Chapter 22. Elementary Graph Algorithms

Graph Search. Adnan Aziz

Design and Analysis of Algorithms

Representations of Graphs

Search Trees. Undirected graph Directed graph Tree Binary search tree

Graph Algorithms Using Depth First Search

Algorithm Design and Analysis

DFS & STRONGLY CONNECTED COMPONENTS

Minimum Spanning Trees Ch 23 Traversing graphs

Graph Algorithms: Chapters Part 1: Introductory graph concepts

22.3 Depth First Search

Graphs. CSE 2320 Algorithms and Data Structures Alexandra Stefan and Vassilis Athitsos University of Texas at Arlington

CS 220: Discrete Structures and their Applications. graphs zybooks chapter 10

CS 303 Design and Analysis of Algorithms

CS521 \ Notes for the Final Exam

Practical Session No. 12 Graphs, BFS, DFS, Topological sort

Proof: if not f[u] < d[v], then u still grey while v is being visited. DFS visit(v) will then terminate before DFS visit(u).

Algorithms. Deleting from Red-Black Trees B-Trees

TIE Graph algorithms

Graph representation

Section Summary. Introduction to Trees Rooted Trees Trees as Models Properties of Trees

Basic Graph Algorithms

Algorithm Design and Analysis

Graph: representation and traversal

Binary Trees. Recursive definition. Is this a binary tree?

Undirected Graphs. DSA - lecture 6 - T.U.Cluj-Napoca - M. Joldos 1

TIE Graph algorithms

Figure 1: A directed graph.

Algorithms. Red-Black Trees

) $ f ( n) " %( g( n)

Graph Algorithms. Definition

CS 341: Algorithms. Douglas R. Stinson. David R. Cheriton School of Computer Science University of Waterloo. February 26, 2019

CS473-Algorithms I. Lecture 14-A. Graph Searching: Breadth-First Search. Cevdet Aykanat - Bilkent University Computer Engineering Department

Suggested Study Strategy

Data Structures and Algorithms CMPSC 465

Properties of red-black trees

W4231: Analysis of Algorithms

Practical Session No. 12 Graphs, BFS, DFS Topological Sort

CMSC 441: Algorithms. Hillol Kargupta, Professor.

Cut vertices, Cut Edges and Biconnected components. MTL776 Graph algorithms

March 20/2003 Jayakanth Srinivasan,

Binary search trees. Support many dynamic-set operations, e.g. Search. Minimum. Maximum. Insert. Delete ...

Announcements. HW3 is graded. Average is 81%

B Tree. Also, every non leaf node must have at least two successors and all leaf nodes must be at the same level.

Direct Addressing Hash table: Collision resolution how handle collisions Hash Functions:

Copyright 2000, Kevin Wayne 1

Binary search trees. Binary search trees are data structures based on binary trees that support operations on dynamic sets.

Scribes: Romil Verma, Juliana Cook (2015), Virginia Williams, Date: May 1, 2017 and Seth Hildick-Smith (2016), G. Valiant (2017), M.

Binary search trees 3. Binary search trees. Binary search trees 2. Reading: Cormen et al, Sections 12.1 to 12.3

Binary Tree. Preview. Binary Tree. Binary Tree. Binary Search Tree 10/2/2017. Binary Tree

13.4 Deletion in red-black trees

Introduction to Algorithms

8. Write an example for expression tree. [A/M 10] (A+B)*((C-D)/(E^F))

Chapter Summary. Introduction to Trees Applications of Trees Tree Traversal Spanning Trees Minimum Spanning Trees

Elementary Graph Algorithms

CS60020: Foundations of Algorithm Design and Machine Learning. Sourangshu Bhattacharya

CS 310 Advanced Data Structures and Algorithms

CS60020: Foundations of Algorithm Design and Machine Learning. Sourangshu Bhattacharya

CHAPTER 23: ELEMENTARY GRAPH ALGORITHMS Representations of graphs

13.4 Deletion in red-black trees

CMSC th Lecture: Graph Theory: Trees.

Definition of Graphs and Trees. Representation of Trees.

12/5/17. trees. CS 220: Discrete Structures and their Applications. Trees Chapter 11 in zybooks. rooted trees. rooted trees

Design and Analysis of Algorithms Lecture- 9: Binary Search Trees

( ) ( ) C. " 1 n. ( ) $ f n. ( ) B. " log( n! ) ( ) and that you already know ( ) ( ) " % g( n) ( ) " #&

Chapter 22: Elementary Graph Algorithms. Definitions:

Computational Optimization ISE 407. Lecture 16. Dr. Ted Ralphs

Graphs. Part I: Basic algorithms. Laura Toma Algorithms (csci2200), Bowdoin College

Algorithms. 演算法 Data Structures (focus on Trees)

Graph Algorithms. Andreas Klappenecker. [based on slides by Prof. Welch]

ץע A. B C D E F G H E, B ( -.) - F I J K ) ( A :. : ע.)..., H, G E (. י : י.)... C,A,, F B ( 2

SFU CMPT Lecture: Week 9

CS 4407 Algorithms Lecture 5: Graphs an Introduction

Introduction to Algorithms. Lecture 11

Trees Algorhyme by Radia Perlman

Introduction to Algorithms

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Balanced search trees

Undirected Graphs. V = { 1, 2, 3, 4, 5, 6, 7, 8 } E = { 1-2, 1-3, 2-3, 2-4, 2-5, 3-5, 3-7, 3-8, 4-5, 5-6 } n = 8 m = 11

Graph Representations and Traversal

Transcription:

Jana Kosecka Red-Black Trees Graph Algorithms Many slides here are based on E. Demaine, D. Luebke slides

Binary Search Trees (BSTs) are an important data structure for dynamic sets In addition to satellite data, eleements have: key: an identifying field inducing a total ordering left: pointer to a left child (may be NULL) right: pointer to a right child (may be NULL) p: pointer to a parent node (NULL for root)

BST property: key[left(x)] key[x] key[right(x)] Example: F B H A D K

An inorder walk prints the set in sorted order: TreeWalk(x) TreeWalk(left[x]); print(x); TreeWalk(right[x]); Easy to show by induction on the BST property Preorder tree walk: print root, then left, then right Postorder tree walk: print left, then right, then root

TreeSearch(x, k) if (x = NULL or k = key[x]) return x; if (k < key[x]) return TreeSearch(left[x], k); else return TreeSearch(right[x], k);

IterativeTreeSearch(x, k) while (x!= NULL and k!= key[x]) if (k < key[x]) x = left[x]; else x = right[x]; return x;

Adds an element x to the tree so that the binary search tree property continues to hold The basic algorithm Like the search procedure above Insert x in place of NULL Use a trailing pointer to keep track of where you came from (like inserting into singly linked list) Like search, takes time O(h), h = tree height

Basic algorithm: Insert elements of unsorted array from 1..n Do an inorder tree walk to print in sorted order Running time: Best case: Ω(n lg n) (it s a comparison sort) Worst case: O(n 2 ) Average case: O(n lg n) (it s a quicksort!)

Average case analysis It s a form of quicksort! 3 1 8 2 6 7 5 for i=1 to n TreeInsert(A[i]); InorderTreeWalk(root); 3 1 2 8 6 7 5 1 8 2 6 7 5 5 7 2 6 5 7

Minimum: Find leftmost node in tree Successor: x has a right subtree: successor is minimum node in right subtree x has no right subtree: successor is first ancestor of x whose left child is also ancestor of x Intuition: As long as you move to the left up the tree, you re visiting smaller nodes. Predecessor: similar to successor

Delete: x has no children: Remove x B x has one child: Splice out x A x has two children: Swap x with successor C Perform case 1 or 2 to delete it D F H Example: delete K or H or B K

Red-black trees: Binary search trees augmented with node color Operations designed to guarantee that the height h = O(lg n) First: describe the properties of red-black trees Then: prove that these guarantee h = O(lg n) Finally: describe operations on red-black trees

The red-black properties: 1. Every node is either red or black 2. Every leaf (NULL pointer) is black Note: this means every real node has children 3. If a node is red, both children are black Note: can t have 2 consecutive reds on a path 4. Every path from node to descendent leaf contains the same number of black nodes 5. The root is always black

Put example on board and verify properties: 1. Every node is either red or black 2. Every leaf (NULL pointer) is black 3. If a node is red, both children are black 4. Every path from node to descendent leaf contains the same number of black nodes 5. The root is always black black-height: # black nodes on path to leaf Label example with h and bh values

What is the minimum black-height of a node with height h? A: a height-h node has black-height h/2 Theorem: A red-black tree with n internal nodes has height h 2 lg(n + 1)

Thus at the root of the red-black tree: n 2 bh(root) - 1 (Why?) n 2 h/2-1 (Why?) lg(n+1) h/2 (Why?) h 2 lg(n + 1) (Why?) Thus h = O(lg n)

Insert 10 Where does it go? What color? 7 5 9 8 12 1. Every node is either red or black 2. Every leaf (NULL pointer) is black 3. If a node is red, both children are black 4. Every path from node to descendent leaf contains the same number of black nodes 5. The root is always black 10 11

Insert 10 Where does it go? 7 What color? A: no color! Tree 5 9 is too imbalanced 8 Must change tree structure to allow recoloring Goal: restructure tree in O(lg n) time 10 11 12

Our basic operation for changing tree structure is called rotation: Operation on BST which preserves BST property A x y B rightrotate(y) C A leftrotate(x) x B y C Does rotation preserve inorder key ordering? What would the code for rightrotate() actually do?

x y C rightrotate(y) A x y A B B C Answer: A lot of pointer manipulation x keeps its left child y keeps its right child x s right child becomes y s left child x s and y s parents change What is the running time?

Rotate left about 9: 7 5 9 8 12 7 5 12 9 11 8 11

Insertion: the basic idea Insert x into tree, color x red Only r-b property 3 might be violated (if p[x] red) If so, move violation up tree until a place is found where it can be fixed Total time will be O(lg n)

rbinsert(x) treeinsert(x); x->color = RED; // Move violation of #3 up tree, maintaining #4 as invariant: while (x!=root && x->p->color == RED) if (x->p == x->p->p->left) y = x->p->p->right; if (y->color == RED) x->p->color = BLACK; y->color = BLACK; x->p->p->color = RED; x = x->p->p; Case 1 else // y->color == BLACK if (x == x->p->right) x = x->p; leftrotate(x); x->p->color = BLACK; x->p->p->color = RED; Case 2 rightrotate(x->p->p); else // x->p == x->p->p->right (same as above, but with Case 3 right & left exchanged)

rbinsert(x) treeinsert(x); x->color = RED; // Move violation of #3 up tree, maintaining #4 as invariant: while (x!=root && x->p->color == RED) if (x->p == x->p->p->left) y = x->p->p->right; if (y->color == RED) x->p->color = BLACK; y->color = BLACK; x->p->p->color = RED; x = x->p->p; else // y->color == BLACK if (x == x->p->right) x = x->p; leftrotate(x); x->p->color = BLACK; x->p->p->color = RED; Case 2 rightrotate(x->p->p); else // x->p == x->p->p->right (same as above, but with Case 3 right & left exchanged) Case 1:uncle is RED

if (y->color == RED) x->p->color = BLACK; y->color = BLACK; x->p->p->color = RED; x = x->p->p; Case 1: uncle is red In figures below, all Δ s are equal-black-height subtrees Δ A Δ B C x Δ Δ D y Δ case 1 Δ A Δ B C Δ new x Change colors of some nodes, preserving #4: all downward paths have equal b.h. The while loop now continues with x s grandparent as the new x Δ D Δ

if (y->color == RED) x->p->color = BLACK; y->color = BLACK; x->p->p->color = RED; x = x->p->p; Case 1: uncle is red In figures below, all Δ s are equal-black-height subtrees A C D y case 1 A C new x D Δ B x Δ Δ Δ Δ Δ B x Δ Δ Δ Δ Same action whether x is a left or a right child

if (x == x->p->right) x = x->p; leftrotate(x); // continue with case 3 code Case 2: Uncle is black Node x is a right child Transform to case 3 via a leftrotation Δ A Δ B C x Δ Δ y case 2 Δ A x Δ B Δ C Δ y Transform case 2 into case 3 (x is left child) with a left rotation This preserves property 4: all downward paths contain same number of black nodes

x->p->color = BLACK; x->p->p->color = RED; rightrotate(x->p->p); Case 3: Uncle is black Node x is a left child Change colors; rotate right Δ A x Δ B Δ C Δ y case 3 Perform some color changes and do a right rotation Again, preserves property 4: all downward paths contain same number of black nodes Δ x A Δ B Δ C Δ

Cases 1-3 hold if x s parent is a left child If x s parent is a right child, cases 4-6 are symmetric (swap left for right)

And you thought insertion was tricky We will not cover RB delete in class You should read section 14.4 on your own Read for the overall picture, not the details

Coming up: Graph Algorithms

Jana Kosecka Graph Algorithms

A graph G = (V, E) V = set of vertices E = set of edges = subset of V V Thus E = O( V 2 )

Variations: A connected graph has a path from every vertex to every other In an undirected graph: Edge (u,v) = edge (v,u) No self-loops In a directed graph: Edge (u,v) goes from vertex u to vertex v, notated u v

More variations: A weighted graph associates weights with either the edges or the vertices E.g., a road map: edges might be weighted w/ distance A multigraph allows multiple edges between the same vertices E.g., the call graph in a program (a function can get called from multiple points in another function)

We will typically express running times in terms of E and V (often dropping the s) If E V 2 the graph is dense If E V the graph is sparse If you know you are dealing with dense or sparse graphs, different data structures may make sense

Assume V = {1, 2,, n An adjacency matrix represents the graph as a n x n matrix A: A[i, j] = 1 if edge (i, j) E (or weight of edge) = 0 if edge (i, j) E

Example: 1 a 2 d 4 b c 3 A 1 2 3 4 1 2 3?? 4

Example: 1 a 2 d 4 b c 3 A 1 2 3 4 1 0 1 1 0 2 0 0 1 0 3 0 0 0 0 4 0 0 1 0

How much storage does the adjacency matrix require? A: O(V 2 ) What is the minimum amount of storage needed by an adjacency matrix representation of an undirected graph with 4 vertices? A: 6 bits Undirected graph matrix is symmetric No self-loops don t need diagonal

The adjacency matrix is a dense representation Usually too much storage for large graphs But can be very efficient for small graphs Most large interesting graphs are sparse E.g., planar graphs, in which no edges cross, have E = O( V ) by Euler s formula For this reason the adjacency list is often a more appropriate respresentation

Adjacency list: for each vertex v V, store a list of vertices adjacent to v Example: Adj[1] = {2,3 1 Adj[2] = {3 Adj[3] = { Adj[4] = {3 Variation: can also keep a list of edges coming into vertex 2 4 3

How much storage is required? The degree of a vertex v = # incident edges Directed graphs have in-degree, out-degree For directed graphs, # of items in adjacency lists is Σ out-degree(v) = E takes Θ(V + E) storage (Why?) For undirected graphs, # items in adj lists is Σ degree(v) = 2 E (handshaking lemma) also Θ(V + E) storage So: Adjacency lists take O(V+E) storage

Given: a graph G = (V, E), directed or undirected Goal: methodically explore every vertex and every edge Ultimately: build a tree on the graph Pick a vertex as the root Choose certain edges to produce a tree Note: might also build a forest if graph is not connected

Explore a graph, turning it into a tree One vertex at a time Expand frontier of explored vertices across the breadth of the frontier Builds a tree over the graph Pick a source vertex to be the root Find ( discover ) its children, then their children, etc.

Again will associate vertex colors to guide the algorithm White vertices have not been discovered All vertices start out white Grey vertices are discovered but not fully explored They may be adjacent to white vertices Black vertices are discovered and fully explored They are adjacent only to black and gray vertices Explore vertices by scanning adjacency list of grey vertices

BFS(G, s) { initialize vertices; Q = {s; // Q is a queue (duh); initialize to s while (Q not empty) { u = RemoveTop(Q); for each v u->adj { if (v->color == WHITE) v->color = GREY; v->d = u->d + 1; v->p = u; Enqueue(Q, v); u->color = BLACK; What does v->d represent? What does v->p represent?

r s t u v w x y

r s t u 0 v w x y Q: s

r s t u 1 0 1 v w x y Q: w r

r s t u 1 0 2 1 2 v w x y Q: r t x

r s t u 1 0 2 2 1 2 v w x y Q: t x v

r s t u 1 0 2 3 2 1 2 v w x y Q: x v u

r s t u 1 0 2 3 2 1 2 3 v w x y Q: v u y

r s t u 1 0 2 3 2 1 2 3 v w x y Q: u y

r s t u 1 0 2 3 2 1 2 3 v w x y Q: y

r s t u 1 0 2 3 2 1 2 3 v w x y Q: Ø

BFS(G, s) { initialize vertices; Q = {s; while (Q not empty) { u = RemoveTop(Q); for each v u->adj { if (v->color == WHITE) So v = every vertex that appears in some other vert s adjacency list v->color = GREY; v->d = u->d + 1; v->p = u; Enqueue(Q, v); u->color = BLACK; Touch every vertex: O(V) u = every vertex, but only once (Why?) What will be the running time? Total running time: O(V+E)

BFS(G, s) { initialize vertices; Q = {s; while (Q not empty) { u = RemoveTop(Q); for each v u->adj { if (v->color == WHITE) v->color = GREY; v->d = u->d + 1; v->p = u; Enqueue(Q, v); u->color = BLACK; What will be the storage cost in addition to storing the tree? Total space used: O(max(degree(v))) = O(E)

BFS calculates the shortest-path distance to the source node Shortest-path distance δ(s,v) = minimum number of edges from s to v, or if v not reachable from s Proof given in the book (p. 472-5) BFS builds breadth-first tree, in which paths to root represent shortest paths in G Thus can use BFS to calculate shortest path from one vertex to another in O(V+E) time

Depth-first search is another strategy for exploring a graph Explore deeper in the graph whenever possible Edges are explored out of the most recently discovered vertex v that still has unexplored edges When all of v s edges have been explored, backtrack to the vertex from which v was discovered

Vertices initially colored white Then colored gray when discovered Then black when finished

source vertex

source vertex d f 1 Green in figure -> gray in code

source vertex d f 1 2

source vertex d f 1 2 3

source vertex d f 1 2 3 4

source vertex d f 1 2 3 4 5

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time;

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; What does u->d represent?

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; What does u->f represent?

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; Will all vertices eventually be colored black?

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; What will be the running time?

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; Running time: O(n 2 ) because call DFS_Visit on each vertex, and the loop over Adj[] can run as many as V times

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; BUT, there is actually a tighter bound. How many times will DFS_Visit() actually be called?

DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time; So, running time of DFS = O(V+E)

This running time argument is an informal example of amortized analysis Charge the exploration of edge to the edge: Each loop in DFS_Visit can be attributed to an edge in the graph Runs once/edge if directed graph, twice if undirected Thus loop will run in O(E) time, algorithm O(V+E) Considered linear for graph, b/c adj list requires O(V+E) storage Important to be comfortable with this kind of reasoning and analysis

source vertex

source vertex d f 1

source vertex d f 1 2

source vertex d f 1 2 3

source vertex d f 1 2 3 4

source vertex d f 1 2 3 4 5

source vertex d f 1 2 3 4 5 6

source vertex d f 1 8 2 7 3 4 5 6

source vertex d f 1 8 2 7 3 4 5 6

source vertex d f 1 8 2 7 9 3 4 5 6 What is the structure of the green vertices? What do they represent?

source vertex d f 1 8 2 7 9 10 3 4 5 6

source vertex d f 1 8 11 2 7 9 10 3 4 5 6

source vertex d f 1 12 8 11 2 7 9 10 3 4 5 6

source vertex d f 1 12 8 11 13 2 7 9 10 3 4 5 6

source vertex d f 1 12 8 11 13 2 7 9 10 3 4 5 6 14

source vertex d f 1 12 8 11 13 2 7 9 10 3 4 5 6 14 15

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15

DFS introduces an important distinction among edges in the original graph: Tree edge: encounter new (white) vertex The tree edges form a spanning forest Can tree edges form cycles? Why or why not?

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15 Tree edges

DFS introduces an important distinction among edges in the original graph: Tree edge: encounter new (white) vertex Back edge: from descendent to ancestor Encounter a grey vertex (grey to grey)

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15 Tree edges Back edges

DFS introduces an important distinction among edges in the original graph: Tree edge: encounter new (white) vertex Back edge: from descendent to ancestor Forward edge: from ancestor to descendent Not a tree edge, though From grey node to black node

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15 Tree edges Back edges Forward edges

DFS introduces an important distinction among edges in the original graph: Tree edge: encounter new (white) vertex Back edge: from descendent to ancestor Forward edge: from ancestor to descendent Cross edge: between a tree or subtrees From a grey node to a black node

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15 Tree edges Back edges Forward edges Cross edges

DFS introduces an important distinction among edges in the original graph: Tree edge: encounter new (white) vertex Back edge: from descendent to ancestor Forward edge: from ancestor to descendent Cross edge: between a tree or subtrees Note: tree & back edges are important; most algorithms don t distinguish forward & cross

Thm 23.9 (22.10 in 3 rd edition): If G is undirected, a DFS produces only tree and back edges Suppose you have u.d < v.d Then search discovered u before v, so first time v is discovered it is white hence the edge (u,v) is a tree edge Otherwise the search already explored this edge in direction from v to u edge must actually be a back edge since u is still gray

Thm 23.9: If G is undirected, a DFS produces only tree and back edges cannot be a forward edge F? source source C?

Thm: An undirected graph is acyclic iff a DFS yields no back edges If acyclic, no back edges (because a back edge implies a cycle If no back edges, acyclic No back edges implies only tree edges (Why?) Only tree edges implies we have a tree or a forest Which by definition is acyclic Thus, can run DFS to find whether a graph has a cycle

How would you modify the code to detect cycles? DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time;

What will be the running time? DFS(G) { for each vertex u G->V { u->color = WHITE; time = 0; for each vertex u G->V { if (u->color == WHITE) DFS_Visit(u); DFS_Visit(u) { u->color = GREY; time = time+1; u->d = time; for each v u->adj[] { if (v->color == WHITE) DFS_Visit(v); u->color = BLACK; time = time+1; u->f = time;

What will be the running time? A: O(V+E) We can actually determine if cycles exist in O(V) time: In an undirected acyclic forest, E V - 1 So count the edges: if ever see V distinct edges, must have seen a back edge along the way

Thm: If G is undirected, a DFS produces only tree and back edges Thm: An undirected graph is acyclic iff a DFS yields no back edges Thus, can run DFS to find cycles

source vertex d f 1 12 8 11 13 16 2 7 9 10 3 4 5 6 14 15 Tree edges Back edges Forward edges Cross edges