CS 270 Algorithms. Oliver Kullmann. Analysing BFS. Depth-first search. Analysing DFS. Dags and topological sorting.

Similar documents
Week 5. 1 Analysing BFS. 2 Depth-first search. 3 Analysing DFS. 4 Dags and topological sorting. 5 Detecting cycles. CS 270 Algorithms.

CS 270 Algorithms. Oliver Kullmann. Breadth-first search. Analysing BFS. Depth-first. search. Analysing DFS. Dags and topological sorting.

Elementary Graph Algorithms

Design and Analysis of Algorithms

Representations of Graphs

Graph Representation

Chapter 22. Elementary Graph Algorithms

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Graph Algorithms. Definition

Graphs. Graph G = (V, E) Types of graphs E = O( V 2 ) V = set of vertices E = set of edges (V V)

Graph representation

Graph: representation and traversal

Proof: if not f[u] < d[v], then u still grey while v is being visited. DFS visit(v) will then terminate before DFS visit(u).

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Graph Algorithms. Chapter 22. CPTR 430 Algorithms Graph Algorithms 1

CS 341: Algorithms. Douglas R. Stinson. David R. Cheriton School of Computer Science University of Waterloo. February 26, 2019

Graph Algorithms: Chapters Part 1: Introductory graph concepts

Minimum Spanning Trees Ch 23 Traversing graphs

Taking Stock. IE170: Algorithms in Systems Engineering: Lecture 16. Graph Search Algorithms. Recall BFS

DFS & STRONGLY CONNECTED COMPONENTS

Graphs. Graphs. Representation of Graphs. CSE 680 Prof. Roger Crawfis. Two standard ways. Graph G = (V, E)» V = set of vertices

CS473-Algorithms I. Lecture 14-A. Graph Searching: Breadth-First Search. Cevdet Aykanat - Bilkent University Computer Engineering Department

COT 6405 Introduction to Theory of Algorithms

Part VI Graph algorithms. Chapter 22 Elementary Graph Algorithms Chapter 23 Minimum Spanning Trees Chapter 24 Single-source Shortest Paths

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Lecture 10. Elementary Graph Algorithm Minimum Spanning Trees

Introduction to Algorithms. Lecture 11

Graph implementations :

Taking Stock. IE170: Algorithms in Systems Engineering: Lecture 17. Depth-First Search. DFS (Initialize and Go) Last Time Depth-First Search

Basic Graph Algorithms

Single Source Shortest Path

Outlines: Graphs Part-2

22.3 Depth First Search

CS 310 Advanced Data Structures and Algorithms

CSI 604 Elementary Graph Algorithms

Unit 2: Algorithmic Graph Theory

Week 12: Minimum Spanning trees and Shortest Paths

Algorithm Design and Analysis

Announcements. HW3 is graded. Average is 81%

CSE 331: Introduction to Algorithm Analysis and Design Graphs

Graph Search. Adnan Aziz

Announcements Problem Set 4 is out!

Data Structures and Algorithms. Chapter 7. Graphs

Data Structures and Algorithms. Werner Nutt

Graph Theory. Many problems are mapped to graphs. Problems. traffic VLSI circuits social network communication networks web pages relationship

Data Structures. Elementary Graph Algorithms BFS, DFS & Topological Sort

CHAPTER 23: ELEMENTARY GRAPH ALGORITHMS Representations of graphs

Topic 12: Elementary Graph Algorithms

Design and Analysis of Algorithms

Figure 1: A directed graph.

Elementary Graph Algorithms. Ref: Chapter 22 of the text by Cormen et al. Representing a graph:

CS 161 Lecture 11 BFS, Dijkstra s algorithm Jessica Su (some parts copied from CLRS) 1 Review

5.4 Shortest Paths. Jacobs University Visualization and Computer Graphics Lab. CH : Algorithms and Data Structures 456

CS Elementary Graph Algorithms

CS Elementary Graph Algorithms

Graph: representation and traversal

Basic Graph Definitions

Graph. Vertex. edge. Directed Graph. Undirected Graph

CISC 320 Introduction to Algorithms Fall Lecture 15 Depth First Search

CS2 Algorithms and Data Structures Note 10. Depth-First Search and Topological Sorting

Directed Graphs. DSA - lecture 5 - T.U.Cluj-Napoca - M. Joldos 1

Unit 5F: Layout Compaction

Unit 3: Layout Compaction

CS 220: Discrete Structures and their Applications. graphs zybooks chapter 10

Graphs. Part I: Basic algorithms. Laura Toma Algorithms (csci2200), Bowdoin College

W4231: Analysis of Algorithms

Tutorial. Question There are no forward edges. 4. For each back edge u, v, we have 0 d[v] d[u].

Undirected Graphs. V = { 1, 2, 3, 4, 5, 6, 7, 8 } E = { 1-2, 1-3, 2-3, 2-4, 2-5, 3-5, 3-7, 3-8, 4-5, 5-6 } n = 8 m = 11

Taking Stock. Last Time Flows. This Time Review! 1 Characterize the structure of an optimal solution

Lecture 3: Graphs and flows

Lecture 11: Analysis of Algorithms (CS ) 1

Graph Algorithms. Andreas Klappenecker. [based on slides by Prof. Welch]

CS200: Graphs. Rosen Ch , 9.6, Walls and Mirrors Ch. 14

Lecture 10: Strongly Connected Components, Biconnected Graphs

Shortest Path Problem

CS 125 Section #6 Graph Traversal and Linear Programs 10/13/14

Trees and Graphs Shabsi Walfish NYU - Fundamental Algorithms Summer 2006

Elementary Graph Algorithms

Algorithm Design and Analysis

Minimum Spanning Trees

Computer Science & Engineering 423/823 Design and Analysis of Algorithms

Review: Graph Theory and Representation

1 Start with each vertex being its own component. 2 Merge two components into one by choosing the light edge

CSC263 Week 8. Larry Zhang.

Chapter 3. Graphs CLRS Slides by Kevin Wayne. Copyright 2005 Pearson-Addison Wesley. All rights reserved.

Algorithms and Data Structures 2014 Exercises and Solutions Week 9

Sample Solutions to Homework #4

CSC 172 Data Structures and Algorithms. Lecture 24 Fall 2017

Copyright 2000, Kevin Wayne 1

Final Exam. EECS 2011 Prof. J. Elder - 1 -

Graph Representations and Traversal

Exam 3 Practice Problems

Graph Traversal CSCI Algorithms I. Andrew Rosenberg

Graphs. A graph is a data structure consisting of nodes (or vertices) and edges. An edge is a connection between two nodes

Graph Algorithms Using Depth First Search

4/8/11. Single-Source Shortest Path. Shortest Paths. Shortest Paths. Chapter 24

CSE 431/531: Analysis of Algorithms. Greedy Algorithms. Lecturer: Shi Li. Department of Computer Science and Engineering University at Buffalo

Module 5 Graph Algorithms

Chapter 9 Graph Algorithms

CSE 431/531: Algorithm Analysis and Design (Spring 2018) Greedy Algorithms. Lecturer: Shi Li

Transcription:

General remarks Week 5 2 We finish, by analysing it. Then we consider the second main graph- algorithm, depth-first (). And we consider one application of, of graphs. Reading from CLRS for week 5 Chapter 22, Sections 22.2, 22., 22.. 5 Recall: Analysis of (G, s) for each u V (G) 2 d[u] = π[s] = nil d[s] = 0 5 Q = (s) 6 while Q () 7 u = Dequeue[Q] 8 for each v Adj[u] 9 if d[v] = 0 d[v] = d[u] + π[v] = u 2 Enqueue(Q, v) Correctness Analysis: At termination of (G, s), for every vertex v reachable from s: Time Analysis: v has been encountered; d[v] holds the length of the shortest path from s to v; π[v] represents an edge on a shortest path from v to s. The initialisation takes time Θ(V ). Each vertex is Enqueued once and Dequeued once; these queueing operations each take constant time, so the queue manipulation takes time Θ(V ) (altogether). The Adjacency list of each vertex is scanned only when the vertex is Dequeued, so scanning adjacency lists takes time Θ(E) (altogether). The overall time of is thus Θ(V + E).

Why do we get shortest paths? The role of the queue Is it really true that we get always shortest paths (that is, using the minimum number of edges)? Let s assume that to some vertex v there exists a shorter path P in G from s to v than found by. Let this length be d < d[v]. v s, since the distance from s to s is zero (using the path without an edge), and this is correctly computed by. 2 Consider the predecessor u on that shorter path P. If also d[u] would be wrong (that is, too big), than we could use u instead of v. Thus w.l.o.g. d[u] is correct. A crucial property of is that the distances set in step 0 of the algorithm are non-decreasing, that is, if we imagine a watch-point set at step 0, and monitor the stream of values d[v], then we will see 0,,...,, 2,..., 2,,...,,.... Why is this the case? This must be due to the queue used whose special properties we haven t yet exploited! Now when exploring the neighbours of u, in case v is still unexplored, it would get the correct distance d = d[u] +. 5 So v must have been explored already earlier (than u). 6 So at the time of determining the distance d[u] d[v] 2, the distance d[v] must have been already set. 7 However the distances d[w] set by are non-decreasing! Since in step 2, directly after setting the distance, we put the vertex into the queue, the above assertion is equivalent to the statement, that the distances of the vertices put into the queue are non-decreasing (in the order they enter the queue). Now why is this the case? The role of the queue (cont.) First we need to notice that once a distance d[v] is set, it is never changed again. The vertices are taken off the queue ( dequeued ) from the left, and are added ( enqueued ) to the right ( first in, first out ). What is added has a distance one more than what was taken away. We start with 0, and add some s. Then we take those s, and add 2 s. Once the s are finished, we take the 2 s, and add s. Running on directed graphs We can run also on a digraph G, with start-vertex s: For digraphs we have directed spanning trees/forests. 2 Only the vertices reachable from s following the directions of the edges are in that directed tree (directed from s towards the leaves). Still the paths in the directed tree, from s to any other vertex, are shortest possible (given that we obey the given directions of the edges). For a graph (i.e., undirected graph), we need to restart (to obtain a spanning forest, not just a spanning tree) only if the graph is disconnected. And so on that why the sequence of distances is non-decreasing. However for a digraph there is a much higher chance, that we need to restart in order to cover all vertices. We thus more often need a directed spanning forest, not just a directed spanning tree.

Restart typically needed for digraphs to cover all vertices Arcs of different lengths Consider the simple digraph 2 In order to cover all vertices, one needs to run at least two times (and if the first time you start it with s =, then you need to run it three times). So even for the above apparently very simple graph we need a directed spanning forest. Note that the obtained directed trees (given by π) overlap. So in such a directed forest there is typically a certain overlap between the directed trees in it. The edges of (di-)graphs have (implicitly) a length of one unit. If arbitrary non-negative lengths are allowed, then we have to generalise to Dijkstra s algorithm. This generalisation must keep the essential properties, that the distances encountered are the final ones and are non-decreasing. But now not all edges have unit-length, and thus instead of a simple queue we need to employ a priority queue. A priority queue returns the vertex in it with the smallest value (distance). The algorithm () is another simple but very important technique for ing a graph. Such a constructs a spanning forest for the graph, called the depth-first forest, composed of several depth-first trees, which are rooted spanning trees of the connected components. recursively visits the next unvisited vertex, thus extending the current path as far as possible; when the gets stuck in a corner it backtracks up along the path until a new avenue presents itself. computes the parent π[u] of each vertex u in the depth-first tree (with the parent of initial vertices being nil), as well as its discovery time d[u] (when the vertex is first encountered, initialised to ) and its finishing time f [u] (when the has finished visiting its adjacent vertices). (G) for each u V (G) 2 d[u] = time = 0 for each u V (G) 5 if d[u] = 6 π[u] = nil 7 -Visit(u) -Visit(u) time = time + 2 d[u] = time for each v Adj[u] if d[v] = 5 π[v] = u 6 -Visit(v) 7 time = time + 8 f [u] = time Analysis: -Visit(u) is invoked exactly once for each vertex, during which we scan its adjacency list once. Hence, like, runs in time Θ(V + E).

illustrated Running on directed graphs /nil / / / / / / 2 6 d[u] 2 6 f [u] /π[u] u / / / (labelling) / / / Stack = () u = Stack = () u = 2 / / / / / / 2 6 2 6 /2 / / /2 / / Stack = (2, ) u = Stack = (, 2, ) u = / / / / / / 2 6 2 6 Again (as with ), we can run on a digraph G. Again, no longer do we obtain spanning trees of the connected components of the start vertex. But we obtain a directed spanning tree with exactly all vertices reachable from the root (when following the directions of the edges). Different from, the root (or start vertex) normally does not play a prominent role for : /2 5 / / /2 5 / 6 /5 Stack = (,, 2, ) u = 5 Stack = (5,,, 2, ) u = 7 / / 9 / /nil 2 / / 9 / 0 2 6 2 6 /2 5 / 6 /5 8 7 2 /2 5 / 6 /5 8 7 Stack = (,, 2, ) u = 6 Stack = () Thus we did not provide the form of with a start vertex as input. 2 But we provided the forest-version, which tries all vertices (in the given order) as start vertices. This will always cover all vertices, via a directed spanning forest. What are the times good for? (Directed) trees do not contain shortest paths to that end their way of exploring a graph is too adventuresomely (while is very cautious ). Nevertheless, the information gained through the computation of discovery and finish times is very valuable for many tasks. We consider the example of scheduling later. In order to understand this example, we have first to gain a better understanding of the meaning of discovery and finishing time. Existence of a path versus finishing times Lemma Consider a digraph G and nodes u, v V (G) with d[u] < d[v]. Then there is a path from u to v in G if and only if f [u] > f [v] holds. In words: If node u is discovered earlier than node v, then we can reach v from u iff v finishes earlier than u. Proof. If there is a path from u to v, then, since v was discovered later than u, the recursive call of -Visit(v) must happen inside the recursive call of -Visit(u) (by construction of the discovery-loop), and thus v must finish before u finishes. In the other direction, if v finishes earlier than u, then the recursive call of -Visit(v) must happen inside the recursive call of -Visit(u) (since the recursion for u must still be running). Thus, when discovering v, we must be on a path from u to v.

Directed acyclic graphs An important applications of digraphs G is with scheduling: The vertices are the jobs (actions) to be scheduled. A directed edge from vertex u to vertex v means a dependency, that is, action u must be performed before action v. Now consider the situation where we have three jobs a, b, c and the following dependency digraph: G = a b Topological Given a dag G modelling a scheduling task, a basic task is to find a linear ordering of the vertices ( actions ) such that all dependencies are respected. This is modelled by the notion of. A sort of a dag is an ordering of its vertices such that for every edge (u, v), u appears before v in the ordering. For example consider Clearly this can not be scheduled! 8 8888888 c In general we require G to by acyclic, that is, G must not contain a directed cycle. G = a 8888888 b 8 c The two possible s of G are a, c, b and c, a, b. A directed acyclic graph is also called a dag. Finishing times in DAGs Topological via Lemma 2 After calling on a dag, for every edge (u, v) we have f [u] > f [v]. Proof. There are two cases regarding the discovery times of u and v: If d[u] < d[v], then by Lemma we have f [u] > f [v] (since there is a path from u to v (of length )). Corollary To ly sort a dag G, we run on G and print the vertices in reverse order of finishing times. (We can put each vertex on the front of a list as they are finished.) 2 Now assume d[u] > d[v]. If f [u] > f [v], then we are done, and so assume that f [u] < f [v]. But then by Lemma there is a path from v to u in G, and this together with the edge from u to v establishes a cycle in G, contradicting that G is a dag.

Topological illustrated Consider the result of running the algorithm on the following dag. /20 2/26 22/25 27/28 / m n o p q r s t u v w (labelling: 2/5 6/9 7/8 0/7 x y z /2 9/8 /5 d[u]/f [u]) 2/2/6 Deciding acyclicity For a graph G (i.e., undirected) detecting the existence of a cycle is simple, via or : G has a cycle (i.e., is not a forest) if and only if resp. will discover a vertex twice. One has to be a bit more careful here, since the parent vertex will always be discovered twice, and thus has to be excluded, but that s it. Listing the vertices in reverse order of finishing time gives the following of the vertices: However for digraphs it is not that simple do you see why? u : p n o s m r y v w z x u q t f [u]: 28 26 25 2 20 9 8 7 6 5 2 8 5 Revisiting the lemma for In Lemma 2 we said: If G has no, then along every edge the finishing times strictly decrease. So if we discover in digraph G an edge (u, v) with f [u] < f [v], then we know there is a cycle in G. Is this criterion also sufficient? YES: just consider a cycle, and what must happen with the finishing times on it. Lemma Consider a digraph G. Then G has a cycle if and only if every run of on G yields some edge (u, v) E(G) with f [u] < f [v].