1 / 25

CSE 421 Algorithms

CSE 421 Algorithms. Richard Anderson Lecture 20 LCS / Shortest Paths. Longest Common Subsequence. C=c 1 …c g is a subsequence of A=a 1 …a m if C can be obtained by removing elements from A (but retaining order) LCS(A, B): A maximum length sequence that is a subsequence of both A and B.

sheila-odom
Download Presentation

CSE 421 Algorithms

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CSE 421Algorithms Richard Anderson Lecture 20 LCS / Shortest Paths

  2. Longest Common Subsequence • C=c1…cg is a subsequence of A=a1…am if C can be obtained by removing elements from A (but retaining order) • LCS(A, B): A maximum length sequence that is a subsequence of both A and B ocurranec occurrence attacggct tacgacca

  3. Optimization recurrence If aj = bk, Opt[ j,k ] = 1 + Opt[ j-1, k-1 ] If aj != bk, Opt[ j,k] = max(Opt[ j-1,k], Opt[ j,k-1])

  4. Dynamic Programming Computation

  5. Storing the path information A[1..m], B[1..n] for i := 1 to m Opt[i, 0] := 0; for j := 1 to n Opt[0,j] := 0; Opt[0,0] := 0; for i := 1 to m for j := 1 to n if A[i] = B[j] { Opt[i,j] := 1 + Opt[i-1,j-1]; Best[i,j] := Diag; } else if Opt[i-1, j] >= Opt[i, j-1] { Opt[i, j] := Opt[i-1, j], Best[i,j] := Left; } else { Opt[i, j] := Opt[i, j-1], Best[i,j] := Down; } b1…bn a1…am

  6. How good is this algorithm? • Is it feasible to compute the LCS of two strings of length 100,000 on a standard desktop PC? Why or why not.

  7. Observations about the Algorithm • The computation can be done in O(m+n) space if we only need one column of the Opt values or Best Values • The algorithm can be run from either end of the strings

  8. Computing LCS in O(nm) time and O(n+m) space • Divide and conquer algorithm • Recomputing values used to save space

  9. Divide and Conquer Algorithm • Where does the best path cross the middle column? • For a fixed i, and for each j, compute the LCS that has ai matched with bj

  10. Divide and Conquer • A = a1,…,am B = b1,…,bn • Find j such that • LCS(a1…am/2, b1…bj) and • LCS(am/2+1…am,bj+1…bn) yield optimal solution • Recurse

  11. Algorithm Analysis • T(m,n) = T(m/2, j) + T(m/2, n-j) + cnm

  12. Memory Efficient LCS Summary • We can afford O(nm) time, but we can’t afford O(nm) space • If we only want to compute the length of the LCS, we can easily reduce space to O(n+m) • Avoid storing the value by recomputing values • Divide and conquer used to reduce problem sizes

  13. Shortest Paths with Dynamic Programming

  14. Shortest Path Problem • Dijkstra’s Single Source Shortest Paths Algorithm • O(mlog n) time, positive cost edges • General case – handling negative edges • If there exists a negative cost cycle, the shortest path is not defined • Bellman-Ford Algorithm • O(mn) time for graphs with negative cost edges

  15. Lemma • If a graph has no negative cost cycles, then the shortest paths are simple paths • Shortest paths have at most n-1 edges

  16. Shortest paths with a fixed number of edges • Find the shortest path from v to w with exactly k edges

  17. Express as a recurrence • Optk(w) = minx [Optk-1(x) + cxw] • Opt0(w) = 0 if v=w and infinity otherwise

  18. Algorithm, Version 1 foreach w M[0, w] = infinity; M[0, v] = 0; for i = 1 to n-1 foreach w M[i, w] = minx(M[i-1,x] + cost[x,w]);

  19. Algorithm, Version 2 foreach w M[0, w] = infinity; M[0, v] = 0; for i = 1 to n-1 foreach w M[i, w] = min(M[i-1, w], minx(M[i-1,x] + cost[x,w]))

  20. Algorithm, Version 3 foreach w M[w] = infinity; M[v] = 0; for i = 1 to n-1 foreach w M[w] = min(M[w], minx(M[x] + cost[x,w]))

  21. Correctness Proof for Algorithm 3 • Key lemma – at the end of iteration i, for all w, M[w] <= M[i, w]; • Reconstructing the path: • Set P[w] = x, whenever M[w] is updated from vertex x

  22. If the pointer graph has a cycle, then the graph has a negative cost cycle • If P[w] = x then M[w] >= M[x] + cost(x,w) • Equal when w is updated • M[x] could be reduced after update • Let v1, v2,…vk be a cycle in the pointer graph with (vk,v1) the last edge added • Just before the update • M[vj] >= M[vj+1] + cost(vj+1, vj) for j < k • M[vk] > M[v1] + cost(v1, vk) • Adding everything up • 0 > cost(v1,v2) + cost(v2,v3) + … + cost(vk, v1) v1 v4 v2 v3

  23. Negative Cycles • If the pointer graph has a cycle, then the graph has a negative cycle • Therefore: if the graph has no negative cycles, then the pointer graph has no negative cycles

  24. Finding negative cost cycles • What if you want to find negative cost cycles?

  25. Foreign Exchange Arbitrage USD 1.2 1.2 EUR CAD 0.6 USD 0.8 0.8 EUR CAD 1.6

More Related