1 / 36

CS4413 Divide-and-Conquer (These materials are used in the classroom only)

CS4413 Divide-and-Conquer (These materials are used in the classroom only). Divide and Conquer. Georgia. Divide and Conquer. The most well known algorithm design strategy: Divide instance of problem into two or more smaller instances Solve smaller instances recursively

micheal
Download Presentation

CS4413 Divide-and-Conquer (These materials are used in the classroom only)

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. CS4413 Divide-and-Conquer (These materials are used in the classroom only)

  2. Divide and Conquer Georgia

  3. Divide and Conquer The most well known algorithm design strategy: • Divide instance of problem into two or more smaller instances • Solve smaller instances recursively • Obtain solution to original (larger) instance by combining these solutions

  4. Divide-and-conquer technique a problem of size n subproblem 1 of size n/2 subproblem 2 of size n/2 a solution to subproblem 1 a solution to subproblem 2 a solution to the original problem

  5. Divide and Conquer Examples • Sorting: mergesort and quicksort • Tree traversals • Binary search • Matrix multiplication-Strassen’s algorithm • Convex hull-QuickHull algorithm

  6. General Divide and Conquer recurrence: Master Theorem T(n) = aT(n/b) + f (n)where f (n)∈Θ(nk) • a < bk T(n) ∈Θ(nk) • a = bk T(n) ∈Θ(nk lg n ) • a > bk T(n) ∈Θ(nlog b a) Note: the same results hold with O and Ω.

  7. Mergesort • Mergesort is a recursive sort algorithm. Algorithm: • Split array A[1...n] in two and make copies of each half in arrays B[1... n/2 ] and C[1... n/2 ] • Sort arrays B and C

  8. Mergesort • Merge sorted arrays B and C into array A as follows: • Repeat the following until no elements remain in one of the arrays: • compare the first elements in the remaining unprocessed portions of the arrays • copy the smaller of the two into A, while incrementing the index indicating the unprocessed portion of that array • Once all elements in one of the arrays are processed, copy the remaining unprocessed elements from the other array into A.

  9. Mergesort Example 7 2 1 6 4 5 3 9

  10. Mergesort … • Algorithm Mergesort (A[0 … n – 1]) • //sorts array A[0 … n – 1] by recursive mergesort • //Input: an array A[0 … n – 1] of orderable elements • //Output: array A[0 … n – 1] sorted in nondecreasing order • If n > 1 • copy A[0 …n/2] to B[0 …n/2 – 1] • copy A[n/2 … n - 1] to C[0 …n/2 – 1] • Mergesort (B[0 …n/2 – 1]) • Mergesort (C[0 …n/2 – 1]) • Merge(B, C, A)

  11. Efficiency of mergesort • C(n) = 2C(n/2) + Cmerge(n) for n > 1, C(1) = 0 • The worst case: • W(n) = 2W(n/2) + n – 1 for n > 1, W(1) = 0

  12. Efficiency of mergesort • All cases have same efficiency: Θ( n log n) • Number of comparisons is close to theoretical minimum for comparison-based sorting: • log n ! ≈ n lg n - 1.44 n • Space requirement: Θ( n ) • Can be implemented without recursion (bottom-up)

  13. Efficiency of Mergesort • Merge analysis (McConnell): • Case 1: where all of the elements of list A are smaller than the first element of list B. • Case 2: what if the first element of A is greater than the first element of B but all of the elements of A are smaller than the second element of B? • Case 3: consider what happens if the elements of A and B are “interleaved” based on their value. In other words, what happens if the value of A[1] is between B[1] and B[2], the value of A[2] is between B[2] and B[3], the value of A[3] is between B[3] and B[4], and so on. • Which is the best case? Worst case?

  14. Exercise • A. Write a pseudocode for a divide-and-conquer algorithm for the exponentiation problem of computing an where a > 0 and n is a positive integer. • B. Set up and solve (for n = 2k) a recurrence relation for the number of multiplications made by this algorithm. • C. How does this algorithm compare with the brute-force algorithm for this problem?

  15. p A[i]≤p A[i]>p Quicksort • Select a pivot (partitioning/pivot element) • Rearrange the list so that all the elements in the positions before the pivot are smaller than or equal to the pivot and those after the pivot are larger than the pivot (See algorithm Partition/PivotList in section 4.2) • Exchange the pivot with the last element in the first (i.e., ≤ sublist) – the pivot is now in its final position • Sort the two sublists

  16. Quicksort algorithm • Algorithm Quicksort (A[l … r]) • If l < r • s ← partition (A[l … r]) // s is a split position • Quicksort (A[l … s – 1]) • Quicksort (A[s + 1 … r])

  17. The partition algorithm

  18. Quicksort … • An efficient method based on two scans of the subarray used in the algorithm. • Three situations may arise: • 1) if scanning indices i and j have not crossed, i.e., i < j, we simply exchange A[i] and A[j] and resume the scans. • 2) if the scanning indices have crossed over, i.e., i > j, we have partitioned the array after exchanging the pivot with A[j]. (n + 1 comparisons) • 3) if the scanning indices stop while pointing to the same element, i.e., i = j, the value they are pointing to must be equal to p. Thus, the array is partitioned. (n comparisons)

  19. Quicksort Example 15 22 13 27 12 10 20 25

  20. Efficiency of Quicksort • PivotList analysis: • Case 1: when PivotList creates two parts that are the same size. • Case 2: when the lists are of drastically different sizes. For example, the largest difference in the size of these two lists occurs if the PivotValue is smaller (or larger) than all of the other values in the list. In that case, we wind up with one part that has no elements and the other that has N – 1 elements. What is the number of comparisons? • W(N) = ? • Which is the best case? Worst case?

  21. Efficiency of quicksort • Best case: split in the middle — Θ( n log n) (why?) • Worst case: sorted array! — Θ( n2) (why?) • Average case: random arrays —Θ( n log n) (why?) • Improvements: • better pivot selection: median of three partitioning avoids worst case in sorted files • switch to insertion sort on small subfiles • elimination of recursion these combine to 20-25% improvement • Considered the method of choice for internal sorting for large files (n ≥ 10000)

  22. Binary Search • Binary search is an efficient algorithm for searching in a sorted array. • Search requires the following steps: 1. Inspect the middle item of an array of size N. 2. Inspect the middle of an array of size N/2. 3. Inspect the middle item of an array of size N/power(2,2) and so on until N/power(2,k) = 1. • This implies k = log2N • k is the number of partitions.

  23. Binary Search • Requires that the array be sorted. • Rather than start at either end, binary searches split the array in half and works only with the half that may contain the value. • This action of dividing continues until the desired value is found or the remaining values are either smaller or larger than the search value.

  24. Binary Search • BinarySearch (List, target, N) • //list: the elements to be searched • //target: the value being searched for • //N: the number of elements in the list • Start = 1 • End = N • While start <= end do • Middle = (start + end) /2 • Select (compare (list [middle], target)) from • Case -1: strat = middle + 1 • Case 0: return middle • Case 1: end = middle – 1 • End select • End while • Return 0

  25. Binary Search • Suppose that the data set is n = 2k - 1sorted items. • Each time examine middle item. If larger, look left, if smaller look right. • Second ‘chunk’ to consider is n/2 in size, third is n/4, fourth is n/8, etc. • Worst case, examine k chunks. n = 2k – 1 so, k = log2(n + 1). (Decision binary tree: one comparison on each level) • Best case, O(1). (Found at n/2). • Thus the algorithm is O(log(n)). • Extension to non-power of 2 data sets is easy.

  26. Binary Search • Best case : O(1) • Worst case : O(log2N) (why?) • Average Case : O(log2N)/2 = O(log2N) (why?)

  27. Strassen’s matrix multiplication • Strassen observed [1969] that the product of two matrices can be computed as follows: C00 C01 A00 A01 B00 B01 = * C10 C11 A10 A11 B10 B11 M1 + M4 - M5 + M7 M3 + M5 = M2 + M4 M1 + M3 - M2 + M6

  28. Submatrices: • M1 = (A00 + A11) * (B00 + B11) • M2 = (A10 + A11) * B00 • M3 = A00 * (B01 - B11) • M4 = A11 * (B10 - B00) • M5 = (A00 + A01) * B11 • M6 = (A10 - A00) * (B00 + B01) • M7 = (A01 - A11) * (B10 + B11)

  29. Efficiency of Strassen’s algorithm • If n is not a power of 2, matrices can be padded with zeros • Number of multiplications: • Number of additions: • Other algorithms have improved this result, but are even more complex

  30. Convex-Hull Problem • There are several divide-and-conquer algorithms for the convex-hull problem of finding the smallest convex polygon that contains n given points in the plane.

  31. QuickHull Algorithm Inspired by Quicksort to compute Convex Hull: • Assume points are sorted by x-coordinate values • Identify extreme points P1 and P2 (part of hull) Pmax P2 P1

  32. QuickHull Algorithm • Compute upper hull: • find point Pmax that is farthest away from line P1P2 • If there is a tie, the point that maximizes the angle P1 PmaxP2 • compute the hull of the points to the left of line P1Pmax • compute the hull of the points to the right of line PmaxP2 • Compute lower hull in a similar manner. Pmax P2 P1

  33. Geometric Operations • If p1 = (x1, y1), p2 = (x2, y2), and p3 = (x3, y3) are three arbitrary points in the plane, then the area of the triangle Δp1 p2 p3 is equal to one half of the magnitude of the determinant. • |x1 y1 1| • |x2 y2 1| = x1 y2 + x3 y1 + x2 y3 – x3 y2 – x2 y1 – x1 y3 • |x3 y3 1| • The sign of this expression is positive if and only if the point p3 = (x3, y3) is to the left of the line p1 p2 . • The formula can be used to check whether a point lies to the left of a line determined by two other points as well as find the distance from the point to the line.

  34. Efficiency of QuickHull algorithm • Finding point farthest away from line P1P2 can be done in linear time. • This gives same efficiency as quicksort: • Worst case: Θ( n2) • Average case: Θ( n log n) • If points are not initially sorted by x-coordinate value, this can be accomplished in Θ( n log n) — no increase in asymptotic efficiency class • Other algorithms for convex hull: • Graham’s scan (Θ( n log n) ) • DCHull (Θ( n log n) )

  35. Sequential (Linear) Searching • Sequential search – starts at the beginning and examines each element in turn. • If we know the array is sorted and we know the search value, we can start the search at the most efficient end. • If the array is sorted, we can stop the search when the condition is no longer valid (that is, the elements are either smaller or larger than the search value).

  36. Sequential Search • Sequential Search: examine each piece of data until the correct one is found. • 1. Worst case: Order O(n). • 2. Best case: O(1) • 3. Average case: O(n/2). • Thus, we say sequential search is O(n).

More Related