Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
courses:cs211:winter2012:journals:virginia:chapter2 [2012/01/18 01:01] โ [Section 2: Asymptotic Order of Growth] lovellv | courses:cs211:winter2012:journals:virginia:chapter2 [2012/01/25 01:53] (current) โ [Section 5: Priority Queues] lovellv | ||
---|---|---|---|
Line 31: | Line 31: | ||
Polynomial, logarithmic and exponential functions are commonly seen. A polynomial' | Polynomial, logarithmic and exponential functions are commonly seen. A polynomial' | ||
+ | |||
+ | Readability: | ||
+ | |||
+ | ===== Section 3: Implementing the Stable Matching Algorithm ===== | ||
+ | |||
+ | To analyze the running time of an algorithm, we need to think about how we would implement it. Two important data structures we might use are arrays and (linked) lists. | ||
+ | |||
+ | We know the G-S algorithm terminates in at most n^2 iterations so if we want O(n^2) running time, each iteration must run in constant time. In each iteration we: | ||
+ | |||
+ | - Identify a free man, m | ||
+ | - Identify m's highest ranked woman w who he has not yet proposed to | ||
+ | - Determine if w is engaged and to whom | ||
+ | - Determine if w prefers her current partner to m | ||
+ | |||
+ | We can implement each of these steps in constant time by: | ||
+ | |||
+ | - Maintaining a list of free men | ||
+ | - Maintaining an array " | ||
+ | - Maintaining an array " | ||
+ | - Creating an array " | ||
+ | |||
+ | This allows us to implement the G-S algorithm in O(n^2) time. | ||
+ | | ||
+ | Readability: | ||
+ | ===== Section 4: A Survey of Common Running Times ===== | ||
+ | |||
+ | Here are some common running times and examples of algorithms with these running times. | ||
+ | |||
+ | * **Linear O(n) time**: Running time is a constant factor times the size of the input. Computing the maximum and merging two sorted lists are O(n) algorithms (p 48). | ||
+ | * **O(nlogn) time**: This is the running time of an algorithm that splits its input in half, recurses over them and combines the solutions in linear time. Mergesort is an O(nlogn) algorithm (p 50). | ||
+ | * **Quadratic O(n^2) time**: This is the running time of an algorithm with two nested loops, such as finding the pair of points that are closest together in the plane (p 51). | ||
+ | * **Cubic O(n^3) time**: This is the running time of an algorithm with three nested loops, such as finding the pairs of disjoint sets when given n sets (p 52). | ||
+ | * **O(n^k) time**: This is the running time of searching over all subsets of size k of a set of size n. The Independent Set problem has O(n^k) running time (p 53). | ||
+ | * **O(2^n) time**: 2^n is the total number of subsets of an n-element set. Finding a independent set of maximum size is an O(2^n) algorithm (p 54). | ||
+ | * **O(n!) time**: O(n!)> | ||
+ | * **Sublinear time**: Binary search has O(logn) running time since it discards a fraction of the input on each iteration. | ||
+ | |||
+ | Readability: | ||
+ | ===== Section 5: Priority Queues ===== | ||
+ | |||
+ | This section explains how to implement priority queue, a data structure that supports addition, deletion, and selection of the highest priority element. | ||
+ | |||
+ | We can implement each queue operation in O(logn) time using a heap. A heap combines the benefits of a sorted array and a list. A heap can be thought of as a balanced binary tree where the key of a parent is always <= the keys of its children. | ||
+ | |||
+ | Thus, if we implement a priority queue with a heap, we get O(n) time to start the queue, O(logn) time to insert, O(1) time to find the minimum key, O(logn) time to delete and O(logn) to extract the value with minimum key. | ||
Readability: | Readability: |