1 / 13

Estimating Growth Rate of Running Time in Algorithm: arrayMax

This article explains how to estimate the growth rate of the running time of the algorithm arrayMax. It covers topics such as big-Oh notation, growth rates of functions, and asymptotic algorithm analysis. It also provides rules of thumb for using big-Oh notation effectively.

dtrent
Télécharger la présentation

Estimating Growth Rate of Running Time in Algorithm: arrayMax

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Algorithm arrayMax executes 3n 1 primitive operations in the worst case Define a Time taken by the fastest primitive operation b Time taken by the slowest primitive operation Let T(n) be the actual worst-case running time of arrayMax. We havea (3n 1) T(n)b(3n 1) Hence, the running time T(n) is bounded by two linear functions Estimating Running Time

  2. Growth Rate of Running Time • Changing the hardware/ software environment • Affects T(n) by a constant factor, but • Does not alter the growth rate of T(n) • The linear growth rate of the running time T(n) is an intrinsic property of algorithm arrayMax

  3. Growth Rates • Growth rates of functions: • Linear  n • Quadratic  n2 • Cubic  n3 • In a log-log chart, the slope of the line corresponds to the growth rate of the function

  4. Constant Factors • The growth rate is not affected by • constant factors or • lower-order terms • Examples • 102n+105is a linear function • 105n2+ 108nis a quadratic function

  5. Big-Oh Notation • Given functions f(n) and g(n), we say that f(n) is O(g(n))if there are positive constantsc and n0 such that f(n)cg(n) for n n0 • Example: 2n+10 is O(n) • 2n+10cn • (c 2) n  10 • n  10/(c 2) • Pick c = 3 and n0 = 10 • Or Pick c=4 and n0 = 5 • Just prove the existence of c and n0

  6. Big-Oh Notation (cont.) • Example: the function n2is not O(n) • n2cn • n c • The above inequality cannot be satisfied since c must be a constant

  7. Big-Oh and Growth Rate • The big-Oh notation gives an upper bound on the growth rate of a function • The statement “f(n) is O(g(n))” means that the growth rate of f(n) is no more than the growth rate of g(n) • We can use the big-Oh notation to rank functions according to their growth rate

  8. Classes of Functions • Let {g(n)} denote the class (set) of functions that are O(g(n)) • We have{n}  {n2}  {n3}  {n4}  {n5}  …where the containment is strict {n3} {n2} {n}

  9. Big-Oh Rules (of Thumb) • If is f(n) a polynomial of degree d, then f(n) is O(nd), i.e., • Drop lower-order terms • Drop constant factors • Use the smallest possible class of functions • Say “2n is O(n)”instead of “2n is O(n2)” • But it is true that 2n is O(n2) • Use the simplest expression of the class • Say “3n+5 is O(n)”instead of “3n+5 is O(3n)” • But it is true that 3n+5 is O(3n)

  10. Asymptotic Algorithm Analysis • The asymptotic analysis of an algorithm determines the running time in big-Oh notation • To perform the asymptotic analysis • We find the worst-case number of primitive operations executed as a function of the input size • We express this function with big-Oh notation • Example: • We determine that algorithm arrayMax executes at most 3n 1 primitive operations • We say that algorithm arrayMax“runs in O(n) time” • Since constant factors and lower-order terms are eventually dropped anyhow in big-Oh, we can disregard them when counting primitive operations

  11. Caution! • Beware of very large constant factors. An algorithm running in time 1,000,000 n is still O(n), but might be less efficient on your “everyday” data set than one running in time 2n2, which is O(n2).

  12. Relatives to Big-Oh’s Big –OmegaW (f(n)) Big-Theta Q (f(n))

More Related