1 / 38

Design and Analysis Algorithm

Design and Analysis Algorithm. Drs. Achmad Ridok M.Kom Fitra A. Bachtiar, S.T., M. Eng Imam Cholissodin , S. Si ., M.Kom Aryo Pinandito , MT. Pertemuan 04. Contents. Asymptotic Notation. Asymptotic Notation.

Télécharger la présentation

Design and Analysis Algorithm

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Design and Analysis Algorithm • Drs. AchmadRidokM.Kom • Fitra A. Bachtiar, S.T., M. Eng • ImamCholissodin, S.Si., M.Kom • AryoPinandito, MT Pertemuan 04

  2. Contents Asymptotic Notation

  3. Asymptotic Notation Think of n as the number of records we wish to sort with an algorithm that takes f(n) to run. How long will it take to sort n records? What if n is big? We are interested in the range of a function as n gets large. Will f stay bounded? Will f grow linearly? Will f grow exponentially? Our goal is to find out just how fast f grows with respect to n.

  4. Asymptotic Notation • Misalkan: T(n) = 5n2 + 6n + 25 T(n) proporsionaluntukordo n2untuk data yang sangatbesar. Memperkirakan formula untuk run-time • Indikasikinerjaalgoritma • (untukjumlah data yang sangatbesar)

  5. Asymptotic Notation • IndikatorefisiensialgoritmabedasarpadaOoGpada basic operation suatualgoritma. • PenggunaannotasisebagaipembandingurutanOoG: • O (big oh) • Ω (big omega) • Ө (big theta) • t(n) : algoritma running time (diindikasikandengan basic operation count (C(n)) • g(n) : simple function to compare the count

  6. Classifying functions by theirAsymptotic Growth Rates (1/2) • asymptotic growth rate, asymptotic order, or order of functions • Comparing and classifying functions that ignores constant factors and small inputs. • O(g(n)), Big-Oh of g of n, the Asymptotic Upper Bound; • W(g(n)), Omega of g of n, the Asymptotic Lower Bound. • Q(g(n)), Theta of g of n, the Asymptotic Tight Bound; and

  7. Example Example: f(n) = n2 - 5n + 13. The constant 13 doesn't change as n grows, so it is not crucial. The low order term, -5n, doesn't have much effect on f compared to the quadratic term, n2. We will show that f(n) = Q(n2) . Q: What does it mean to say f(n) = Q(g(n)) ? A: Intuitively, it means that function f is the same order of magnitude as g.

  8. Example (cont.) Q: What does it mean to say f1(n) = Q(1)? A: f1(n) = Q(1) means after a few n, f1 is bounded above & below by a constant. Q: What does it mean to say f2(n) = Q(n log n)? A: f2(n) = Q(n log n) means that after a few n, f2 is bounded above and below by a constant times nlog n. In other words, f2 is the same order of magnitude as nlog n. More generally, f(n) = Q(g(n)) means that f(n) is a member of Q(g(n)) where Q(g(n)) is a set of functions of the same order of magnitude.

  9. Big-Oh The O symbol was introduced in 1927 to indicate relative growth of two functions based on asymptotic behavior of the functions now used to classify functions and families of functions

  10. Upper Bound Notation • We say Insertion Sort’s run time is O(n2) • Properly we should say run time is in O(n2) • Read O as “Big-O” (you’ll also hear it as “order”) • In general a function • f(n) is O(g(n)) if  positive constants c and n0such that f(n)  c  g(n)  n  n0 • e.g. if f(n)=1000n and g(n)=n2, n0> 1000 andc = 1 then f(n0) < 1.g(n0) and we say that f(n) = O(g(n))

  11. Asymptotic Upper Bound c g(n) • f(n)  c g(n) for all nn0 • g(n) is called an • asymptotic upper bound of f(n). • We write f(n)=O(g(n)) • It reads f(n) is big oh of g(n). f(n) g(n) n0

  12. Big-Oh, the Asymptotic Upper Bound This is the most popular notation for run time since we're usually looking for worst case time. If Running Time of Algorithm X is O(n2) , then for any input the running time of algorithm X is at most a quadratic function, for sufficiently large n. e.g. 2n2 = O(n3) . From the definition using c = 1 and n0 = 2. O(n2) is tighter than O(n3).

  13. Example 1 g(n) f(n) 6 for all n>6, g(n) > 1 f(n). Thus the function f is in the big-O of g. that is, f(n) in O(g(n)).

  14. Example 2 g(n) f(n) 5 There exists a n0=5 s.t. for all n>n0, f(n) < 1 g(n). Thus, f(n) is in O(g(n)).

  15. Example 3 3.5 h(n) f(n) h(n) 5 There exists a n0=5, c=3.5, s.t. for all n>n0, f(n) < c h(n). Thus, f(n) is in O(h(n)).

  16. Example of Asymptotic Upper Bound 4g(n)=4n2 • 4 g(n) = 4n2 • = 3n2 + n2 •  3n2 + 9 for all n 3 • > 3n2 + 5 • = f(n) • Thus, f(n)=O(g(n)). f(n)=3n2+5 g(n)=n2 3

  17. Exercise on O-notation • Show that 3n2+2n+5 = O(n2) 10 n2 = 3n2 + 2n2 + 5n2  3n2 + 2n + 5 for n  1 c = 10, n0 = 1

  18. Tugas 1 : O-notation • O(n2) • O(n log n) • O(n2) • O(n log n) f1(n) = 10 n + 25 n2 f2(n) = 20 n log n + 5 n f3(n) = 12 n log n + 0.05 n2 f4(n) = n1/2 + 3 n log n

  19. Classification of Function : BIG O (1/2) A function f(n) is said to be of at most logarithmic growth if f(n) = O(log n) A function f(n) is said to be of at most quadratic growth if f(n) = O(n2) A function f(n) is said to be of at most polynomial growth if f(n) = O(nk), for some natural number k > 1 A function f(n) is said to be of at most exponential growth if there is a constant c, such that f(n) = O(cn), and c > 1 A function f(n) is said to be of at most factorial growth if f(n) = O(n!).

  20. Classification of Function : BIG O (2/2) • A function f(n) is said to have constant running time if the size of the input n has no effect on the running time of the algorithm (e.g., assignment of a value to a variable). The equation for this algorithm is f(n) = c • Other logarithmic classifications: • f(n) = O(n log n) • f(n) = O(log log n)

  21. Lower Bound Notation • We say InsertionSort’s run time is (n) • In general a function • f(n) is (g(n)) if  positive constants c and n0such that 0  cg(n)  f(n)  n  n0 • Proof: • Suppose run time is an + b • Assume a and b are positive (what if b is negative?) • an  an + b

  22. Big  Asymptotic Lower Bound • f(n) c g(n) for all nn0 • g(n) is called an • asymptotic lower bound of f(n). • We write f(n)=(g(n)) • It reads f(n) is omega of g(n). f(n) c g(n) n0

  23. Example of Asymptotic Lower Bound g(n)=n2 g(n)=n2 • g(n)/4 = n2/4 • = n2/2 – n2/4 • n2/2 – 9 for all n 6 • < n2/2 – 7 • Thus, f(n)= (g(n)). f(n)=n2/2-7 c g(n)=n2/4 6

  24. Example: Big Omega Example: n 1/2 = W( log n) . Use the definition with c = 1 and n0 = 16. Checks OK. Let n ≥ 16 : n 1/2≥ (1) log n if and only if n = ( log n )2by squaring both sides. This is an example of polynomial vs. log.

  25. Big Theta Notation Definition: Two functions f and g are said to be of equal growth, f = Big Theta(g) if and only if both f=Q(g) and g = Q(f). Definition: f(n) =Q(g(n)) means positive constants c1, c2, and n0 such that c1 g(n)  f(n)  c2 g(n)  n  n0 If f(n) = O(g(n)) and f(n) = W(g(n)) then f(n) = Q(g(n)) (e.g. f(n) = n2 and g(n) = 2n2)

  26. Theta, the Asymptotic Tight Bound Theta means that f is bounded above and below by g; BigTheta implies the "best fit". f(n) does not have to be linear itself in order to be of linear growth; it just has to be between two linear functions,

  27. Asymptotically Tight Bound • f(n) = O(g(n)) and f(n) = (g(n)) • g(n) is called an • asymptotically tight bound of f(n). • We write f(n)=(g(n)) • It reads f(n) is theta of g(n). c2 g(n) f(n) c1 g(n) n0

  28. Other Asymptotic Notations • o() is like < • O() is like  • () is like > • () is like  • () is like = A function f(n) is o(g(n)) if  positive constants c and n0such that f(n) < c g(n)  n  n0 A function f(n) is (g(n)) if  positive constants c and n0such that c g(n) < f(n)  n  n0 Intuitively,

  29. Examples 1. 2n3 + 3n2 + n = 2n3 + 3n2 + O(n) = 2n3 + O( n2 + n) = 2n3 + O( n2 ) = O(n3 ) = O(n4) 2. 2n3 + 3n2 + n = 2n3 + 3n2 + O(n) = 2n3 + Q(n2 + n) = 2n3 + Q(n2) = Q(n3)

  30. Tugas 2 : Examples (cont.) 3. Suppose a program P is O(n3), and a program Q is O(3n), and that currently both can solve problems of size 50 in 1 hour. If the programs are run on another system that executes exactly 729 times as fast as the original system, what size problems will they be able to solve?

  31. Example (cont.) n3 = 503* 729 3n = 350* 729 n = n = log3 (729 * 350) n = n = log3(729) + log3 350 n = 50 * 9 n = 6 + log3 350 n = 50 * 9 = 450 n = 6 + 50 = 56 Improvement: problem size increased by 9 times for n3 algorithm but only a slight improvement in problem size (+6) for exponential algorithm.

  32. More Examples 0.5n2- 5n + 2 = W( n2). Let c = 0.25 and n0 = 25. 0.5 n2 - 5n + 2 = 0.25( n2) for all n = 25 (b) 0.5 n2 - 5n + 2 = O( n2). Let c = 0.5 and n0 = 1. 0.5( n2) = 0.5 n2 - 5n + 2 for all n = 1 (c) 0.5 n2- 5n + 2 = Q( n2) from (a) and (b) above. Use n0 = 25, c1 = 0.25, c2 = 0.5 in the definition.

  33. More Examples (d) 6 * 2n + n2 = O(2n). Let c = 7 and n0 = 4. Note that 2n = n2 for n = 4. Not a tight upper bound, but it's true. (e) 10 n2 + 2 = O(n4). There's nothing wrong with this, but usually we try to get the closest g(n). Better is to use O(n2 ).

  34. Practical Complexity t < 250

  35. Practical Complexity t < 500

  36. Practical Complexity t < 1000

  37. Practical Complexity t < 5000

  38. Thank You ! Click to edit subtitle style

More Related