1 / 75

Figure 1.1 Rules for the contact lens data.

Figure 1.1 Rules for the contact lens data. Figure 1.2 Decision tree for the contact lens data. (a). (b). Figure 1.3 Decision trees for the labor negotiations data. Figure 2.1 A family tree and two ways of expressing the sister-of relation. Figure 2.2 ARFF file for the weather data.

whitby
Télécharger la présentation

Figure 1.1 Rules for the contact lens data.

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Figure 1.1 Rules for the contact lens data.

  2. Figure 1.2 Decision tree for the contact lens data.

  3. (a) (b) Figure 1.3 Decision trees for the labor negotiations data.

  4. Figure 2.1 A family tree and two ways of expressing the sister-of relation.

  5. Figure 2.2 ARFF file for the weather data.

  6. Figure 3.1 Decision tree for a simple disjunction.

  7. If x=1 and y=0 then class = a If x=0 and y=1 then class = a If x=0 and y=0 then class = b If x=1 and y=1 then class = b Figure 3.2 The exclusive-or problem.

  8. If x=1 and y=1 then class = a If z=1 and w=1 then class = a Otherwise class = b Figure 3.3 Decision tree with a replicated subtree.

  9. Default: Iris-setosa 1 • except if petal-length  2.45 and petal-length < 5.355 2 • and petal-width < 1.75 3 • then Iris-versicolor 4 • except if petal-length  4.95 and petal-width < 1.55 5 • then Iris-virginica 6 • else if sepal-length < 4.95 and sepal-width  2.45 7 • then Iris-virginica 8 • else if petal-length  3.35 9 • then Iris-virginica 10 • except if petal-length < 4.85 and sepal-length < 5.95 11 • then Iris-versicolor 12 Figure 3.4 Rules for the Iris data.

  10. Shaded: standing Unshaded: lying Figure 3.5 The shapes problem.

  11. PRP = - 56.1 + 0.049 MYCT + 0.015 MMIN + 0.006 MMAX + 0.630 CACH - 0.270 CHMIN + 1.46 CHMAX Figure 3.6(a) Models for the CPU performance data: linear regression.

  12. Figure 3.6(b) Models for the CPU performance data: regression tree.

  13. Figure 3.6(c) Models for the CPU performance data: model tree.

  14. (a) (b) (c) (d) Figure 3.7 Different ways of partitioning the instance space.

  15. (a) (b) (c) 1 2 3 a 0.4 0.1 0.5 b 0.1 0.8 0.1 c 0.3 0.3 0.4 d 0.1 0.1 0.8 e 0.4 0.2 0.4 f 0.1 0.4 0.5 g 0.7 0.2 0.1 h 0.5 0.4 0.1 … (d) Figure 3.8 Different ways of representing clusters.

  16. Figure 4.1 Pseudo-code for 1R.

  17. (a) (b) Figure 4.2 Tree stumps for the weather data. (c) (d)

  18. (a) (b) (c) Figure 4.3 Expanded tree stumps for the weather data.

  19. Figure 4.4 Decision tree for the weather data.

  20. Figure 4.5 Tree stump for the ID code attribute.

  21. (a) (b) Figure 4.6 (a) Operation of a covering algorithm; (b) decision tree for the same problem.

  22. Figure 4.7 The instance space during operation of a covering algorithm.

  23. Figure 4.8 Pseudo-code for a basic rule learner.

  24. Figure 5.1 A hypothetical lift chart.

  25. Figure 5.2 A sample ROC curve.

  26. Figure 5.3 ROC curves for two learning schemes.

  27. (a) (b) Figure 6.1 Example of subtree raising, where node C is “raised” to subsume node B.

  28. Figure 6.2 Pruning the labor negotiations decision tree.

  29. Figure 6.3 Generating rules using a probability measure.

  30. p number of instances of that class that the rule selects; • t total number of instances that the rule selects; • p total number of instances of that class in the dataset; • t total number of instances in the dataset. Figure 6.4 Definitions for deriving the probability measure.

  31. Figure 6.5 Algorithm for forming rules by incremental reduced error pruning.

  32. Figure 6.6 Algorithm for expanding examples into a partial tree.

  33. (a) (c) (b) Figure 6.7 Example of building a partial tree.

  34. (d) (e) Figure 6.7 (continued) Example of building a partial tree.

  35. Exceptions are represented as Dotted paths, alternatives as solid ones. Figure 6.8 Rules with exceptions, for the Iris data.

  36. Figure 6.9 A maximum margin hyperplane.

  37. Figure 6.10 A boundary between two rectangular classes.

  38. Figure 6.11 Pseudo-code for model tree induction.

  39. Figure 6.12 Model tree for a dataset with nominal attributes.

  40. (a) (b) (c) Figure 6.13 Clustering the weather data.

  41. (d) (e) Figure 6.13 (continued) Clustering the weather data.

  42. (f) Figure 6.13 (continued) Clustering the weather data.

  43. (a) Figure 6.14 Hierarchical clusterings of the Iris data.

  44. (b) Figure 6.14 (continued) Hierarchical clusterings of the Iris data.

  45. data A 51A 43B 62B 64A 45A 42A 46A 45A 45 B 62A 47A 52B 64A 51B 65A 48A 49 A 46 B 64A 51A 52B 62A 49A 48B 62A 43A 40 A 48B 64A 51B 63A 43B 65B 66 B 65A 46 A 39B 62B 64A 52B 63B 64A 48B 64A 48 A 51A 48B 64A 42A 48A 41 model A=50, A =5, pA=0.6 B=65, B =2, pB=0.4 Figure 6.15 A two-calss mixture model.

  46. Figure 7.1 Attribute space for the weather dataset.

  47. Figure 7.2 Discretizing temperature using the entropy method.

  48. 64 65 68 69 70 71 72 75 80 81 83 85 no yes yes no yes yes yes no no yes yes no yes yes F E D C B A 66.5 70.5 73.5 77.5 80.5 84 Figure 7.3 The result of discretizing temperature.

  49. Figure 7.4 Class distribution for a two-class, two-attribute problem.

  50. Figure 7.5 Number of international phone calls from Belgium, 1950–1973.

More Related