1 / 18

Multi-Granularity Classi fi cation Rule Discovery Using ERID

Multi-Granularity Classi fi cation Rule Discovery Using ERID. by Seunghyun Im, Zbigniew Ras , Li-Shiang Tsay University of Pittsburgh at Johnstown University of North Carolina North Carolina A&T University. Outline. 1. Introduction 2. Backgrounds 3. Algorithm 4. Experiment

terri
Télécharger la présentation

Multi-Granularity Classi fi cation Rule Discovery Using ERID

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Multi-Granularity Classification Rule DiscoveryUsing ERID by Seunghyun Im, Zbigniew Ras, Li-Shiang Tsay University of Pittsburgh at Johnstown University of North Carolina North Carolina A&T University

  2. Outline 1. Introduction 2. Backgrounds 3. Algorithm 4. Experiment 5. Conclusion

  3. 1. Introduction • Discoveringclassification rules at multiple (relevant) levels ofabstraction specified by the user • e.g. Find out the dependency betweenthe flight schedule and ticket sales to increase the profit. • 10:00 AM and PA  High sales (Rule from given data set) • Morning and PA  High Sales (General) • 10:00 AM and Pittsburgh  Medium sales (Specific)

  4. 2. Backgrounds and Contribution • Related Work • Generalized association rules • Multilevel Classification rules using ontology • Rule based attribute oriented induction • Multiple-level decision tree, ontology-driven decision tree learning • Feature hierarchies, Structured attributes, Variant levels of precision

  5. 2. Backgrounds and Contribution • Proposed Method • Discover more specific and generalclassification rules from an incomplete information systemwith hierarchical attributes using ERID • * An incomplete informationsystem enables to contain multiple weighted values as a single attribute value. • * ERID (Extracting Rules from Incomplete Decision Table) algorithm isdesigned to extract classification rules from (partially) incomplete informationsystems.

  6. 3. Algorithm (By example) Incomplete Information System S Information System S = (X,A,V) X is a set of objects= {x1, x2, x3, x4, x5, x6} A is a set of attributes= {a, b, c, d} Va is a set of values of attribute a, where a  A, and V = {Va : a  A},

  7. 3. Algorithm (By example) Incomplete Information System S Assumption

  8. 3. Algorithm (By example) Incomplete Information System S e.g. An attribute value for an object in an incomplete informationsystem can be {(dark black,0.5),(light black, 0.5)}, meaning that the object maybe dark black or light black with equal weight. e.g. If {(v1, 0.2), (v2, 0.4), (v3, 0.4)}is being assigned asingle attribute value in S with 1= 0.3 then, it is automatically converted to{(v2,0.5),(v3,0.5)}because the confidence of v1< 0.3, and sum ofthe weights of viis 1 by definition. Weights for v2 and v3 are calculated by theratio of its weight and the total weight after eliminating v1, that is p2 = (0.4)/(0.4+0.4), and p3 = (0.4)/(0.4+0.4)

  9. 3. Algorithm Attribute Value Hierarchy (AVH) for S An AVH t = {(a[p])}is the set ofnode values, where a is an attribute name and p is the path fromthe root to the node. (e.g. b[1,1] is the value of the left most element at depth2). a, where is the empty sequence, is equivalent to the attributename.

  10. 3. Algorithm Replacement of Attribute Value and Weight 1. The sum of the weights for a'(x) = 1 (unchanged by the definition of anincomplete information system) 2. If ai(x) = {(v,w)}is generalized to level k, a'(x) = {(v',w)}, where v'is anancestor node of v at depth k, and w is carried over. 3. If ai(x) = {(v,w)}is specified to level k, each v is replaced by one or moredescendant nodes {v'1, v'2,..., v'n}, and w'for each v'is 1 /n,where n is thenumber of sibling nodes at depth k. 4. If ti T is not balanced, and the granularity of ai(x) is given to a levelwhere no node exists, choose the closest parent node.

  11. 3. Algorithm Creating a new information system S' with specified granularity e.g. If an attribute value {(a[2,1], 0.5), (a[2,2], 0.5)} is generalized to one level higher  (a[2], 1). e.g. If {(a[2],1)}is specified to the values in depth 3, {(a[2,1,1],1/6),(a[2,1,2],1/6 ),(a[2,3],1/3 ),(a[2,3],1/3 )}. S' with granularity at a = 1b = 2c = 2d = 1

  12. 3. Algorithm Extract Classification Rules from S' using ERID Support = 1 confidence = 0.75 Granules associated with a,b,c a[1]= {(x1,1/2), (x3,1), (x4,1), (x5,1)} a[2]= { (x1,1/2), (x6,1)} b[1,1]= {(x2,1/2), (x4,1), (x5,1)} b[1,2]= {(x2,1/2) } b[2,1]= {(x1,1/2), (x3,1/2), (x6,1/2) } b[2,1]= {(x1,1/2), (x3,1/2), (x6,1/2) } c[1,1]= {(x1,1), (x2,1), (x3,1/2), (x6,1)} c[1,2]= {(x3,1/2), (x4,1)} Granules associated with d (decision) d1= {(x1,1), (x4,1), (x5,1) } d2= {(x2,1), (x3,1), (x6,1) }

  13. 3. Algorithm Single Element Term Rules a[1]* d[1]* sup = 5/2 , conf = 0.72 + a[1]*  d[2]* sup = 1, conf = 0.29 a[2]*  d[1]* sup = 1/2 - a[2]*  d[2]* sup = 1, conf = 0.67 b[1,1]*  d[1]* sup = 2, conf = 0.8 + b[1,1]*  d[2]* sup = 1/2 - b[2,1]*  d[1]* sup = 1/2 - b[2,1]*  d[2]* sup = 1, conf = 0.67 b[2,2]*  d[1]* sup = 1/2 - b[2,2]*  d[2]* sup = 1, conf = 0.67 c[1,1]*  d[1]* sup = 3, conf = 0.67 c[1,1]*  d[2]* sup = 1/2 - c[1,2]*  d[1]* sup = 2, conf = 0.8 + c[1,2]*  d[2]* sup = 1/2 - rule sup> conf> no further consideration because sup < will be used in 2element terms

  14. 3. Algorithm Two Element Term Rules (a[1], b[2,1])* d[2]* sup = 1 , conf = 0.67 (a[1], b[2,1])* d[2]* sup = 1 , conf = 0.67 + (a[1], b[2,2])* d[1]* sup = 0 - (a[1], c[1,1])* d[2]* sup = 1 , conf = 2 + (b[1,1], c[1,1])* d[1]* sup = 2 , conf = 0.8 - (b[2,1], c[1,1])* d[2]* sup = 1/2

  15. 4. Implementation and Experiment • Data Set • Census bureau database (obtained from UCI ML repository) • 4884 objects and 12 Attributes • AVH(one to three levels of depth) was built based on thedescription given by the data provider Levels of attribute granularity for the test data set Rules at different granularity

  16. 4. Implementation and Experiment • Implementation • Platform : Pentium M 1.6, Windows XP • Programming Language : Matlab 7.1, Visual Basic .Net

  17. 5. Conclusion • Described the use of ERID for discovering multi-granularity classification rules from an incomplete information system having attribute value hierarchies. • ERID is especially suitable to extract rules from information systems containing multiple weighted values, and this feature enables to handle partial incompleteness of attribute values when the levels of abstraction are changed.

  18. Questions? Thank You

More Related