html5-img
1 / 57

Rough Sets

Rough Sets. Basic Concepts of Rough Sets. Information/Decision Systems (Tables) Indiscernibility Set Approximation Reducts and Core. IS is a pair ( U, A ) U is a non-empty finite set of objects. A is a non-empty finite set of attributes such that for every

noah
Download Presentation

Rough Sets

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Rough Sets

  2. Basic Concepts of Rough Sets • Information/Decision Systems (Tables) • Indiscernibility • Set Approximation • Reducts and Core

  3. IS is a pair (U, A) U is a non-empty finite set of objects. A is a non-empty finite set of attributes such that for every is called the value set of a. Information Systems/Tables Age LEMS x1 16-30 50 x2 16-30 0 x3 31-45 1-25 x4 31-45 1-25 x5 46-60 26-49 x6 16-30 26-49 x7 46-60 26-49

  4. DS: is the decision attribute(instead of one we can consider more decision attributes). The elements of A are called the condition attributes. Decision Systems/Tables Age LEMS Walk x1 16-30 50 yes x2 16-30 0 no x3 31-45 1-25 no x4 31-45 1-25 yes x5 46-60 26-49 no x6 16-30 26-49 yes x7 46-60 26-49 no

  5. Indiscernibility • The equivalence relation A binary relation which is reflexive (xRx for any object x) , symmetric (if xRy then yRx), and transitive (if xRy and yRz then xRz). • The equivalence class of an element consists of all objects such that xRy.

  6. Indiscernibility (2) • Let IS = (U, A) be an information system, then with any there is an associated equivalence relation: where is called the B-indiscernibility relation. • If then objects x and x’ are indiscernible from each other by attributes from B. • The equivalence classes of the B-indiscernibility relation are denoted by

  7. The non-empty subsets of the condition attributes are {Age}, {LEMS}, and {Age, LEMS}. IND({Age}) = {{x1,x2,x6}, {x3,x4}, {x5,x7}} IND({LEMS}) = {{x1}, {x2}, {x3,x4}, {x5,x6,x7}} IND({Age,LEMS}) = {{x1}, {x2}, {x3,x4}, {x5,x7}, {x6}}. An Example of Indiscernibility Age LEMS Walk x1 16-30 50 yes x2 16-300 no x3 31-45 1-25 no x4 31-451-25 yes x5 46-60 26-49 no x6 16-30 26-49 yes x7 46-60 26-49 no

  8. Set Approximation • Let T = (U, A) and let and We can approximate X using only the information contained in B by constructing the B-lower and B-upper approximations of X, denoted and respectively, where

  9. Set Approximation (2) • B-boundary region of X, consists of those objects that we cannot decisively classify into X in B. • A set is said to be roughif its boundary region is non-empty, otherwise the set is crisp.

  10. Let W = {x | Walk(x) = yes}. The decision class, Walk, is rough since the boundary region is not empty. An Example of Set Approximation Age LEMS Walk x1 16-30 50 yes x2 16-30 0 no x3 31-45 1-25 no x4 31-45 1-25 yes x5 46-60 26-49 no x6 16-30 26-49 yes x7 46-60 26-49 no

  11. An Example of Set Approximation (2) {{x2}, {x5,x7}} {{x3,x4}} yes AW {{x1},{x6}} yes/no no

  12. Lower & Upper Approximations U U/R R : subset of attributes setX

  13. Lower & Upper Approximations (2) Upper Approximation: Lower Approximation:

  14. Lower & Upper Approximations (3) The indiscernibility classes defined by R = {Headache, Temp.} are {u1}, {u2}, {u3}, {u4}, {u5, u7}, {u6, u8}. X1 = {u | Flu(u) = yes} = {u2, u3, u6, u7} RX1 = {u2, u3} = {u2, u3, u6, u7, u8, u5} X2 = {u | Flu(u) = no} = {u1, u4, u5, u8} RX2 = {u1, u4} = {u1, u4, u5, u8, u7, u6}

  15. Lower & Upper Approximations (4) R = {Headache, Temp.} U/R = { {u1}, {u2}, {u3}, {u4}, {u5, u7}, {u6, u8}} X1 = {u | Flu(u) = yes} = {u2,u3,u6,u7} X2 = {u | Flu(u) = no} = {u1,u4,u5,u8} X2 X1 RX1 = {u2, u3} = {u2, u3, u6, u7, u8, u5} u2 u7 u5 u1 RX2 = {u1, u4} = {u1, u4, u5, u8, u7, u6} u6 u8 u3 u4

  16. Accuracy of Approximation where |X| denotes the cardinality of Obviously If X is crisp with respect to B. If X is rough with respect to B.

  17. Issues in the Decision Table • The same or indiscernible objects may be represented several times. • Some of the attributes may be superfluous (redundant). That is, their removal cannot worsen the classification.

  18. Reducts • Keep only those attributes that preserve the indiscernibility relation and, consequently, set approximation. • There are usually several such subsets of attributes and those which are minimal are called reducts.

  19. Dispensable & Indispensable Attributes Let Attribute c is dispensable in T if , otherwise attribute c is indispensable in T. The C-positive region of D:

  20. Independent • T = (U, C, D) is independent if all are indispensable in T.

  21. Reduct & Core • The set of attributes is called a reduct of C, if T’ = (U, R, D) is independent and • The set of all the condition attributes indispensable in T is denoted by CORE(C). where RED(C) is the set of all reducts of C.

  22. An Example of Reducts & Core Reduct1 = {Muscle-pain,Temp.} Reduct2 = {Headache, Temp.} CORE = {Headache,Temp} {MusclePain, Temp} = {Temp}

  23. Discernibility Matrix (relative to positive region) • Let T = (U, C, D) be a decision table, with By a discernibility matrix of T, denoted M(T), we will mean matrix defined as: for i, j = 1,2,…,n such that or belongs to the C-positive region of D. • is the set of all the condition attributes that classify objects ui and uj into different classes.

  24. Discernibility Matrix (relative to positive region) (2) • The equation is similar but conjunction is taken over all non-empty entries of M(T) corresponding to the indices i, j such that or belongs to the C-positive region of D. • denotes that this case does not need to be considered. Hence it is interpreted as logic truth. • All disjuncts of minimal disjunctive form of this function define the reducts of T (relative to the positive region).

  25. Discernibility Function (relative to objects) • For any where (1) is the disjunction of all variables a such that if (2) if (3) if Each logical product in the minimal disjunctive normal form (DNF) defines a reduct of instance

  26. Examples of Discernibility Matrix In order to discern equivalence classes of the decision attribute d, to preserve conditions described by the discernibility matrix for this table No a b c d u1 a0 b1 c1 y u2 a1 b1 c0 n u3 a0 b2 c1 n u4 a1 b1 c1 y u1 u2 u3 C = {a, b, c} D = {d} u2 u3 u4 a,c b c a,b Reduct = {b, c}

  27. Examples of Discernibility Matrix (2) u1 u2 u3 u4 u5 u6 u2 u3 u4 u5 u6 u7 b,c,d b,c b b,d c,d a,b,c,d a,b,c a,b,c,d a,b,c,d a,b,c a,b,c,d a,b,c,d a,b c,d c,d Core = {b} Reduct1 = {b,c} Reduct2 = {b,d}

  28. Discretization • In the discretization of a decision table T = where is an interval of real values, we search for a partition of for any • Any partition of is defined by a sequence of the so-called cuts from • Any family of partitions can be identified with a set of cuts.

  29. Discretization (2) In the discretization process, we search for a set of cuts satisfying some natural conditions. U a b d P P U a b d x1 0.8 2 1 x2 1 0.5 0 x3 1.3 3 0 x4 1.4 1 1 x5 1.4 2 0 x6 1.6 3 1 x7 1.3 1 1 x1 0 2 1 x2 1 0 0 x3 1 2 0 x4 1 1 1 x5 1 2 0 x6 2 2 1 x7 1 1 1 P = {(a, 0.9), (a, 1.5), (b, 0.75), (b, 1.5)}

  30. A Geometrical Representation of Data b x3 x6 3 x1 x5 2 x7 1 x4 x2 0.5 a 0 0.8 1 1.3 1.4 1.6

  31. A Geometrical Representation of Data and Cuts b x3 x6 3 x1 x5 2 x7 x4 1 x2 0.5 a 0 0.8 1 1.3 1.4 1.6

  32. Discretization (3) • The sets of possible values of a and b are defined by • The sets of values of a and b on objects from U are given by a(U) = {0.8, 1, 1.3, 1.4, 1.6}; b(U) = {0.5, 1, 2, 3}.

  33. Discretization Based on RSBR (4) • The discretization process returns a partition of the value sets of condition attributes into intervals.

  34. A Discretization Process • Step 1: define a set of Boolean variables, where corresponds to the interval [0.8, 1) of a corresponds to the interval [1, 1.3) of a corresponds to the interval [1.3, 1.4) of a corresponds to the interval [1.4, 1.6) of a corresponds to the interval [0.5, 1) of b corresponds to the interval [1, 2) of b corresponds to the interval [2, 3) of b

  35. 0.8 1.0 1.3 1.4 1.6 a The Set of Cuts on Attribute a

  36. A Discretization Process (2) • Step 2: create a new decision table by using the set of Boolean variables defined in Step 1. Let be a decision table, be a propositional variable corresponding to the interval for any and

  37. A Sample Defined in Step 2 U* 1 0 0 0 1 1 0 1 1 0 0 0 0 1 1 1 1 0 0 0 0 0 1 1 0 1 0 0 0 0 1 0 0 1 1 0 0 0 0 0 1 0 0 1 1 1 1 1 1 0 0 1 1 0 0 0 0 0 0 1 0 0 1 0 1 0 0 1 0 0 0 0 0 0 0 1 0 0 0 1 0 0 1 0 (x1,x2) (x1,x3) (x1,x5) (x4,x2) (x4,x3) (x4,x5) (x6,x2) (x6,x3) (x6,x5) (x7,x2) (x7,x3) (x7,x5)

  38. The Discernibility Formula • The discernibility formula means that in order to discern object x1 and x2, at least one of the following cuts must be set, a cut between a(0.8) and a(1) a cut between b(0.5) and b(1) a cut between b(1) and b(2).

  39. The Discernibility Formulae for All Different Pairs

  40. The Discernibility Formulae for All Different Pairs (2)

  41. A Discretization Process (3) • Step 3: find the minimal subset of p that discerns all objects in different decision classes. The discernibility boolean propositional formula is defined as follows,

  42. The Discernibility Formula in CNF Form

  43. The Discernibility Formula in DNF Form • We obtain four prime implicants, is the optimal result, because it is the minimal subset of P.

  44. The Minimal Set Cuts for the Sample DB b x3 x6 3 x1 x5 2 x7 x4 1 x2 0.5 a 0 0.8 1 1.3 1.4 1.6

  45. A Result U a b d P P U a b d x1 0.8 2 1 x2 1 0.5 0 x3 1.3 3 0 x4 1.4 1 1 x5 1.4 2 0 x6 1.6 3 1 x7 1.3 1 1 x1 0 1 1 x2 0 0 0 x3 1 1 0 x4 1 0 1 x5 1 1 0 x6 2 1 1 x7 1 0 1 P = {(a, 1.2), (a, 1.5), (b, 1.5)}

  46. Searching for Reducts Condition Attributes: a: Va = {1, 2} b: Vb = {0, 1, 2} c: Vc = {0, 1, 2} d: Vd = {0, 1} Decision Attribute: e: Ve = {0, 1, 2}

  47. Searching for CORE Removing attribute a Removing attribute a does not cause inconsistency. Hence, a is not used as CORE.

  48. Searching for CORE(2) Removing attribute b Removing attribute b cause inconsistency. Hence, b is used as CORE.

  49. Searching for CORE(3) Removing attribute c Removing attribute c does not cause inconsistency. Hence, c is not used as CORE.

  50. Searching for CORE(4) Removing attribute d Removing attribute d does not cause inconsistency. Hence, d is not used as CORE.

More Related