1 / 18

Decision Trees with Numeric Tests

Decision Trees with Numeric Tests. Industrial-strength algorithms. For an algorithm to be useful in a wide range of real-world applications it must: Permit numeric attributes Allow missing values Be robust in the presence of noise

dahlia
Download Presentation

Decision Trees with Numeric Tests

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Decision Trees with Numeric Tests

  2. Industrial-strength algorithms • For an algorithm to be useful in a wide range of real-world applications it must: • Permit numeric attributes • Allow missing values • Be robust in the presence of noise • Basic schemes need to be extended to fulfill these requirements

  3. C4.5 History • ID3, CHAID – 1960s • C4.5 innovations (Quinlan): • permit numeric attributes • deal sensibly with missing values • pruning to deal with for noisy data • C4.5 - one of best-known and most widely-used learning algorithms • Last research version: C4.8, implemented in Weka as J4.8 (Java) • Commercial successor: C5.0 (available from Rulequest)

  4. Numeric attributes • Standard method: binary splits • E.g. temp < 45 • Unlike nominal attributes,every attribute has many possible split points • Solution is straightforward extension: • Evaluate info gain (or other measure)for every possible split point of attribute • Choose “best” split point • Info gain for best split point is info gain for attribute • Computationally more demanding

  5. Example • Split on temperature attribute: • E.g. temperature  71.5: yes/4, no/2 temperature  71.5: yes/5, no/3 • Info([4,2],[5,3])= 6/14 info([4,2]) + 8/14 info([5,3]) = 0.939 bits • Place split points halfway between values • Can evaluate all split points in one pass!

  6. Example • Split on temperature attribute: infoGain 0

  7. X Speeding up • Entropy only needs to be evaluated between points of different classes (Fayyad & Irani, 1992) value class Potential optimal breakpoints Breakpoints between values of the same class cannot be optimal

  8. Missing as a separate value • Missing value denoted “?” in C4.5 • Simple idea: treat missing as a separate value • Q: When is this not appropriate? • A: When values are missing due to different reasons • Example 1: gene expression could be missing when it is very high or very low • Example 2: field IsPregnant=missing for a male patient should be treated differently (no) than for a female patient of age 25 (unknown)

  9. Missing values - advanced Split instances with missing values into pieces • A piece going down a branch receives a weight proportional to the popularity of the branch • weights sum to 1 • Info gain works with fractional instances • use sums of weights instead of counts • During classification, split the instance into pieces in the same way • Merge probability distribution using weights

  10. Application: Computer Vision 1

  11. Application: Computer Vision 2 feature extraction • color (RGB, hue, saturation) • edge, orientation • texture • XY coordinates • 3D information

  12. Application: Computer Vision 3 how grey? below horizon?

  13. Application: Computer Vision 4 prediction

  14. Application: Computer Vision 4 inverse perspective

  15. Application: Computer Vision 5 inverse perspective path planning

  16. B + + + + + - - - - A Quiz 1 Q: If an attribute A has high info gain, does it always appear in a decision tree? A: No. If it is highly correlated with another attribute B, and infoGain(B) > infoGain(A), then B will appear in the tree, and further splitting on A will not be useful.

  17. + + + - - + + - - A Quiz 2 Q: Can an attribute appear more than once in a decision tree? A: Yes. If a test is not at the root of the tree, it can appear in different branches. Q: And on a single path in the tree (from root to leaf)? A: Yes. Numeric attributes can appear more than once, but only with very different numeric conditions.

  18. B - + - - + + + - - + - + the XOR problem: - + - + A Quiz 3 Q: If an attribute A has infoGain(A)=0, can it ever appear in a decision tree? A: Yes. • All attributes may have zero info gain. • info gain often changes when splitting on another attribute.

More Related