navneet goyal n.
Download
Skip this Video
Loading SlideShow in 5 Seconds..
Active Learning PowerPoint Presentation
Download Presentation
Active Learning

Loading in 2 Seconds...

play fullscreen
1 / 19

Active Learning - PowerPoint PPT Presentation


  • 146 Views
  • Uploaded on

Navneet Goyal. Active Learning. Instance Based Learning. Rote Classifier K- nearest neighbors (K-NN) Case Based Resoning (CBR). Classification: Eager & Lazy Learners. Decision Tree classifier is an example of an “eager learner”

loader
I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.
capcha
Download Presentation

Active Learning


An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.


- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
    Presentation Transcript
    1. NavneetGoyal Active Learning

    2. Instance Based Learning • Rote Classifier • K- nearest neighbors (K-NN) • Case Based Resoning (CBR)

    3. Classification: Eager & Lazy Learners • Decision Tree classifier is an example of an “eager learner” • Because they are designed to learn a model that maps the input attributes to the class label as soon as the training data becomes available • An opposite strategy would be to delay the process of modeling the training data until it is needed to classify the test examples • LAZY Learners

    4. Classification: Eager & Lazy Learners • Rote classifier is an example of lazy learner, which memorizes the entire training data & performs the classification only if the attributes of a test instance matches exactly with one of the training examples • Drawback: Cannot classify a new instance if it does match any training example

    5. Classification: Nearest Neighbors • To overcome this drawback, we find all training examples that are relatively ‘similar’ to the test example • Examples, which are known as ‘Nearest Neighbors’ can be used to determine the class label of the test example • If it walks like a duck, quacks like a duck, and looks like a duck, then it is probably a duck

    6. Compute Distance Test Record Training Records Choose k of the “nearest” records Nearest Neighbor Classifiers

    7. Classification: Nearest Neighbors • Each test example is represented as a point in a d-dimensional space • For east test example we use a proximity measure • K-nearest neighbors of a given example z refer to the k points that are closest to z

    8. Classification Using Distance • Place items in class to which they are “closest”. • Must determine distance between an item and a class. • Classes represented by • Centroid: Central value. • Medoid: Representative point. • Individual points • Algorithm: KNN

    9. Definition of Nearest Neighbor K-nearest neighbors of a record x are data points that have the k smallest distance to x

    10. Nearest-Neighbor Classifiers • Requires three things • The set of stored records • Distance Metric to compute distance between records • The value of k, the number of nearest neighbors to retrieve • To classify an unknown record: • Compute distance to other training records • Identify k nearest neighbors • Use class labels of nearest neighbors to determine the class label of unknown record (e.g., by taking majority vote)

    11. Nearest Neighbor Classification • Compute distance between two points: • Euclidean distance Determine the class from nearest neighbor list • take the majority vote of class labels among the k-nearest neighbors • Weigh the vote according to distance • weight factor, w = 1/d2

    12. Nearest Neighbor Classification… • Choosing the value of k: • If k is too small, sensitive to noise points • If k is too large, neighborhood may include points from other classes

    13. Nearest Neighbor Classification… • Problem with Euclidean measure: • High dimensional data • curse of dimensionality • Can produce counter-intuitive results 1 1 1 1 1 1 1 1 1 1 1 0 1 0 0 0 0 0 0 0 0 0 0 0 vs 0 1 1 1 1 1 1 1 1 1 1 1 0 0 0 0 0 0 0 0 0 0 0 1 d = 1.4142 d = 1.4142 • Solution: Normalize the vectors to unit length

    14. Nearest neighbor Classification… • Part of more general technique called as ‘Instance-based learning’ • Require a proximity measure • k-NN classifiers are lazy learners • It does not build models explicitly • Unlike eager learners such as decision tree induction and rule-based systems • Classifying unknown records are relatively expensive

    15. Nearest neighbor Classification… • Make their classification based on local information, where as DT & rule-based classifiers attempt to find global model that fits the entire input space • As decisions are made locally, they are quite susceptible to noise (for small k) • Can produce wrong results unless the appropriate proximity measure and data preprocessing steps are taken

    16. Nearest Neighbor Classification… • Scaling issues • Attributes may have to be scaled to prevent distance measures from being dominated by one of the attributes • Example: • height of a person may vary from 1.5m to 1.8m • weight of a person may vary from 90lb to 300lb • income of a person may vary from $10K to $1M • Proximity measure may be dominated by differences in weights and income of a person

    17. Example: PEBLS • PEBLS: Parallel Examplar-Based Learning System (Cost & Salzberg) • Works with both continuous and nominal features • For nominal features, distance between two nominal values is computed using modified value difference metric (MVDM) • Each record is assigned a weight factor • Number of nearest neighbor, k = 1

    18. Example: PEBLS Distance between nominal attribute values: d(Single,Married) = | 2/4 – 0/4 | + | 2/4 – 4/4 | = 1 d(Single,Divorced) = | 2/4 – 1/2 | + | 2/4 – 1/2 | = 0 d(Married,Divorced) = | 0/4 – 1/2 | + | 4/4 – 1/2 | = 1 d(Refund=Yes,Refund=No) = | 0/3 – 3/7 | + | 3/3 – 4/7 | = 6/7

    19. Example: PEBLS Distance between record X and record Y: where: wX 1 if X makes accurate prediction most of the time wX> 1 if X is not reliable for making predictions