1 / 23

Machine Learning II - Outline

Machine Learning II - Outline. Instance Based Learning k-Nearest Neighbour Learning Case-based Reasoning Unsupervised Learning Numeric Clustering Methods Conceptual Clustering Other Machine Learning Paradigms Artificial Neural Networks Evolutionary Techniques. Instance-Based Learning.

elisha
Download Presentation

Machine Learning II - Outline

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Machine Learning II - Outline • Instance Based Learning • k-Nearest Neighbour Learning • Case-based Reasoning • Unsupervised Learning • Numeric Clustering Methods • Conceptual Clustering • Other Machine Learning Paradigms • Artificial Neural Networks • Evolutionary Techniques J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  2. Instance-Based Learning • Lazy methods • do not estimate the target function once for the entire instance space • estimate it locally and differently for each new instance • generalising beyond the training examples is postponed • Instances represented as points in a Euclidean space as well as more complex symbolic representations J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  3. Instance-Based Learning • Straightforward approaches to approximating real-valued or discrete-valued target functions • Key idea: just store all training examples xi, f(xi) • Given new query instance xq, • first locate set of similar related training examples, • then use them to classify the new instance • Construct a local approximation to the target function that applies in the neighbourhood of the new instance • Even very complex target function can be described by a collection of less complex approximations J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  4. k-Nearest Neighbour Learning • Instances map to points in the n-dimensional space Rn • The distance between two instances is defined as the static Euclidean distance • Given xq, take vote among its k nearest nbrs (if discrete-valued target function) • Take mean of f values of k nearest nbrs (if real-valued) • ; where (a, b)=1 if a=b and (a, b)=0 otherwise. J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  5. Operation of the k-NN algorithm Voronoi diagram - the shape of the decision surface induced by 1-NN over the entire instance space J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  6. Distance-Weighted k-NN Algorithm • Weights the contributions of the neighbours according to their distance to the query point where • Note now it makes sense to use all training examples instead of just k (Shepard’s method) J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  7. Remarks on k-NN Algorithm • Robust to noisy training data • Effective for large set of training data • The cost of classification of a new instances can be high - techniques for efficiently indexing training examples • All attributes are considered of the instance when calculating a distance • dangerous, if the concept depends on only a few relevant attributes • curse of dimensionality J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  8. Curse of Dimensionality • Imagine instances described by 20 attributes, but only 2 are relevant to target function • Curse of dimensionality: nearest neighbours is easily mislead when high-dimensional • The distance between two neighbours will be dominated by the large number of irrelevant attributes • One approach to remedy: • to weight each attribute differently when calculating the distance • shortening (lengthening) the axes that correspond to more (less) relevant attributes J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  9. IBL: Case StudyPrediction of Gas Consumption J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  10. IBL: Case StudyPrediction of Gas Consumption J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  11. Case-Based Reasoning • Symbolic representation of instances • function graphs, logical descriptions, sequence (or tree) of operations • Methods for retrieving similar instances are more elaborate • size of the largest shared subgraph between two function graphs • Multiple cases retrieved, modified and combined to form solution to new problem • Applications • storing and reusing experience at a help desk • conceptual design of mechanical devices • solving complex planning and scheduling problems by reusing relevant portions of previously solved problems J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  12. Case-Based Reasoning in CADET • CADET: • Conceptual Design of simple mechanical devices- water faucets • 75 stored examples of mechanical devises • CBR: • training example: qualitative function, mechanical structure • new query: desired function • target value: mechanical structure for this function • distance metric: match qualitative function descriptions J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  13. Case-Based Reasoning in CADET J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  14. Unsupervised Learning • No teacher - a method of classifying training examples • Learner forms and evaluates concepts on its own • must discover patterns, features, regularities, correlation, or categories in the input data • Scientists: • do not have benefit of a teacher, • propose hypotheses to explain observations, • evaluate & test hypotheses • Clustering problem • Numeric taxonomy • Conceptual clustering J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  15. Conceptual Clustering • Objects represented as sets of symbolic rather than numeric features • Measuring the similarity of objects • proportion of common features similarity({small,red,rubber,ball},{small,blue,rubber,ball}) = 3/4 • Background knowledge used to formation of categories • Intensional definition rather than extensional enumerative representation {X | X has been elected secretary-general of the United Nations} vs. list of particular individuals • Machine learning techniques used to produce general concepts J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  16. Conceptual Clustering - CLUSTER/2 • CLUSTER/2 (Michalski and Stepp, 1983) - forms k categories 1. Select k seeds from the set of observed objects. 2. For each seed, using that seed as a positive instance and all other seed as negative instances, produce a maximally general definition covering only positive and none of the negative instances. 3. Classify all objects in the sample according to these descriptions. Replace each description with a maximally specific description that covers all objects in the category. 4. Adjust overlapping definitions. 5. Select an element closest to the center of each class. 6. Using these central elements as new seeds, repeat steps 1-5. Stop when clusters are satisfactory - complexity of the general descriptions of classes. 7. If unsatisfactory clusters and no improvement occurs select the new seeds closest to the edge of the cluster. J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  17. Steps of a CLUSTER/2 run J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  18. Numeric Taxonomy • Objects represented as a collection of numeric-valued features • Similarity measure - similarity of two objects • Euclidean distance in the n-dimensional space • Distance between two clusters • minimal, maximal, average, mean measure • Criterion of the quality of clustering • maximising the similarity of objects in the same class • Methods: • iterative optimisation • agglomerative clustering J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  19. Distribution of Training Examples a) proper distribution b) improper distribution J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  20. Iterative Optimisation: Algorithm 1. Split the space by the randomly chosen border line so that two non-empty sets of objects are created. 2. For each set of objects calculate a vector of mean values of each attribute - ethalons. Classify all objects according to those ethalons. 3. New separation line is given by the ethalons. If the classification according to the ethalons changed at least in one object then continue with step 2 else end. J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  21. Iterative Optimisation J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  22. Agglomerative Clustering - Dendrogram J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

  23. Agglomerative Clustering - Algorithm 1. Examining all pairs of objects, select the pair with the highest degree of similarity and make that pair a cluster • replace the component members of the cluster with the cluster definition 2. Repeat the step 1 on the collection of objects and clusters until all objects have been reduced to a single cluster. • Result: A binary tree whose leaf nodes are instances and internal nodes are clusters of increasing size J. Kubalík, Gerstner Laboratory for Intelligent Decision Making and Control

More Related