1 / 15

Introduction to Machine Learning

Introduction to Machine Learning. Dmitriy Dligach. Representations. Objects Real-life phenomena viewed as objects and their properties (features) Feature Vectors Examples Text classification Face recognition WSD. Supervised Learning. Vector-value pair

kane-burton
Download Presentation

Introduction to Machine Learning

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Introduction to Machine Learning Dmitriy Dligach

  2. Representations • Objects • Real-life phenomena viewed as objects and their properties (features) • Feature Vectors • Examples • Text classification • Face recognition • WSD

  3. Supervised Learning • Vector-value pair • (x0, y0),(x1, y1), … , (xn, yn) • Task: learn function y = f(x) • Algorithms • KNN • Decision Trees • Neural Networks • SVM

  4. Issues in Supervised Learning • Training data • Why are we learning? • Test data • Unseen data • Overfitting • Fitting noise reduces performance

  5. Unsupervised Learning • Only feature vectors are given • x0, x1,…, xn • Task: group feature vectors into clusters • Algorithms • Clustering • k-means • mixture of gaussians • Principal Component Analysis • Sequence labeling • HMMs

  6. Supervised Example: Decision Trees

  7. A Tree

  8. Word Sense Disambiguation (WSD) • bat (noun) • http://wordnet.princeton.edu/perl/webwn • http://verbs.colorado.edu/html_groupings/

  9. Another DT Example • Word Sense Disambiguation • Given an occurrence of a word, decide which sense, or meaning, was intended. • Example, run • run1: move swiftly ( I ran to the store.) • run2: operate (I run a store.) • run3: flow (Water runs from the spring.) • run4: length of torn stitches (Her stockings had a run.)

  10. WSD • Word Sense Disambiguation • Categories • Use word sense labels (run1, run2, etc.) • Features – describe context of word • near(w) : is the given word near word w? • pos: word’s part of speech • left(w): is word immediately preceded by w? • etc.

  11. Using a decision Tree • Given an event (=list of feature values): • Start at the root. • At each interior node, follow the outgoing arc for the feature value that matches our event • When we reach a leaf node, return its category. run4 pos verb noun “I saw John run a race by a river.” near(stocking) near(race) yes no yes no run1 near(river) yes run3 4pm

  12. WSD: Sample Training Data

  13. Unsupervised Example: K-Means • Distance between two objects • Cosine distance • Euclidean distance • Algorithm • Pick cluster centers at random • Assign the data points to the nearest clusters • Re-compute the cluster centers • Re-assign the data points • Continue until the clusters settle • Hard clustering vs. soft clustering

  14. Interactive Demos • K-Means • http://home.dei.polimi.it/matteucc/Clustering/tutorial_html/AppletKM.html • SVMs • http://www.csie.ntu.edu.tw/~cjlin/libsvm/#GUI

  15. ML Reference • Tom Mitchell “Machine Learning” • http://home.dei.polimi.it/matteucc/Clustering/tutorial_html/AppletKM.html • http://www.aaai.org/AITopics/html/machine.html

More Related