1 / 12

Improving the Graph Mincut Approach to Learning from Labeled and Unlabeled Examples

Improving the Graph Mincut Approach to Learning from Labeled and Unlabeled Examples. Avrim Blum, John Lafferty, Raja Reddy, Mugizi Rwebangira. Outline. Often have little labeled data but lots of unlabeled data

Download Presentation

Improving the Graph Mincut Approach to Learning from Labeled and Unlabeled Examples

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Improving the Graph Mincut Approach to Learning from Labeled and Unlabeled Examples Avrim Blum, John Lafferty, Raja Reddy, Mugizi Rwebangira

  2. Outline • Often have little labeled data but lots of unlabeled data • Graph mincuts: based on a belief that most ‘close’ examples have same classification • Problem: -Does not say where it is most confident • Our approach: Add noise to edges to extract confidence scores

  3. Learning using Graph Mincuts:Blum and Chawla (ICML 2001)

  4. Construct a Graph

  5. Add sink and source + -

  6. Obtain s-t mincut + - Mincut

  7. Classification + - Mincut

  8. Goal • To obtain a measure of confidence on each classification Our approach • Add random noise to the edges • Run min cut several times • For each unlabeled example take majority vote

  9. Experiments • Digits data set (each digit is a 16 X 16 integer array) • 100 labeled examples • 3900 unlabeled examples • 100 runs of mincut

  10. Results

  11. Conclusions • 3% error on 80% of the data • Standard mincut only gives us 6% error on all the data • Future Work • Conduct further experiments on other data sets • Compare with similar algorithm of Jerry Zhu • Investigate the properties of the distribution we get by selecting minimum cuts in this way

  12. Questions?

More Related