If P(  m | x)  1 , then the nearest neighbor selection is almost always the same as the Bayes selection

1 / 8

# If P(  m | x)  1 , then the nearest neighbor selection is almost always the same as the Bayes selection - PowerPoint PPT Presentation

Pattern Classification All materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000 with the permission of the authors and the publisher.

I am the owner, or an agent authorized to act on behalf of the owner, of the copyrighted work described.

## If P(  m | x)  1 , then the nearest neighbor selection is almost always the same as the Bayes selection

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

Pattern ClassificationAll materials in these slides were taken from Pattern Classification (2nd ed) by R. O. Duda, P. E. Hart and D. G. Stork, John Wiley & Sons, 2000with the permission of the authors and the publisher

If P(m | x)  1, then the nearest neighbor selection is almost always the same as the Bayes selection
The k – nearest-neighbor rule
• Goal: Classify x by assigning it the label most frequently represented among the k nearest samples and use a voting scheme
Example:

k = 3 (odd value) and x = (0.10, 0.25)t

Closest vectors to x with their labels are:

{(0.10, 0.28, 2); (0.12, 0.20, 2); (0.15, 0.35,1)}

One voting scheme assigns the label 2 to x since 2 is the most frequently represented