1 / 20

Lecture 14: Classification

Thursday 18 February 2010. Lecture 14: Classification. Reading: Ch. 7.13 – 7.19. Last lecture: Spectral Mixture Analysis. Classification vs. Spectral Mixture Analysis.

Download Presentation

Lecture 14: Classification

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Thursday 18 February 2010 Lecture 14: Classification Reading: Ch. 7.13 – 7.19 Last lecture: Spectral Mixture Analysis

  2. Classification vs.Spectral Mixture Analysis In SMA, image pixels were regarded as being mixed from various proportions of common materials. The goal was to find what those materials were in an image, and what their proportions were pixel by pixel. In classification, the image pixels are regarded as grouping into thematically and spectrally distinct clusters (in DN space). Each pixel is tested to see what group it most closely resembles. The goal is to produce a map of the spatial distribution of each theme or unit. Water – group 1 Forests – group 2 Desert – group 3

  3. AVHRR Images with pixels similar to vegetation flagged according to distance at different tolerances D D=2 D=48 D=60 Multi-unit veg map

  4. What is spectral similarity? Spectral distance: d Spectral angle: f x A dA dB B Y fA fB X Spectral contrast between similar objects is small

  5. Manual Classification • association by spectral similarity • of pixels into units • 2) naming those units, generally using independent information • - reference spectra • - field determinations • - photo-interpretation Seattle

  6. Basic steps in image classification: 1) Data reconnaissance and self-organization 2) Application of the classification algorithm 3) Validation

  7. Reconnaissance and data organization Reconnaissance What is in the scene? What is in the image? What bands are available? What questions are you asking of the image? Can they be answered with image data? Are the data sufficient to distinguish what’s in the scene? Organization of data How many data clusters in n-space can be recognized? What is the nature of the cluster borders? Do the clusters correspond to desired map units?

  8. Classification algorithms Unsupervised Supervised Form Images Of Data Separate Data Into Groups With Clustering Choose Training Pixels For Each Category Classify Data Into Groups Calculate Statistical Descriptors Assign Name To Each Group Satisfactory ? No Satisfactory ? No Yes Classify Data Into Categories Defined Yes

  9. Unsupervised Classification: K-Means algorithm Pick number of themes; set distance tolerance ° 1st pixel defines 1st theme ° is 2nd pixel within tolerance? - YES: redefine theme - NO: define 2nd theme ° Interrogate 3rd pixel… ° Iterate, using “found” themes as the new seed How do you estimate the number of themes? - can be greater than number of bands

  10. Supervised Classification: What are some algorithms? • Parallelipiped • Minimum Distance • Maximum Likelihood • Decision-Tree x + “Hard” vs. “soft” classification Hard: winner take all Soft: “answer” expressed as probability x belongs to A, B “Fuzzy” classification is very similar to spectral unmixing

  11. Parallelepiped Classifier Assigns a DN range in each band for each class (parallelepiped) Advantages: simple Disadvantages: low accuracy - especially when the distribution in feature space has covariance or dependency with oblique axes.

  12. Minimum-Distance Classifier Uses only the mean of each class. The unknown pixel is classified using its distance to each of the class means. The shortest distance wins. Decision boundaries

  13. Maximum Likelihood The most commonly used classifier used. A pixel is assigned to the class based on statistical probability. Based on statistics (mean; covariance) A (Bayesian) probability function is calculated from the inputs for classes established from training sites. Each pixel is then judged as to the class to which it most probably belong.

  14. Maximum Likelihood For each DN ntuple in the image, 1) calculate the distance to each cluster mean 2) scale by the number of standard deviations in the direction of the ntuple from the mean 3) construct rule images, pixel by pixel for each cluster, in which the number of std dev’s is recorded 4) threshold the rule images (null pixels too far from a cluster) 5) pick best match (least number of std dev’s and record it in the appropriate pixel of the output , image or map

  15. Decision-Tree Classifier Hierarchical classifier compares the data sequentially with carefully selected features. Features are determined from the spectral distributions or separability of the classes. There is no general procedure. Each decision tree or set of rules is custom-designed. A decision tree that provides only two outcomes at each stage is called a “binary decision tree” (BDT) classifier.

  16. Pre-processing - dimension transformation One goal: reduce impact of topography on outcome ratioing NDVI Spectral angle Line of constant ratio y B x/y A x y/z  B A

  17. Validation Photointerpretation Look at the original data: does your map make sense to you?

  18. Confusion matrices Well-named. Also known as contingency tables or error matrices Here’s how they work… All non diagonal elements are errors Row sums give “commission” errors Column sums give “omission” errors Overall accuracy is the diagonal sum over the grand total This is the assessment only for the training areas What do you do for the rest of the data? Training areas A B C D E F Row sums A 0 5 0 0 0 485 480 B 0 52 0 20 0 72 0 C 0 0 Classified data D 0 16 E 0 0 F 0 0 Grand sum 480 68 1992 p 586, LKC 6th Col sums

  19. A basic problem with classification We tend to want to classify by land use, and from the remote sensing perspective this may lead to ambiguity = bear habitat meadow Theme A = grass ≠ bear habitat is found to be spectrally similar to: golf course Mystery pixel X ≠ bear habitat cemetary What’s actually on the ground – all three look similar (A) because they are grassy I want to find bears. Bears like meadows. I train on a meadow (Theme A) and classify an image to see where the bears are. Pixel X is classified as similar to A. Will I find bears there? Maybe not. 1) they might be somewhere else even they like the meadow. 2) What a meadow is from the RS perspective is a high fraction of GV. Other things share this equivalence. Therefore, X may indeed belong to A spectrally, but not according to use. Thought exercise: what would you need to do in order to classify by land use?

  20. Next class: Radar

More Related