1 / 42

Principal Components

Principal Components . Karl Pearson. Principal Components (PC). Objective : Given a data matrix of dimensions nxp (p variables and n elements) try to represent these data by using r variables (r<p) with minimum lost of information .

pepper
Download Presentation

Principal Components

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Principal Components

  2. Karl Pearson

  3. Principal Components (PC) • Objective: Given a data matrix of dimensions nxp (p variables and n elements) try to represent these data by using r variables (r<p) with minimum lost of information

  4. We want to find a new set of p variables, Z, which are linear combinations of the original X variable such that : • r of them contains all the information • The remaining p-r variables are noise

  5. First interpretation of principal components Optimal Data Representation

  6. Proyection of a point in direction a: minimize the squared distance Implies maximizing the variance (assuming zero mean variables) ri xi zi xiTxi = riT ri+ zTi zi a

  7. Second interpretation of PC: Optimal Prediction Find a new variable zi =a’Xi which is optimal to predict The value of Xi in each element . In general, find r variables, zi =Ar Xi , which are optimal to forecast All Xi with the least squared error criterion It is easy to see that the solution is that zi =a’Xi must have maximum variance

  8. Third interpretation of PC Find the optimal direction to represent the data. Axe of the ellipsoid which contains the data The line which minimizes the orthogonal distance provides the axes of the ellipsoid This is idea of Pearson orthogonal regression

  9. Second component

  10. Properties of PC

  11. Standardized PC

  12. Example Inves

  13. Example Inves

  14. Example Medifis

  15. Example mundodes

  16. Example Mundodes

  17. Example for image analysis

  18. The analysis have been done with 16 images. PC allows that Instead of sending 16 matrices of N2 pixels we send a vector 16x3 with the values of the components and a matrix 3xN2 with the values of the new variables. We save If instead of 16 images we have 100 images we save 95%

More Related