1 / 68

Recent developments in nonlinear dimensionality reduction

Recent developments in nonlinear dimensionality reduction. Josh Tenenbaum MIT. Collaborators. Vin de Silva John Langford Mira Bernstein Mark Steyvers Eric Berger. Outline. The problem of nonlinear dimensionality reduction The Isomap algorithm Development #1: Curved manifolds

Download Presentation

Recent developments in nonlinear dimensionality reduction

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Recent developments in nonlinear dimensionality reduction Josh Tenenbaum MIT

  2. Collaborators • Vin de Silva • John Langford • Mira Bernstein • Mark Steyvers • Eric Berger

  3. Outline • The problem of nonlinear dimensionality reduction • The Isomap algorithm • Development #1: Curved manifolds • Development #2: Sparse approximations

  4. Learning an appearance map • Given input: . . . • Desired output: • Intrinsic dimensionality: 3 • Low-dimensional representation:

  5. Linear dimensionality reduction: PCA, MDS • PCA dimensionality of faces: • First two PCs:

  6. Linear manifold: PCA • Nonlinear manifold: ?

  7. Previous approaches to nonlinear dimensionality reduction • Local methods seek a set of low-dimensional models, each valid over a limited range of data: • Local PCA • Mixture of factor analyzers • Global methods seek a single low-dimensional model valid over the whole data set: • Autoencoder neural networks • Self-organizing map • Elastic net • Principal curves & surfaces • Generative topographic mapping

  8. A generative model • Latent space Y _Rd • Latent data {yi} _Y generated from p(Y) • Mapping f: Yd RN for some N > d • Observed data {xi = f (yi)} _ RN Goal: given {xi}, recover f and {yi}.

  9. Chicken-and-egg problem • We know {xi} . . . • . . . and if we knew{yi}, could estimate f. • . . . or if we knew f, could estimate {yi}. • So use EM, right? Wrong.

  10. The problem of local minima GTM SOM • Global nonlinear dimensionality reduction + local optimization = severe local minima

  11. A different approach • Attempt to infer {yi} directly from {xi}, without explicit reference to f. • Closed-form, non-iterative, globally optimal solution for {yi}. • Then can approximate f with a suitable interpolation algorithm (RBFs, local linear, ...). • In other words, finding f becomes a supervised learning problem on pairs {yi ,xi}.

  12. When does this work? • Only given some assumptions on the nature of f and the distribution of the {yi}. • The trick: exploit some invariant of f, a property of the {yi} that is preserved in the {xi}, and that allows the {yi} to be read off uniquely*. * up to some isomorphism (e.g., rotation).

  13. iii) ii) i) The assumptions behind three algorithms Distribution: p(Y) Mapping: f Algorithm No free lunch: weaker assumptions on f u stronger assumptions on p(Y). i) arbitrary linear isometric Classical MDS ii) convex, dense isometric Isomap iii) convex, uniformly dense conformal C-Isomap

  14. i) The assumptions behind three algorithms Distribution: p(Y) Mapping: f Algorithm i) arbitrary linear isometric Classical MDS ii) convex, dense isometric Isomap iii) convex, uniformly dense conformal C-Isomap

  15. Classical MDS • Invariant: Euclidean distance • Algorithm: • Calculate Euclidean distance matrix D • Convert D to canonical inner product matrix B by “double centering”: • Compute {yi} from eigenvectors of B.

  16. ii) The assumptions behind three algorithms Distribution: p(Y) Mapping: f Algorithm i) arbitrary linear isometric Classical MDS ii) convex, dense isometric Isomap iii) convex, uniformly dense conformal C-Isomap

  17. Isomap • Invariant: geodesic distance

  18. The Isomap algorithm • Construct neighborhood graph G. • e method • K method • Compute shortest paths in G, with edge ij weighted by the Euclidean distance |xi - xj|. • Floyd • Dijkstra (+ Fibonacci heaps) • Reconstruct low-dimensional latent data {yi}. • Classical MDS on graph distances • Sparse MDS with landmarks

  19. Illustration on swiss roll

  20. MDS / PCA Isomap Discovering the dimensionality • Measure residual variance in geodesic distances . . . • . . . and find the elbow.

  21. Theoretical analysis of asymptotic convergence • Conditions for PAC-style asymptotic convergence • Geometric: • Mapping f is isometric to a subset of Euclidean space (i.e., zero intrinsic curvature). • Statistical: • Latent data {yi} are a “representative” sample* from a convex domain. • * Minimum distance from any point on the manifold to a • sample point < e (e.g., variable density Poisson process).

  22. Theoretical results on the rate of convergence • Upper bound on the number of data points required. • Rate of convergence depends on several geometric parameters of the manifold: • Intrinsic: • dimensionality • Embedding-dependent: • minimal radius of curvature • minimal branch separation

  23. MDS / PCA Isomap Face under varying pose and illumination • Dimensionality • picture

  24. MDS / PCA Isomap Hand under nonrigid articulation • Dimensionality • picture

  25. Apparent motion

  26. MDS / PCA Isomap Digits • Dimensionality • picture.

  27. Summary of Isomap A framework for global nonlinear dimensionality reduction that preserves the crucial features of PCA and classical MDS: • A noniterative, polynomial-time algorithm. • Guaranteed to construct a globally optimal Euclidean embedding. • Guaranteed to converge asymptotically for an important class of nonlinear manifolds. Plus, good results on real and nontrivial synthetic data sets.

More Related