1 / 21

The Beauty of Local Invariant Features

The Beauty of Local Invariant Features. Svetlana Lazebnik Beckman Institute, University of Illinois at Urbana-Champaign. IMA Recognition Workshop University of Minnesota May 22, 2006. What are Local Invariant Features?.

rgarver
Download Presentation

The Beauty of Local Invariant Features

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The Beauty of Local Invariant Features Svetlana Lazebnik Beckman Institute, University of Illinois at Urbana-Champaign IMA Recognition Workshop University of Minnesota May 22, 2006

  2. What are Local Invariant Features? • Descriptors of image patches that are invariant to certain classes of geometric and photometric transformations Lowe (2004)

  3. A Historical Perspective Appearance-based methods: global appearance, no local shape Model-based methods: local shape, no appearance information ACRONYM: Brooks and Binford (1981)Alignment: Huttenlocher & Ullman (1987)Invariants: Rothwell et al. (1992) Eigenfaces: Turk & Pentland (1991)Appearance manifolds: Murase & Nayar (1995)Color histograms: Swain & Ballard (1990) Local invariant features: local shape + appearance pattern +

  4. Feature Detection and Description 3. Compute appearancedescriptors 2. Normalize regions SIFT: Lowe (2004) invariant description 1. Detect regions covariant detection

  5. Advantages • Locality • Robustness to clutter and occlusion • Repeatability • The same feature occurs in multiple images of the same scene or class • Distinctiveness • Salient appearance pattern that provides strong matching constraints • Invariance • Allow matching despite scale changes, rotations, viewpoint changes • Sparseness • Relatively few features per image, compact and efficient representation • Flexibility • Many existing types of detectors, descriptors

  6. Scale-Covariant Detectors • Laplacian, Hessian, Difference-of-Gaussian(blobs)Lindeberg (1998), Lowe (1999, 2004) • Harris-Laplace(corners)Mikolajczyk & Schmid (2001)

  7. Scale-Covariant Detectors • Salient (high entropy) regionsKadir & Brady (2001) • Circular edge-based regionsJurie & Schmid (2003)

  8. Affine-Covariant Detectors • Laplacian, Hessian-Affine (blobs) Gårding & Lindeberg (1996), Mikolajczyk et al. (2004) • Harris-Affine (corners)Mikolajczyk & Schmid (2002)

  9. Affine-Covariant Detectors • Edge- and intensity-based regionsTuytelaars & Van Gool (2004) • Maximally stable extremal regions (MSER)Matas et al. (2002)

  10. Types of Descriptors Johnson & Hebert (1999)Lazebnik, Schmid & Ponce (2003) Belongie, Malik & Puzicha (2002) Lowe (1999, 2004) PCA-SIFT: Ke & Sukthankar (2004)GLOH: Mikolajczyk & Schmid (2004) • Differential invariantsKoenderink & Van Doorn (1987), Florack et al. (1991) • Filter banks: complex, Gabor, steerable, … • Multidimensional histograms

  11. Applications (1) • Wide-baseline matching and recognition of specific objects Tuytelaars & Van Gool (2004) Ferrari, Tuytelaars & Van Gool (2005) Lowe (2004) Rothganger, Lazebnik, Schmid & Ponce (2005)

  12. Applications (2) • Category-level recognition based on geometric correspondence Lazebnik, Schmid & Ponce (2004) Berg, Berg & Malik (2005)

  13. Applications (3) Bag of features Csurka, Dance, Fan, Willamowski & Bray (2004)Dorko & Schmid (2005)Sivic, Russell, Efros, Zisserman & Freeman (2005)Sivic & Zisserman (2003) • Learning parts and visual vocabularies Constellation model Fergus, Perona & Zisserman (2003)Weber, Welling & Perona (2000)

  14. Applications (4) • Building global image models invariant to a wide range of deformations Lazebnik, Schmid & Ponce (2005)

  15. Comparative Evaluations • Flat scenesMikolajczyk & Schmid (2004), Mikolajczyk et al. (2004) • MSER and Hessian regions have the highest repeatability • Harris and Hessian regions provide the most correspondences • SIFT (GLOH, PCA-SIFT) descriptors have the highest performance • 3D objectsMoreels & Perona (2006) • Features on 3D objects are much more unstable than on planar objects • All detectors and descriptors perform poorly for viewpoint changes > 30° • Hessian with SIFT or shape context perform best

  16. Comparative Evaluations • Object classesMikolajczyk, Liebe & Schiele (2005) • Hessian regions with GLOH perform best • Salient regions work well for object classes • Texture and object classes Zhang, Marszalek, Lazebnik & Schmid (2005) • Laplacian regions with SIFT perform best • Combining multiple detectors and descriptors improves performance • Scale+rotation invariance is sufficient for most datasets

  17. Sparse vs. Dense Features: UIUC texture dataset 25 classes, 40 samples each Lazebnik, Schmid & Ponce (2005)

  18. Sparse vs. Dense Features: UIUC texture dataset Multi-class classification accuracy vs. training set size • A system with intrinsically invariant features can learn from fewer training examples Invariant local features SVM Non-invariant dense patches NN Baseline(global features) SVM NN Zhang, Marszalek, Lazebnik & Schmid (2005)

  19. Sparse vs. Dense Features: CUReT dataset Dana, van Ginneken, Nayar, and Koenderink (1999) 61 classes, 92 samples each, 43 training Non-invariant features (SVM) Non-invariant features (NN) Invariant local features (SVM) Baseline – global features Invariant local features (NN) Relative Strengths

  20. Anticipating Criticism • Existing local features are not ideal for category-level recognition and scene understanding • Designed for wide-baseline matching and specific object recognition • Describe texture and albedo pattern, not shape • Do not explain the whole image • A little invariance goes a long way • It is best to use features with the lowest level of invariance required by a given task • Scale+rotation is sufficient for most datasets Zhang, Marszalek, Lazebnik & Schmid (2005) • Denser sets of local features are more effective • Hessian detector produces the most regions and performs best in several evaluations • Regular grid of fixed-size patches is best for scene category recognitionFei-Fei & Perona (2005)

  21. Future Work • Systematic evaluation of sparse vs. dense features • Combining sparse and dense representations, e.g., keypoints and segments Russell, Efros, Sivic, Freeman & Zisserman (2006) • Learning detectors and descriptors automatically • Developing shape-based features

More Related