1 / 59

Shared features and Joint Boosting

Shared features and Joint Boosting. Sharing visual features for multiclass and multiview object detection A. Torralba, K. P. Murphy and W. T. Freeman PAMI. vol. 29, no. 5, pp. 854-869, May, 2007. Yuandong Tian. Outline. Motivation to choose this paper Motivation of this paper

linnea
Download Presentation

Shared features and Joint Boosting

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Shared features and Joint Boosting Sharing visual features for multiclass and multiview object detectionA. Torralba, K. P. Murphy and W. T. Freeman PAMI. vol. 29, no. 5, pp. 854-869, May, 2007. Yuandong Tian

  2. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  3. Motivation to choose this paper Axiom: Computer vision is hard. Assumption: (smart-stationary) Equally smart people are equally distributed over time. Conjure: If computer vision cannot be solved in 30 years, it won’t be solved forever!

  4. Wrong! Because we are standing on the Shoulder of Giants.

  5. Where are the Giants? • More computing resources? • Lots of data? • Advancement of new algorithm? • Machine Learning? What I believe

  6. Cruel Reality Why ML seems not to help much in CV (at least for now)? My answer: CV and ML are weakly coupled

  7. A typical question in CV Q: Why do we use feature A instead of feature B? A1: Feature A gives better performance. A2: Feature A has some fancy properties. A3: The following step requires the feature to have a certain property that only A has. A strongly-coupled answer

  8. Typical CV pipeline Preprocessing Steps (“Computer Vision”) Feature/Similarity ML black box Have some domain-specific structures Design for generic structures

  9. Contribution of this paper • Tune the ML algorithm in a CV context • A good attempt to break the black box and integrate them together

  10. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  11. This paper • Object Recognition Problem • Many object category. • Few images per category • Solution—Feature sharing • Find common features that distinguish a subset of classes against the rest.

  12. Feature sharing Concept of Feature Sharing

  13. Typical behavior of feature sharing Template-like features 100% accuracy for a single object But too specific. Wavelet-like features, weaker discriminative power but shared in many classes.

  14. Result of feature sharing

  15. Why feature sharing? • ML: Regularization—avoid over-fitting • Essentially more positive samples • Reuse the data • CV: Utilize the intrinsic structure of object category • Use domain-specific prior to bias the machine learning algorithm

  16. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  17. Basic idea in Boosting • Concept: Binary classification • samples, labels(+1 or -1) • Goal: Find a function (classifier) H which • maps positive samples to the positive value • Optimization: Minimize the exponential loss w.r.t the classifier H

  18. Basic idea in boosting(2) • Boosting: Assume H is additive • Each is a “weak” learner (classifier). • Almost random but uniformly better than random • Example: • Single feature classifier: make decision only on a single dimension

  19. How weak learner looks like Key point: The addition of weak classifiers gives a strong classifier!

  20. Basic idea in boosting(3) • How to minimize? • Greedy Approach • Fix H, add one h in each iteration • Weighting samples • After each iteration, wrongly classified samples (difficult samples) get higher weights

  21. Technical parts • Greedy -> Second-order Taylor Expansion in each iteration weights The weak learner to be optimized in this iteration labels Solved by Least Square

  22. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  23. Joint Boost—Multiclass • We can minimize a similar function using one-vs-all strategy • This doesn’t work very well, since it is separable in c. • Put constraints. -> shared features!

  24. Joint Boost (2) • In each iteration, choose • One common feature • A subset of classes that use this feature • So that the objective decreases most

  25. Sharing Diagram #Iteration #class I II III IV V Features 1 3 4 5 2 1 4 6 2 7 3

  26. Key insight • Each class may have its own favorite feature • a common feature may not be any of them, however it simultaneously decreases errors of many classes.

  27. Joint Boost – Illustration

  28. Computational issue • Choose the best subset is prohibitive • Use greedy approach • Choose one class and one feature so that the objective decreases the most • Iteratively add more classes until the objective increases again • Note the common feature may change • From O(2^C) to O(C^2)

  29. #features = O(log #class) (greedy)

  30. 0.95 ROC 29 objects, average over 20 training sets

  31. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  32. Feature they used in the paper • Dictionary • 2000 random sampled patches • Of size from 4x4 to 14x14 • no clustering • Each patch is associated with a spatial mask

  33. The candidate features position template Dictionary of 2000 candidate patches and position masks, randomly sampled from the training images

  34. Features • Building feature vectors • Normalized correlation with each patch to get response • Raise the response to some power • Large value gets even larger and dominate the response (max operation) • Use spatial mask to align the response to the object center (voting) • Extract response vector at object center

  35. Results • Multiclass object recognition • Dataset: LabelMe • 21 objects, 50 samples per object • 500 rounds • Multiview car recognition • Train on LabelMe, test on PASCAL • 12 views, 50 samples per view • 300 rounds

  36. 70 rounds, 20 training per class, 21 objects

  37. 12 views 50 samples per class 300 features

  38. Outline • Motivation to choose this paper • Motivation of this paper • Basic ideas in boosting • Joint Boost • Feature used in this paper • My results in face recognition

  39. Simple Experiment • Main point of this paper • They claimed shared feature helps in the situation of • many categories, only a few samples in each category. • Test it! • Dataset: face recognition • “Face in the wild” dataset. • Many famous figures

  40. Experiment configuration • Use Gist-like feature but • Only Gabor response • Use finer grid to gather histogram • Face is aligned in the dataset. • Feature statistics • 8 orientation, 2 scale, 8x8 grid • 1024 dimension

More Related