1 / 59

Contextual models for object detection using boosted random fields

Contextual models for object detection using boosted random fields. by Antonio Torralba, Kevin P. Murphy and William T. Freeman. Quick Introduction. What is this? Now can you tell?. Belief Propagation (BP). Network (Pairwise Markov Random Fields) observed nodes ( y i ).

Download Presentation

Contextual models for object detection using boosted random fields

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Contextual models for object detection using boosted random fields by Antonio Torralba, Kevin P. Murphy and William T. Freeman

  2. Quick Introduction • What is this? • Now can you tell?

  3. Belief Propagation (BP) • Network (Pairwise Markov Random Fields) • observed nodes (yi)

  4. Belief Propagation (BP) • Network (Pairwise Markov Random Fields) • observed nodes (yi) • hidden nodes (xi)

  5. Belief Propagation (BP) • Network (Pairwise Markov Random Fields) • observed nodes (yi) • hidden nodes (xi) Statistical dependency, called local evidence: Shord-hand

  6. Belief Propagation (BP) Statistical dependency: Local evidence Shord-hand Statistical dependency: Compatibility function

  7. Belief Propagation (BP) • Joint probability

  8. Belief Propagation (BP) • Joint probability x x1 x2 y1 y2 xi …. yi x5 x3 x1 x4 x12 xj

  9. Belief Propagation (BP) • Joint probability x x1 x2 y1 y2 xi …. yi x5 x3 x1 x4 x12 xj

  10. Belief Propagation (BP) • The belief b at a node i is represented by • the local evidence of the node • all the messages coming in from neighbors Ni xi xj ∏ yi

  11. Belief Propagation (BP) • The belief b at a node i is represented by • the local evidence of the node • all the messages coming in from neighbors Ni xi xj ∏ yi

  12. Belief Propagation (BP) • Messages m between hidden nodes • How likely node j thinks it is that node i will be in the corresponding state. mji(xi) xi xj

  13. Belief Propagation (BP) mji(xi) xi xj xi xj xk

  14. Conditional Random Field • Distribution of the form:

  15. Conditional Random Field • Distribution of the form:

  16. Boosted Random Field • Basic Idea: • Use BP to estimate P(x|y) • Use boosting to maximize Log Likelihood of each node wrt to

  17. Algorithm: BP • Minimize negative log likelihood of training data (yi). Label Loss function to minimize:

  18. Algorithm: BP • Minimize negative log likelihood of training data (yi). Label Loss function to minimize:

  19. Algorithm: BP • Minimize negative log likelihood of training data (yi). Label Loss function to minimize:

  20. Algorithm: BP Ni xi xj ∏ yi

  21. Algorithm: BP Ni xi xj ∏ yi

  22. Algorithm: BP Ni xi xj ∏

  23. Algorithm: BP xi xj

  24. Algorithm: BP xi yi F: a function of the input data

  25. Algorithm: BP yi xj xi with

  26. Algorithm: BP yi xj xi with

  27. Function F yi • Boosting! • f is the weak learner:weighted decision stumps. xi

  28. Minimization of loss L

  29. Minimization of loss L

  30. Minimization of loss L where

  31. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  32. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  33. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  34. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  35. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  36. Local Evidence: algorithm yi • For t=1..T • Iterate Nboost times • find the best basis function h • update local evidence with • update the beliefs • update the weights • Iterate NBP times • update messages • update the beliefs xj xi

  37. Function G • By assuming that the graph is densely connected we can make the approximation: • Now G is a non-linear additive function of the beliefs:

  38. Function G • Instead of learning the function can be learnt with an additive model: weighted regression stumps

  39. Function G • The weak learner is chosen by minimizing the loss:

  40. The Boosted Random Field Algorithm • For t=1..T • find the best basis function h for f • find the best basis function for • compute local evidence • compute compatibilities • update the beliefs • update weights xj xi yi

  41. The Boosted Random Field Algorithm • For t=1..T • find the best basis function h forf • find the best basis function for • compute local evidence • compute compatibilities • update the beliefs • update weights b1 xi b2 … bj

  42. Final classifier • For t=1..T • update local evidences F • update compatibilities G • compute current beliefs • Output classification:

  43. Multiclass Detection • U: Dictionary of ~2000 images patches • V: Same number of image masks

  44. Multiclass Detection • U: Dictionary of ~2000 images patches • V: Same number of image masks • At each round t, for each class c for each dictionary entry d there is a weak learner:

  45. Function f • To take into account different sizes, we first downsample the image and then upsample and OR the scales: • which is our function for computing the local evidence.

  46. Function g • The compatibily function has a similar form:

  47. Function g • The compatibily function has a similar form: • W represent a kernel with all the messages directed to node x,y,c

  48. Kernels W • Example of incoming messages:

  49. Function G • The overall incoming messages function is given by:

  50. Learning… • Labeled dataset of office and street scenes, with each ~100 images • In the first 5 round updated only the local evidence • After the 5th iteration update also the compatibility functions • At each round update onlyF and G of the single object class that reduces the most the multiclass cost.

More Related