1 / 11

Belief Networks in Computer Vision Applications

Alex Yakushev CMPS 290C final project Winter 2006. Belief Networks in Computer Vision Applications. Kalman Filters. Predicted position = (Previous Position)* (Motion model ) + Noise. True position = observed position + Noise. Extrapolate object position by taking

moana
Download Presentation

Belief Networks in Computer Vision Applications

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Alex Yakushev CMPS 290C final project Winter 2006 Belief Networks in Computer Vision Applications

  2. Kalman Filters Predicted position = (Previous Position)* (Motion model ) + Noise True position = observed position + Noise Extrapolate object position by taking a weighted average of the two. Only depends on current observation and previous state

  3. Kalman Filter as a Belief Network Motion model Prediction covariance Model Noise xk xk+1 zk zk+1 State model Observation noise

  4. Kalman Filters Becomes a simple update rule

  5. Something completely different • Are old fashioned Bayesian networks of any use? • What if the object we are trying to track does not have a constant velocity or acceleration? • Sherrah and Gong have designed a BN to track hand and face movements.

  6. Something completely different (Sherrah, Gong)

  7. Results (Sherrah, Gong) • For their setup they had 70% of frames classified correctly without any contextual knowledge • This improved to 78% with a Kalman filter • 87% with their method... ...but Assumptions are unreasonable: 1. the subject is oriented roughly towards the camera for most of the time 2. the subject is wearing long sleeves 3. reasonably good color segmentation of the head and hands is possible, and 4. the head and hands are the larges moving skin color clusters in the image

  8. “Dynamic Bayesian Network” State variables st belong to a set of S discrete symbols. (Pavlovic, Rehg, Cham, Murphy)

  9. Dynamic Bayesian Network • Two hidden variables (xt, st) • If x0 is Gaussian then x1 is a mixture of S Gaussians, x2 is a mixture of S2 Gaussians... • No “simple” update rule • Exact inference is expensive (Pavlovic, Rehg, Cham, Murphy)

  10. Conclusions • As usual, increase in expressive power costs computation time • People create custom networks, and either give a solution for the specific structure, or use approximate inference • No silver bullet, good results possible with both simple and complex network structures • Using a simpler structure (Markov Network) may work for many applications

  11. References • Jamie Sherrah, Shaogang Gong, Tracking Discontinuous Motion Using Bayesian Inference, Lecture Notes in Computer Science, Volume 1843,Jan 2000, Pages 150 – 166 • V. Pavlović, J. M. Rehg, T. J. Cham, and K. P. Murphy, A dynamic Bayesian network approach to figure tracking using learned dynamic models, in International Conference on Computer Vision, Corfu, Greece, September 1999. • Murphy, K., Dynamic Bayesian Networks: Representation, Inference and Learning. PhD thesis, UC Berkeley, Computer Science Division (2002) • Hai Tao, Object Tracking and Kalman Filtering, CMPE264 Lecture notes, http://www.soe.ucsc.edu/classes/cmpe264/Winter06/Lec15.pdf

More Related