1 / 37

Loss-based Visual Learning with Weak Supervision

Loss-based Visual Learning with Weak Supervision. M. Pawan Kumar. Joint work with Pierre-Yves Baudin , Danny Goodman , Puneet Kumar, Nikos Paragios , Noura Azzabou , Pierre Carlier. SPLENDID. Self-Paced Learning for Exploiting Noisy, Diverse or Incomplete Data . Machine Learning.

thom
Download Presentation

Loss-based Visual Learning with Weak Supervision

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Loss-based Visual Learning with Weak Supervision M. Pawan Kumar Joint work with Pierre-Yves Baudin, Danny Goodman, Puneet Kumar, Nikos Paragios, NouraAzzabou, Pierre Carlier

  2. SPLENDID Self-Paced Learning for Exploiting Noisy, Diverse or Incomplete Data Machine Learning Weak Annotations Noisy Annotations Applications Computer Vision Medical Imaging Nikos Paragios Equipe Galen INRIA Saclay Daphne Koller DAGS Stanford 2 Visits from INRIA to Stanford 2012 ICML 1 Visit from Stanford to INRIA 3 Visits Planned 2013 MICCAI

  3. Medical Image Segmentation MRI Acquisitions of the thigh

  4. Medical Image Segmentation MRI Acquisitions of the thigh Segments correspond to muscle groups

  5. Random Walks Segmentation Probabilistic segmentation algorithm Computationally efficient Interactive segmentation Automated shape prior driven segmentation L. Grady, 2006 L. Grady, 2005; Baudin et al., 2012

  6. Random Walks Segmentation x: Medical acquisition y(i,s): Probability that voxel ‘i’ belongs to segment ‘s’ miny E(x,y) = yTL(x)y + wshape||y-y0||2 Positive semi-definite Laplacian matrix Convex Shape prior on the segmentation Parameter of the RW algorithm Hand-tuned

  7. Random Walks Segmentation Several Laplacians L(x) = Σα wαLα(x) Several shape and appearance priors Σβ wβ||y-yβ||2 Hand-tuning large number of parameters is onerous

  8. Parameter Estimation Learn the best parameters from training data Σα wαyTLα(x)y+ Σβ wβ||y-yβ||2

  9. Parameter Estimation Learn the best parameters from training data wTΨ(x,y) w is the set of all parameters Ψ(x,y) is the joint feature vector of input and output

  10. Outline • Parameter Estimation • Supervised Learning • Hard vs. Soft Segmentation • Mathematical Formulation • Optimization • Experiments • Related and Future Work in SPLENDID

  11. Supervised Learning Dataset of segmented fMRIs Sample xk, voxel i Probabilistic segmentation?? 1, s is ground-truth zk(i,s) = 0, otherwise

  12. Supervised Learning minwΣkξk+ λ||w||2 wTΨ(xk,ŷ) - wTΨ(xk,zk) Δ(ŷ,zk) - ξk ≥ Energy of Segmentation Energy of Ground-truth Δ(ŷ,zk) = Fraction of incorrectly labeled voxels Structured-output Support Vector Machine Taskar et al., 2003; Tsochantardis et al., 2004

  13. Supervised Learning Convex with several efficient algorithms No parameter provides ‘hard’ segmentation We only need a correct ‘soft’ probabilistic segmentation

  14. Outline • Parameter Estimation • Supervised Learning • Hard vs. Soft Segmentation • Mathematical Formulation • Optimization • Experiments • Related and Future Work in SPLENDID

  15. Hard vs. Soft Segmentation Hard segmentation zk Don’t require 0-1 probabilities

  16. Hard vs. Soft Segmentation Soft segmentation yk Compatible with zk Binarizingyk gives zk

  17. Hard vs. Soft Segmentation Soft segmentation yk Compatible with zk yk C(zk) Which yk to use?? yk provided by best parameter Unknown

  18. Outline • Parameter Estimation • Supervised Learning • Hard vs. Soft Segmentation • Mathematical Formulation • Optimization • Experiments • Related and Future Work in SPLENDID

  19. Learning with Hard Segmentation minwΣkξk+ λ||w||2 wTΨ(xk,ŷ) - wTΨ(xk,zk) Δ(ŷ,zk) - ξk ≥

  20. Learning with Soft Segmentation minwΣkξk+ λ||w||2 wTΨ(xk,ŷ) - wTΨ(xk,yk) Δ(ŷ,zk) - ξk ≥

  21. Learning with Soft Segmentation minwΣkξk+ λ||w||2 wTΨ(xk,ŷ) - minyk wTΨ(xk,yk) Δ(ŷ,zk) - ξk ≥ yk C(zk) Latent Support Vector Machine Smola et al., 2005; Felzenszwalb et al., 2008; Yu et al., 2009

  22. Outline • Parameter Estimation • Optimization • Experiments • Related and Future Work in SPLENDID

  23. Latent SVM minwΣkξk + λ||w||2 wTΨ(xk,ŷ) – minyk wTΨ(xk,yk) ≥Δ(ŷ,zk)– ξk yk C(zk) Difference-of-convex problem Concave-Convex Procedure (CCCP)

  24. CCCP Estimate soft segmentation yk* = minyk wTΨ(xk,yk) s.t.yk C(zk) Efficient optimization using dual decomposition Update parameters minwΣkξk + λ||w||2 wTΨ(xk,ŷ) – wTΨ(xk,yk*) ≥Δ(ŷ,zk)– ξk Convex optimization Repeat until convergence

  25. Outline • Parameter Estimation • Optimization • Experiments • Related and Future Work in SPLENDID

  26. Dataset 30 MRI volumes of thigh Dimensions: 224 x 224 x 100 4 muscle groups + background 80% for training, 20% for testing

  27. Parameters 4 Laplacians 2 shape priors 1 appearance prior Baudin et al., 2012 Grady, 2005

  28. Baselines Hand-tuned parameters Structured-output SVM Hard segmentation Soft segmentation based on signed distance transform

  29. Results Small but statistically significant improvement

  30. Outline • Parameter Estimation • Optimization • Experiments • Related and Future Work in SPLENDID

  31. Loss-based Learning x: Input a: Annotation

  32. Loss-based Learning x: Input a: Annotation h: Hidden information h h = “soft-segmentation” a = “jumping”

  33. Loss-based Learning Annotation Mismatch minΣkΔ(correct ak, predicted ak) x: Input a: Annotation h: Hidden information h h = “soft-segmentation” a = “jumping”

  34. Loss-based Learning Annotation Mismatch minΣkΔ(correct ak, predicted ak) Small improvement using small medical dataset

  35. Loss-based Learning Annotation Mismatch minΣkΔ(correct ak, predicted ak) Large improvement using large vision dataset

  36. Loss-based Learning Output Mismatch Modeled using a distribution minΣkΔ(correct {ak,hk}, predicted {ak,hk}) Inexpensive annotation No experts required Richer models can be learnt Kumar, Packer and Koller, ICML 2012

  37. Questions?

More Related