1 / 31

Bayesian Methods in Particle Physics: From Small-N to Large

Bayesian Methods in Particle Physics: From Small-N to Large. Harrison B. Prosper Florida State University SCMA IV 12-15 June, 2006. Outline. Measuring Zero Bayesian Fit Finding Needles in Haystacks Summary. Measuring Zero. Measuring Zero – 1. In the mid-1980s, an experiment at the

jtucker
Download Presentation

Bayesian Methods in Particle Physics: From Small-N to Large

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Bayesian Methods in Particle Physics: From Small-N to Large Harrison B. Prosper Florida State University SCMA IV 12-15 June, 2006 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  2. Outline • Measuring Zero • Bayesian Fit • Finding Needles in Haystacks • Summary From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  3. Measuring Zero

  4. Measuring Zero – 1 In the mid-1980s, an experiment at the Institut Laue Langevin (Grenoble, France) searched for evidence of neutron antineutron oscillations, a characteristic prediction of certain Grand Unified Theories. From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  5. Neutron gas CRISP Experiment Institut Laue Langevin Field-off: -> N Field-on: -> B Magnetic shield From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  6. Measuring Zero – 2 Count number of signal + background events N. Suppress putative signal and count background events B, independently. Results: N = 3 B = 7 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  7. Measuring Zero – 3 Classic 2-Parameter Counting Experiment N ~ Poisson(s+b) B ~ Poisson(b) Infer a statement of form: Pr[s < u(N,B)] ≥ 0.9 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  8. Measuring Zero – 4 In 1984, no exact solution existed in the particle physics literature! Moreover, calculating exact confidence intervals is, according to Kendal and Stuart, “a matter of very considerable difficulty” From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  9. Measuring Zero – 5 Exact in what way? Over some ensemble of statements of the form 0 < s < u(N,B) at least 90% of them should be true whatever the true values of s and b. Neyman (1937) From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  10. Measuring Zero - 6 Tried a Bayesian approach: f(s, b|N) = f(N|s, b) p(s, b) / f(N) = f(N|s, b) p(b|s) p(s) / f(N) Step 1. Compute the marginal likelihood f(N|s) = ∫f(N|s, b) p(b|s) db Step 2. f(s|N) = f(N|s) p(s) / ∫f(N|s) p(s) ds From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  11. Discrepancy “Distance” between models But is there a signal? 1. Hypothesis testing (J. Neyman) H0: s = 0 H1 : s > 0 2. p-value (R.A. Fisher) H0: s = 0 3. Decision theory (J.M. Bernardo, 1999) From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  12. Bayesian Fit

  13. Bayesian Fit Problem: Given counts Data: N = N1, N2,..,NM Signal model: A = A1, A2,..,AM Background model: B = B1, B2,..,BM where M is number of bins (or pixels) find the admixture of A and B that best matches the observations N. From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  14. Problem (DØ, 2005) Observations = Background + Signal model model (M) From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  15. Bayesian Fit - Details Assume model of the form Marginalize over a and b From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  16. Bayesian Fit – Pr(Model) Moreover,… One can compute f(N|pa, pb) for different signal models M, in particular, for models M that differ by the value of a single parameter. Then compute the probability of model M Pr(M|N) = ∫dpa ∫dpb f(N|pa, pb, M) p(pa,pb|M) p(M) / p(N) From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  17. 0.3 0.2 P(M|N) 0.1 0 130 140 150 160 170 180 190 200 210 220 230 Top quark mass hypothesis (GeV) Bayesian Fit – Results (DØ, 1997) mass = 173.5 ± 4.5 GeV signal = 33 ± 8 background = 50.8 ± 8.3 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  18. Finding Needles in Haystacks

  19. The Needles single top quark events 0.88 pb 1.98 pb From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  20. The Haystacks W boson events 2700 pb signal : noise = 1 : 1000 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  21. The Needles and the Haystacks From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  22. Finding Needles - 1 The optimal solution is to compute p(S|x) = p(x|S) p(S) / [p(x|S) p(S) + p(x|B) p(B)] Every signal/noise discrimination method is ultimately an algorithm to approximate p(S|x), or a function thereof. From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  23. Finding Needles - 2 Problem: Given D = x (= x1,…xN), y (= y1,…yN) of N labeled events. x are the data, y are the labels. Find A function f(x, w), with parameters w, that approximates p(S|x): p(w|x, y) = p(x, y|w) p(w) / p(x, y) = p(y|x, w) p(x|w) p(w) / p(y|x) p(x) = p(y|x, w) p(w) / p(y|x) assuming p(x|w) = p(x) From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  24. Finding Needles - 3 Likelihood for classification: p(y|x, w) = Pi f(xi, w)y [1 – f(xi, w)]1-y where y = 0 for background events y = 1 for signal events If f(x, w) flexible enough, then maximizing p(y|x, w) with respect to w yields f = p(S|x), asymptotically. From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  25. Finding Needles - 4 However, in a Bayesian calculation it is more natural to average with respect to the posterior density f(x|D) = ∫ f(x, w) p(w|D) dw Questions: 1. Do suitably flexible functions f(x, w) exist? 2. Is there a feasible way to do the integral? From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  26. u, a x1 v, b f(x,w) x2 Answer 1: Yes! A neural network is an example of a Kolmogorov function, that is, a function capable of approximating arbitrary mappings f:Rn -> R The parameters w = (u, a, v, b) are called weights From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  27. Answer 2: Yes! Computational Method Generate a Markov Chain (MC) of K points {w}, whose stationary density is p(w|D), and average over the stationary part of the chain. Map problem to that of a “particle” moving in a spatially-varying “potential” and use methods of statistical mechanics to generate states (p, w) with probability ~ exp(-H), where H is the “Hamiltonian” H = p2 + log p(w|D), with “momentum” p. From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  28. Hybrid Markov Chain Monte Carlo Computational Method… For a fixed H traverse space (p, w) using Hamilton’s equations, which guarantees that all points consistent with H will be visited with equal probability. To allow exploration of states with differing values of H one introduces, periodically, random changes to the momentum p. Software Flexible Bayesian Modeling by Radford Neal http://www.cs.utoronto.ca/~radford/fbm.software.html From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  29. Example - Finding SUSY! Transverse momentum spectra Signal: black curve Signal:Noise 1:25,000 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  30. Example - Finding SUSY! Distribution of f(x|D) beyond 0.9 Assuming L = 10 fb-1 Cut S B S/√B 0.99 1x103 2x104 7.0 Signal:Noise 1:20 From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

  31. Summary • Bayesian methods have been at the heart of several important results in particle physics. • However, there is considerable room for expanding their domain of application. • A couple of current issues: • Is there a signal? Is the Bernardo approach useful in particle physics? • Fitting: Is there a practical (Bayesian?) method to test whether or not an N-dimensional function fits an N-dimensional swarm of points? From Small-N to Large Harrison B. Prosper SCMA IV, June 2006

More Related