1 / 21

Statistical Data Analysis: Lecture 7

Statistical Data Analysis: Lecture 7. 1 Probability, Bayes’ theorem 2 Random variables and probability densities 3 Expectation values, error propagation 4 Catalogue of pdfs 5 The Monte Carlo method 6 Statistical tests: general concepts 7 Test statistics, multivariate methods

cstapleton
Download Presentation

Statistical Data Analysis: Lecture 7

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Statistical Data Analysis: Lecture 7 1 Probability, Bayes’ theorem 2 Random variables and probability densities 3 Expectation values, error propagation 4 Catalogue of pdfs 5 The Monte Carlo method 6 Statistical tests: general concepts 7 Test statistics, multivariate methods 8 Goodness-of-fit tests 9 Parameter estimation, maximum likelihood 10 More maximum likelihood 11 Method of least squares 12 Interval estimation, setting limits 13 Nuisance parameters, systematic uncertainties 14 Examples of Bayesian approach Lectures on Statistical Data Analysis

  2. Nonlinear test statistics The optimal decision boundary may not be a hyperplane, →nonlinear test statistic H1 Multivariate statistical methods are a Big Industry: Neural Networks, Support Vector Machines, Kernel density methods, ... H0 accept Particle Physics can benefit from progress in Machine Learning. Lectures on Statistical Data Analysis

  3. Introduction to neural networks Used in neurobiology, pattern recognition, financial forecasting, ... Here, neural nets are just a type of test statistic. logistic sigmoid Suppose we take t(x) to have the form This is called the single-layer perceptron. s(·) is monotonic →equivalent to linear t(x) Lectures on Statistical Data Analysis

  4. The multi-layer perceptron Generalize from one layer to the multilayer perceptron: The values of the nodes in the intermediate (hidden) layer are and the network output is given by weights (connection strengths) Lectures on Statistical Data Analysis

  5. Neural network discussion Easy to generalize to arbitrary number of layers. Feed-forward net: values of a node depend only on earlier layers, usually only on previous layer (“network architecture”). More nodes → neural net gets closer to optimal t(x), but more parameters need to be determined. Parameters usually determined by minimizing an error function, where t(0) , t(1) are target values, e.g., 0 and 1 for logistic sigmoid. Expectation values replaced by averages of training data (e.g. MC). In general training can be difficult; standard software available. Lectures on Statistical Data Analysis

  6. Neural network example from LEP II Signal: e+e-→ W+W- (often 4 well separated hadron jets) Background: e+e-→ qqgg (4 less well separated hadron jets) ← input variables based on jet structure, event shape, ... none by itself gives much separation. Neural network output does better... (Garrido, Juste and Martinez, ALEPH 96-144) Lectures on Statistical Data Analysis

  7. Some issues with neural networks In the example with WW events, goal was to select these events so as to study properties of the W boson. Needed to avoid using input variables correlated to the properties we eventually wanted to study (not trivial). In principle a single hidden layer with an sufficiently large number of nodes can approximate arbitrarily well the optimal test variable (likelihood ratio). Usually start with relatively small number of nodes and increase until misclassification rate on validation data sample ceases to decrease. Usually MC training data is cheap -- problems with getting stuck in local minima, overtraining, etc., less important than concerns of systematic differences between the training data and Nature, and concerns about the ease of interpretation of the output. Statistical methods for particle physics

  8. Probability Density Estimation (PDE) techniques Construct non-parametric estimators of the pdfs and use these to construct the likelihood ratio (n-dimensional histogram is a brute force example of this.) More clever estimation techniques can get this to work for (somewhat) higher dimension. See e.g. K. Cranmer, Kernel Estimation in High Energy Physics, CPC 136 (2001) 198; hep-ex/0011057; T. Carli and B. Koblitz, A multi-variate discrimination technique based on range-searching, NIM A 501 (2003) 576; hep-ex/0211019 Lectures on Statistical Data Analysis

  9. Kernel-based PDE (KDE, Parzen window) Consider d dimensions, N training events, x1, ..., xN, estimate f (x) with bandwidth (smoothing parameter) kernel Use e.g. Gaussian kernel: Need to sum N terms to evaluate function (slow); faster algorithms only count events in vicinity of x (k-nearest neighbor, range search). Lectures on Statistical Data Analysis

  10. Product of one-dimensional pdfs First rotate to uncorrelated variables, i.e., find matrix A such that for we have Estimate the d-dimensional joint pdf as the product of 1-d pdfs, (here x decorrelated) This does not exploit non-linear features of the joint pdf, but simple and may be a good approximation in practical examples. Lectures on Statistical Data Analysis

  11. Decision trees A training sample of signal and background data is repeatedly split by successive cuts on its input variables. Order in which variables used based on best separation between signal and background. Iterate until stop criterion reached, based e.g. on purity, minimum number of events in a node. Resulting set of cuts is a ‘decision tree’. Tends to be sensitive to fluctuations in training sample. Example by Mini-Boone, B. Roe et al., NIM A 543 (2005) 577 Lectures on Statistical Data Analysis

  12. Boosted decision trees Boosting combines a number classifiers into a stronger one; improves stability with respect to fluctuations in input data. To use with decision trees, increase the weights of misclassified events and reconstruct the tree. Iterate → forest of trees (perhaps > 1000). For the mth tree, Define a score am based on error rate of mth tree. Boosted tree = weighted sum of the trees: Algorithms: AdaBoost (Freund & Schapire), e-boost (Friedman). Lectures on Statistical Data Analysis

  13. Multivariate analysis discussion For all methods, need to check: Sensitivity to statistically unimportant variables (best to drop those that don’t provide discrimination); Level of smoothness in decision boundary (sensitivity to over-training) Given the test variable, next step is e.g., select n events and estimate a cross section of signal: Now need to estimate systematic error... If e.g. training (MC) data ≠ Nature, test variable is not optimal, but not necessarily biased. But our estimates of background b and efficiencies would then be biased if based on MC. (True also for ‘simple cuts’.) Lectures on Statistical Data Analysis

  14. Overtraining If decision boundary is too flexible it will conform too closely to the training points →overtraining. Monitor by applying classifier to independent test sample. training sample independent test sample

  15. N(y) f(y) y y Using classifier output for discovery signal search region background background excess? ycut Normalized to expected number of events Normalized to unity Discovery = number of events found in search region incompatible with background-only hypothesis. p-value of background-only hypothesis can depend crucially distribution f(y|b) in the "search region".

  16. Single top quark production (CDF/D0) Top quark discovered in pairs, but SM predicts single top production. Use many inputs based on jet properties, particle i.d., ... Pair-produced tops are now a background process. signal (blue + green) SUSSP65, St Andrews, 16-29 August 2009 / Statistical Methods 3

  17. Different classifiers for single top Also Naive Bayes and various approximations to likelihood ratio,.... Final combined result is statistically significant (>5s level) but not easy to understand classifier outputs. Statistical methods for particle physics

  18. Statistical methods for particle physics

  19. Comparing multivariate methods (TMVA) Choose the best one! Lectures on Statistical Data Analysis

  20. Statistical methods for particle physics

  21. Wrapping up lecture 7 We looked at statistical tests and related issues: discriminate between event types (hypotheses), determine selection efficiency, sample purity, etc. Some modern (and less modern) methods were mentioned: Fisher discriminants, neural networks, PDE, KDE, decision trees, ... Next we will talk about goodness-of-fit tests: p-value expresses level of agreement between data and hypothesis Lectures on Statistical Data Analysis

More Related