230 likes | 368 Views
Continuous simulation of Beyond-Standard-Model processes with multiple parameters. Jiahang Zhong (University of Oxford * ) Shih-Chang Lee (Academia Sinica) ACAT 2011, 5-9 September, London. * Was in Academia Sinica and Nanjing University. Motivation.
E N D
Continuous simulation of Beyond-Standard-Model processes with multiple parameters Jiahang Zhong (University of Oxford*) Shih-Chang Lee (Academia Sinica)ACAT 2011, 5-9 September, London * Was in Academia Sinica and Nanjing University
Motivation • Many Beyond Standard Model (BSM) processes are defined by more than one free parameters • Masses of hypothetical particles • Coupling constants • … • Grid Scan • Scan the parameter spacewith grid points • Simulate a sample of events on each point Var2 Var1 ACAT 2011, 5-9 September, London
Motivation • The difficulties of the grid-scan approach: • Curse of dimensionality • Npoints~Nd • Hard to go beyond 2D • Costly for finer granularity Var2 Var1 ACAT 2011, 5-9 September, London
Motivation • The difficulties of the grid-scan approach: • Curse of dimensionality • Npoints~Nd • Hard to go beyond 2D • Costly for finer granularity • Large statistics required • Samples at different points are treated independently • Considerable statistics neededwithin each sample Var2 Pass Fail ~10k evts Var1 ACAT 2011, 5-9 September, London
Motivation • The difficulties of the grid-scan approach: • Curse of dimensionality • Npoints~Nd • Hard to go beyond 2D • Costly for finer granularity • Large statistics required • Samples at different points are treated independently • Considerable statistics neededwithin each sample • Discreteness • Considerable space between points • Smoothing/interpolation needed • Consequent systematic uncertainties Var2 ~TeV ~100GeV Var1 ACAT 2011, 5-9 September, London
Motivation • Grid-scan: • Curse of dimensionality • Large statistics needed • Discreteness • The aim of Continuous MC • Competent for multivariate parameter space • Less events to be simulated • Continuous estimation of signal yield over the parameter space ACAT 2011, 5-9 September, London
Motivation • The usage of multivariate BSM simulation is to estimate signal yields over the parameter space. • Yields: N(x)=L* σ(x) * ε(x) • L: Luminosity.Irrelevant to x (the free parameters) • σ: Cross section, branching ratio. Easy to calculate with event generators • ε: Detector acceptance, offline efficiencyNeed large amount and expensive detector simulation • Therefore, our method is focused on easing the estimation of ε ACAT 2011, 5-9 September, London
The procedure Event generation Grid Scan Continuous MC Var2 Var2 O(10d) space points O(100k) space points O(10k) events/point O(1) events/point Var1 Var1 ACAT 2011, 5-9 September, London
The procedure • Bayesian Neural Network (BNN) is used to fit the efficiency ε • Desirable features of NN fitting • Non-parametric modeling • Smooth over the parameter space • Unbinned fitting • Suffer less from dimensionality • Correlation between the variables Unbinned fitting vs. BinnedHistogram ACAT 2011, 5-9 September, London
The procedure • Bayesian implementations of NN further provide • Automatic complexity control of NN topology during training • Probabilistic output • Uncertainty estimation of the output • Uncertainty of the output estimated based on the p.d.f. of the NN parameters. • Statistical fluctuation of the training sample • Choice of NN topology • Impact of fitting goodness at certain space point x ACAT 2011, 5-9 September, London
Demo • Production of right-handed W boson and Majorana neutrino • Di-lepton final state • 2 leptons (e,μ) • pT>20GeV, |eta|<2.5 • cone20/pT<0.1 • Two free parameters • WR mass [500GeV,1500GeV] • NR mass [0, M(WR)] • Affect both the cross-section and efficiency
Demo • Continuous Simulation • Generated 100k events, each with random { M(WR), M(NR) } • Put each event through the selection criteria, and assign target value 1/0 if it pass/fail • Feed all events to a BNN, with { M(WR), M(NR) } as the input variables • Use the trained BNN as a function to provide ε±σε • Reference grid-scan • A grid with 100GeV step in M(WR) and 50GeV step in M(NR) (171 samples in total) • Sufficient statistics in each sample to achieve precise reference values ACAT 2011, 5-9 September, London
Demo The BNN fitted efficiency Reference from grid-scan ACAT 2011, 5-9 September, London
Demo The difference between fitted values and reference values ACAT 2011, 5-9 September, London
Demo Uncertainty estimated by the BNN. ACAT 2011, 5-9 September, London
Demo The real deviations vs. estimated uncertainties (Nσ) ACAT 2011, 5-9 September, London
Summary • New approach to simulate multivariate BSM processes • More space points, less events • Use BNN fitting to obtain smooth yield estimation • Performance tested by • The deviation between BNN and reference values • This deviation vs. BNN uncertainty • Limitation: the assumption of smooth distribution • Not sensitive to local abrupt changes • Less performance across physics boundary. ACAT 2011, 5-9 September, London
完Thank you! ACAT 2011, 5-9 September, London
Backup More detailed documentation of this methodhttp://arxiv.org/abs/1107.0166 The Bayesian Neural Network in TMVA/ROOThttp://www.sciencedirect.com/science/article/pii/S0010465511002682 Links ACAT 2011, 5-9 September, London
Backup How does BNN fitting work • A black-box of discriminatorA white-box of non-parametric fitting tool • A multivariate function y(x) • Generic function approximator (analog to polynomial in 1D) • Training unbinned MLE fitting y: NN output, a probability, [0,1]t: Target value, 1=pass, 0=fail ACAT 2011, 5-9 September, London
Backup: Bayesian implementation of NN(I) • Probability fitting • Unbinned fitting • Full usage of every event • Extrapolation/Interpolation • Fit y as probability function • Bernoulli likelihood Histogram BNN ACAT 2011, 5-9 September, London
Backup: Bayesian implementation of NN (II) • Uncertainty estimation • Training: • Most probable value wMP • P(w|D)Probability of other w • Prediction • Probability • Uncertainty of y • Avoid excessive extrapolation (non-trivial for multivariate analysis) Histogram BNN ACAT 2011, 5-9 September, London
Backup: Bayesian implementation of NN (III) Early stop • Regulator • Overtraining is possible due to excessive complexity of NN • Early stop • Use half input sample as monitor • Manual decision of when to stop excessive fitting • Regulator • Prior knowledge that “simpler” model is preferred • Adaptive during training • Save the monitor sample!!! Regulator ACAT 2011, 5-9 September, London