1 / 38

Is there evidence for a peak in this data?

. Is there evidence for a peak in this data?. Is there evidence for a peak in this data?. “Observation of an Exotic S=+1 Baryon in Exclusive Photoproduction from the Deuteron” S. Stepanyan et al, CLAS Collab, Phys.Rev.Lett. 91 (2003) 252001

diella
Download Presentation

Is there evidence for a peak in this data?

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. .

  2. Is there evidence for a peak in this data?

  3. Is there evidence for a peak in this data? “Observation of an Exotic S=+1 Baryon in Exclusive Photoproduction from the Deuteron” S. Stepanyan et al, CLAS Collab, Phys.Rev.Lett. 91 (2003) 252001 “The statistical significance of the peak is 5.2 ± 0.6 σ”

  4. Is there evidence for a peak in this data? “Observation of an Exotic S=+1 Baryon in Exclusive Photoproduction from the Deuteron” S. Stepanyan et al, CLAS Collab, Phys.Rev.Lett. 91 (2003) 252001 “The statistical significance of the peak is 5.2 ± 0.6 σ” “A Bayesian analysis of pentaquark signals from CLAS data” D. G. Ireland et al, CLAS Collab, Phys. Rev. Lett. 100, 052001 (2008) “The ln(RE) value for g2a (-0.408) indicates weak evidencein favour of the data model without a peak in thespectrum.” Comment on “Bayesian Analysis of Pentaquark Signals from CLAS Data” Bob Cousins, http://arxiv.org/abs/0807.1330

  5. PARADOX Histogram with 100 bins Fit 1 parameter Smin: χ2 with NDF = 99 (Expected χ2 = 99 ± 14) For our data, Smin(p0) = 90 Is p1 acceptable if S(p1) = 115? • YES. Very acceptable χ2 probability • NO. σp from S(p0 +σp) = Smin +1 = 91 But S(p1) – S(p0) = 25 So p1 is 5σ away from best value

  6. Perspective on Cosmostats from Particle Physics Louis Lyons Particle Physics, Oxford and IC (CMS experiment, CERN CDF experiment, Fermilab) l.lyons@physics.ox.ac.uk Cosmostats Ascona July 2009

  7. Topics • Similarities, differences, and what you do particularly well • What Particle Physicists have learnt • Discovery issues • Wish list and Conclusions Thanks to Stefen Lauritzen, CDF + CMS St Cttees (esp Bob Cousins, Luc Demortier, Joel Heinrich, Tom Junk), Subir Sarkar, Roberto Trotta, many speakers here.

  8. Similarities between Cosmology and Particle Physics Differences What Astro/Cosmo do well

  9. Similarities • Large data sets {LHC: event = Mbyte; total = 20 Pbytes} • Experimental resolution • Acceptance • Systematics • Separating signal from background (Glen Cowan, Ofer Lahav) • Parameter estimation (Fraction of Dark Energy) • Testing models (Is there Dark Energy?) • Search for signals: Setting limits or Discovery • {SM; Concordance} v {SUSY; Dark energy eos} • {Cosmostats, SCMA} and PHYSTAT

  10. Differences • Background • Bayes or Frequentism? • Specific Astrophysics issues Time dependence Spatial structures Correlations Non-parametric methods Visualisation Cosmic variance • Blind analyses `

  11. Bayes versus Frequentism Cosmologists use Bayes, unless forced to use Frequentism, while Particle Physicists ……… Goldstein at Durham PHYSTAT: Particle Physicists are last living dinosaurs still doing frequentist analysis Preprint: “As with any frequentist result, the use of priors means that the exact figures should be taken with a pinch of salt.” Talk (not here): We used a non-informative prior for the 12 parameters, and 7 pieces of data, to determine their posteriors Argument against constant priors: Fit y = a + b*x, with constant priors for a and b. Favours vertical line. Many talks here: “Marginalise over models”

  12. BLIND ANALYSES Why blind analysis? Selections, corrections, method Methods of blinding Add random number to result * Study procedure with simulation only Look at only first fraction of data Keep the signal box closed Keep MC parameters hidden Keep fraction visible for each bin hidden After analysis is unblinded, …….. * Luis Alvarez suggestion re “discovery” of free quarks • David Friedenberg’s source detection

  13. What Astrophysicists/Cosmologists do well • Glorious pictures  • Scientists + Statisticians working together • Sharing data • Making data publicly available • Dealing with large data sets • Visualisation • Funding for Astrostatistics • Statistical software

  14. Whirlpool Galaxy Width of Z0 3 light neutrinos

  15. What Particle Physicists now know • (ln L) = 0.5 rule (Coverage) • Unbinned Lmax and Goodness of fit • Prob (data | hypothesis) ≠ Prob (hypothesis | data) • Comparing 2 hypotheses • Δ(χ2) ≠χ2 (Wilks’ Theorem) • Frequentism for bounded parameters: Feldman and Cousins • Use correct L (Punzi effect)

  16. Statistical issues forDiscovery PHYSTAT-LHC Upper limits (e.g. WIMP flux) H0, or H0 versus H1 p-values: What p is not Combining p’s Methods for selecting between H0 and H1 Why 5σ? Look-elsewhere effect

  17. Ilya Narsky, FNAL CLW 2000

  18. Comparing Hypotheses: H0 versus H1 “How do you tell who is a Statistician, who is a Cosmologist and who is a Particle Physicist?” J.B: “Very hard for Bayesian, but worse for Frequentists” • Checking H0 (G of F), or H0 versus H1 (HT)? • p-values p-values or likelihood-ratios? Misunderstanding p-values Combining p-values • Look elsewhere effect: Why 5σ? How wide? Why not for exclusion? • Methods: Standard Frequentist Frequentist with ‘No choice’: p0 versus p1; Conditional Frequentism CLs Δχ2 Bayesian methods:Posterior odds, Bayes’ factor, BIC, AIC,… • Treatment of systematics for p-values Profile likelihood Frequentist methods Bayesian marginalisation Mixed

  19. p-values for non-Gaussian distributions e.g. Poisson counting experiment, bgd = b P(n) = e-b* bn/n! {P = probability, not prob density} b=2.9 P 0 n 10 For n=7, p = Prob( at least 7 events) = P(7) + P(8) + P(9) +…….. = 0.03

  20. Poisson p-values n = integer, so p has discrete values So p distribution cannot be uniform Replace Prob{p≤p0} = p0, for continuous p by Prob{p≤p0} ≤ p0, for discrete p (equality for possible p0) p-values often converted into equivalent Gaussian σ e.g. 3*10-7 is “5σ” (one-sided Gaussian tail)

  21. p-value is not …….. Does NOT measure Prob(H0 is true) i.e. It is NOT P(H0|data) It is P(data|H0) N.B. P(H0|data) ≠ P(data|H0) P(theory|data) ≠ P(data|theory) “Of all results with p ≤ 5%, half will turn out to be wrong” Supposed to be funny, but is perfectly OK e.g. 1000 tests of energy conservation ~50 should have p ≤ 5%, and so reject H0 = energy conservation Of these 50 results, all are likely to be “wrong” Are p-values so misunderstood that it is unwise to use them?

  22. p-values or L? • If statistic for p is L-ratio, can regard p-values as: • p-value method with L-ratio just as statistic; or • ii) L-ratio method, with pdf’s just as calibration

  23. Combining different p-values Several results quote p-values for same effect: p1, p2, p3….. e.g. 0.9, 0.001, 0.3 …….. What is combined significance? Not just p1*p2*p3….. If 10 expts each have p ~ 0.5, product ~ 0.001 and is clearly NOT correct combined p S = z * (-ln z)j/j! , z = p1p2p3……. (e.g. For 2 measurements, S = z * (1 - lnz) ≥ z ) Slight problem: Formula is not associative Combining {{p1 and p2}, and then p3} gives different answer from {{p3 and p2}, and then p1} , or all together Due to different options for “more extreme than x1, x2, x3”. (χ2, ndf) = (100.0, 100) and (15.0, 1)

  24. Combining different p-values Conventional: Are set of p-values consistent with H0? p2 SLEUTH: How significant is smallest p? 1-S = (1-psmallest)n p1 p1 = 0.01 p1 = 10-4 p2 = 0.01 p2 = 1 p2 = 10-4 p2 = 1 Combined S Conventional 1.0 10-3 5.6 10-2 1.9 10-7 1.0 10-3 SLEUTH 2.0 10-2 2.0 10-2 2.0 10-4 2.0 10-4

  25. Why 5σ? • Past experience with 3σ, 4σsignals (but……) • Look elsewhere effect: Different bins (and binning) of this histogram Different cuts to produce data Different distributions Collaboration did/could look at Defined in SLEUTH • Bayesian priors: P(H0|data)P(data|H0) * P(H0) P(H1|data)P(data|H1) * P(H1) Bayes posteriorsLikelihoodsPriors Prior for {H0 = S.M.} >>> Prior for {H1 = New Physics}

  26. Why 5σ? BEWARE of tails, especially for nuisance parameters Same criterion for all searches? {Cousins} Single top production Higgs Highly speculative particle Energy non-conservation

  27. Methods for H0 versus H1 Standard Frequentist Frequentist with ‘No choice’ (Want to exclude H1) (Berger conditioning) CLs (for H1 exclusion) Δχ2 Bayesian methods: Posterior odds, Bayes’ factor, BIC, AIC,… {J.B: “….no agreement on a reasonable method of model selection.”} {Cousins: Almost no use of Bayes for model selection in HEP} {Lahav, and others}

  28. Exclusion No decision Discovery (a) (b) H0 H1 t0 t t tcrit t0 p1 p0 Why ‘no decision’? When we accept H0, Need to know whether we can reject H1. Depends on separation of pdf’s H0 H1 (c) t1 t0 t

  29. p0 v p1 plots: Decision regions

  30. p0 v p1 contours, decision regions, CLs = p1/(1-p0), Punzi sensitivity

  31. Luc Demortier,“p-values: What they are and how we use them”, CDF memo June 2006 http://www-cdfd.fnal.gov/~luc/statistics/cdf0000.ps Includes discussion of several ways of incorporating nuisance parameters Desiderata: Uniformity of p-value (averaged over ν, or for each ν?) p-value increases as σν increases Generality Maintains power for discovery

  32. Ways to incorporate nuisance params in p-values • Supremum Maximise p over all ν. Very conservative • Conditioning Good, if applicable • Prior PredictiveBox. Most common in HEP p = ∫p(ν) π(ν) dν • Posterior predictiveAverages p over posterior • Plug-in Uses best estimate of ν, without error • L-ratio • Confidence intervalBerger and Boos. p = Sup{p(ν)} + β, where 1-β Conf Int for ν • Generalised frequentistGeneralised test statistic Performances compared by Demortier

  33. Particle Physicist’s wish list • Is absolute coverage really necessary? • Criteria for ‘best’ intervals • Software for Neyman construction in several dimensions • When data to be combined are a bit inconsistent Measurements of a single quantity (0±1 and 4±1) Fitting 3000 data points with 24 params (χ2/ndf OK; Δχ2=100 for params) Non-param? bsγ and g-2 (Roberto Trotta) • Look elsewhere effect • G of F for unbinned multivariate data • Choosing between 2 hypotheses Quoting signif of weak data No decision • When Wilks’ Th is not satisfied (for distribution of Δχ2) Deviations small? Simulation for 5σ (p = 3*10-7) • Expected or observed errors? • Nuisance parameters

  34. Observed or expected uncertainty? Observed: Benefit by fortunate fluctuations Cox’s choice of measuring device ALEPH measurement of m(τ) Efron and Hinkley 0±1 and 4±1 Expected: Better for combining, or choosing better method n ±n; τ± τ/n (and better than estimating from spread); binomial param Error on mean of Gaussian (use known rather than estimated width)

  35. Conclusions • Common problems: scope for learning from each other Large data sets Separating signal Testing models Signal’s Discovery • Targetted Workshops e.g. CERN: PDF4LHC (Continuing) Banff: July 2010 Discovery issues PHYSTAT: CERN 2011? Discovery issues • Summer schools: CERN, FNAL, Durham, Spain,….

  36. Excellent Workshop: Great talks Vigorous interactions Bayesian ideas and detailed implementations Statisticians: Non-parametric approaches Let’s have more!

  37. Excellent Workshop: Great talks Vigorous interactions Bayesian ideas and detailed implementations Statisticians: Non-parametric approaches Let’s have more! Very big THANK YOU to Roberto, Martin and Uros !

More Related