1 / 22

P-values and their limitations & Type I and Type II errors

P-values and their limitations & Type I and Type II errors. Stats Club 5 Marnie Brennan. What do you know about P-values?. References. Petrie and Sabin - Medical Statistics at a Glance: Chapter 17 & 18 Good Petrie and Watson - Statistics for Veterinary and Animal Science: Chapter 6 Good

rhian
Download Presentation

P-values and their limitations & Type I and Type II errors

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. P-values and their limitations &Type I and Type II errors Stats Club 5 Marnie Brennan

  2. What do you know about P-values?

  3. References • Petrie and Sabin - Medical Statistics at a Glance: Chapter 17 & 18 Good • Petrie and Watson - Statistics for Veterinary and Animal Science: Chapter 6 Good • Kirkwood and Sterne – Essential Medical Statistics: Chapter 8 & 35 • Dohoo, Martin and Stryhn – Veterinary Epidemiologic Research: Chapter 2 & 6

  4. Interesting reads! • Sterne, JAC and Davey-Smith, G (2001) Sifting the evidence – what’s wrong with significance tests? British Medical Journal, Vol. 322, 226-231. - Good • Altman, DG and Bland, JM (1995) Absence of evidence is not evidence of absence. British Medical Journal, Vol. 311, 485. • Nakagawa, S and Cuthill, IC (2007) Effect size, confidence interval and statistical significance: a practical guide for biologists. Biological Reviews, Vol. 82, 591-605. – I’ve not read this, but it has been recommended

  5. Differences between groups • Many different tests to measure the difference between two or more groups of subjects/animals/patients • We will cover these individually in subsequent weeks • How do we know whether they are truly different from each other? • i.e. There is truly a difference between groups or not?

  6. Hypothesis (significance) testing • You have a scientific question you want to answer • You construct a hypothesis to test your question • You have to have an alternative hypothesis to test it against

  7. Null and alternative hypotheses • Null hypothesis – No difference between groups/no association between variables • Sometimes written as H0 • Alternative hypothesis – There is a difference between groups/an association between variables • Sometimes written as H1 • These hypotheses relate to the population of interest, not your sample of the population

  8. What is a P-value? • We do our study, run our statistical tests, and come up with a P-value or Probability • ‘The P-value is the probability of obtaining our results or something more extreme, if the null hypothesis is true’ (Petrie and Sabin) • ‘The probability of getting a difference at least as big as that observed if the null hypothesis is true’ (Kirkwood and Sterne) • ‘The chance of getting the observed effect (or one or more extreme) if the null hypothesis is true’ (Petrie and Watson)

  9. What does this mean??!! • Basically the probability of getting what you have got with your study results if the null hypothesis is true! • If the difference between our groups is large • The probability would be small, therefore unlikely the null hypothesis is true (and you usually reject the null hypothesis as there is evidence against it) • If the difference between our groups is small • The probability would be large, therefore likely the null hypothesis is true (there is not enough evidence to reject the null hypothesis) • Bad to say you accept the null hypothesis! • ‘Absence of evidence is not evidence of absence’

  10. Not significant at the 5% level Significant at the 5% level A value of the test statistic which gives P>0.05 A value of the test statistic which gives P<0.05

  11. Using P-values • Usually you set your ‘significance’ level before you collect your data – this should be stated in the methods • e.g. ‘We set the significance level at P=0.01 for our analysis’ • P<0.05 is a fairly arbitrary level (one guy’s ponderings!) • Read the article by Sterne and Davey Smith • Bottom line - the smaller the P-value, the more evidence against the null hypothesis

  12. A sliding scale.......

  13. How does this fit with what you do or have seen/experienced?

  14. P-value etiquette! • Always quote the exact P-value • E.g. P = 0.032, not P<0.05 • Display P-values accurate to two significant figures • E.g. P=0.032, or 0.17 • When P-values become very small, acceptable to display as P<0.001

  15. Limitations of using just P-values • By just using P-values, you lose a lot of information • Doesn’t tell you about the magnitude of the effect observed • Often researchers only talk about P-values, and nothing else • I am certainly guilty of this! • It is also important to determine whether your result is biologically or clinically important (not only that it is ‘significant’) – if you just use a number to interpret outcomes, it may not ‘mean’ anything • You can use Confidence Intervals to quantify the effect of interest • Gives you a range of values which represent the difference between your groups • This is our next Stats Club session

  16. Interpretation of research

  17. Errors in hypothesis testing • Our decision to reject the null hypothesis or not can be wrong sometimes • Petrie and Watson

  18. Type I error • When we reject the null hypothesis and it is actually true • Affected by: • Significance level chosen (becomes the maximum chance of making a Type I error) • e.g. If significance level chosen is P<0.05, we have a 1 in 20 chance that a test will be significant by chance; if P<0.01 is chosen, we have a 1 in 100 chance the test is significant by chance • Number of comparisons – the greater number of comparisons carried out, the more likely you will get a ‘positive’ result that is spurious • Comes back to whether the result is biologically or clinically important • Can adjust for this using post-hoc analysis e.g. Bonferroni correction

  19. Type II error • We don’t reject the null hypothesis when it is actually false • Affected by: • Small sample sizes – more chance of getting Type II errors • Precision of the measurements – if measurements are precise, less chance of getting Type II errors • Effect of interest – the larger the difference between the groups, the less likely that a Type II error will occur

  20. Type I and Type II error - relationship • These two things are related, generally as one increases, the other decreases • Bottom line – if your study design is correct, you have carried out a sample size calculation and have recruited the right number of subjects, then the chances of error decrease hugely as the power of your study will be sufficient • Sample size calcs and power will be discussed in later Stats Club sessions

  21. Summary • Set your study up right to decrease the chances of Type I and Type II errors • Use P-values but also CI’s to get an idea of the magnitude of the difference between groups • Set your significance level BEFORE you start your data collection, and don’t just go automatically for P<0.05! • Display your P-values correctly!

  22. Next month • Confidence intervals beware……

More Related