1 / 47

HSS4303B – Intro to Epidemiology Feb 8, 2010 - Agreement

HSS4303B – Intro to Epidemiology Feb 8, 2010 - Agreement. Answers from Thursday’s Homework. Compute: Prevalence of cancer 44% Sensitivity & specificity 93.3% and 96.1% % of false positives 532/ (56+532) % of false negatives 4/(4+13194)

isleen
Download Presentation

HSS4303B – Intro to Epidemiology Feb 8, 2010 - Agreement

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. HSS4303B – Intro to Epidemiology Feb 8, 2010 - Agreement

  2. Answers from Thursday’s Homework • Compute: • Prevalence of cancer 44% • Sensitivity & specificity 93.3% and 96.1% • % of false positives 532/ (56+532) • % of false negatives 4/(4+13194) • PV+ and PV- 9.5% and 100%

  3. Last Time… • Screening Tests • Validity and Reliability • Specificity and Sensitivity • Pos Predictive Value and Neg Predictive Value

  4. Sensitivity = a/(a+c) PV+ = a/(a+b) Specificity = d/(b+d) PV- = d/(c+d)

  5. Ultimately, What Do All These Indicators Want To Tell Us? “What is the likelihood is it that you have the disease?”

  6. Likelihood Ratio • A way of using the sensitivity and specificity of a test to see if a positive or negative result usefully changes the probability of having the disease • Assesses the value of performing the screening test at all • Who is this useful for?

  7. Likelihood Ratio • LR+ (positive likelihood ratio) • The probability of a positive test result for a person who really has the disease divided by the probability of a positive test result for someone who doesn’t really have the disease • i.e. “P(true positives)” / “P(false positives)” = sensitivity / (1 − specificity)

  8. Likelihood Ratio • LR- (negative likelihood ratio) • The probability of a negative test result for a person who really has the disease divided by the probability of a negative test result for someone who doesn’t really have the disease • i.e. “P(false negatives)” / “P(true negatives)” = (1 − sensitivity) / specificity

  9. Sensitivity = a/(a+c) PV+ = a/(a+b) Specificity = d/(b+d) PV- = d/(c+d) True positives True negatives False positives False negatives a d b c LR+ = P (true +ve)/ P(false +ve) =(a/(a+c)) / (b/(b+d)) =(a/(a+c))/(1-(d/(b+d)) =sensitivity / (1-specificity)

  10. Interpreting the LR • A likelihood ratio of >1 indicates the test result is associated with the disease • A likelihood ratio <1 indicates that the result is associated with absence of the disease • In other words • High LR+ means strong suspicion that a +ve test result means the person has the disease • Low LR- means strong suspicion that a –ve test result means the person doesn’t have disease • What about “1”?

  11. Interpreting the LR • Arbitrary cutoffs: • LR+ >10 means strong diagnostic value • LR- <0.1 means strong diagnostic value • (Some literature suggests 5 and 0.2 are more appropriate cutoffs) The likelihood ratio, which combines information from sensitivity and specificity, gives an indication of how much the odds of disease change based on a positive or a negative result

  12. LR+ • The smallest possible value of the LR+ is zero, when sensitivity is zero. • The maximum possible value of the LR+ is infinity when the denominator is minimized (specificity = 1, so 1 - specificity = 0). • LR+ = 1: indicates a test with no value in sorting out persons with and without the disease of interest, since the probability of a positive test result is equally likely for affected and unaffected persons.

  13. LR- • The smallest value of the LR– occurs when the numerator is minimized (sensitivity = 1, so 1 - sensitivity = 0), resulting in an LR– of zero. • The largest value of the LR– occurs when the denominator is minimized (specificity = 0), resulting in an LR– of positive infinity. • LR– = 1: indicates a test with no value in sorting out persons with and without the disease of interest, as the probability of a negative test result is equally likely among persons affected and unaffected with the disease of interest.

  14. FNA test (fine needle aspiration)

  15. FNA test (fine needle aspiration) LR+ = sensitivity / (1-specificity) = 0.93 / (1-0.92) = 11.63 <- FNA test has high diagnostic value

  16. Probability of presence of disease • Pretest probability of disease - the likelihood that a person has the disease of interest before the test is performed. • Pretest odds of disease are defined as the estimate before diagnostic testing of the probability that a patient has the disease of interest divided by the probability that the patient does not have the disease of interest. • Posttest odds of disease are defined as the estimate after diagnostic testing of the probability that a patient has the disease of interest divided by the probability that the patient does not have the disease of interest. • Posttest probability of disease – the likelihood that a person has the disease of interest post the test is performed.

  17. Pretest probability and pretest odds Pretest probability = Pretest odds = pretest probability / (1-pretest probability) = = 0.15

  18. Pretest probability and pretest odds Pretest probability = 15/114 = 0.13 Pretest odds = pretest probability / (1-pretest probability) = 0.13/0.87 = 0.15

  19. What does this have to do with LR? • LR = post test odds / pre test odds • So now we can compute the odds of having the disease after applying the test and computing LR

  20. Pretest probability and pretest odds Pretest odds = 0.15 Sensitivity = 93% Specificity = 92% Compute LR+ and LR-: LR+ = 0.93/0.08 = 11.63 LR- = 0.07/0.92 = 0.08

  21. So… • Knowing pretest odds and LR+, what are the posttest odds ? (i.e., odds of having the disease after positive test result)? Post test odds = LR x pre=test odds = 11.63 x 0.15 = 1.74 NB, textbook (p.99) multiplies 11.63 by 0.15 and gets 1.76, which is wrong

  22. And then…. • Can you now compute post-test probability? • (do you remember the difference between probability and odds?) Post test prob = post test odds / (1 -+ post test odds) = 1.74 / 2.74 = 0.64

  23. LR vs PV • Positive predictive value is the proportion of patients withpositive test results who are correctly diagnosed. • The likelihood ratio indicates the value of the test for increasingcertainty about a positive diagnosis • Relates to a comparison between pre-test odds of having the disease vs post-test odds of having the disease LR+ = post-test odds / pre-test odds

  24. LR vs PV • Remember that PV varies with prevalence of the disease • LR is independent of prevalence

  25. Pretest odds = 0.15 Sensitivity = 93% Specificity = 92% LR+ = 11.63 LR- = 0.08 Post test odds = 1.74 Post test prob = 64% Similar thing can be done with LR-, but in general we don’t bother

  26. Performance Yield True Disease Status - + Results of Screening Test 400 995 + - 98905 100 Sensitivity: a / (a + c) = 400 / (400 + 100) = 80% Specificity: d / (b + d) = 98905 / (995 + 98905) = 99% PV+: a / (a + b) = 400 / (400 + 995) = 29% PV-: d / (c + d) = 98905 / (100 + 98905) = 99% Prevalence: (a+c)/(a+b+c+d) = 500/100400 = 0.5% LR+ = sens / (1-spec) = 0.8/(1-0.99) = 80

  27. Comparing LR and PV True Disease Status - + Results of Screening Test 400 995 + - 98905 100 PV+=29% Among persons who screen positive, 29% are found to have the disease. LR+ = 80 A positive test result increases your odds of having the disease by 80 fold

  28. Homework #1 • Geenberg p. 105, question 1-13: • 13786 Japanese patients underwent CT scans to detect first signs of cancer, then had pathology tests 2 years later to confirm whether or not they actually had cancer • Compute: • LR+ • LR- • Pre-test probability of cancer • Pre-test odds of cancer • Post-test odds of cancer • Post-test probability of cancer (Answers are in the notes section of this slide)

  29. What if you have a continuous variable? • What kind of variableis cancer vs no cancer? • What is a continuous diagnostic variable? • Examples: • Body temperature • Blood pressure • Height • Weight • etc

  30. signal noise Receiver Operator Curve (ROC)

  31. Useful for comparing 2 diagnostic tests. The greater the area under the curve, the better signal-to-noise ratio and the better the test

  32. See article on website called “Kappa.pdf” Agreement

  33. Remember Reliability? • The extent to which the screening test will produce the same or very similar results each time it is administered. • Inter-rater reliability is “the variation in measurements when taken by a different persons but with the same method or instruments” Also called CONCORDANCE

  34. Inter-rater Reliability • Is a measurement of Agreement • A score of how much consensus there is among judges, observers, technicians or any number of people who are using the same instrument(s) to measure the same data. Eg: • Judges scoring a beauty pageant contestant from 1-10 • Several psychologists using a PTSD scale to assess a patient • Different devices measuring body temperature simultaneously on same patient

  35. How Do We Measure Agreement? • Lots of stats available to us: • Inter-rater correlation coefficient • Intra-class correlation coefficient • Concordance correlation coefficient • Fleiss’s kappa • Cohen’s kappa

  36. Kappa (κ) • Cohen • Two raters • Fleiss • Adaptation of Cohen, applicable to multiple raters • Kappa is generally thought to be a more robust measure than simple percent agreement calculation since κ takes into account the agreement occurring by chance

  37. Cohen’s Kappa Cohen the Barbarian

  38. Cohen’s Kappa • Κ = {Pr(a) – Pr(e)} / {1-Pr(e)} Pr(a) = relative observed agreement Pr(e) = prob that agreement is due to chance Results in a ratio from 0 to 1

  39. Two Judges Decide Whether Or Not 75 Beauty Pageant Contestants Are Hot Judge #1 = Hasselhoff Judge #2 = Shatner

  40. The Data

  41. The Data Pr(a) = relative observed agreement = (41 + 27 )/ 75 = 90.7%

  42. The Data Pr(a) = relative observed agreement = (41 + 27 )/ 75 = 90.7% Pr(e) = prob that agreement is due to chance = (44x45/752 + (31x30)/752 = 0.352 + 0.165 = 51.7% (multiply marginals and divide by total squared)

  43. Compute Kappa • K = [ Pr(a) – Pr(e) ] / 1 – Pr(e) • = (0.907 – 0.517) / (1-0.517) • = 0.81 How do we interpret this?

  44. Interpreting Kappa Hasselhoff and Shatner are in almost perfect agreement over who is hot and who is not.

  45. What if….? • There are >2 raters? • There are >2 categories? • Eg, “ugly, meh, hmm, pretty hot, very hot, smokin’” • Eg, “don’t like, somewhat like, like” • Then it is possible to apply kappa, but only to determine complete agreement. So? • Dichotomize variables • Weighted kappa

  46. Homework #2 Compute Cohen’s Kappa in both cases and interpret. (The answers are in the notes section of this slide)

  47. So When/Why Use Screening Tests?

More Related