1 / 35

Analysis of Variance

1 Overview 2 One-way ANOVA. Analysis of Variance. Ronald Aylmer Fisher (1890-1962). Where ANOVA Comes From. An introduction of a procedure for testing the hypothesis that three or more population means are equal. For example: H 0 : µ 1 = µ 2 = µ 3 = . . . µ j

mieko
Download Presentation

Analysis of Variance

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. 1 Overview 2 One-way ANOVA Analysis of Variance

  2. Ronald Aylmer Fisher (1890-1962) Where ANOVA Comes From

  3. An introduction of a procedure for testing the hypothesis that three or more population means are equal. For example: H0: µ1 = µ2 = µ3 = . . . µj H1: At least one mean is different Overview

  4. Definition Analysis of Variance (ANOVA) a method of testing the equality of three or more population means by analyzing sample variations Overview

  5. 1. The F-distribution is not symmetric; it is skewed to the right. 2. The values of F can be 0 or positive, they cannot be negative. 3. There is a different F-distribution for each pair of degrees of freedom for the numerator and denominator. ANOVA methods require the F-distribution

  6. F - distribution Not symmetric (skewed to the right)  nonnegative values only

  7. Assumptions 1. The populations have normal distributions. 2. The populations have the same variance 2 (or standard deviation ). 3. The samples are simple random samples. 4. The samples are independent of each other. One-Way ANOVA

  8. Treatment (or factor) a property or characteristic that allows us to distinguish the different populations from another Definition

  9. Relationships Among Components of ANOVA

  10. Estimate the common value of 2 using 1. The variance between samples (also called variation due to treatment) is an estimate of the common population variance 2 that is based on the variability among the sample means. 2. The variance within samples (also calledvariation due to error) is an estimate of the common population variance 2 based on thesample variances. ANOVA Fundamental Concept

  11. Test Statistic for One-Way ANOVA variance between samples F = variance within samples ANOVA Fundamental Concept A excessively large F test statistic is evidence against equal population means.

  12. Right-tailed test Degree of freedom with j samples of the same size n numerator df = j -1 denominator df = n - j Critical Value of F

  13. SS(total),or total sum of squares, is a measure of the total variation (around Y) in all the sample data combined. Key Components of ANOVA Method

  14. SS(total),or total sum of squares, is a measure of the total variation (around Y) in all the sample data combined. Key Components of ANOVA Method SS(total) = (Yij - Y)2

  15. SS(treatment) is a measure of the variation between the samples. In one-way ANOVA, SS(treatment) is sometimes referred to as SS(factor). Because it is a measure of variability between the sample means, it is also referred to as SS (betweengroups) or SS (between samples). Key Components of ANOVA Method

  16. SS(treatment) is a measure of the variation between the samples. In one-way ANOVA, SS(treatment) is sometimes referred to as SS(factor). Because it is a measure of variability between the sample means, it is also referred to as SS (betweengroups) or SS (between samples). Key Components of ANOVA Method SS(treatment) = n1(Y1 - Y)2 + n2(Y2 - Y)2 + . . . nk(Yk - Y)2 = nj(Yj - Y)2

  17. SS(within or error) is a sum of squares representing the variability that is assumed to be common to all the populations being considered. Key Components of ANOVA Method

  18. SS(within or error) is a sum of squares representing the variability that is assumed to be common to all the populations being considered. SS(within) = (Yij - Yj)2 Key Components of ANOVA Method

  19. SS(total) = SS(treatment) + SS(error) SS(total) = SS(between) + SS(within) Key Components of ANOVA Method

  20. Sum of Squares SS(treatment)andSS(error) divided by corresponding number of degrees of freedom. MS (treatment) is mean square for treatment, obtained as follows: Mean Squares (MS)

  21. Sum of Squares SS(between)andSS(within) divided by corresponding number of degrees of freedom. MS (between or treatment) is mean square for treatment, obtained as follows: Mean Squares (MS) SS (between) MS (between) = j - 1

  22. MS (within or error) is mean square for error, obtained as follows: Mean Squares (MS) SS (within) MS (within) = n - j

  23. MS (within or error) is mean square for error, obtained as follows: Mean Squares (MS) SS (within) MS (within) = n - j SS (total) MS (total) = n - 1

  24. The ANOVA Table

  25. Example Grand Mean= (70+79+76)/3 = 75

  26. Each observation is different from the Grand (total sample) Mean by some amount • There are two sources of variance from the mean • 1) That due to the treatment (COLOUR) or independent variable • 2) That which is unexplained by our treatment H0: µ1 = µ2 = µ3 H1: At least one mean is different

  27. (Each Observation-Grand Mean)2 =(60-75)2 + (68-75)2+ (63-75)2 + (76-75)2 +(73-75)2 + (75-75)2+ (80-75)2 + (77-75)2 + (85-75)2+(78-75)2 + (79-75)2 + (72-75)2 +(80-75)2 + (73-75)2 + (76-75)2 = 416 Total Sum of Squares

  28. (Average treatment outcome-Grand Mean)2•number of observations in treatment (70-75)2•5 + (79-75)2 •5 + (76-75)2 •5=210 Explained variance Treatment Sum of Squares

  29. (Each observation-treatment mean)2 =(68-70)2 + (70-70)2 + (63-70)2 + (76-70)2 + (73-70)2 + (75-79)2 + (80-79)2 + (77-79)2 + (85-79)2 + (78-79)2 + (79-76)2 + (72-76)2 + (80-76)2 + (73-76)2 + (76-76)2 = 206 Within SS or Unexplained Variance Unexplained Sum of Squares

  30. df = j-1, n-j (3-1), (15-3) Tabled value for F(2;12) = 3,885 (=0.05) Fcomp must equal or exceed 3,885 for the means to be significantly different Calculating F

  31. Calculating F

  32. Conclusion • Computed value F = 6,18 • Tabled value for F(2;12) = 3,885 (=0.05) • H0 should be rejected. Population means are NOT equal. • If the null hypothesis is rejected and you are interested in finding the differences you shoud run Fisher’s Least Significance Difference (LSD).

  33. Least Significance Difference LSD is a method to compare pairwise comparisons between quantitative variables coming from three or more independent groups. Two treatment means are declared to be significantly different (using LSD) if wherendenotes the number of observations involved in the computation of a treatment mean.

  34. Least Significance Difference • df = j(n-1)=3(5-1)=12 • Tabled value for t(0,1;12) =2,179 • must equal or exceed 5,71 for the means to be significantly different

  35. Least Significance Difference • |blue – white| = |70-79| = 9* • |blue – green| = |70-76| = 6* • |white – green| = |79-76| =3

More Related