1 / 32

LOGLINEAR MODELS FOR INDEPENDENCE AND INTERACTION IN THREE-WAY TABLES

LOGLINEAR MODELS FOR INDEPENDENCE AND INTERACTION IN THREE-WAY TABLES. BY ENI SUMARMININGSIH, SSI, MM. Table Structure For Three Dimensions. When all variables are categorical, a multidimensional contingency table displays the data We illustrate ideas using thr three-variables case.

tien
Download Presentation

LOGLINEAR MODELS FOR INDEPENDENCE AND INTERACTION IN THREE-WAY TABLES

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. LOGLINEAR MODELS FOR INDEPENDENCE ANDINTERACTION IN THREE-WAY TABLES BY ENI SUMARMININGSIH, SSI, MM

  2. Table Structure For Three Dimensions • When all variables are categorical, a multidimensional contingency table displays the data • We illustrate ideas using thr three-variables case. • Denote the variables by X, Y, and Z. We display the distribution of X-Y cell count at different level of Z using cross sections of the three way contingency table (called partial tables)

  3. The two way contingency table obtained by combining the partial table is called the X-Y marginal table (this table ignores Z)

  4. Death Penalty Example

  5. Marginal table

  6. Partial and Marginal Odd Ratio Partial Odd ratio describe the association when the third variable is controlled The Marginal Odd ratio describe the association when the Third variable is ignored (i.e when we sum the counts over the levels of the third variable to obtain a marginal two-way table)

  7. Types of Independence A three-way IXJXK cross-classification of response variables X, Y, and Z has several potential types of independence We assume a multinomial distribution with cell probabilities {i jk}, and The models also apply to Poisson sampling with means }. The three variables are mutually independent when

  8. Similarly, X could be jointly independent of Y and Z, or Z could be jointly independent of X and Y. Mutual independence (8.5) implies joint independence of any one variable from the others. X and Y are conditionally independent, given Z when independence holds for each partial table within which Z is fixed. That is, if then

  9. Homogeneous Association and Three-Factor Interaction

  10. Marginal vs Conditional Independence • Partial association can be quite different from marginal association • For further illustration, we now see that conditional independence of X and Y, given Z, does not imply marginal independence of X and Y • The joint probability in Table 5.5 show hypothetical relationship among three variables for new graduate of a university

  11. Table 5.5 Joint Probability

  12. The association between Y=income at first job(high, low) and X=gender(female, male) at two level of Z=major discipline (liberal art, science or engineering) is described by the odd ratios Income and gender are conditionally independent, given major

  13. Marginal Probability of Y and X The odd ratio for the (income, gender) from marginal table =2 The variables are not independent when we ignore major

  14. Suppose Y is jointly independent of X and Z, so Then And summing both side over i we obtain = Therefore

  15. So X and Y are also conditionally independent. In summary, mutual indepedence of the variables implies that Y is jointly independent of X and Z, which itself implies that X and Y are conditionaaly independent. • Suppose Y is jointly independent of X and Z, that is . Summing over k on both side, we obtain Thus, X and Y also exhibit marginal independence

  16. So, joint independence of Y from X and Z (or X from Y and Z) implies X and Y are both marginally and condotionally independent. Since mutual independence of X, Y and Z implies that Y is jointly independent of X and Z, mutual independence also implies that X and Y are both marginally and conditionally independent However, when we know only that X and Y are conditionally independent, Summing over k on both sides, we obtain

  17. All three terms in the summation involve k, and this does not simplify to marginal independence

  18. A model that permits all three pairs to be conditionally dependent is Model 8.11. is called the loglinear model of homogeneous association or of no three-factor interaction.

  19. Loglinear Models for Three Dimensions • Hierarchical Loglinear Models Let {ijk} denote expected frequencies. Suppose all ijk>0 and let ijk = log ijk. A dot in a subscript denotes the average with respect to that index; for instance, We set , ,

  20. The sum of parameters for any index equals zero. That is

  21. The general loglinear model for a three-way table is This model has as many parameters as observations and describes all possible positive i jk Setting certain parameters equal to zero in 8.12. yields the models introduced previously. Table 8.2 lists some of these models. To ease referring to models, Table 8.2 assigns to each model a symbol that lists the highest-order term(s) for each variable

  22. Interpreting Model Parameters Interpretations of loglinear model parameters use their highest-order terms. For instance, interpretations for model (8.11). use the two-factor terms to describe conditional odds ratios At a fixed level k of Z, the conditional association between X and Y uses (I- 1)(J – 1). odds ratios, such as the local odds ratios Similarly, (I– 1)(K – 1) odds ratios {i (j)k} describe XZ conditional association, and (J – 1)(K – 1) odds ratios {(i)jk} describe YZ conditional association.

  23. Loglinear models have characterizations using constraints on conditional odds ratios. For instance, conditional independence of X and Y is equivalent to {ij(k)} = 1, i=1, . . . , I-1, j=1, . . . , J-1, k=1, . . . , K. substituting (8.11) for model (XY, XZ, YZ)into log ij(k)yields Any model not having the three-factor interaction term has a homogeneous association for each pair of variables.

  24. For 2x2x2 tables

  25. Alcohol, Cigarette, and Marijuana Use Example Table 8.3 refers to a 1992 survey by the Wright State University School of Medicine and the United Health Services in Dayton, Ohio. The survey asked 2276 students in their final year of high school in a nonurban area near Dayton, Ohio whether they had ever used alcohol, cigarettes, or marijuana. Denote the variables in this 222 table by A for alcohol use, C for cigarette use, and M for marijuana use.

  26. Table 8.5 illustrates model association patterns by presenting estimated conditional and marginal odds ratios For example, the entry 1.0 for the AC conditional association for the model (AM, CM)of AC conditional independence is the common value of the AC fitted odds ratios at the two levels of M,

  27. The entry 2.7 for the AC marginal association for this model is the odds ratio for the marginal AC fitted table Table 8.5 shows that estimated conditional odds ratios equal 1.0 for each pairwise term not appearing in a model, such as the AC association in model (AM, CM). For that model, the estimated marginal AC odds ratio differs from 1.0, since conditional independence does not imply marginal independence. Model (AC, AM, CM)permits all pairwise associations but maintains homogeneous odds ratios between two variables at each level of the third. The AC fitted conditional odds ratios for this model equal 7.8. One can calculate this odds ratio using the model’s fitted values at either level of M, or from (8.14) using

  28. INFERENCE FOR LOGLINEAR MODELS Chi-Squared Goodness-of-Fit Tests As usual, X 2 and G2 test whether a model holds by comparing cell fitted values to observed counts =2 Where nijk = observed frequency and =expected frequency Here df equals the number of cell counts minus the number of model parameters. For the student survey (Table 8.3), Table 8.6 shows results of testing fit for several loglinear models.

  29. Models that lack any association term fit poorly The model ( AC, AM, CM)that has all pairwise associations fits well (P= 0.54) It is suggested by other criteria also, such as minimizing AIC= - 2(maximized log likelihood - number of parameters in model) or equivalently, minimizing [G2- 2(df)].

More Related