1 / 22

Causal Analysis and Model Confirmation Dr. Larry J. Williams Dean ’ s Research Chair and Professor of Management/Psycho

Causal Analysis and Model Confirmation Dr. Larry J. Williams Dean ’ s Research Chair and Professor of Management/Psychology CARMA Director, Wayne State University April 26, 2013: Honoring the Career Contributions of Professor Larry James. SEM: A Tool for Theory Testing. x 1. l x 11. δ 1.

deion
Download Presentation

Causal Analysis and Model Confirmation Dr. Larry J. Williams Dean ’ s Research Chair and Professor of Management/Psycho

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Causal Analysis and Model Confirmation Dr. Larry J. Williams Dean’s Research Chair and Professor of Management/Psychology CARMA Director, Wayne State University April 26, 2013: Honoring the Career Contributions of Professor Larry James

  2. SEM: A Tool for Theory Testing x1 lx11 δ1 lx21 x1 LV1 x2 δ2 lx31 x3 δ3 11 φ21 z1 z2 x4 φ31 lx42 δ4 21 12 lx52 h2 LV5 x2 LV2 h1 LV4 x5 δ5 lx62 x6 δ6 ly31 ly62 ly11 ly42 ly21 ly52 13 φ32 y2 y3 y5 y6 y1 y4 x7 lx73 δ7 lx83 x3 LV3 e1 e3 e4 e6 e2 e5 x8 δ8 lx93 x9 δ9 Figure 1. Example Composite Model (Includes Path and Measurement Models)

  3. Who is familiar with this type of model? • Was not always the case- Larry’s workshops His journal publications • A cynical view of contemporary SEM practices A parameter estimate represents a “causal” effect A good fitting model is “the” model • Fancy software/fast processors does NOT mean good science! • Science would be improved by revisiting Larry’s work

  4. Key principles of model confirmation/disconfirmation Models lead to predictions Confirmation occurs if predictions match data Confirmation implies support for a model, but does not prove it is THE model • Today: towards improved causal analysis Revisit key work on model comparisons and the role of the measurement model Share results from two recent publications based on that work Provide recommendations based on old and new

  5. James, Mulaik, and Brett (1982) • Conditions 1-7: appropriateness of theoretical models Formal statement of theory via structural model Theoretical rationale for causal hypotheses Specification of causal order Specification of causal direction Self contained functional equations Specification of boundaries Stability of structural model • Condition 8: Operationalization of variables • Condition 9 and 10: Empirical confirmation of predictions Not well understood by many management researchers: Do models with good fit have significant paths? Do models with bad fit have non-significant paths?

  6. Condition 9: Empirical support for functional equations Tests whether causal parameters in model are different from zero Implemented by comparing theoretical model to a restricted path (structual model), with paths in the model forced to zero Hoping for a significant chi-square difference test (rejecting null H0) This is global test, individual parameter tests also important ***Weighted in favor of confirming predictions- power • Condition 10: Fit of model to data Tests whether paths left out of model are important Implemented by comparing theoretical model to a less restricted model that adds omitted paths Hoping to fail to reject the null that these paths are zero Fit index: non-statisicalassessment of model, practical judgment ***The omitted parameter test saves considerable fuss, while the fit approach is more subtle

  7. Mulaik, James, Van Alstine, Bennett, Lind, & Stilwell (1989) • Proposed that goodness-of-fit indices usually heavily influenced by goodness of fit of measurement model (foreshadowed in James, Mulaik, & Brett, 1982) • And indices reflect, to a much lesser degree, the goodness of fit of the causal portion of the model • Thus, possible to have models with overall GFI values that are high, but for which causal portion and relations among latent variables are misspecified • Introduced RNFI; demonstrated it with artificial data

  8. Williams & O’Boyle (2011) • Can lack of focus of global fit indices like CFI and RMSEA on path component lead to incorrect decisions about models – models with adequate values but misspecified path components? • Requires simulation analysis- know “true model”, evaluate misspecified models, examine their CFI and RMSEA values • Examined four representative composite models, two of these were examined using both 2 and 4 indicators per latent variable, results in 6 examples • Focus: what types of models have CFI and RMSEA values that meet the “gold standards”?

  9. ξ1 ξ2 η1 ξ3 ξ4 η2 ξ5 ξ6 Ex.1: Duncan, et al., 1971

  10. Ex. 2: Ecob (1987)

  11. ξ1 η1 ξ2 η3 ξ3 η2 ξ4 Ex. 3: Mulaik et al., (1989)

  12. ξ1 ξ2 η1 η2 ξ3 Ex. 4: MacCallum (1986)

  13. Results: Are there misspecified models with CFI>.95 or RMSEA<.08? Example 1: MT-4 Example 2: MT-6 Example 3 (2 indicators): MT-5 Example 3 (4 indicators): MT-7 Example 4 (2 indicators): MT-2 Example 4 (4 indicators): MT-4 Conclusion: multiple indicator models with severe misspecifications can have CFI and RMSEA values that meet the standards used to indicate adequate fit, this creates problems for testing path model component of composite model and evaluating theory

  14. Also investigated two alternative indices based on previous work by James, Mulaik & Brett (1982), Sobel and Bohrnstedt (1985), Williams & Holahan (1994) , Mulaik et al. (1989), and McDonald & Ho (2002) • Goal was to develop indices that more accurately reflect the adequacy of the path model component of a composite model • NSCI-P an extension of the RNFI of Mulaik et al. (1989), RMSEA-P originally proposed by McDonald & Ho (2002) • We calculated values for correctly and incorrectly specified models presented earlier

  15. RMSEA-P: Remember null H0 is that MT is true, hope is to not reject MT should have value <.08 (if not Type I error), MT-X should have values >.08 (if not Type II error) Results- All MT had values < .08 (no Type I errors) Misspecified models (MT-x): success (all values >.08) with 5 of 6 examples Example 1: MT-2 and MT-1 retained (RMSEA-P<.08) other 5 examples: MT-1 and all other MT-x had RMSEA-P>.08

  16. Notes: Example 1 had #indicators/latent variable ratio of 1.25, all others > 2.0 Results with confidence intervals also supportive • NSCI-P values for MT approached 1.0 in all examples (within rounding error) average drop in value in going from MT to MT-1 was .043

  17. O’Boyle & Williams (2011) • Of 91 articles in top journals, 45 samples with information needed to do RMSEA decomposition to yield RMSEA-P, examine with real data if models with good composite fit have good path model fit Mean value of RMSEA-P = .111 Only 3 < .05, 15 < .08 recommended values RMSEA-P Confidence intervals (CI): 19 (42%) have lower bound of CI >.08 Reject close and reasonable fit, models bad Only 5 (11%) have upper bound of CI <.08 Fail to reject close fit, models good • Possibilities/suspicions of James and colleagues supported

  18. What is role of fit indices in broader context of model evaluation? • James, Mulaik, & Brett (1982) and Anderson & Gerbing (1988) stress MT comparison with MSS via χ2 difference test • Fit indices were supposed to supplement this model comparison as test of constraints in path model (former more subtle, latter saves considerable fuss) Only 3 of 43 did this test 30 were significant (MT includes significant misspecifications) For 26 of these, CFI and RMSEA values met criteria “good” fit conclusion, counter to χ2 difference test • Researchers need to save the fuss, be less subtle

  19. Conclusions:Toward Improved Practices • Use composite model fit indices (CFI, RMSEA), cutoffs? • Use indicator residuals • Include focus on path model relationships Chi-square difference test: MSS -MT Specialized fit indices (NSCI, RMSEA-P) • Do not forget other model diagnostics • Remember that a path left out is a H0 and should be supported theoretically , always alternative models

  20. And we should remember that scholars like Larry (and his collaborators) created the “path” for many of us who followed by: (a) helping create the discipline of organizational research methods, (b) providing educational leadership, (c) conducting exemplary scholarship to “model” our work after, and (d) supporting many others and helping them “fit” in as their careers developed • On behalf of all of us, and CARMA, thanks!

  21. Key References • Anderson, J. C., & Gerbing, D. W. (1988). Structural equation modeling in practice: A review and recommended two-step approach. Psychological Bulletin, 103, 411-423 • James, L. R., Mulaik, S. A., & Brett, J. (1982). Causal analysis: Models, assumptions, and data. Beverly Hills, CA: Sage Publications. • McDonald, R. P., & Ho, M. H. R. (2002). Principles and practice in reporting structural equation analyses. Psychological Methods, 7, 64-82. • Mulaik, S. A., James, L. R., Van Alstine, J., Bennett, N., Lind, S., & Stilwell, C. D. (1989). Evaluation of goodness-of-fit indices for structural equation models. Psychological Bulletin, 105, 430-445. • O’Boyle & Williams (2011). Decomposing model fit: Measurement versus theory in organizational research using latent variables. Journal of Applied Psychology, 96, 1, 1-12. • Williams & O’Boyle (2011). The myth of global fit indices and alternatives for assessing latent variable relations. Organizational Research Methods,14, 350-369.

More Related