1 / 20

Part II S igma Freud & Descriptive Statistics

Part II S igma Freud & Descriptive Statistics. Chapter 6    Just the Truth: An Introduction to Understanding Reliability and Validity. Why Measurement?. What is measurement ? You need to know that the data you are collecting represents what it is you want to know about.

minh
Download Presentation

Part II S igma Freud & Descriptive Statistics

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Part IISigma Freud & Descriptive Statistics Chapter 6    Just the Truth: An Introduction to Understanding Reliability and Validity

  2. Why Measurement? • What is measurement? • You need to know that the data you are collecting represents what it is you want to know about. • How do you know that the instrument you are using to collect data works every time (reliability) and measures what it is supposed to (validity)?

  3. Scales of Measurement • Measurement is the assignment of values to outcomes following a set of rules • There are four types of measurement scales • Nominal • Ordinal • Interval • Ratio

  4. Nominal Level of Measurement • Characteristics of an outcome that fits one and only one category • Mutually exclusive categories such as • Male or Female • Democrat, Republican, or Independent • Categories cannot be ordered meaningfully • Least precise level of measurement

  5. Ordinal Level of Measurement • Characteristics being measured are ordered • Rankings such as #1, #2, #3 • You know that a higher rank is better, but not by how much

  6. Interval Level of Measurement • Test or tool is based on an underlying continuum that allows you to talk about how much higher one score is than another • Intervals along the scale are equal to one another • Example: “Rate your restaurant experience on a scale of 1-7 with 1 = unsatisfactory and 7 = excellent”

  7. Ratio Level of Measurement • Characterized by the presence of absolute zero on the scale • An absence of any of the trait being measured • Examples: • How many kids do you have? (can have 0) • Scores on a test (0 is possible!)

  8. Things to Remember • Any outcome can be assigned one of four scales of measurement • Scales of measures have an order • The “higher” up the scale of measurement, the more precise (and useful) the data are • Use the scale most appropriate for the research task at hand

  9. Classical Test Theory: Os = Ts + E • Observed score • the actual score on a test, scale or measure • True score • theoretical reflection of the actual amount of a trait or characteristic an individual possesses • Error score • part of the score that is random, or the difference between the Observed and True scores Reliability = True Score / (True Score + Error)

  10. Types of Reliability • Test-Retest • Measure of Stability • Parallel Forms • Measure of Equivalence • Internal Consistency • Measure of Consistency • Cronbach’s Alpha (coefficient alpha) • Inter-Rater • Measure of Agreement

  11. Using the Computer • SPSS and Cronbach’s Alpha

  12. How Big is Good Enough? • Reliability coefficients should be positive • 0.0 to 1.0 • General Rules of Thumb… • Test-Retest = .60-1.0 • Inter-Rater = 85% agreement or better • Internal Consistency alpha = .70 – 1.0 High Reliability alone DOES NOT mean you are testing or measuring the right thing!!

  13. Establishing Reliability • Make sure instructions are standardized across all settings • Increase number of items or observations • Delete unclear items • Moderate easiness or difficulty of tests (“middle-of-the-road” strategy) • Minimize the effect of external events

  14. What is the Truth? • Validity • The extent to which inferences made from a test are… • Appropriate • Meaningful • Useful (American Psychological Association & the National Council on Measurement) • Does the test measure what it is supposed to measure?

  15. Types of Validity • Three types of validity: • Content Validity • Criterion Validity • Predictive Criterion validity • Concurrent Criterion validity • Construct Validity

  16. Content Validity • Property of a test such that the test items sample the universe of items for which the test is designed. • How to Establish… • Content Expert • Do items represent all possible items? • How well do the number of items reflect what was taught?

  17. Criterion Validity • Assesses whether a test reflects a set of abilities in a current (concurrent) or future (predictive) setting as measured by some other test. • Concurrent Validity • How well does my test correlate with the outcomes of a similar test right now? • Predictive Validity • How well does my test predict performance on a similar measure in the future?

  18. Construct Validity • Most difficult source of validity to establish • Construct = group of interrelated variables such as... • Aggression • Hope • Intelligence (Verbal, Quantitative, Emotional) • Want your construct to correlate with related behaviors and not correlate with behaviors that are not related.

  19. All About Validity

  20. Validity & Reliability • The “Kissing Cousins” • A test can be reliable but not valid • A test cannot be valid unless it is reliable because… • “A test cannot do what it is supposed to do (validity) until it does what it is supposed to do consistently (reliability).”

More Related