1 / 36

Chapter 7

Chapter 7. Flashcards.

harmon
Download Presentation

Chapter 7

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Chapter 7 Flashcards

  2. overall plan that describes all of the elements of a research or evaluation study, and ideally the plan allows the researcher or evaluator to reach valid conclusions (e.g., questions or hypotheses to be addressed, number and types of participants to be included, number and types of variables to be studied, collection and analysis of data) Evaluation/research design

  3. family of research and evaluation designs characterized by the systematic repeated measurement of a client’s outcome(s) at regular, frequent, predesignated intervals under different conditions (baselineandintervention), and the evaluation of outcomes over time and under different conditions in order to monitor client progress, identify intervention effects, and more generally, learn when, why, how, and the extent to which client change occurs. Also known as single-subject designs, single-system designs, N = 1 designs, or sometimes time series or interrupted time series designs Single-case design

  4. period of time during which an outcome is measured repeatedly in the absence of an intervention in order to (1) describe the naturally occurring pattern of outcome data (e.g., level, trend, variability) and (2) determine the effect of an intervention on that outcome. Typically symbolized by the letter A Baseline phase

  5. period of time during which an intervention is implemented while an outcome is measured repeatedly Intervention phase

  6. period of time after an intervention has ended during which outcome data are collected to determine the extent to which a client’s progress has been maintained. Also known as a maintenance phase Follow-up phase

  7. a variable (e.g.,intervention) that produces an effector is responsible for events or results (e.g., outcome) Cause

  8. change in one variable (e.g., outcome) that occurred at least in part as the result of another variable (e.g., intervention) Effect

  9. measure of the strength of the relationship between variables (e.g., effect of an intervention on an outcome, as quantified by any one of a number of different statistics). Effect size

  10. conclusion based on evidence and reasoning that one variable (e.g., intervention) causes another (e.g., outcome) Causal inference

  11. accuracy of conclusions based on evidence and reasoning about the presence, direction, and strength of relationships between variables (e.g., outcome is different during baseline than intervention). The ability to establish that one variable (e.g., intervention) is related to another (e.g., outcome) is a requirement for inferring that one variable caused another Statistical conclusion validity

  12. accuracy of conclusions based on evidence and reasoning about causal relationships between variables (e.g., extent to which an intervention, as opposed to other factors, caused a change in an outcome). Internal validity

  13. uncertainty about which of several events or processes caused an outcome Causal ambiguity

  14. variable that is associated with the independent variable inadvertently influences the outcome, and consequently makes it difficult to determine the effect of the intervention on the outcome (e.g., an unknown event that occurs during intervention but not baseline and causes the pattern of outcome data to change from baseline to intervention). Also known as a confound or confounding variable, and such a result is said to be confounded Extraneous variable

  15. plausible reasons for a relationship between an intervention and an outcome, other than that the intervention caused the outcome. Also known as alternative hypotheses. The ability to rule out alternative hypotheses is a requirement for inferring that one variable (intervention) caused another (outcome) Alternative explanations

  16. reasons why it might be partly or completely wrong (i.e., invalid) to conclude that one variable (e.g., an intervention) caused another (e.g., an outcome). See also Ambiguous temporal precedence, History effect, Instrumentation effect, Maturation effect, Regression effect, and Testing effect Threats to internal validity

  17. potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by an external event that occurs at the same time as the intervention (e.g., a student who has trouble completing his homework for a new teacher improves his performance as he becomes accustomed to her and her expectations, and the improvement is misinterpreted as being due to the rewards he earns working with a social worker). History effect

  18. potential threat to internal validity in which an apparent change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by a change in how the outcome is measured (e.g., an older man’s weight stabilizes when he begins weighing at his physician’s office rather than at home). Instrumentation effect

  19. potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by naturally occurring changes in clients over time (e.g., a toddler’s tantrums diminish not in response to an intervention, but due to maturing out of the terrible twos, or a child outgrows enuresis). Maturation effect

  20. potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by repeated measurement of the outcome (e.g., a pretest about health behaviors may sensitize a client to the need to make changes in diet and exercise). See also Fatigue effect and Practice effect. Testing effect

  21. deterioration in an outcome caused by fatigue associated with repeated measurement of the outcome (e.g., a mother who is self-recording each instance of time-out with her preschooler reduces those time-outs to avoid recording the behavior) Fatigue effect

  22. improvement in an outcome caused by repeated measurement of the outcome (e.g., taking multiple practice exams may improve a student’s score on the Graduate Record Exam simply because he or she becomes familiar with the format of the exam) Practice effect

  23. potential threat to internal validity in which change in an outcome could be misinterpreted as an intervention effect, when in fact it is caused by the tendency of an individual with unusually high or low scores on a measure to subsequently have scores closer to the mean (e.g., clients who are depressed frequently seek help when they have hit bottom, and their scores are likely to improve somewhat in the following weeks, even without intervention). Also known as regression toward the mean Regression effect

  24. result due to the order in which different interventions are administered (e.g., a couple may be more successful in an intervention designed to increase their pleasant time together each day if they first complete an intervention designed to increase their reflective listening). Also known as a sequence effect Order effect

  25. single-casedesign(arguably) consisting of aninterventionphase(B) during which the outcome is measured repeatedly B-only design

  26. single-casedesign(arguably, since there isn’t repeated measurement during baseline) consisting of one pre-interventionoutcome measurement followed by an interventionphase (B) during which the outcome is measured repeatedly B+ design

  27. two-phase single-case design consisting of a pre-intervention baseline phase (A) followed by an intervention phase (B) A-B design

  28. three-phase single-case design consisting of a pre-intervention baseline phase (A1); an intervention phase (B); and a second baseline phase (A2) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern A-B-A design

  29. three-phase single-case design beginning with the intervention phase (B1), followed by the withdrawal of the intervention (A) to determine if the outcome changes in the absence of the intervention, and reintroduction of the intervention (B2) to see whether the initial intervention effects are replicated B-A-B design

  30. four-phase single-case design consisting of a pre-intervention baseline phase (A1); an intervention phase (B1); a second baseline phase (A2) in which the intervention is withdrawn to determine if the outcome reverses to the initial baseline pattern; and a reintroduction of the intervention (B2) to see whether the initial intervention effects are replicated. Also known as a reversal or withdrawal design A-B-A-B design

  31. single-casedesignthat begins with a baseline during which the sameproblemis measured for a single client in two or more settings at the same time. Baseline is followed by the application of theinterventionin one setting while baseline conditions remain in effect for other settings, then the intervention is applied sequentially across the remaining settings to see whetherinterventioneffectsare replicated across different settings Multiple baseline across settings design

  32. single-casedesign that begins with a baseline during which the same problem is measured for two or more clients at the same time in a particular setting. Baseline is followed by the application of the intervention to one client while baseline conditions remain in effect for other clients, then the intervention is applied sequentially to remaining clients to see whether interventioneffects are replicated across different clients Multiple baseline across subjects (clients) design

  33. single-casedesign that begins with a baseline during which two or more problems are measured at the same time for a single client in a particular setting. Baseline is followed by the application of the intervention to one problem with baseline conditions remaining in effect for other problems, then the intervention is applied sequentially to the remaining problems to see whether interventioneffects are replicated across different problems Multiple baseline across behaviors (problems) design

  34. three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase (C) in which a new intervention is introduced in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-C design

  35. three-phase single-case design consisting of a pre-intervention baseline (A); an intervention phase (B); and a second intervention phase in which a new intervention (C) is added to the first intervention in response to the failure of the first intervention to produce sufficient improvement in the outcome A-B-BC design

  36. accuracy of conclusions based on evidence and reasoning about how well a causal relationship applies across or beyond the people, settings, treatment variables, and measurement variables that were studied (e.g., extent to which a causal relationship between an intervention and outcome is the same with different people) External validity

More Related