- 56 Views
- Uploaded on
- Presentation posted in: General

Renewing Health Concertation Meeting September 9, 2011 Brussels Gold or Bronze?

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

Renewing Health Concertation Meeting

September 9, 2011

Brussels

Gold or Bronze?

JRC â€“ IPTS â€“ IS Unit

Cristiano Codagnone

The views expressed in this presentation are those of the author and do not (necessarily) reflect the position of the European Commission. Neither the Commission nor any person acting on behalf of the Commission can be hold responsible for the use which is made of this presentation

- Mr. Smith suffering from a headache and taking an aspirin
- What happens after taking aspirin: Y1
- And What, had he not taken the aspirin: Y0 (counterfactual)
- Causal effect under counterfactual â€˜successionistâ€™ concept of causation: Y1- Y0 with ways to deal with the fact that Y0 is not observable

- Let us look in formalised but simplified notation to the question, first using a typical regression equation used for estimation:
- Yi =Î±Xi +Î²Ti +Îµi. (1) where:
- T is a dummy variable with value 1 for individuals receiving treatment and 0 for those not receiving.
- X indicate what is generally referred to as â€˜a bunch of Xs observed characteristicsâ€™ (confounders)
- whereas Îµ is an error term reflecting unobserved characteristics that also affect Y

- The problem for the estimation of this equation springs from the fact:
- Intervention can entail a purposive definition of who is eligible for the intervention and who is not;
- Individuals can self-select into the intervention

- These two are sources of unobserved feeding the error term Îµ, which ends up containing variables that are correlated with the dummy variable T, in formal notation cov (T, Îµ) =Ì¸ 0
- This is, however, a clear violation of the orthogonality assumption of ordinary least square regression: independence of regressors from the error term.
- Without further elements addressing this problem, the causal parameter Î² cannot be estimated.

- Now let us look at the same problem using a more conceptual notation of equation (2):
- D = E(Yi(1) | Ti= 1) â€“ E(Yi(0) | Ti= 0) (2) where:
- E indicate the average (linear operator), Yi is the general outcome variable for treated individuals (Ti = 1), and the value of Yi under treatment is represented as Yi (1). For non-treated, Ti = 0, and Yi can be represented as Yi (0).
- Then in (2) the average effect D of the intervention is posited equal to the differences between the averages outcomes for treated and non-treated.
- Because treated and non-treated might not be similar prior to the intervention and the difference above cannot be attributed solely to the intervention.

- Now let us both add and subtract (nothing change) in equation (2) the expected outcome for non-treated had they been treated â€” E (Yi(0) / Ti = 1):
- D = E(Yi(1) | Ti= 1) â€“ E(Yi(0) | Ti= 0) + [E(Yi(0) | Ti = 1) â€“ E(Yi(0) | Ti = 1)]
- Or= [E(Yi(1) | Ti= 1) â€“ E(Yi(0) | Ti = 1)] + [E(Yi(0) | Ti = 1) â€“ E(Yi(0) | Ti = 0)]
- So D= ATE+B
- ATE is the average treatment effect and B is the selection bias. Because one does not know E (Yi(0) | Ti = 1), one cannot calculate the magnitude of selection bias.

- As a result, if one does not know the extent to which selection bias makes up D, one may never know the exact difference in outcomes between the treated and the control groups.
- Perfect randomisation solves the selection bias because it ensures perfect comparability between treated and non-treated and they would have the same expected outcome in the absence of the intervention.
- This means that E[Yi(0)|Ti = 1] is equal to E[Yi(0)|Ti = 0] that applied to the equation above lead B to be equal to zero.
- This is the main reason why experimental evaluation design through randomisation has been considered for a long time the golden standard

- Since Fisher (1926), randomisation as golden standard:
- Experimental (control / manipulability of selection process)
- Treated and non treated are equivalent in all relevant respects but different wrt the probability of being exposed to the intervention
- Selection bias eliminated and internal validity of estimates ensured

- Randomized experiments as the bronze standard, Journal of Experimental Criminology, (2005) 1:417-433
- RCT balance treatment and control groups on potential confounders to produce unbiased estimates of causal effects. No other research design confer this benefit so reliably
- Random assignment: on average balance two groups on all possible confounders, this means that the estimation of the term X can be un biased.
- Average effects are substituted to individual effects: unlikely that every unit would have precisely the same response to an intervention. Scientifically it makes sense to look at averages
- Both critics and supporter are biased: expectations on RCT unrealistic

- Yet, there is a catch. You must meet the conditions of no interference or SUTVA: treatment and condition to which a unit is assigned has no impact on the response of another unit.
- Policy problem: key is to unambiguously manipulate: if response depends on the pattern of assignment even if it is random then you lose control
- If SUTVA is violated there is no statistical fix.
- A number of benefits but only if conditions are met and on important assumptions
- Easy to explain
- Face validity
- Easy to analyse

â€œRandomised experiments rest on more complicated, subtle, and fragile foundations than some researchers appreciate. Proper implementation of randomized experiments is demanding. Textbook requirements are rarely met. Thus, randomized experiments are not the gold standard. But if the truth be told, there is no gold standard.

Conclusions

- Are Renewing Health et al interested only in M&OE?
- I YES, THEN:
- MAST and current work is more than enough, Implementation readiness assessment is straightforward, mostly â€˜yesâ€™ answers to brainstorming questions

- IF NO (ALSO IE), THEN:
- A lot is still missing, choices are needed, answers to brainstorming questions change radically

- I YES, THEN:
- Impact Evaluation (IE) is not simply about data and indicators: define, identify, and estimate counterfactuals

- Which is the unit of analysis (i) ?
- Individuals/families?
- Health and social care producing units?
- Administrative territorial units?
- All of them? That is we have (i), (j), and (k) units of analysis

- Which is the time unit (t)
- When is t+1? (first time period after which intervention should have had an effect)
- Is there also t+2, t+3? (continuous treatment with longitudinal explication of effects)
- How do we define t â€“ 1? ( the pre-intervention period on which baseline data are available)

- Which is/are the outcome variable(s) Y [ issue not orthogonal to (i)]: so, are we estimating Yit or Y1-nijkt?
- Are the SUTVA (Stable Unit Treatment Value Assumption ) conditions met?
- Only one form of treatment and of non treatment for each unit
- No interference across units (no contagion effect)

- Is the CS (Common Support) condition (comparing the comparable) respected across time and space ?

- Everything is possible, complex methodological solutions are available
- Counterfactual IE set up can be extended to situations where:
- Time is individual specific
- Units can be observed at two and ,possibly more, different times
- There are multiple treatment states, corresponding to different types of treatment or levels of treatment intensity, or where treatment is continuous

- Counterfactual IE set up can be extended to situations where:
- Yet, they are demanding and increased technical complexity may come at the cost of internal or external validity (back to â€˜castles in the airâ€™?)
- Counterfactual IE should be applied only to max 2 well defined outcoems.:
- The most strategic outcome aimed at (addressing the direst societal challenge): Policy-driven
- An outcome whose analysis would greatly advance our knowledge of what works: Science-driven

- This requires thinking bigger than simply indicators and data: make strategic choices followed by theoretical and methodological reasoning
- Avoid repeating the â€˜one mile wide but one inch deepâ€™ mistake over and over again (especially if time and budget constraints are cogent)

- Convergence Telecare and Telehealth?
- Yes for M&OE with limited obstacles
- It depends on the selection of the outcome variable(s) for IE

- Single methodology independent of duration and technology used?
- Probably yes for M&OE
- Very challenging, if not impossible, for IE

- eReadiness self-assessment mandatory?
- Very feasible and unproblematic

- Is evaluation worth pursuing? Yes, as long as it also include some IE
- What can be done within the current funding of running projects?
- Extensive M&OE (but less than the full list from MAST)
- Very selective and well defined IE, recalling that the best is the enemy of the good