1 / 10

Thoughts on reconciling international comparative studies and national assessments

Thoughts on reconciling international comparative studies and national assessments. Harvey Goldstein Centre for Multilevel Modelling Graduate School of Education University of Bristol. A paradox?. An apparent fundamental paradox of comparative international achievement studies:

freira
Download Presentation

Thoughts on reconciling international comparative studies and national assessments

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Thoughts on reconciling international comparative studies and national assessments Harvey Goldstein Centre for Multilevel Modelling Graduate School of Education University of Bristol

  2. A paradox? • An apparent fundamental paradox of comparative international achievement studies: • They strive to achieve cross-country test ‘equivalence’ • Unique country content (items) down-weighted or excluded • This implies less relevance for each individual country compared to a national assessment designed to measure response to a country’s curriculum • This down-weighting can also result from translational problems to which there may beno formal solution • Thus, the more ‘comparable’ the test in the sense of the items performing similarly (difficulty, discrimination) across countries,the less relevant they may become for comparisons in terms of any given country curriculum, and therefore less useful for policy purposes. • It would appear that no formal solution possible – unless we look at the wider context that is exemplified by claims for PISA.

  3. PISA and policy. • OECD suggests that PISA is useful for informing educational policy decisions and that governments (who are funders of OECD) can draw lessons from PISA comparisons. • PISA also states that it seeks to measure performance that is not tied to any particular curriculum, and this is consistent with the item selection procedure. • These two positions are reconcilable if (and only if) the aim of OECD is to move towards a common international curriculum. Encouraging governments to optimise their country’s performance on PISA tests then becomes an indirect attempt to move in this direction. • If successful OECD would indeed become a kind of World Ministry of Education. • In fact it already promotes policies across countries that, it suggests, are based upon PISA results in terms of ‘skills’ etc.

  4. Relating the national to the international • Insofar as national assessments do reflect the diversity of national curricula, how can we further explore the actual relevance of PISA for each country? • Utilise curriculum ‘experts’ to describe the extent of the test relevance to each curriculum to inform comparisons. • Worthwhile if systematic, but can be subjective and contestable • Tends to describe relevance of the items as designed rather that as they perform in practice. • If you could link national (locally relevant) to international assessments you could then sensibly use the information to compare countries in terms of those components which might be comparable.

  5. Another perspective Suppose we took each PISA test item or group of items and deliberately redesigned it specifically to be relevant to a given target country, for example by using a context familiar in that country, so possibly changing its relative difficulty and also in terms of its relationship with other items (discrimination) or variables. Remembering also that there are typically several dimensions underlying each PISA (and national assessment) scale. We could in principle do this experimentally for, say, a test of reading comprehension in PISA. While an interesting experiment it seems unlikely, not least because of the effort and expense involved. Nevertheless, it would throw interesting light on the inherent variability of the item responses that may be imposing ‘random’ over time fluctuations in country scores. Note that this is not the same as taking PISA items to use within a national assessment framework.

  6. Linking methods An alternative to studying items is to utilise existing national assessments by linking to international studies. I look at different models for such linkage then and their relative strengths and benefits. Simplest and most efficient is linking via students: A suitably large subset of students responding to PISA and national assessment. ISSUES Timing differences between testings Numbers of schools and students ADVANTAGES The individual relationships between items and scales can be studied in relation to difficulty and associations with other variables measured at individual and school level. PISA results can be related to assessment s made at same and prior and later ages.

  7. Linking by Schools • We may not be able to link at student level for practical reasons = e.g confidentiality, but could be overcome: • If not, next best is to use some of the same schools in PISA as in national assessments (NA) – straightforward if all schools are in NA. • Relationships can be measured at school level but limited. Thus, overall difficulties can be compared, school rankings compared, but student level relationships cannot be inferred. • What analyses can we use given linking is avalable?

  8. Example I • Standardised PISA reading score (y) related to National Assessment (NA) reading attainment (x) and gender (z) at similar times • - country 1 • - country 2 • So if or non-zero then PISA differentially applicable to boys and girls. • Additional if y, then implies PISA gender differences are not the same after adjusting for each country – could then explore detailed responses and corresponding relationships at subtest level and to explore items to see whether this can be explained in terms of item wording, contexts etc. • We can study other factors, separately or in combination.

  9. Example II • We can also see whether relative amount of between-school or between-student variation is different for NA and PISA. If e.g. between school variation is less for PISA than NA then it implies that the more generic ‘skills’ being measured by PISA show less diversity across the system than the responses to the actual curriculum. • This can then be further studied to see if this is the case for subgroups (gender, ethnic group etc.) These approaches provides a robust empirical approach to the comparability problem.

  10. Some implications for PISA, TIMSS, etc • Full openness of item test pool to researchers and the public so that proper comparisons can be made. This would be welcome anyway since it is just good practice. • Planned coordination of national with international assessments. • Resources to be made available for such analyses • If PISA itself were longitudinal this would be even more interesting, but does require extra resources

More Related