140 likes | 268 Views
REAQ. Report on E-Assessment Quality A JISC project in eLearning. Lester Gilbert Gary Wills Bill Warburton Veronica Gale. March 2009. The big picture. Report process. The team. Management group Lester Gilbert, ECS, University of Southampton , PI
E N D
REAQ Report on E-Assessment Quality A JISC project in eLearning Lester Gilbert Gary Wills Bill Warburton Veronica Gale March 2009
The team • Management group • Lester Gilbert, ECS, University of Southampton , PI • Dr Gary Wills, ECS, University of Southampton • Expert consultants group • Cliff Beevers, Heriot-Watt • Paul Booth, Question Tools • John Kleeman & Greg Pope, Questionmark • Harvey Mellor, IoE • Chris Ricketts, Plymouth • Denise Whitelock, OU • The workers • Consultant researcher (Veronica Gale) • Bill Warburton, iSolutions, University of Southampton
Interviewees • HEIs • Heriot-Watt • University of Southampton • Newcastle University • University of Plymouth • The Open University • Edinburgh University • Institute of Education • Cambridge Assessment • SQA • Question Tools (Network Rail) • VrijeUniversiteit Amsterdam
Questions asked (1) • What denotes ‘high quality’ in summative e-assessment? • What steps do you follow to create and use summative e-assessment? • How do you ensure e-assessment: • reliability, • validity, • security, and • accessibility? • How does the process of creating good quality e-assessment differ from the process of creating traditional assessment?
Questions asked (2) • Please give us examples of good e-assessment; why are these ‘good’? • When you have heard of poor e-assessment, what has made it ‘poor’? • What feedback have you received from students who have taken e-assessments? • What advice would you give to others using summative e-assessment? • What research or other work has informed your thinking about summative e-assessment? • What further research would you like to see conducted?
Expectations… We thought the interviewees would talk about: … psychometric measures… … intended learning outcomes … … appropriate standards … … delivery issues… … and if we were lucky, capability maturity …
We heard a lot about… Delivery issues: infrastructure, support, and how things go wrong …
But not much about… Point biserials, Cronbach Alphas, Kuder-Richardsons Practice standards Content validity, Conformance to ILOs Metrics, Capability maturity
And when we did hear… Difficulty coefficients / facility values were often inappropriately used…
Possible recommendations • Toolkit for end-to-end e-assessment creation, delivery, and evaluation • Guidelines & workshops: • Psychometrics • Validity through ILOs • Capability maturity • Research on the psychometrics of item banks • Research on quality metrics for multi-staged and stepped constructed response questions • Exemplars repository
Thanks! Questions, comments, discussion…