1 / 19

What Does That Mean? Author Selection Of Virtual Patient Metrics

What Does That Mean? Author Selection Of Virtual Patient Metrics. Rachel Ellaway, David Topps, Richard Witham Northern Ontario School of Medicine. Designs for learning. PBL Simulation CAL OSCE Games Virtual Patients Design using patterns, methods, templates. Virtual Patients.

ellery
Download Presentation

What Does That Mean? Author Selection Of Virtual Patient Metrics

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. What Does That Mean?Author Selection Of Virtual Patient Metrics Rachel Ellaway, David Topps, Richard Witham Northern Ontario School of Medicine

  2. Designs for learning • PBL • Simulation • CAL • OSCE • Games • Virtual Patients Design using patterns, methods, templates

  3. Virtual Patients • “an interactive computer simulation of real-life clinical scenarios for the purpose of health professions training, education, or assessment. Users may be learners, teachers, or examiners” Ellaway, Candler et al. 2006 • Response to changing needs • Technological possibilities • Breadth of applications • Current themes

  4. Problem Statement “virtual patients can … be used in learner assessment, but scoring rubrics should emphasize non-analytical clinical reasoning rather than completeness of information or algorithmic approaches. Potential variations in VP design are practically limitless, yet few studies have rigorously explored design issues” Cook, D. and Triola M. (2009) Virtual Patients: a critical literature review and proposed next steps. Medical Education (in press)

  5. Metrics, assessment and feedback • Educational game = rules + experience + simulation + educational design • Educational design: objectives/intervention/feedback/outcomes • Formative: feedback at decision point, outcomes guided • Summative: feedback at end of activity, outcomes unguided • Game rules: agency, feedback, assessment

  6. OpenLabyrinth • Pattern-based: medical model • Narrative-based: timeline, character, motive, causality • Game-based: branching, strategy, scores, counters and rules OpenLabyrinth: open source VP authoring, delivery and feedback system – supports all three forms although without strong templating is more useful for narrative and game based VPs

  7. Diff kinds of OL VP metrics • Reaching end point(s) • Time taken • Number of steps taken • Patient model – survival, pulse, BP • Professional model – DDx, Rx • Other counters (keys, strength, chance factors) • Steps/areas visited or avoided • Sequence of steps • Confidence of decision • Aggregate/function of some or all of the above

  8. OL metrics (time and sequence) time per decision key nodes sequence

  9. OL metrics (counters and sequence) counter value end values change over time: trend, max, min

  10. OpenLabyrinth Design OL three design dimensions: narrative, simulation, game OL three implementation dimensions: topology, rules, content Authoring process: • Deductive – objectives > key points > narrative > CSP > branching > rules > media • Inductive – narrative > CSP > branching > key points > objectives > rules > media Recurring issues with best use of metrics

  11. Author selection of metrics • Focus on critical story path, • Branching distractors, or multiple clues/resources • Typically one successful endpoint + several failure endpoints – few mazes or phases • Counters – time, patient health (typically not money or reputation) • Conditionals to regulate flow rather than measure • Largely formative – summative have tended to resort to tried and tested e.g. key feature problem modes

  12. Four dimensions of Validation • Face Validity - presentation and interface – not a metrics issue • Content Validity - relation to domain and context – not a metrics issue • Predictive Validity - functions as predicted by practitioners/experts • Convergent and Discriminant Validity - performance correlates with other measures as predicted by practitioners/experts

  13. Predictive Validity Techniques • design: review overall options and design • standard setting (modified Angoff – probabilistic expert estimate of performance of minimally passing learners) • suitability: pilot with 3-5 representative candidates to evaluate understandability, accessibility, performability, usability and applicability • runtime: review and validate different ways VP can be executed by a candidate • Experts/authors (problems with non-linear extrapolation of expertise) • Excellent candidates • Minimally passing candidates

  14. Metrics and Standards • SCORM1 has basic tracking (pilot Peter- MedBiq 2005), expectation of more in SCORM2 – watch this space • MedBiquitous VP is an extension of SCORM – tracking is a base functional requirement – nodes visited and counter values in a session • OL implements MVP (at least two variants) • However, no standard currently exists that models objective measures of performance in VPs – tracking is purely for runtime

  15. R&D: telemetric research VERSE (Virtual Educational Research Services Environment) Remote telemetrics tracking and database for Second Life, haptics (Omni), OpenLabyrinth, Mitsubishi light surfaces Generic data tracking model Requires major storage and parsing Creates new opportunities for metrics development and modeling

  16. R&D: telemetric research Network enabled platform Edge services = device + wrapper Heterogenous devices: virtual patients (OpenLabyrinth), mannequins (LaerdalSimMan 3G), light fields (virtualised cameras), 3D visualization (RSV and Volseg), multiple data sources (CMA, Medline) Integrated service model for connecting, controlling and intertwining heterogenous devices (physical, online, endpoint, model, source, renderer, aggregator)

  17. HSVO Service Architecture

  18. Alien VPs and metrics • Physiognomic models – physiognomes • Ontology and AI • Persistent avatars • EHRs as VPs • Data shadows Increasing convergence – augmented by new mashups – geotagging, transponder feeds, twitters

  19. Where next? Current VP models provide rich but relatively unused metrics New VP models produce rapidly increasing dimensions and details Key issues: • testing of validation methods • testing of correlation between VP and real world performance • Identify and separate causal and coincident factors • Develop predictive mapping between VP design, selection and use of metrics and reliability of conclusions Platform and design development still in flux – creativity and opportunity ahead of knowledge Standards follow once evidence base is established and validated

More Related