1 / 25

ENSURING ACCURATE FEEDBACK FROM OBSERVATIONS

ENSURING ACCURATE FEEDBACK FROM OBSERVATIONS. Great Lakes TURN Forum Presentation by : Craig D. Jerald Break the Curve Consulting May 4, 2012. Two Kinds of “Feedback” from Observation. End of Year Based on MULTIPLE Observations Quantitative data (scores) Inform major decisions

Download Presentation

ENSURING ACCURATE FEEDBACK FROM OBSERVATIONS

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. ENSURING ACCURATE FEEDBACK FROM OBSERVATIONS Great Lakes TURN Forum Presentation by: Craig D. Jerald Break the Curve Consulting May 4, 2012

  2. Two Kinds of “Feedback” from Observation • End of Year Based on MULTIPLE Observations • Quantitative data (scores) • Inform major decisions • Teacher: Choose PD to improve an aspect of instruction • Administration: Personnel decisions • Throughout Year Based on SINGLE Observations • Qualitative “coaching conversation” during a “post conference” with teacher following the observation • Informs ongoing improvement of practice by identifying: • Effective practices to extend into future lessons • Less effective practices to improve in future lessons

  3. Post-Observation Feedback Can Be Powerful • Cognitive Science • Feedback is critical for improving practice in any field of endeavor, from music to sports to professions • Recent Education Studies • Taylor and Tyler study of Cincinnati TES • Students of mid-career teachers scored significantly better the year following teacher’s participation • Experimental study of My Teaching Partner • Substantial impact on secondary students’ performance on Virginia assessments year following teacher’s participation • Equivalent of moving from 50th to 59th percentile

  4. Post-Observation Feedback Can Be Powerful “My Teaching Partner” professional development based on the Classroom Assessment Scoring System (CLASS) observation instrument

  5. Post-Observation Feedback Must Be Accurate • Inaccurate Feedback Costs Teachers and Students • If weak practice misclassified as strong, teacher will extend the practice into future lessons • If strong practice misclassified as weak, teacher will not be investing precious time/energy most efficiently • The Challenge: • How can school systems ensure that feedback avoids major errors in classification and instead reflects accurate judgments of practice based on evidence collected during a classroom observation?

  6. Knowledge Development Project • Gates Foundation Partnership Sites • Atlanta • Denver • Hillsborough CO, FL • Los Angeles CMOs • Memphis • Pittsburgh • Prince George’s CO, MD • Tulsa • Additional organizations • DC Public Schools • Tennessee • American Federation of Teachers • National Institute for Excellence in Teaching (TAP System) • University of Virginia (CLASS)

  7. A New Job for Staff: How to Ensure Success? • QUESTION: If you gave someone on your team a critical but very challenging new job to perform, one they had never attempted before, how would you help them succeed in it? • Train them well and give them the right tools for the job • Remove obstacles to success in the field • Monitor their work, and if necessary solve problems

  8. Ensuring Accuracy: Three Key “Action Areas” • 1) Build Observers’ Capacity • Provide observers with the necessary knowledge, skills, and tools to observe accurately • 2) Create Conducive Conditions • Even well-equipped observers can fail if they encounter significant obstacles “in the field” • 3) Monitor and Ensure Quality • Taking the extra step to audit results in order to identify and remediate problems

  9. Building Observers’ Capacity: Training • Understanding the observation instrument Trainers show short video segments to illustrate practice at different performance levels

  10. Building Observers’ Capacity: Training • Typical “flow” of observation training • Understanding the observation instrument • Techniques for collecting appropriate evidence • Aligning evidence to the observation instrument • Practicing evidence collection and scoring • Sources of bias and how to mitigate them • Lessons learned by school systems • Collaborative, with plenty of opportunity for dialogue • Lots of practice with video-recorded lessons • Some practice observing “live” lessons

  11. Building Observers’ Capacity: Certification • Certification assessment following training • Scoring “normed” video-recorded lessons • Memphis: Certification Committee of teachers, principals, and administrators to establish “gold standard” scores for certification videos • Live observations with trainers or experts • Hillsborough: Successfully complete two live observation cycles to satisfaction of trainer • Pass/Fail only or “conditionally certified” category? • Periodic re-certification assessments

  12. Building Observers’ Capacity: Tools • Most important tool: The observation instrument • Must support reliable observation through • Clarity of language • Descriptors and examples • Manageable number of dimensions • Pilot testing, gathering feedback from teachers and observers, refining observation instrument • Other tools for capturing and aligning evidence • PUC Schools: • LiveScribe pens • Evidence Guide

  13. Building Observers’ Capacity: Reinforcement • Even highly calibrated observers can experience “rater drift” over time • Examples of reinforcement include: • Deep-dive trainings for groups of observers focused on critical dimensions of instrument • One-on-one coaching • Paired observations (video or live) • Group calibration sessions (video or live) • P.G. County: Videoconferencing enables large groups of up to 40 observers to co-observe and score live lessons

  14. Conducive Conditions: Manageable Caseload

  15. Conducive Conditions: Manageable Caseload

  16. Conducive Conditions: Manageable Caseload • Decreasing top of ratio (observation workload) • Decreasing number of teachers • Pittsburgh: One-third of experienced teachers per year participate in alternative observation system using peer observation and coaching • Decreasing number of observations • Hillsborough County: “Proportionate” approach requires teachers to be observed from 3 to 11 times per year depending on prior year’s evaluation results • Decreasing minimum time per observation or number of dimensions to be scored

  17. Conducive Conditions: Manageable Caseload • However, keep in mind this MET Project finding:

  18. Conducive Conditions: Manageable Caseload

  19. Conducive Conditions: Manageable Caseload • Increasing bottom of ratio (observation capacity) • Training and certifying more administrators • School-level (assistant principals, etc.) • District-level • Training and certifying a cadre of teacher-leaders • Hillsborough County: Rotating peer and mentor evaluators • DC: Permanent master educators • In both cases, costs less than 2% of overall personnel budget • Other advantages • Leadership opportunity for teachers • Can match observers to teachers’ subject area and grade level • Feedback from observers with recent classroom experience • Taylor & Tyler study of Cincinnati: Net student learning gain

  20. Conducive Conditions: Positive Culture • Communication: Helping teachers and administrators understand observation system and how it can support improvement • Collaboration: Inviting some teachers to help develop or select observation instrument or establish “gold standard” scores for observers’ certification assessments • Calibration: Providing all teachers with opportunities to reach a deeper understanding of the observation instrument so they can begin to calibrate their own vision for effective instruction against it • Coaching & Professional Development: Providing teachers with meaningful opportunities to improve on the practices measured by the observation instrument

  21. Conducive Conditions: Positive Culture • Calibration • TAP System: • During first year, regular PLC meetings focus on understanding the observation instrument (TAP Rubric) • Master and mentor teachers model effective practices in observation instrument in PLC meetings and in teachers’ own classrooms with teachers’ own students • Teachers score their own observed lessons using instrument, with self scores counting toward final evaluation score • Coaching & Professional Development • Hillsborough: Office of Staff Development offers PD courses aligned with specific dimensions of instrument

  22. Monitor & Ensure Accuracy • Analyze data from observations to flag patterns that suggest problems with accuracy • Inter-rater reliability of scores • Anticipated distribution of scores • Alignment with other measures such as value-added • Audit evidence collected by observers to examine alignment with scores • Double score a sample of teachers using impartial observers • Described in the MET Project reports

  23. Monitor & Ensure Accuracy: NIET Example Observers will “re-calibrate” on the Questioning dimension using normed videos

  24. Wrapping Around the School Year

  25. Additional Information • Written report is available available on the Bill & Melinda Gates Foundation website at: www.gatesfoundation.org/college-ready-education/Pages/college-ready-resources.aspx • Craig Jerald can be reached at: • (202) 232-5109 • craig@breakthecurve.com

More Related