1 / 14

Session 1: Responding to Unwarranted Clinical Variation: A Case Jean-Frederic Levesque

Session 1: Responding to Unwarranted Clinical Variation: A Case Jean-Frederic Levesque Chief Executive, Bureau of Health Information. Session 1: Responding to Unwarranted Clinical Variation – A Case Study Measuring Stroke Mortality Variation: What We Learned. 19 June 2014

yepa
Download Presentation

Session 1: Responding to Unwarranted Clinical Variation: A Case Jean-Frederic Levesque

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Session 1: Responding to Unwarranted Clinical Variation: A Case Jean-Frederic Levesque Chief Executive, Bureau of Health Information

  2. Session 1: Responding to Unwarranted Clinical Variation – A Case Study Measuring Stroke Mortality Variation: What We Learned 19 June 2014 NSW Health Symposium • Jean-Frederic Levesque, MD, PhDChief Executive

  3. December 5th release: 30-day mortality Insights into Care report • Acute myocardial infarction, ischaemic stroke, haemorrhagic stroke, pneumonia, hip fracture surgery • NSW results and variation within the state Hospital profiles • Up to 21 pages of content for each hospital Spotlight on measurement • Discussion of the approach and sensitivity analyses Technical supplement

  4. Why are we reporting this measure? Mortality following hospitalisation is reported internationally as part of performance assessments. Mortality ratios provide a piece of the picture and are complementary to other quality and safety measures. Mortality data is influenced by the performance of local systems, not just hospitals, and of multidisciplinary teams. RSMRs are screening tools that provide an indication of where further assessment may be needed. Public reporting of mortality results can catalyse improvements in comprehensiveness and appropriateness of care for patients.

  5. What is this measure about? It compares the ‘observed’ number of deaths in the 30-days following admission with the number of ‘expected’ deaths. • Deaths occurring in-hospital and after discharge are counted using linked data sets. • A statistical model is used to calculate the ‘expected’ number of deaths based on the age, sex and comorbidities of patients. • Cases are attributed to their first presenting hospital during an episode of care. The findings are not appropriate for comparing or ranking hospitals or for identifying avoidable deaths.

  6. Overview of the results

  7. Overview of the results

  8. Funnel plot Ischaemic Stroke • Ischaemic stroke 30-day risk-standardised mortality ratio, NSW public hospitals, July 2009 – June 2012 ∆ μ (∆) Patients are assigned to the first admitting hospital in their last period of care. Data for hospitals with an expected mortality of < 1 are supressed. (μ) Hospitals with < 50 patients are not reported publicly. Deaths are from all causes, in or out of hospital. Data exclude AMI STEMI-not specified (ICD-10-AM I21.9). Source: SAPHaRI, Centre for Epidemiology and Evidence NSW Ministry of Health.

  9. Analysis by peer group Ischaemic stroke Ischaemic stroke 30-day risk-standardised mortality ratio, by peer group July 2009 – June 2012

  10. Hospital profile: Summary dashboard

  11. Hospital profile: Stability of results

  12. Following the report’s release • Various hospitals have considered the results and identified where improvements could be made. • Results were reviewed alongside other quality and safety measures, such as clinical audit and review panels. • Eight clinical settings contacted us to obtain clarifications on the measures or more detailed information. • Future provision of updates and complementary analyses will be provided.

  13. What have we learned? • Mortality is an easily understood outcomes that requires very sophisticated methods to be reported in a fair way • Understanding contextual and system-level factors is important in developing mortality measures (e.g. impact of transfers) • Partnering with clinical leaders supports the development of clinically-relevant measures and the rigorous validation • Internal and public reporting of information acts in synergy to raise awareness and catalyse assessments • Outcomes that are sensitive but crucial for patients can be part of the range of measures reported publicly to support accountability and meaningful discussion

  14. Acknowledgements Kim Sutherland, Director, System and Thematic Reports, Bureau of Health Information Doug Lincoln, Lead Analyst, Bureau of Health Information SadafMarashi-Poor, Senior Analyst, Bureau of Health Information KerrinBleicher, Analyst, Bureau of Health Information Sally Prisk, Graphic Designer, Bureau of Health Information All BHI staff

More Related