Metrics for 211 Centers and Systems
This presentation is the property of its rightful owner.
Sponsored Links
1 / 22

Metrics for 211 Centers and Systems Policy Dialogue with Maribel Marin PowerPoint PPT Presentation

  • Uploaded on
  • Presentation posted in: General

Metrics for 211 Centers and Systems Policy Dialogue with Maribel Marin Executive Director, 211 LA County – CAIRS President 29 th I&R Annual Training and Education Conference May 7, 2007. Policy Objectives. Identify a specific set of metrics that all 211 centers can measure and track.

Download Presentation

Metrics for 211 Centers and Systems Policy Dialogue with Maribel Marin

An Image/Link below is provided (as is) to download presentation

Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author.While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server.

- - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - -

Presentation Transcript

Metrics for 211 centers and systems policy dialogue with maribel marin

Metrics for 211 Centers and Systems

Policy Dialogue with Maribel Marin

Executive Director, 211 LA County – CAIRS President

29th I&R Annual Training and Education Conference

May 7, 2007

Policy objectives

Policy Objectives

  • Identify a specific set of metrics that all 211 centers can measure and track.

  • Provide a clear definition of each metric so that they measure “apples to apples.”

  • Communicate a clear scope for 211 service to the public and for funders.

  • Enable regional, statewide, and national calibration, evaluation, and assessment of 211 services.

What is a metric and what is it s value

What is a metric and what is it’s value?

  • A metric is a measure of activity or performance that enables the assessment of outcomes.

  • Metrics can help to answer key questions about operational effectiveness:

    • Are long term goals and objectives being achieved?

    • What does success look like?

    • How satisfied are callers with services?

    • How important is the service to the community?

    • How effective are managers and specialists?

What are the benefits

What are the Benefits?

  • Enhanced decision making – goals can be set for desired results, results can be measured, outcomes can be clearly articulated.

  • Improved internal accountability – more delegation and less “micro-management” when individuals are clear about responsibilities and expectations.

  • Goals and strategic objectives are meaningful – tracking progress enables the evaluation of planning efforts and can aid in determining whether a plan is good or not.

    Source: Wayne Parker, Strategic Planning 101 – Why Measure Performance? Workstar Library 2003

Why metrics matter for 211

Why Metrics Matter for 211?

  • Most 211s collect or have the ability to collect voluminous amounts of data through their Call Management Systems (CMS), Automated Call Distribution systems (ACDs), and/or their I&R software.

  • Much data is collected for purposes of complying with AIRS standards, particularly for agency accreditation or reporting to boards, funders, and/or contracts.

  • We have yet to realize the full value of this data for benchmarking (comparing performance against goals/mission and/or industry peers) and for its potential for aggregation across regional, state, and national levels.

Why metrics matter for 2111

Why Metrics Matter for 211?

  • Through benchmarking and aggregation of data across the field, a story can be told about the value of 211 service as a social safety net for the entire nation on a daily basis and during times of crisis and disaster.

  • The national 211 business plan clearly calls out the need for creating a unified system through the development of industry standards in order to avoid misuse of 211 (too broadly or too narrowly defined) and create “sustainability for the total system.”

  • Funding, funding, funding – aggregate data and demonstrable performance outcomes enable pursuit of system-wide funding strategies.

Airs data requirements reports and measures

AIRS Data Requirements:Reports and Measures

  • Service Requests

  • Referrals Provided

  • Service Gaps

  • Demographic Data

    • Zip Code

    • City

    • Age

    • Gender

    • Language

    • Target Population

    • First Time/Repeat Caller

  • Follow up

Airs standards national reporting

AIRS StandardsNational Reporting

  • Total Calls Answered/Handled

  • Services Requested

Commercial call center metrics

Commercial Call Center Metrics

Call center trade journals* consistently identify the following Key Performance Indicators (KPIs) as leading performance metrics or benchmarks:

  • Service Level (% of calls responded to within a specified timeframe such as 80% in under 60 seconds)

  • Speed of Answer (time caller waits in queue before live answer)

  • First-Call Resolution (% of callers helped with one call not requiring repeat call on same issue)

  • Adherence to Schedule (actual and scheduled work by time of day and type of work – handling calls, attending meetings, coaching, breaks, etc.)

  • Forecasting Accuracy (measured on 2 levels: actual vs. forecasted call volume for hiring/recruiting and for existing staff schedules)

  • Handle Time (includes talk time, hold time, and after call work)

  • Customer Satisfaction

  • Cost per Call

  • Abandonment Rate (calls in queue disconnected by caller)

    *ICMI Call Center Magazine, Contac Professional, Benchmark Portal, Customer Operations Performance Center Inc. (COPC Inc.),

Govt non profit gold standard

Govt/Non-Profit Gold Standard

Benchmark Portal (2005) Govt & NonProfit industry benchmark report: Best-in-class call center performance”

Benchmark Portal (2003) Govt & NonProfit Industry Benchmark Report

Best practices study on customer service city of los angeles

Best Practices Study on Customer Service – City of Los Angeles

Common 211 metrics

Quanity of Service

Number of Calls Handled/Answered

Referrals Provided to Callers

Call Length

Service Gaps

Quality of Service

Speed of Answer/Service Level

Abandonment Rate/Dropped Calls

Follow up Rate

Caller Satisfaction

Common 211 Metrics

211 recommended best practices

211 Recommended Best Practices

  • Agency Accreditation (AIRS Standards compliance)

  • CIRS/CRS certification

  • 24/7 service

  • Universal Access – cell, TTY, cable, web

  • Service provided by a live, trained I&R Specialist without being required to leave a message or hang up and dial a separate number.

  • Multilingual service

  • Call monitoring

211 la county metrics

211 LA County Metrics

Standardizing 211 metrics

Standardizing 211 Metrics

Attempts to generate national reports on performance and service outcomes to support funding requests have been challenged by the lack of common definitions for basic measurements.

Many 211 systems are attempting to bridge the differences through data standardization processes – just a few referenced here:

  • Texas I&R Network: Dr. Sherry Bame, Texas A&M University

  • United Ways of Ontario, Canada: Michael Milhoff, consultant

  • WIN 211 (Washington and Oregon): Karen Fisher, Associate Professor, University of Washington Information School (211 Outcomes Registry project)

  • IN 211 (Indiana): through it’s 211 operations manual development process

  • AIRS Accreditation Committee: through 2007 revised standards led by Faed Hendry, chair and Manager of Training and Outreach at Findhelp Information Services of Toronto.

Challenges to standardization

Challenges to Standardization

  • Limited data collection/reporting capability

  • Lack of common terms and definitions

  • Lack of standard call types/needs lists

  • Inconsistent measurement/reporting frequencies

  • Too much variation in data collection fields among I&R software systems

  • Varying data sources: I&R software vs. CMS/ACD

  • Too much data collection not enough data analysis

Key consistency questions

Key Consistency Questions

  • How is a call defined?

  • Are demographics taken on caller or client?

  • Are multiple clients on one call counted as a single transaction? A single call?

  • Is age data collected by number of years, age range/group, birth date?

  • Is location data related to caller or client? Is it location of call or location of residence?

  • Data reported daily, monthly, quarterly, annually?

Why metrics matter

Why Metrics Matter

  • What you don’t measure doesn’t count.

  • What gets measured gets done.

  • If you don’t measure results, you can’t tell success from failure.

  • If you can’t see success, you can’t reward it.

  • If you can’t reward success, you’re probably rewarding failure.

  • If you can’t see success, you can’t learn from it.

  • If you can’t recognize failure, you can’t learn from it.

  • If you can demonstrate results, you can win public support.

    Source: Aiming to Improve, Audit Commission, Reinventing Government, Osborne and Gaebler

Next steps

Next Steps

  • Join 211 North America Metrics Project

  • Share findings of your region’s work

  • Weigh in on which metrics should be collected by 211s

  • Voice your position on how the scope of 211 service should be defined

  • Participate in discussions on:

    • Common definitions and terms

    • Standard call types/needs lists

    • Measurement/reporting frequencies

    • Demographic data best practices

Sources and references

Sources and References




  • (Queue Tips)

  • Customer Operations Performance Center Inc. (COPC Inc.) is a leading authority on operations management and performance improvement for buyers and providers of customer contact center services.

  • - The Interagency Working Group (IAWG) on U.S. Government-Sponsored International Exchanges and Training created in 1997 for improving the coordination, efficiency, and effectiveness of United States Government-sponsored international exchanges and training.

  • Section on 211 Standards

Sources and references1

Sources and References


  • National 211 Benchmark Survey – conducted by 211 San Diego ( in 211 standards section)

  • Devising, Implementing, and Evaluating Customer Service Initiatives (1/30/2007) – Office of Citizen Services & Communications: US GSA

  • Agency Experiences with Outcomes Measurement – United Way of America (2000)

  • WIN 2-1-1 Performance Evaluation and Cost-Benefit Analysis of 2-1-1 I&R Systems – University of Washington, Information School – Karen Fisher/Matt Saxton (2005)

  • National 211 Business Plan – AIRS/UWA (2002)

  • 211 Data Entry & Database Coding for TIRN – Texas A&M University – Dr. Sherry I. Bame

  • 211 Across California by 2010 – Business Plan: 211 California Partnership with Sadlon & Associates, Inc.

  • AIRS Standards for Professional Information and Referral, Version 5.1 (2006)

  • Operations Manual – Indiana 211 Partnership, Inc. - First Approved 7/09/02; last revision 5/9/06

  • COPC-2000 CSP Gold Standard Release 4.1 (January 2007)

Contact information maribel marin executive director 211 la county 626 350 1841 mmarin@211la org

Contact Information:Maribel MarinExecutive Director211 LA County(626) [email protected]

Real People. Real Answers. Real Help.


Serving Los Angeles County since 1981

  • Login