1 / 40

Path to Accelerated Completion and Employment

Path to Accelerated Completion and Employment. Evaluation Meeting July 31, 2012. New Growth Group. New Growth is a full-service evaluation firm specializing in postsecondary education and workforce development. Christopher Spence, Evaluation Project Manager

Download Presentation

Path to Accelerated Completion and Employment

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Path to Accelerated Completion and Employment Evaluation Meeting July 31, 2012

  2. New Growth Group • New Growth is a full-service evaluation firm specializing in postsecondary education and workforce development. • Christopher Spence, Evaluation Project Manager • Joel Elvery, PhD, Data Analysis • Partnering with Corporation for a Skilled Workforce for implementation assessment • Holly Parker, Ed Strong, LeiseRosman

  3. Goals • Measure the impact of strategies on student outcomes • Capture the variety of approaches implemented for each strategy in the state • Contribute to continuous improvement • Comply with USDOL evaluation requirements

  4. Approach • Two parts: • Impact assessment • Implementation assessment • Two audiences • USDOL – Reporting and compliance • PACE colleges – More detail to ID best practices • Approach tailored for each strategy in the original proposal

  5. Impact Assessment • Before-and-after research design Measures - Before Measures - After • Key Measures • Academic Progress Measures • Program Completion Rates • Employment Outcomes

  6. Implementation Assessment • Documentation of approaches at each college and how implemented • Interviews, questionnaires, etc. • Informs initiative continuous improvement efforts in later stages • Sets the stage for future student success agenda

  7. The Strategies

  8. First Year Timeline Aug - 13 Nov - 12 Feb - 13 Sep - 12 May - 13 • 6/21 College data reports • 8/14 Quarterly report • 8/16 Second semester roll-up • 9/21 College data reports • 5/15 Quarterly report • 8/3 Surveys Employer Engagement, Developmental Education, Streamlining • 9/1 Comparison cohorts defined • Program Launch • 11/14 Quarterly Report • 11/ 14 Annual Report • 1/21 College data reports • 2/14 Quarterly report • 3/15 First semester roll-up • Notes: • Colleges still provide monthly progress reports to NWACC • Expect implementation assessment activities closer to end of first semester

  9. Questions?

  10. Contact Information • Project Manager: Chris Spence, 216.767.6262, cspence@newgrowthplanners.com • Impact Assessment: Joel Elvery, PhD: 216.375.6777, jelvery@newgrowthplanners.com • Implementation Assessment: Holly Parker, 734.769.2900, hparker@skilledwork.org

  11. PACE Impact Assessment

  12. Data plan • Quantitative evaluation design • Comparison cohort plan • Data requested • Data submission

  13. Quantitative evaluation design • Two purposes • Meet DOL requirements • Inform stakeholders whether new approaches are increasing student success • Using before-and-after comparison • Focusing on cohorts engaged in targeted programs in Fall 2012 vs. those in Fall 2010 • More than what’s needed for DOL requirements

  14. Comparison cohort plan • Where possible, will use past cohorts from targeted programs as comparison group • Gathering comparison data from ADHE • Except for some developmental education metrics not in ADHE data • New programs or dramatically shortened programs will have to be matched to other similar programs • DOL convening in early August

  15. Comparison cohort plan • Next steps on comparison cohort plan • Learn about targeted programs & their duration • Develop groupings of programs • Write up cohort strategy for DOL • Get DOL approval • Inform colleges of any additional data need to provide

  16. Means of Data Collection

  17. Data required from colleges • Test scores & placement of students involved in assessment test preparation • Prior Learning Assessments • Demographics of students in targeted programs of study • Completion of developmental education requirements for students in targeted programs • Historic data on developmental education progress for past cohorts

  18. Data on PREP Participants • Who should be included • Every student who uses assessment test preparation provided in conjunction with PACE grant, regardless of whether in targeted program • What need to know • Identifying variables • Type of assessment test, placement before & after readiness course • For math, reading, & English assessments

  19. Data on PLA Participants • Who should be included • Everyone who attempts to get credit through a prior learning assessment • What need to know • Identifying variables • Total credit hours earned through PLA • Credit hours earned through each of the following • Portfolio • Standardized test • Local test • Training

  20. Questions on PREP or PLA data?

  21. 1 time student data • Who should be included • All students enrolled in a targeted program of study • Includes students who began prior to Fall 2012 who are still enrolled • What need to know • Identifying variables • Student demographics from intake form • Developmental ed. placements • Whether have completed developmental ed.

  22. Term data • Data to be reported each term for each student • Who should be included • All students enrolled in a targeted program of study • Includes students who began prior to Fall 2012 who are still enrolled • What need to know • Identifying variables • Whether taking Technical Math & number of modules have to take • Whether changed program of study & what new program of study is • Whether completed developmental ed. requirements

  23. Program-level data • What programs should be included • Each targeted program included in PACE • A separate row for each different one • What need to know • Identifying variables • Credit hours before & after redesign • 2-year dev. ed. math, reading, & English completion rates for cohorts from Fall 2008, Fall 2009, & Fall 2010 • 2-year college-level math, reading, & English completion rates fro cohorts from Fall 2008, Fall 2009, & Fall 2010

  24. Developmental education worksheet • One for each targeted program of study • Need to know course numbers for • Redesigned dev. ed. classes • Technical math • Past courses that students would have taken in place of these courses • Will be used to gather data on student progress through developmental courses

  25. How is your college using technical math? • Will you have a modular technical math course this Fall? • Is it replacing only developmental math? • Is it replacing only college-level math? • Is it replacing both? • If it is replacing both, will some students have to do remediation prior to Technical Math? • Do your programs have additional math requirements on top of Technical Math?

  26. Questions on targeted program participant & program data?

  27. Spreadsheets • 1st sheet has list of variables, their definitions, & required format • Other sheets are data table shells to be completed by colleges

  28. Submission • Data will contain confidential data • Each college will be given a password & will use password protection built into Excel • Submission via secure Drop Box • Timing of submissions • Fall semester data – January 21 • Spring semester data – June 21 • Summer semester data – September 21

  29. Wrap up data plan • Only asking colleges for information can not get from other sources • Your help is crucial because changes to dev. ed. large part of PACE initiative • Especially true of historic dev. ed. completion data & PLA data

  30. PACE Implementation Assessment

  31. Why do an Implementation Evaluation? • Tell the story behind the data • Contribute to Continuous Improvement • Share learning across locations • Stay on track with goals and funding requirements • USDOL requirement

  32. Overall Objectives • Ultimate objective: capture lessons and best practices from your experiences that contribute to your ongoing efforts and the field in general • Understand how you plan to implement the strategies • Track early outcomes (findings and challenges) from initial implementation • Describe and share adaptations made in response to these early outcomes • Document lessons learned from modifications and final outcomes

  33. Our Approach to Evaluating Implementation • Greater focus on qualitative information • Evaluation plan must be fluid and responsive • Each phase builds on the prior phase • Start up and end of grant period usually reflects heaviest information gathering push • Timing is frequently subject to course corrections

  34. Key Topics of Inquiry • For each of the three strategies outlined to USDOL: how has the strategy been implemented and how have students utilized/experienced it? • Describe key redesign features and approaches used in implementing them, for example: • Personnel changes/additions • Professional development and peer learning activities • Specific models employed (i.e., CAEL, El Paso PREP, etc.) • Curricula and/or delivery innovations • New uses of technology • Involvement of external partners (employers, WIBs, etc.) • New roles for staff or faculty

  35. Methods of Evaluating Implementation • Document review • Relevant institutional policies • Curricula materials • Scheduling information • Informational/outreach materials • Surveys • Interviews • Phone and/or in person • On-site observation • Focus Groups • On site

  36. Implementation Evaluation Information Gathering Timeline • Fall 2012 semester • Analyze information from initial surveys (due Aug. 3) • Document review • Winter 2013 semester • Second round of surveys on planning progress (first half of semester) and early lessons/challenges • Site visits (end of semester) • Academic year 2013-14 • Surveys to track implementation progress, adaptations • Phone interviews or other follow up if needed • Fall 2014 semester • Final document review • Final surveys and close-out site visit

  37. Before we go to lunch… • Any questions about the implementation evaluation approach? • Lunch discussion topics: • Reflect on data plan • What are the key student success priorities at your institution? • What would be most useful (for your institution) to learn during and after PACE implementation?

  38. Next Steps • Updates based on today’s discussion • Questions and clarifications • Cohort definitions • Rolling out analyses during the semester

  39. Contact Information • Project Manager: Chris Spence, 216.767.6262, cspence@newgrowthplanners.com • Impact Assessment: Joel Elvery, PhD: 216.375.6777, jelvery@newgrowthplanners.com • Implementation Assessment: Holly Parker, 734.769.2900, hparker@skilledwork.org

  40. We look forward to working with you!

More Related