1 / 15

Using Mixed-Methods and Survey Research in Project Evaluation/Assessment Pat Campbell QEM Workshop on Assessment and Eva

Using Mixed-Methods and Survey Research in Project Evaluation/Assessment Pat Campbell QEM Workshop on Assessment and Evaluation August 7-8, 2009 . Evaluation Basics: Soup, Cooks, Guests & Improvement.

feng
Download Presentation

Using Mixed-Methods and Survey Research in Project Evaluation/Assessment Pat Campbell QEM Workshop on Assessment and Eva

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Using Mixed-Methods and Survey Research in Project Evaluation/Assessment Pat Campbell QEM Workshop on Assessment and Evaluation August 7-8, 2009

  2. Evaluation Basics: Soup, Cooks, Guests & Improvement When cooks taste the soup, it’s formative evaluation; the collection of information that can be used to improve the soup. If necessary, the cook’s next step is to explore strategies to fix the problem. The cook makes some changes and then re-tastes the soup, collecting more formative evaluation data. When the guests taste the soup at the table, they’re doing summative evaluation. They are collecting information to make a judgment about the overall quality and value of soup. Once the soup is on the table and in the guests’ mouths, there is little that can be done to improve that soup. Thanks to Bob Stake for first introducing this metaphor.

  3. Challenging Assumptions When I was a physicist people would often come and ask me to check their numbers, which were almost always right. They never came and asked me to check their assumptions, which were almost never right. Eli Goldratt

  4. Pat’s Evaluation Assumptions • The core evaluation question is “What works for whom in what context?” • “Black hole” evaluations are bad. • A bad measure of the right thing is better than a good measure of the wrong thing. • Acknowledging WIIFM increases response rates. • Process is a tool to help understand outcomes. • Outcomes are at the core of accountability.

  5. Hierarchy of Study Designs

  6. The Right Design for the Question

  7. The Right Design for the Question

  8. Compared to What? Evaluation Designs • Experimental designs • Quasi-experimental designs • Mixed methods designs • Case studies NSF does not promote one design, rather it wants the design that will do the best job answering your evaluation questions!

  9. Compared to What? Why Bother

  10. Web-based Sources of Comparisons WebCASPAR database (http://caspar.nsf.gov) provides free access to institutional level data on students from surveys as Integrated Postsecondary Education Data System (IPEDS) and the Survey of Earned Doctorates. The Engineering Workforce Commission (http://www.ewc-online.org/) provides institutional level data (for members) on bachelors, masters and doctorate enrollees and recipients by sex by race/ethnicity for US students and by sex for foreign students. Comparison institutions can be selected from the Carnegie Foundation for the Advancement of Teaching’s website, (http://www.carnegiefoundation.org/classifications/) based on Carnegie Classification, location, private/public designation, size and profit/nonprofit status.

  11. Some Web-based Sources of Resources OERL, the Online Evaluation Resource Library. http://oerl.sri.com/home.html User Friendly Guide to Program Evaluationhttp://www.nsf.gov/pubs/2002/nsf02057/start.htm AGEP Collecting, Analyzing and Displaying Datahttp://www.nsfagep.org/CollectingAnalyzingDisplayingData.pdf American Evaluation Association http://www.eval.org/resources.asp

  12. OERL, the Online Evaluation Resource Library. http://oerl.sri.com/home.html Includes NSF project evaluation plans, instruments, reports and professional development modules on: • Designing an Evaluation • Developing Written Questionnaires • Developing Interviews • Developing Observation Instruments • Data Collection • Instrument Triangulation and Adaptation.

  13. User Friendly Guide to Program Evaluationhttp://www.nsf.gov/pubs/2002/nsf02057/start.htm • Introduction • Section I - Evaluation and Types of Evaluation • Section II - The Steps in Doing an Evaluation • Section III - An Overview of Quantitative and Qualitative Data Collection Methods • Section IV - Strategies That Address Culturally Responsive Evaluations • Other Recommending Reading, Glossary, and Appendix A: Finding An Evaluator

  14. AGEP Collecting, Analyzing and Displaying Datahttp://www.nsfagep.org/CollectingAnalyzingDisplayingData.pdf I. Make Your Message Clear II. Use Pictures, Where Appropriate III. Use Statistics and Stories IV. Be Responsive to Your Audience. V. Make Comparisons VI. Find Ways To Deal With Volatile Data VII. Use the Results

  15. Some Things to Consider When Choosing a Design The balance between the level of investment in the evaluation and the level of investment in and the intensity of the intervention (they should be roughly commensurate). The appropriateness of the fit between the design of the program or “intervention” and the requirements of more rigorous evaluation methodologies. The timing of the evaluation also has an impact the design to be selected. The level of evidence expected given the nature of the intervention.

More Related