Evaluation of Digital library - PowerPoint PPT Presentation

adrian-gonzales
evaluation of digital library n.
Skip this Video
Loading SlideShow in 5 Seconds..
Evaluation of Digital library PowerPoint Presentation
Download Presentation
Evaluation of Digital library

play fullscreen
1 / 22
Download Presentation
Evaluation of Digital library
178 Views
Download Presentation

Evaluation of Digital library

- - - - - - - - - - - - - - - - - - - - - - - - - - - E N D - - - - - - - - - - - - - - - - - - - - - - - - - - -
Presentation Transcript

  1. Evaluation of Digital library Anna Maria Tammaro University of Parma University of Tbilisi, 5-15 July 2010

  2. Outline • What is the digital library? • Why evaluate? • Evaluation cycle • What? • How? • Good practices University of Tbilisi, 5-15 July 2010

  3. What is a digital library? • What is encompassed? • Visions of library • What elements to take? • What is critical? University of Tbilisi, 5-15 July 2010

  4. Evaluation is a fact finding, Evidence based value measuring, Integrated in the management process of digital libraries Accountability: evidence of resources spent Effectiveness: understanding basic phenomena (as information seeking) Impact: as increased learning, research, dissemination Why evaluate? University of Tbilisi, 5-15 July 2010

  5. 4 major questions for evaluation • What actually occurred? • How can it be improved? • Did it accomplish the objectives? • What impact did it have? University of Tbilisi, 5-15 July 2010

  6. Documentation Evaluation, MIS Description • What actually occurred? University of Tbilisi, 5-15 July 2010

  7. How can it be improved? Formative evaluation Improvement University of Tbilisi, 5-15 July 2010

  8. Did it accomplish its objectives? Effectiveness evaluation Fit for purposes? University of Tbilisi, 5-15 July 2010

  9. Impact of digital library • What impact did it have? • The ultimate question for evaluation is: “How are digital libraries transforming research,education, learning and living?” (Saracevic 2002, p. 368) University of Tbilisi, 5-15 July 2010

  10. What evaluate? • Content • Services/system • Users and uses University of Tbilisi, 5-15 July 2010

  11. Content evaluation • Content quality (subject coverage, relevance) • Content scope (what is included? Online journals, ebook) • Content organisation (metadata, bibliographic organisation,indexing) • Effectiveness (management, user support) • Efficiency (cost) University of Tbilisi, 5-15 July 2010

  12. System interface • Interface (usability, design, accessibility) • System performance (interactivity,algorithms for searching, processing time) • System configuration (networks, security,authentication) University of Tbilisi, 5-15 July 2010

  13. Outcomes • The ways in which library users are changed as a result of their contact with the library resources and programs (ARL 1998) University of Tbilisi, 5-15 July 2010

  14. Outcomes based evaluation • Have audiences been sufficiently identified? • Are outcomes clearly written? • Are outcomes sufficient to describe what you hope will happen? • Are data collection methods cost efficient? Add: Do they provide the data you want and need? University of Tbilisi, 5-15 July 2010

  15. Users • Who are they? (researchers, students, remote, etc.? What is their context?) • How do they access the digital library? (infomation seeking behviour, usability) • Why do they need the digital library? (activities, expectations) • What type of resources do they need? (subject, etc.) • What is the value of digital library? (impact, outcomes, potential for community building) University of Tbilisi, 5-15 July 2010

  16. European Minerva Project • Minerva • Handbook on cultural web user interaction • http://www.minervaeurope.org/publications/handbookwebusers.htm University of Tbilisi, 5-15 July 2010

  17. Survey Focus group Interviews Transaction logs Observation Ethnographic evaluation Usability Combined methods Longitudinal studies Crosscultural assessment Benchmarking How to evaluate? University of Tbilisi, 5-15 July 2010

  18. Standard – COUNTER, SUSHI (NISO standard usage statistics harvesting initiative) • No benchmarking or longitudinal studies (for the rate of change) University of Tbilisi, 5-15 July 2010

  19. Good practice • DigiQual – http:// www. digiqual. org/ • PEAK – http:// www. dlib. org/ dlib/ june99/ 06bonn. html • E- valued – http:// www. evalued. uce. ac. uk University of Tbilisi, 5-15 July 2010

  20. Bad news • There is no single, easy to administer, inexpensive, reliable, and valid approach to evaluating interactive learning from DLs. University of Tbilisi, 5-15 July 2010

  21. Good news There are practical strategies for documenting the development and use of interactive learning, improving it, and building a case for its effectiveness and impact. University of Tbilisi, 5-15 July 2010

  22. Questions? • Thanks of attention! Annamaria.tammaro@unipr.it University of Tbilisi, 5-15 July 2010