1 / 23

E-Book Usage Statistics

E-Book Usage Statistics. Data Collection & Assessment Maryland Library Association Technical Service Division E-Resources Boot Camp August 4, 2014 Randy Lowe – Collection Development, Acquisition & Serials Librarian, Frostburg State University. Overview. Why E-Resources (E-Book) Assessment?

Download Presentation

E-Book Usage Statistics

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. E-Book Usage Statistics Data Collection & Assessment Maryland Library Association Technical Service Division E-Resources Boot Camp August 4, 2014 Randy Lowe – Collection Development, Acquisition & Serials Librarian, Frostburg State University

  2. Overview • Why E-Resources (E-Book) Assessment? • Usage Statistics – Characteristics, Reports, Data Elements, Collection • Assessment – Applying usage statistics to collection management decision-making

  3. Why E-Resource Assessment? • Libraries have historically measured use of services (circulation statistics, re-shelving counts, gate counts, etc.) • The technology upon which e-resources reside inherently allows for extensive collection of usage data – and assessment of that use • Assessment of use data supports evidence-based collection management • Libraries operate in a challenging fiscal environment – demonstrating e-resource value and fiscal responsibility is a must

  4. Effective E-Resources Assessment • Two essential elements in conducting effective e-resource assessments: • Efficient and Accurate Data Collection • Clear and Succinct Analysis • E-Resource assessment is more than just collecting usage statistics – it is applying them in the making of sound management decisions regarding library resources

  5. E-Resource Usage Statistics – Characteristics • Statistics usually available at platform, package and title levels • Useful measures include Sessions, Searches, Accesses, Linking data (sources & targets) • Usage statistics measure number of times resources are used, but also can be extremely useful in identifying user behavior trends • Focus here will be e-books, but basic concepts apply to all e-resource types

  6. What Can You Do with Usage Statistics? • Track usage / Assess overall collection use • Track expenditures / Figure cost-per-use • Track turnaways • Assess title, subject, publisher and other usage elements • Assist in making collection development decisions, including acquisition model selection • Effectively advocate for resources – especially if assessment is tied to institutional goals/strategic plan

  7. Types of Usage Statistics Reports You can only work with the data that is provided to you • Vendor-Defined • Range widely, depending on the vendor • Some vendors provide great detail in one or two reports, which can be very useful in assessing various use elements [Example 1] [Example 2] • Other vendors provide basic reports only [Example] • Useful when assessing data from a single vendor • Financial transaction/cost information obtained through these reports • Usage comparison across vendors may be problematic

  8. Types of Usage Statistics Reports • COUNTER-Compliant • Project COUNTER (http://www.projectcounter.org) – Establishes standards and protocols for consistent, credible and comparable vendor-generated usage statistics • Especially useful when collecting and analyzing e-resource statistics across vendors • Vendor diligence and application varies • Does not include cost data

  9. Types of Usage Statistics Reports • COUNTER-Compliant (continued) • COUNTER Reports for E-Books (Release 4 – http://www.projectcounter.org/r4/COPR4.pdf) • Book Report 1 – Number of Successful Title Requests by Month & Title [Example] – Useful for tracking individual and aggregate e-book use; a book is a single file on vendor platform • Book Report 2 – Number of Successful Section Requests by Month & Title [Example] – Useful for tracking section/chapter use, especially for reference and edited works; book sections comprise the files on vendor platform

  10. Types of Usage Statistics Reports • COUNTER-Compliant (continued) • COUNTER Reports for E-Books (Release 4 – http://www.projectcounter.org/r4/COPR4.pdf) • Book Report 3 – Access Denied to Content Items by Month, Title & Category [Example] – Turnaways at the title level; helpful in assessing acquisition model/simultaneous user limits/needed copies • Book Report 4 – Access Denied to Content Items by Month, Platform & Category –Turnaways at the platform level • Book Report 5 – Total Searches by Month & Title – Total searches and sessions at the title level

  11. When to Use Various Usage Reports • Vendor-Defined • Analyzing usage data from a single vendor • Obtaining cost information • Comprehensive data files make it easy to analyze combinations of various data elements [Example] • COUNTER-Compliant • Analyzing usage data across multiple vendors • Ensuring data integrity though adherence to recognized standards

  12. Data Elements Common to E-Book Assessment • Primary Elements (Examples) • Title • Publisher • Cost/Expenditures • Transaction type (purchase or short-term loan/rental) • Subject, LCC, Dewey (if available) • Situational Elements (Examples) • Campus or Library (consortial, library system) • User/User Type (if ID is established – Example)

  13. Collecting Usage Data • Define Objectives • What you need to know or are trying to find out should drive your data collection decisions • Collecting Usage Statistics can be a major time commitment • Use your assessment objectives to help you to not only determine what data to collect, but when you have collected enough data to analyze • Properly balancing time and resources dedicated to both data collection and analysis is vital

  14. Collecting Usage Data • Manual Collection • Automated Collection Harvesting Tools • SUSHI (Standardized Usage Statistics Harvesting Initiative) – http://www.niso.org/workrooms/sushi/ • Next generation ILS tools • Third-party Usage Collection

  15. Collecting Usage Data • Various vendors present data differently – this can present a challenge not only across vendors, but even with combining data elements from a single vendor • Manipulation / Formatting of raw data will likely be necessary • Schedule time(s) to collect data – vendors’ archival policies for maintaining usage statistics vary • Example – COUNTER BR1 Report + Acquisition Type Data + Cost Data Compiled Manually = Data for Assessment

  16. Assessing Usage Data You have usage data – What do you do with it? • It is easy to get overwhelmed in usage data – analysis should be guided by your assessment objectives • What do you want/need to assess? • What questions are you trying to answer? • Who is your audience?

  17. Assessing Usage Data • E-Book Assessment Objective Examples • Frostburg State University: Report overall use and expenditures of e-books over time; implement the most cost effective DDA acquisition model(s) • USMAI Consortial E-Book Pilot: Assess the effectiveness of a specific DDA acquisition model for the consortium; use and expenditures by consortium members and user types; identification of possible future program funding models • There is no single method for assessing usage statistics in every case – the “right data” to analyze and include in your report is that which will support your assessment objectives

  18. Usage Data Analysis • Data analysis should be thorough, but presented succinctly • Conclusions, trends, etc. should be clear and verifiable • Beware of pre-conceived notions, perceptions or opinions – hypotheses can be both proven and refuted • State known limitations of the data you have collected and how they may affect your analysis

  19. Writing Your Report • Know your audience • Include a brief purpose/introduction • Write clearly and succinctly • Reported usage data should support the purpose of the assessment • Only include data that supports your stated objectives – don’t include all collected data; it won’t be read by administrators

  20. Writing Your Report • Reported usage data should support the purpose of the assessment (continued) • Include data within the text of your report where it is necessary and provides clear evidence for the points you are making • It is usually more effective to include visual representations of (charts, graphs) rather than just figures within the text of reports • Larger tables and data sets, if necessary to include, are best placed in appendices • Conclusions and recommendations should be easily identified and based on the evidence presented

  21. E-Book Usage Assessment Examples • Frostburg State University: Objective - Report overall use and expenditures of e-books; implement the most cost effective DDA acquisition model(s) [Report] • USMAI Consortial E-Book Pilot: Objective - Assess the effectiveness of a specific DDA acquisition model for the consortium; use and expenditures by consortium members and user types; identification of possible future program funding models • [Report]

  22. Other Examples • Examples of reports from many college and university libraries as well as library systems, consortia and organizations may be found via a quick Internet Search. • One of my favorite E-Book usage studies is: • Fischer, K., Wright, M., Clatanoff, K., Barton, H., & Shreeves, E. (2012). Give 'Em What They Want: A One-Year Study of Unmediated Patron-Driven Acquisition of e-Books. College & Research Libraries, 73(5), 469-492. http://crl.acrl.org/content/73/5/469

  23. Thank You • Questions? • Contact Information: Randy Lowe Frostburg State University rlowe@frostburg.edu

More Related