1 / 51

Central R&D assessment indicators: Scientometric and Webometric Methods

Central R&D assessment indicators: Scientometric and Webometric Methods. Peter Ingwersen Royal School of LIS - 2010 Denmark – pi@iva.dk http://www.iva.dk/pi Oslo University College, Norway. Agenda. Scientific Communication: Classic & present models Scientometrics:

jacklyns
Download Presentation

Central R&D assessment indicators: Scientometric and Webometric Methods

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Central R&D assessment indicators: Scientometric and Webometric Methods Peter Ingwersen Royal School of LIS - 2010 Denmark – pi@iva.dk http://www.iva.dk/pi Oslo University College, Norway

  2. Agenda • Scientific Communication: • Classic & present models • Scientometrics: • Publication Analyses • Publication Point evaluation (’Norwegian Model’) • Citation Analyses • Crown Indicators (research profile weighting) • Hirsch Index (h-Index) • Webometrics • Web Impact Factors; Issue tracking - mining • Concluding remarks 2011

  3. Peers TechnicalResearch report • Un-published • non-peer review • informal com. Archive Library index Conf. Papers (Peer reviewed) Journal Articles (Peer reviewed) Research idea & activities Domain databases Citation database Time Scientific communication 1 – Classic Model (prior to Web / open access) 2011

  4. Scientific communication 1 – Present Model (incl. Web / open access) Peers • TechnicalResearch reports • Working • papers • Un-published • public • Non Peer review Inst. Repositories Open access journals Full text Domain database Journal Articles (Peer reviewed) Conf. Papers (Peer reviewed) Research idea & activities - Web of Science - Scopus Google (Scholar) Academic Web Search Engines Time 2011

  5. Confidence in information source? Authoritative source Student output Open Access - Journals (peer reviewed) - Inst. Repositories (Duplicates/versions) Restricted Access - Journal articles (peer reviewed) Working papers Research reports Blogs … Conference Papers Posters, Abstracts (peer reviewed) Collaboratory round tables Restricted Access - Research Monographs (peer reviewed) Teaching material Searchable on Open Web Partly searchable on Open Web Qualified knowledge source (Domain dependent ) Scientific communication 2 – What ’is’ scientific information? 2011

  6. Examples of Publication analysis • Ranking most productive • Countries in a field • Journals in a field • Institutions or universities; departments or goups • (Exponential Bradford-like distributions) • Counting scientific publications in • Academic fields / disciplines • Countries, regions, universities, departments Counting number of articles OVER TIME • Time series 2011

  7. Typical time series 1981-2005 2011

  8. Productivity Growth 2011

  9. Publication Growth – all fields 1981-20061981-85 = index 1: China=14,114 p.; EU=736,616 p; USA=887,039 p. India = 65,250 (98,598) publ. 2011

  10. Publication success ‘points’ • As done in Norway: • Articles from the journals considered among the 20 % best journals in a field: 3 points • Articles from other (peer reviewed) journals: 1 point • Conference papers (peer reviewed): .7 points • Monographs (int.publisher): 8 points • Monographs (other publishers): 5 points • Fractional counts; points used for funding distribution • Covers all research areas, incl. humanities, for all document types 2011

  11. One cannot use the publication points for DIRECTcomparison • Between universities or countries • Or applied to individual researchers • Recent detailed article on the system: • Schneider, J.W. (2009) An outline of the bibliometric indicator used for performance-based funding of research institutions in Norway. European Political Science, 8(3), p. 364-378. India 2010

  12. However: Publication Point Indicators established! • Elleby, A., & Ingwersen, P. Publication point indicators: A comparative case study of two publication point systems and citation impact in an interdisciplinary context. Journal of Informetrics, 4(2010): 512-523. • doi:10.1016/j.joi.2010.06.001 India 2010

  13. Publication Point Indicators 2 • Comparing the vectors of ideal cumulated PP (= expected success gain) with the actually obtained PP, for the same publication (types), providing a ratio that can be normalized: nPPI:The normalized Publication Point Index • Comparisons between institutionas can be done at specific ranges of publication vector values through their nPPI. India 2010

  14. Cumulated Publ. Point Indicatorthe DIIS example (n=70) Tutorial 2011

  15. Citation Analyses • Diachronic (forward in time) … or Synchronous (back in time – like ISI-JIF) • Observing: how older research is received by current research (ISI+Scopus: always peer reviewed sources) • Citation indicators: • Time series (like for publications) • Citation Impact (Crown Indicators) • Citedness 2011

  16. Absolute Citation Impact 2011

  17. ‘Crown indicators’ • Normalized impact-indicators for one unit (center/university/country) in relation to research field globally: • JCI : Journal Crown Indicator • FCI : Field Crown Indicator – both provide an index number 2011

  18. Journal Crown Indicator • The ratio between: - the real numberof citations received for all journal articles in a unit from a year, and - the diachronic citation impact of the same journals used by the unit, covering the same period (= the expected impact). ONE WOULD NEVER APPLY THE ISI-JIF!! Since it only signifies the AVERAGE (international) impact of an article made in a synchronous way 2011 2010 2009 2007 2008 2011

  19. Journal Impact Factor - ISI • Synchroneous method: • For 2010: Analysis done in Febr-April, 2011 for … 1) all citations given in 2010 to journal X for articles+notes+letters in journal X, 2) Published in previous two years: 2008-2009 2011 2010 2009 2007 2008 2011

  20. Field Crown Indicator - FCI • Normalisationmust be weighted in relation to the observed unit’s publication profile: • Like a ’shadow’ unit (country) • An example of this weighting for India: 2011

  21. Research profile as weight for impact calculation (as ‘shadow country’) 2011

  22. Research profile (China) as weight for impact calculation (as ‘shadow country’) 2011

  23. A small European country with very different profile 2011

  24. Example of research profile with FCI-index score 2011

  25. Summary: Different indicators – one given period • Σc/Σp / ΣC/ΣP – Globally normalized impact: • For single fields it is OK to use! • If averaged over all subject areas: quick’n dirty!: all areas have the same weight! – thus: • Σc / Σ(C/Parea x parea ) = FCI: StandardField Crown Indicator (FCI) for ’profile’ of subject areas for a local unit (country/university) – via applying it as global profile, like a kind of ’shadow unit’. Made as ratio of sums of citations over publications (weights) (If done as sum of rations divided by fields: all fields equal) 2011

  26. Ageing of journals or articles • Cited half-life - diachronic: • Acumulate citations forward in time by year: • 1990 91 92 93 94 95 96 97 98 99 00 01 02 - yrs • 2 12 20 25 30 17 12 10 0 3 10 0 - Citations Acum: 2 14 34 59†89106 118 128 128 131 132 • 1/2 life= 132/2 = 66 = ca. 4,2years Tutorial 2011

  27. Ageing of journals or articles – 2 Tutorial 2011

  28. Hirsch Index (2005) • A composite index of publications and citations for a unit (person, group, dept. …) in a novel way: • H is the number of articles given a number of citations larger or equal to h. • A person’s h-index of 13 implies that he/she among all his/her publications has 13, that at least each has obtained 13 citations. • The index is dependent on research field and age of researcher. Can be normalized in many ways. Tutorial 2011

  29. Criticism of Citation Analyses • Formal influences not cited • Biased citing • Informal influences not cited • Self-citing – may indeed improve external cits.! Different types of citations – • Variations in citation rate related to type of publication, nationality, time period, and size and type of speciality – normalization? • Technical limitations of citation indexes and domain databases • Multiple authorship – fractional counting/article level Tutorial 2011

  30. Reply: van Raan (1998) • Different biases equalizes each other • If researchers simply do most of their referencing ”in a reasonable way” and if a sufficient number of citations are counted, realiable patterns can be observed. • It is very unlikely that all researchers demonstrate the same biases (e.g. all researchers consciously cite research, which does not pertain to their field) Tutorial 2011

  31. Google Scholar • Does not apply PageRank for ranking but citations • Contains conference papers and journal articles (??) • Workable for Computer Science and Engineering (and Inf. Sc.) • Requires a lot of clean-up! • Apply http://www.harzing.com/pop.htmfor (Publish or Perish) for better analysis on top of GS • Google Scholar may provide the h-index for persons 2011

  32. L. Björneborn & P. Ingwersen 2003 infor-/biblio-/sciento-/cyber-/webo-/metrics informetrics bibliometrics scientometrics cybermetrics webometrics 2011

  33. L. Björneborn & P. Ingwersen 2003 Link terminologybasic concepts A B E G C D F • B has an outlink to C; outlinking : ~ reference • B has an inlink from A; inlinked : ~ citation • B has a selflink; selflinking : ~ self-citation • A has no inlinks; non-linked: ~ non-cited • E and F are reciprocally linked • A is transitively linked with H via B – DH is reachable from A by a directed link path • A has a transversal link to G : short cut • C and D are co-linked from B, i.e. have co-inlinks orshared inlinks: co-citation • B and E are co-linking to D, i.e. have co-out-links orshared outlinks: bibliog.coupling H co-links 2011

  34. 2011

  35. www.internetworldstats.com 2011

  36. Search engine analyses • See e.g. Judith Bar-Ilan’s excellent longitudinal analyses • Mike Thelwall et al. in several case studies • Scientific material on the Web: • Lawrence & Giles (1999):approx. 6 % of Web sites contains scientific or educational contents • Increasingly:the Web is a web of uncertainty • Allen et al. (1999) – biology topics from 500 Web sites assessed for quality: • 46 % of sites were ”informative” – but: • 10-35 % inaccurate; 20-35 % misleading • 48 % unreferenced 2011

  37. The Web-Impact Factor Ingwersen, 1998 • Intuitively (naively?) believed as similar to the Journal ImpactFactor • Demonstraterecognition by other web sites - or simply impact – notnecessarilyquality • Central issue: are web sites similar to journals and web pages similar to articles? • Arein-linkssimilar to citations – orsimplyroadsigns? • What is really calculated? • DEFINE WHAT YOU ARE CALCULATING: site or page IF 2011

  38. The only valid webometric tool: Site Explorer Yahoo Search … • If one enters (old valid) commands like: • Link:URL or Domain: topdomain (edu, dk) or Site:URL you are transferred to: http://siteexplorer.search.yahoo.com/new/ • Or find it via this URL • The same facilities are available in click-mode, as one starts with a given URL: • Finding ‘all’ web pages in a site • Finding ‘all’ inlinks to that site/those pages • Also without selflinks! – this implies … Ingwersen

  39. … to calculate Web Impact Factors • But one should be prudent in interpretations. • Note that external inlinks is the best indicator of recognition • Take care of how many sub-domains (and pages) that are included in the click analysis. • Results can be downloaded Ingwersen

  40. Possible types of Web-IF: • E-journal Web-IF • Calculated by in-links • Calculated as traditional JIF (citations) • Scientific web site – IF (by link analyses) • National – regional (some URL-problems) • Institutions – single sites • Other entities, e.g. domains • Best nominator: no. of staff – or simply use external inlinks 2011

  41. Web-links like citations? • Kleinberg (1998) between citation weights and Google’s PageRank: Hubs~ review article: have many outlinks (refs) to: Authority pages~ influential (highly cited) documents: have many inlinks fromHubs! Typical: Web index pages =homepage with self-inlinks = Table of contents 2011

  42. Reasons for outlinking … • Out-linksmainly for functionalpurposes • Navigation– interestspaces… • Pointing to authority in certaindomains? (Latour:rhetoricreasons for references-links) • Normativereasonsfor linking? (Merton) • Dowehavenegativelinks? • Wedohavenon-linking(commercialsites) 2011

  43. Some additional reasons for providing links • In part analogous to providing references (recognition) • And, among others, • emphasising the own position and relationship (professional, collaboration, self-presentation etc.) • sharing knowledge, experience, associations … • acknowledging support, sponsorship, assistance • providing information for various purposes (commercial, scientific, education, entertainment) • drawing attention to questions of individual or common interest and to information provided by others (the navigational purpose) 2011

  44. Other differences between references, citations & links • The time issue: • Agingof sources are different on the Web: • Birth, Maturity & Obsolescence happens faster • Decline & Death of sources occur too– but • Mariages – Divorse – Re-mariage – Death & Resurrection … & alike liberalphenomena are found on the Web! (Wolfgang Glänzel) 2011

  45. Issue tracking – Web mining • Adequatesamplingrequiresknowledge of the structure and properties of the population - the Web space to besampled • Issuetrackingof knownproperties / issuesmay help • Web mining the unknown is moredifficult, due to • the dynamic, distributed & diversenature • the variety of actors and minimum of standards • the lack of qualitycontrol of contents • Web archeology – study of the past Web 2011

  46. Nielsen Blog Pulse • Observes blogs worldwide by providing: • Trend search – development over time of terms/concepts – user selection! • Featured trends – predefined categories • Coversation tracker – blog conversations • BlogPulse profiles – blog profiles • Look into: http://www.blogpulse.com/tools.html 2011

  47. Home > ToolsTrend Search 2011

  48. Concluding remarks: Future • With open access we can foresee a nightmare as concerns tracking qualified and authoritative scientific publications, aside from the citation indexes because of • Lack of Bibliographic control (what is original – vs. parallel and spin-off versions & crab?) over many institutional repositories – and mixed on the web with all other document types incl. Blogs (web 2.0) – Google Scholar(?) … Google Books (?) 2011

  49. Concluding remarks • One may be somewhat cautious on Web-IF applications without careful sampling via robotsdue to its incomprehensiveness and what it actually signifies • One might also try to investigate more the behavioural aspects of providing and receiving linksto understand what the impact might mean and how/whylinksare made • Understand the Web space structurebetter • Design workable robots, downloading & local analyses 2011

  50. References • Allen, E.S., Burke, J.M., Welch, M.E., Rieseberg, L.H. (1999). How reliable is science information on the Web? Science, 402, 722. • Björneborn, L., Ingwersen, P. (2004). Towards a basic framework for webometrics. Journal of American Society for Information Science and Technology, 55(14): 1216-1227. • Brin, S., Page, L. (1998). The anatomy of a large scale hypertextual web search engine. Computer Networks and ISDN Systems, 30(1-7), 107-117. • Elleby, A., Ingwersen, P. Publication Point Indicators: A Comparative Case Study of two Publication Point Systems and Citation Impact in an Interdisciplinary Context. Journal of Informetrics, 2010, 4, p. 512-523. • Hirsch, J.E. (2005): An index to quantify an individual’s scientific research output. PNAS, 102: 16569-16572. 2011

More Related