This site has been permanently archived. This is a static copy provided by the University of Southampton.

The Use And Misuse Of Bibliometric Indices In Evaluating Scholarly Performance

From: Stevan Harnad <amsciforum_at_GMAIL.COM>
Date: Mon, 11 Aug 2008 23:23:09 -0400

    [ The following text is in the "WINDOWS-1252" character set. ]
    [ Your display is set for the "iso-8859-1" character set. ]
    [ Some characters may be displayed incorrectly. ]

Ethics In Science And Environmental Politics (ESEP)�

ESEP Theme Section:�The Use And Misuse Of Bibliometric Indices In
Evaluating Scholarly Performance�+ accompanying�Discussion Forum

Editors:�Howard I. Browman, Konstantinos I. Stergiou
            Quantifying the relative performance of
            individual scholars, groups of scholars,
            departments, institutions,
            provinces/states/regions and countries has
            become an integral part of decision-making
            over research policy, funding allocations,
            awarding of grants, faculty hirings, and
            claims for promotion and tenure. Bibliometric
            indices (based mainly upon citation counts),
            such as the h-index and the journal impact
            factor, are heavily relied upon in such
            assessments. There is a growing consensus,
            and a deep concern, that these indices ?
            more-and-more often used as a replacement for
            the informed judgement of peers ? are
            misunderstood and are, therefore, often
            misinterpreted and misused. The articles in
            this ESEP Theme Section present a range of
            perspectives on these issues. Alternative
            approaches, tools and metrics that will
            hopefully lead to a more balanced role for
            these instruments are presented.

      Browman HI, Stergiou KI�INTRODUCTION: Factors and indices
      are one thing, deciding who is scholarly, why they are
      scholarly, and the relative value of their scholarship is
      something else entirely�
      ESEP 8:1-3�

      Campbell P�Escape from the impact factor�
      ESEP 8:5-7�

      Lawrence PA�Lost in publication: how measurement harms
      science�
      ESEP 8:9-11�

      Todd PA, Ladle RJ�Hidden dangers of a 'citation culture'�
      ESEP 8:13-16�

      Taylor M, Perakakis P, Trachana V�The siege of science�
      ESEP 8:17-40�

      Cheung WWL�The economics of post-doc publishing�
      ESEP 8:41-44�

      Tsikliras AC�Chasing after the high impact�
      ESEP 8:45-47�

      Zitt M, Bassecoulard E�Challenges for scientometric
      indicators: data demining, knowledge flows measurements
      and diversity issues�
      ESEP 8:49-60�

      Harzing AWK, van der Wal R�Google Scholar as a new source
      for citation analysis�
      ESEP 8:61-73�

      Pauly D, Stergiou KI�Re-interpretation of 'influence
      weight' as a citation-based Index of New Knowledge (INK)�
      ESEP 8:75-78�

      Giske J�Benefitting from bibliometry�
      ESEP 8:79-81�

      Butler L�Using a balanced approach to bibliometrics:
      quantitative performance measures in the Australian
      Research Quality Framework�
      ESEP 8:83-92�
      Erratum�

      Bornmann L, Mutz R, Neuhaus C, Daniel HD�Citation counts
      for research evaluation: standards of good practice for
      analyzing bibliometric data and presenting and
      interpreting results�
      ESEP 8:93-102�

      Harnad S�Validating research performance metrics against
      peer rankings�
      ESEP 8:103-107
Received on Tue Aug 12 2008 - 04:29:11 BST

This archive was generated by hypermail 2.3.0 : Fri Dec 10 2010 - 19:49:26 GMT