Comment on:
Richard Monastersky, The Number That's Devouring Science,
Chronicle of Higher Education (CHE), October 1, 2005
(truncated version of reply below appeared in
CHE November 14, 2005)
Although Richard Monasterky describes a real problem -- the abuse of journal impact factors -- its solution is so obvious one hardly required so many words on the subject:
A journal's citation impact factor (CIF) is the average number of citations received by articles in that journal (
ISI -- somewhat arbitrarily -- calculates CIFs on the basis of the preceding two years, although
other time-windows may also be informative.)
There is an undeniable relationship between the usefulness of an article and how many other articles use and hence cite it. Hence CIF does measure the average usefulness of the articles in a journal. But there are three problems with the way CIF itself is used, each of them readily correctable:
(1) A measure of the average usefulness of the articles in the journal in which a given article appears is no substitute for the actual usefulness of each article itself: In other words, the journal CIF is merely a crude and indirect measure of usefulness; each article's own citation count is the far more direct and accurate measure. (Using the CIF instead of an article's own citation count [or the average citation count for the author] for evaluation and comparison is like using the average marks for the school from which a candidate graduated, rather than the actual marks of the candidate.)
(2) Whether comparing CIFs or direct article/author citation counts, one must always compare like with like. There is no point comparing either CIFs between journals in different fields, or citation counts for articles/authors in different fields. (A normalised citation count can also be used, adjusting for different baseline citation levels and variability in different fields.)
(3) Both CIFs and citation counts can be distorted and abused. Authors can self-cite, or cite their friends; some journal editors can and do encourage self-citing their journal. These malpractices are deplorable, but most are also detectable, and then name-and-shame-able and correctable. ISI could do a better job policing them, but soon the playing field will widen, for as authors make their articles open access online, other harvesters -- such as citebase and citeseer and even google scholar -- will be able to harvest and calculate citation counts, and average, compare, expose, enrich and correct them in powerful ways that were inconceivable in the Gutenberg era.
So, yes, CIFs are being misused and abused currently, but the cure is already obvious -- and a wealth of powerful new resources are on the way for measuring and analyzing research usage and impact online, including (1) download counts, (2) co-citation counts (co-cited with, co-cited by), (3) hub/authority ranks (authorities are highly cited papers cited by many highly cited papers; hubs cite many authorities), (4) download/citation correlations and other time-series analyses, (5) download growth-curve and peak latency scores, (6) citation growth-curve and peak-latency scores, (7) download/citation longevity scores, (8) co-text analysis (comparing similar texts, extrapolating directional trends), and much more. It will no longer be just CIFs and citation counts but a rich multiple regression equation, with many weighted predictor variables based on these new measures. And they will be available for both navigators and evaluators online, and based not just on the current ISI database but on all of the peer-reviewed research literature.
Meanwhile, use the direct citation counts, not the CIFs.
Some self-citations follow:
Brody, T. (2003)
Citebase Search: Autonomous Citation Database for e-print Archives, sinn03 Conference on Worldwide Coherent Workforce, Satisfied Users - New Services For Scientific Information, Oldenburg, Germany, September 2003
Brody, T. (2004)
Citation Analysis in the Open Access World Interactive Media International
Brody, T. , Harnad, S. and Carr, L. (2005)
Earlier Web Usage Statistics as Predictors of Later Citation Impact. Journal of the American Association for Information Science and Technology (JASIST, in press).
Hajjem, C., Gingras, Y., Brody, T., Carr, L. & Harnad, S. (2005)
Across Disciplines, Open Access Increases Citation Impact. (manuscript in preparation).
Hajjem, C. (2005)
Analyse de la variation de pourcentages d'articles en accès libre en fonction de taux de citations
Harnad, S. and Brody, T. (2004a)
Comparing the Impact of Open Access (OA) vs. Non-OA Articles in the Same Journals. D-Lib Magazine, Vol. 10 No. 6
Harnad, S. and Brody, T. (2004)
Prior evidence that downloads predict citations. British Medical Journal online.
Harnad, S. and Carr, L. (2000) Integrating, Navigating and Analyzing
Eprint Archives Through Open Citation Linking (the OpCit Project). Current Science 79(5):pp. 629-638.
Harnad, S. , Brody, T. , Vallieres, F. , Carr, L. , Hitchcock, S. , Gingras, Y. , Oppenheim, C. , Stamerjohanns, H. and Hilf, E. (2004)
The Access/Impact Problem and the Green and Gold Roads to Open Access. Serials Review, Vol. 30, No. 4, 310-314
Hitchcock, S. , Brody, T. , Gutteridge, C. , Carr, L. , Hall, W. , Harnad, S. , Bergmark, D. and Lagoze, C. (2002)
Open Citation Linking: The Way Forward. D-Lib Magazine 8(10).
Hitchcock, S. , Carr, L. , Jiao, Z. , Bergmark, D. , Hall, W. , Lagoze, C. and Harnad, S. (2000)
Developing services for open eprint archives: globalisation, integration and the impact of links. In Proceedings of the 5th ACM Conference on Digital Libraries, San Antonio, Texas, June 2000., pages pp. 143-151.
Hitchcock, S. , Woukeu, A. , Brody, T. , Carr, L. , Hall, W. and Harnad, S. (2003)
Evaluating Citebase, an open access Web-based citation-ranked search and impact discovery service. Technical Report ECSTR-IAM03-005, School of Electronics and Computer Science, University of Southampton
Stevan Harnad