Quantitative deconstruction of citation impact indicators: Waxing field impact but waning journal impact

A.J. Nederhof (Center for Science and Technology Studies (CWTS), Leiden University, Leiden, The Netherlands)
M.S. Visser (Center for Science and Technology Studies (CWTS), Leiden University, Leiden, The Netherlands)

Journal of Documentation

ISSN: 0022-0418

Article publication date: 1 December 2004

674

Abstract

In two case studies of research units, reference values used to benchmark research performance appeared to show contradictory results: the average citation level in the subfields (FCSm) increased world‐wide, while the citation level of the journals (JCSm) decreased, where concomitant changes were expected. Explanations were sought in: a shift in preference of document types; a change in publication preference for subfields; and changes in journal coverage. Publishing in newly covered journals with a low impact had a negative effect on impact ratios. However, the main factor behind the increase in FCSm was the distribution of articles across the five‐year block periods that were studied. Publication in lower impact journals produced a lagging JCSm. Actual values of JCSm, FCSm, and citations per publication (CPP) values are not very informative either about research performance, or about the development of impact over time in a certain subfield with block indicators. Normalized citation impact indicators are free from such effects and should be consulted primarily in research performance assessments.

Keywords

Citation

Nederhof, A.J. and Visser, M.S. (2004), "Quantitative deconstruction of citation impact indicators: Waxing field impact but waning journal impact", Journal of Documentation, Vol. 60 No. 6, pp. 658-672. https://doi.org/10.1108/00220410410568142

Publisher

:

Emerald Group Publishing Limited

Copyright © 2004, Emerald Group Publishing Limited


1. Introduction

1.1 Background

In recent years, we have seen in several studies that worldwide impact levels in a number of subfields appeared to be on the increase (e.g. Adair and Vohra, 2003; Nederhof et al., 1999). This phenomenon, although by no means general, could be witnessed most clearly in the average field citation scores (FCSm) figures in several of those studies (e.g. Nederhof et al., 1999).

The producer of the Science Citation Indices, Thomson ISI, has categorized journals in Subject Categories, such as “Physics, Particles and Fields” or “Economics”. The average citation rate of all articles in these journals can be computed, and is called “mean field citation score”, or “FCSm” (Moed et al., 1995; see ). The FCSm is used as a reference level against which research performance is benchmarked (e.g. Moed et al., 1995; Nederhof and Van Wijk, 1999; Van Raan, 1996). Then, a higher average impact of journal articles in a subfield (a higher FCSm score) means that research groups operating in that field who produce work with a constant citation impact are seen as declining when their stable impact score is compared with the higher FCSm score. Even research groups who actually improved their average number of citations per paper (CPP) may then be perceived as being in decline. In most older studies using the FCSm indicator (and many recent studies as well), FCSm scores tended to be more or less stable over time (e.g. Moed et al., 1995). Usually, only when groups switched between research fields, or witnessed a change in their production of document types (that tend to have different impact levels), appreciable shifts in FCSm scores did appear.

An increase in the average field citation score (FCSm) is by no means an impossible or even an unlikely phenomenon. For example, if reference lists of articles increase only slightly on average combined with a stable percentage of recent references, or the percentage of recent articles in references increases somewhat, the impact of (recent) articles increases on average as a result. It is well‐known that subfields differ in impact levels, and this is even one of the principal reasons behind the normalization of impact scores by comparison with a citation average based on journals or subfields (e.g. Moed et al., 1995; Zitt et al., 2003). Nevertheless, until recently, large shifts in average citation rates were not frequently observed on the subfield level, or were mainly limited to subfields with already high impact levels.

Of course, it is theoretically possible that an observed increase in the impact of subfields is artifactual, for example due to a database effect or an error in computation. This would affect the normalized impact scores of evaluated research units. However, previous tentative investigations of this phenomenon in several subfields suggested that the increase in FCSm in these subfields is autonomous, and not an artifact of ISI Citation Index coverage of journals (Nederhof et al., 1999; see Adair and Vohra, 2003). Until now, a full and thorough investigation, in which citation impact indicators are disassembled, has not been conducted. The present study intends to fill this gap. We study the increase of FCSm as visible in the bibliometric results of two academic institutes active in veterinary and animal sciences (VAS) (see below).

1.2 Science indicators

In CWTS research performance studies, two international reference values are computed. A first value represents the mean citation rate of the journals in which the research unit has published (the mean journal citation score (JCSm)), taking into account both the type of paper (e.g. normal article, review, and so on), as well as the specific years in which the research unit's papers were published. To give an example, the number of citations received during the period 1995‐1998 by a letter published by a research unit in 1995 in journal X is compared to the average number of citations received during the same period (1995‐1998) by all letters published in the same journal (X) in the same year (1995). The latter score is the journal citation score (JCS) of the letter. Generally, a research unit publishes its papers in several journals rather than one. Therefore, we calculated a weighted average JCS indicated as JCSm, with the weights determined by the number of papers published in each journal. Self‐citations are excluded from the computation of JCSm. A unit U that has published two articles in journal Y in 1996 (JCS=3) and one letter in journal X in 1995 (JCS=0.3) obtains a JCSm of Σ(JCS)/ ΣN =(3+3+0.3)/(1+1+1) (where N designates the number of papers)or 6.3/3 is 2.1 citations per publication.

The second reference value presents the mean citation rate of the subfields (journal categories) in which the research unit is active (FCSm). Our definition of subfields is based on a classification of scientific journals into categories developed by ISI. In calculating FCSm, we used the same procedure as the one we applied in the calculation of JCSm, with journals replaced by subfields. In most cases, a research unit is active in more than one subfield (i.e. journal category). In those cases, we calculate a weighted average value, the weights being determined by the total number of papers the research unit has published in each subfield. Suppose that journal X belongs to subfield Z, and that all 1995 articles in subfield Z are cited 1.5 times on average in 1995‐1998, while journal Y belongs to subfield A where all 1996 letters are cited 0.6 times on average in 1996‐1998. Then, the unit U mentioned before obtains an FCSm score of Σ(FCS)/ΣN=1.5+1.5+0.6)/(1+1+1) or 1.2 citations per publication (where N designates the number of papers). If a paper appears in a journal that is classified in more than one subject category, the paper (and its citations) is distributed over the subject categories. Thus, a paper with seven citations published in a journal categorized in three subject categories is counted as a 0.33 publication (with 2.33 citations) in each subject category. As a reference value, the FCSm values of the three subfields are used, similarly divided by three.

Two important indicators compare the average number of citations to the oeuvre of a research unit (CPP) to the two international reference values, namely the corresponding journal and field mean citation scores (JCSm and FCSm, respectively), by calculating the ratio for both. Self‐citations are excluded in the calculation of the ratios CPP/FCSm and CPP/JCSm, to prevent that ratios are affected by divergent self‐citation behavior. To maintain a balanced result, self‐citations are removed from both the numerator and the denominator.

If the ratio CPP/FCSm is above 1.0, this means that the oeuvre of the research unit is cited more frequently than an “average” publication in the subfield(s) in which the research unit is active. FCSm constitutes a world subfield average in a specific (combination of) subfield(s). In this way, one may obtain an indication of the international position of a research unit, in terms of its impact compared to a “world” average. This world average is calculated for the total population of articles published in CI journals assigned to a particular subfield or journal category. If the ratio CPP/JCSm is above 1.0, the mean impact of a research unit's papers exceeds the mean impact of all articles published in the journals in which the particular research unit has published its papers (the research unit's journal set). We apply a statistical test to establish whether the average impact of a research unit's publication oeuvre (CPP) differs significantly from the average impact of all papers in the research unit's journal set (JCSm) or from the world subfield average (FCSm) in the subfield(s) in which the research unit is active The test is based on the negative binomial distribution (see Schubert and Glänzel, 1983). If a research unit has a citation per publication ratio (CPP) significantly above (below) the average field (FCSm) or journal citation score (JCSm), this is indicated in the tables by means of a “+” (“−”) symbol directly after the numerical value of the indicators CPP/FCSmand CPP/JCSm.

So‐called “block indicators” were computed, with equal sets of publication years and citation years (see Schubert et al., 1989). For example, for the publications appearing in 1994‐1998, citations were counted also in 1994‐1998.

2. Effect of document types

In a study of VAS research units at two Dutch universities, Utrecht (UU) and Wageningen (WU), a significant increase (p < 0.001) was observed in the average field citation score (FCSm) between 1989‐1993 and 1994‐1998 (Nederhof et al., 1999). Relevant findings are shown in Table I.

Modest increases in JCSm values are obtained for both UU (12 percent; p<0.001) and WU (5 percent; NS), while FCSm values increase much more strongly for both institutes (respectively, 24 percent (p < 0.001) and 33 percent (p < 0.001)). As a result, the JCSm/FCSm indicator declines with 10 percent for UU (p < 0.05) and even with 21 percent for WU (p < 0.01).

A first explanation for the increase in FCSm might be a change in preference for document types as publication outlets. For example, while UU published 22 percent more articles in 1994‐1998, it published 33 percent more reviews, whereas the number of notes (−62 percent) and letters (−45 percent) decreased strongly (see Table II). To gauge the effect of this shift in document preference on FCSm, we multiplied for each of the four document types their 1994‐1998 publication numbers with the corresponding average 1989‐1993 FCSm values. Then we are able to observe the effect of the shift in document types, while holding FCSm constant. This resulted in a modest increase of 2.5 percent in FCSm for UU, while the increase of 0.8 percent in FCSm is hardly noticeable for WU. Apparently, the substantive increases in FCSm over time are only for a small part due to a shifting preference for document types. It should be noted that the observed change in use of document types as publication outlets is at least partly the result of a database effect rather than the preference of individual researchers, as the document type “notes” is merged with “articles” by ISI since 1996.

In order to eliminate disturbances resulting from the effect of document type on FCSm, we decided to look from now on only at the single most important document type, articles, which accounts for 91.4 percent of the total production in Dutch VAS (Nederhof et al., 1999, p. 13).

If we select articles only in order to gauge the effect of a shift in document type over time, the increase of FCSm is reduced from 24 percent to 18 percent (p < 0.001) for UU, and from 33 percent to 24 percent (p < 0.001) for WU (see Table I). Thus, in both cases about a quarter of the increase in FCSm is associated with notes, letters, and reviews, although they only represent about 9 percent of the publications. As the impact of letters is low in the present study (see Table II), and notes tend to resemble articles, it seems likely that reviews, which have a high average impact, contribute above average to the observed increase in FCSm.

The findings show that FCSm is subject to more important influences than just the distribution of papers over document types. Below, we assess the effect of several of these other influences, although only for articles.

3. Changes in publication preference between subfields

A second possible explanation for the observed shift in FCSm is a change in publication preference for subfields. For instance, increasing FCSm scores coupled with stable or declining JCSm scores might be observed if later papers are published in a set of subfields with higher FCSm scores than the earlier papers, while these later papers are also published in journals with a lower or equal JCSm score than before. Such a hypothetical shift in preference for subfields and journals reproduces the pattern of shifts in JCSm and FCSm values observed for UU and WU.

To test this hypothesis, we examined the pattern of publication of articles in the main subfields in terms of number of publications (P). Coverage of WU was at least 91 percent of P in both periods, while that of UU was at least 97 percent. To monitor the effect of publication shifts in subfields, the following procedure was applied. In each of the main subfields, we multiplied P in both periods with the corresponding FCSm score of 1989‐1993. In both cases, a (slight) decrease in FCSm resulted for the 1994‐1998 papers (see Table III). Similar results were obtained, if P was multiplied with the FCSm score of 1994‐1998 in both periods. These figures indicate a small shift towards lower impact subfields rather than a massive switch to subfields with a higher average impact. This is consistent with the fact that the main changes in publication activity (all increases) occurred for UU in the subfield “Veterinary Science” and for WU in that subfield as well as in “Agriculture, Diary and Animal Products”. Both of these subfields are characterized by a relatively low impact level compared to impact levels of other subfields in which these institutes publish. So, we have to conclude that a shift in preference for subfields fails to explain the observed increase in FCSm scores.

4. Changes in FCSm

4.1 Changes in FCSm within subfields

If the increase in FCSm is not to a large extent due to changes in preference between subfields, it may be related to changes in impact within subfields. An increase in FCSm might be either obtained by the addition of high impact journals to the subfield, or by the elimination of low impact journals (or by some combination of both). Research groups publishing in a stable set of journals containing neither the new high impact journals nor many of the eliminated low impact journals may then have a stable JCSm score, whereas their FCSm score increases. Of course, parallel results would be expected for both institutes across subfields.

To examine this hypothesis, we first identified the main subfields (in terms of number of publications) that showed the overall pattern at the level of institutes: for these, FCSm increased considerably, but JCSm to a much lesser extent, or even declined, while the JCSm/FCSm ratio declined substantially. Here, as before, only articles were considered. As Table IV shows, for WU this pattern of JCSm and FCSm changes was observed in three of their 15 main subfields, but these do not include their four most important subfields in terms of publications (Agriculture, Diary and Animal products, Veterinary Science, Genetics and; Heredity, and Nutrition and Dietetics).

For UU, a relatively minor change occurred in their main subfield Veterinary Science, where FCSm increased a little, but JCSm dropped, resulting in a JCSm/FCSm decline (see Table IV). In four other subfields, JCSm increased, but less steeply than FCSm. For both institutes, it is notable that six out of eight of the largest increases in FCSm occur in fields with an already high impact level, thus providing extra leverage on overall FCSm scores.

If we compare the results in subfields of both institutes, few parallel developments are visible. In the four main subfields listed in Table IV that UU and WU share (Immunology, Genetics and Heredity, Veterinary Science, and Endocrinology and Metabolism) under the listed conditions, only one parallel case is found. It seems that FCSm of articles does not generally increase with a concomitant falling behind of JCSm. It should be taken into account that both institutes differ in focus and publish differently across subfields. Even then, however, one might have expected, as visible for UU, a shift in the second major subfield of WU, Veterinary Science. Here, in contrast to UU findings, the WU results show stable JCSm, FCSm, and JCSm/FCSm figures. Only in one subfield, Immunology, both institutes do show parallel findings in accord with the hypothesis, at least for FCSm and JCSm/FCSm values. Therefore, we decided to examine this subfield in detail. Also, we took a closer look at the Genetics and Heredity subfield where FCSm as well as JCSm strongly increased for UU.

4.2 Changes in coverage of journals and FCSm

To allow a detailed examination of changes in journal coverage in both subfields, journals were classified into four categories according to their status as source journal for the citation indices (CI):

  1. 1.

    “Dropped” journals (i.e. after initial CI coverage somewhere in 1989‐1997, the journal is no longer a CI source journal in subsequent years).

  2. 2.

    “Whole period” journals (CI source journal each year during 1989‐1998).

  3. 3.

    “Interrupted” journals (i.e. irregular CI coverage in 1989‐1998; at least one year not covered in between two periods of coverage).

  4. 4.

    “New” journals (i.e. CI coverage of journal starts somewhere in 1990‐1998).

Recently, the importance of new journals to citation patterns, particularly in the Science Citation Index, has been emphasized by Leydesdorff (2003).

Some “dropped” journals may have reappeared as “new journals” after a change of name or a split‐up (e.g. Biochimica et Biophysica Acta, American Journal of Physiology). As both of these events frequently influence the coverage of topics and make‐up of journals, and thus may have an impact on the citation scores of the journal, we decided not to classify these journals under “whole period”. We computed the worldwide JCSm/FCSm scores for the four types of journals for all publications in each subfield, irrespective of UU and WU publication patterns. In this analysis, a journal is characterized by a single JCSm/FCSm score for each five‐year period (either 1989‐1993 or 1994‐1998), which covers all articles in that journal published during the five‐year period, as well as all citations to these articles received during that five‐year period.

It is clear that in the two subfields, both the dropped journals and the interrupted journals, have an impact below the subfield average, as evident from their JCSm/FCSm score (see Table V). The new journals have the highest average impact of the four categories in 1994‐1998, although in Immunology, the impact level of new journals that were covered already in 1989‐1993 was lower than average, whereas in Genetics and Heredity the new journals started with a very high average impact. By definition, none of the new journals obtained full five‐year impact scores in the first period (otherwise they would not have been “new”), and in many cases, papers of new journals were followed for considerably shorter periods. If the new journals reached their impact peaks only after several years, this may have contributed to a lower initial impact, as observed in one subfield (Immunology) in the first period.

However, in the other subfield (Genetics and Heredity), the initial impact of newly covered journals was higher in the earlier period than in the more recent one. Here, journals may have reached their impact peaks in a shorter period, or may have had top contributions in the first issues, which generated a high citation level from the start. In the two subfields, continuously covered journals declined a bit in impact. In both subfields, neglecting publication in newly covered journals may lead to a declining impact compared to the world subfield average. In Genetics and Heredity, the increase in impact of the “whole period” journals has not matched that of FCSm. Moreover, their article share dropped from 77 percent to 65 percent, whereas the article share of new journals increased from 12 percent to 32 percent. Note, however, that the impact of new journals can not explain the increase of FCSm in Immunology, as the impact of new journals is similar to that of the established, “whole period” journals, which, apparently, have kept up pace with the increased FCSm, and saw their article share decline from 86 percent to 82 percent. Thus, while new journals are important in both subfields, only in one subfield do they contribute substantially to differences in impact development of JCSm and FCSm. The next step is to assess the effects of journal change in the impact figures of the two universities (see Tables VI and VII).

Table VI shows the relative importance of the four types of journals for UU and WU across all their subfields. Also, a combined score for VAS in The Netherlands has been computed. Both UU and WU tend to publish the bulk of their articles in continuously covered journals (see Table VI). Publications in newly covered journals represent 8 percent‐10 percent of their total output, while about 4 percent appears in journals that were dropped by ISI. Publication in irregularly covered journals is of some importance only for UU, who publish frequently in the journal Veterinary Quarterly, which accounts for most of the publications in the “Interrupted” journal category.

To maximize reliability, results for all VAS articles are presented in Table VII. Then it becomes clear that on average, continuously covered ISI source journals (“whole period journals”) drop only slightly in impact between 1989‐1993 and 1994‐1998. The impact of irregularly covered journals (mostly Veterinary QuarterlyUU articles) drops precipitously, while in the most recent period, mostly relatively new journals with a low impact are selected as outlet. The impact of CI journals for which coverage is discontinued remains stable and below average. In 1994‐1998, few VAS papers appear in dropped journals, which effectively reduces the contribution of this category to the overall impact.

In general, impact of VAS journals declines in the major journal categories. Apparently, instead of selecting high impact new journals as outlet, relatively low impact new journals were chosen by VAS researchers. This lowers JCSm scores, and as a result, JCSm/FCSm scores. Following the same procedures as earlier in this article, we can compute that the combined effect of publishing changes between the two periods in dropped, new, and interrupted journals causes the JCSm/FCSm five‐year score of VAS to decrease by 4 percent (p < 0.05). This decline is largely due to the strongly increased number of VAS articles in both interrupted and new journals in 1994‐1998. Although the decline in JCSm/FCSm score due to this factor is by no means negligible, it still does not account for the major part of the observed variance in FCSm and JCSm scores on the level of articles. So, further study is needed. In the following section, results of a case study allow examination on an even more detailed level.

5. Performance in immunology

5.1 Overall impact scores

As a case study, we examine the performance of WU and UU in the subfield of Immunology. Here, the performance of both institutes declined significantly compared to the world subfield level (see Table VIII).

It is interesting to note that, even though many factors have been held constant, the JCSm and FCSm values for the two institutes differ considerably. For example, the JCSm for UU for 1994‐1998 is more than twice as high as that of WU in the same period. Although the FCSm value for both institutes is similar, but not equal, in 1989‐1993, the gap between the FCSm values of both institutes widens to almost a full point in 1994‐1998. While the difference in JCSm values may at least be partly due to a simple difference in the impact level of the UU and WU journals sets that are involved, the differences in FCSm values are less easily explained. The figures illustrate strikingly that it is dangerous to interpret the FCSm value, as constructed in the present set of performance indicators, as a simple indicator of the general impact level of a subfield.

The performance figures (CPP/FCSm) are about average for both institutes in 1989‐1993, but significantly below average in 1994‐1998. For both institutes, the FCSm value increases strongly: with 38 percent for WU and with 27 percent for UU. At the first blush, this seems to indicate that worldwide, Immunology articles are cited much more heavily in the recent period than in 1989‐1993. However, alternative explanations should be considered, as we shall see. First, we note that the JCSm value increases only moderately for UU (6 percent), while it even declines with more than a third for WU (−35 percent). In both cases, but strongly so for WU, the JCSm/FCSm indicator shows a considerable decline in the impact level of the journals. As the value for P shows, fractional counting has been applied, in which a paper in a journal that is classified in, for example, three subject categories is counted as 0.33 (P). Here, the FCSm value concerns only immunology citation scores.

5.2 The increase in FCSm scores

Especially the WU shows large and contrasting fluctuations in JCSm and FCSm over time. Therefore, we look at those results in more detail. Table IX presents the WU performance data for each year in the five‐year period (see column A), counting from −4 (1989, 1994 respectively) to 0 (1993, 1998 respectively). Columns B and C contain the number of publications for each year, while columns D and E present the matching FCS values for Immunology articles.

As Table IX shows, the oldest papers in both periods (respectively from 1989 (see column D) and 1993 (see column E) database years; indicated as year −4) are the most heavily cited ones, with an average of, respectively, 10.8 and 11.7 citations received in five years. Not surprisingly, the most recent papers in each period (respectively from 1993 and 1998; indicated as year 0) are hardly cited at all in the year of publication, the only year for which data have been collected.

If we compare the FCS data in columns D and E, an increase in FCS is visible. The FCS in 1994‐1998 is higher in each of the five years than the FCS score in corresponding years in 1989‐1993, although to a negligible extent for the 1991 and 1996 papers (indicated as year −2). On average, the FCS increases with 10.5 percent between 1989‐1993 (total: 26.4) and 1994‐1998 (total: 29.2).

Multiplying the number of publications (P) with the FCS scores renders the figures in columns headed by B*D and C*E. If these numbers are totaled and divided by the total number of articles in the corresponding periods, we obtain the FCSm scores already presented in Table VIII 5.2 for 1989‐1993 and 7.2 for 1994‐1998.

To isolate the effect of the change in FCSm score over time, the papers from one period can be weighted with the corresponding FCSm scores from the other period. If the WU papers from 1989‐1993 are weighted with the 1994‐1998 FCS weights, their average FCSm score would increase from 5.20 to 5.72, which accounts only for about a quarter of the 5.20 to 7.20 hike that was actually found (see column B*E).

Another factor is even more important. This concerns the distribution of papers across the five‐year periods. Older papers are weighted much more heavily on FCSm than more recent ones, as Table IX shows. Indeed, on the one hand, the WU has more papers in the first three years of the 1994‐1998 period (82 percent of its total output) than in the corresponding period of 1989‐1993 (only 63 percent of its total output). On the other hand, the WU counts only 18 percent of its papers in 1994‐1995, whereas it has 37 percent of its papers in the corresponding period of 1992‐1993. Thus, in the 1994‐1998 period, a much higher percentage of papers is published in “old” years with a high FCSm than in the 1989‐1993 period, which has an overrepresentation of papers in “young” years with a low FCSm value. As a result, when the 1994‐1998 papers are weighted with the 1989‐1993 FCS values, the FCSm increases substantially from 5.20 to 6.16 (see column C*D). Thus, there is a large effect of the distribution of articles over the period, accounting for nearly half of the observed FCSm increase.

However, only when we combine the multiplicative effects of increasing FCSm and distribution of articles over the period, the great increase in FCSm from 5.20 to 7.20 results. Similar factors are behind the less pronounced FCSm increase for UU.

The large power of the distributional effect can be illustrated by a hypothetical example. If another group had 20 Immunology papers published in 1993, and 20 also in 1994 (but none in other years), the FCSm for 1989‐1993 would amount to 4.40, whereas the FCSm for 1994‐1998 would have been 234.8, using the figures from Table IX, columns D and E. The resulting more than 53‐fold increase in FCSm would have been due for only 8 percent to an inherent increase in FCSm (p < 0.001), but for 92 percent to the distributional effect (p < 0.001).

5.3 The change in JCSm scores

However, although this does explain the observed increase in FCSm, it does not account for the meager trend results witnessed for the JCSm scores. If similar forces were active there as with FCSm, one would expect an increase of about 38 percent in JCSm for WU and of 27 percent for UU. However, as noted in , the JCSm score actually declines for WU, and it increases only modestly for UU.

Once again, we take a closer look at the most divergent case, the WU findings. Table X provides the distribution of JCSm scores of journals in which the WU published during 1989‐1993 and 1994‐1998. Note that JCSm scores involve whole counts instead of fractional counts, as used with FCSm scores, as publications need not be divided across subfields in JCSm scores. Notwithstanding the inclusion of a relatively large number of older publications, few articles from 1994‐1998 obtain a high JCS score (see Table X). Only one 1994‐1998 article obtains a JCSm higher than 3.5, whereas the earlier period counts seven articles surpassing this threshold. The normalized JCSm/FCSm scores show that nine articles in 1989‐1993 have appeared in journals with close to (five) or above average impact, versus just one in 1994‐1998. The bulk of the 1994‐1998 articles appears in journals with an impact that represents only 20 percent‐40 percent of the average value in the subfield. Apparently, the WU published its papers in 1994‐1998 in Immunology journals with both a much lower absolute and relative impact level than in 1989‐1993, resulting in the decrease witnessed in its JCSm score in 1994‐1998.

The average decrease in JCSm of the journal set in which WU published during 1989‐1998 was 11 percent between the earlier and the later period. This is considerably lower than the 35 percent decline in JCSm that was actually observed for the WU (see ). Therefore, the decrease in JCSm that is observed in Immunology is only for a minor part due to a decline in impact of the journals in which it publishes. The decrease in JCSm of WU is mostly due to a selection of lower impact journals in 1994‐1998 compared to 1989‐1993.

6. Summary and conclusions

Because of their importance as reference values, an observed increase in citation rates, in particular within a subfield, may be unwelcome. Such an increase means that one has to improve one's impact just to keep up with the rising impact level of one's subfields. It gets worse when, in addition to a rising FCSm, the impact of one's journals, as reflected in one's JCSm value, is decreasing. If one feels that the set of journals in which one is publishing is a reasonably good reflection of one's subfield(s), it may be baffling to see the average citation level in the subfield increase, while the citation level of the journals in which one publishes decreases. This is just what we have encountered in two case studies.

Changes in average citation scores, such as FCSm and JCSm, are hard to trace due to the multitude of factors impinging upon them. First, as the various document types that are included in our research performance studies differ in impact on both journal and subfield level, a shift in a research unit's preference of document types may conceivably lead to a considerable change in FCSm and JCSm values. However, in both institutes that were sampled, a shift in preference for document types led only to a minor increase in FCSm scores.

Second, a change in publication preference for subfields may affect JCSm and FCSm values, for example, as one changes from a low impact subfield to a high impact subfield. However, in the present study, the shift amounted mainly to intensification of publication in subfields characterized by a relatively low impact. This resulted in a (small) decrease rather than an increase in FCSm. Note that this shift in preference for subfields does not constitute a negative (nor positive) assertion about the research performance of research units, as impact levels across subfields cannot be compared directly. In the present cases, the observed shifts represent an intensification of publishing in the one or two main subfields of the research units. Thus, this factor does not help to explain a rise in FCSm, but may contribute to a modest decrease in JCSm scores.

Third, changes in overall JCSm and FCSm scores of the research units occurred in only a minority of their subfields rather than in all. In several subfields, FCSm scores increased for the two institutes while their JCSm scores increased to a lesser extent or declined. Findings failed to replicate across the institutes, with only one exception, the Immunology subfield.

Fourth, changes in journal coverage by the CI may affect citation impact averages of journals (JCSm) and subfields (FCSm). In two subfields with strongly increasing FCSm scores, Immunology and Genetics and Heredity, the impact of continuously covered journals declined compared to the subfield average over time, whereas the impact of newly covered CI journals was well above average in the most recent period. However, articles in new journals contributed only in one of the two subfields substantially to differences in impact development of JCSm/FCSm. These findings cannot be generalized, as results for all subfields show, when we retrieved the journal status of all articles of the institutes. Then, for both institutes, impact of continuously covered journals remained stable compared to the subfield average, while articles in newly covered journals had an impact below the subfield average. Publishing in newly covered journals with a below average impact level, as well as publishing in journals of which CI coverage has been interrupted had a significant negative effect (in all −4 percent) on JCSm/FCSm ratios of the two institutes. Both of these publication strategies resulted in articles in below average impact journals. The mixed findings for the two subfields and all articles show that it is important to monitor closely both current and new journals for impact developments in order to optimize one's publication strategy in different subfields.

At last, we managed to identify the main factor behind the increase in FCSm. A detailed case study showed the large effect of the distribution of articles across the five‐year periods that were studied. A relatively large percentage of papers published in the later years of a publication period leads to a lower FCSm score than average, as these publications have few years to get cited, and so their expected citation rate (FCSm) is low. In contrast, a high percentage of papers published in the first years of the period that is studied leads to a relatively highly FCSm score, as articles in the early years of a citation window have a longer period to get cited. A combination of the latter publication pattern with a tendency to publish in lower impact journals gives the pattern observed in the animal science and veterinary data: the combination of an increasing FCSm and a lagging JCSm, with strongly declining JCSm/FCSm scores. This distributional effect creates considerable variations in FCSm scores. An example using observed FCSm values showed a 53‐fold increase in FCSm, almost completely due to distributional effects. In general, this study shows that distributional effects associated with the use of block indicators can seriously affect citation impact indicators such as JCSm and FCSm.

In the present study, although the field citation score (FCSm) in the subfield that was studied showed a considerable inherent increase over time even after discounting distributional effects, the contribution of this to the actually obtained FCSm scores of papers was minor compared to the effect of the distribution of papers within a period. This distributional effect explained on its own nearly half of the increase in FCSm between 1989‐1993 and 1994‐1998, while its interaction with the general inherent increase of the impact level in Immunology accorded for another 25 percent. Thus, the effect of distribution of papers over the years is so large that it overrides all other, more modest, effects in determining FCSm values.

In general, distributional effects tend to be larger for smaller data sets. Indeed, we found that the effects tended to be stronger for the smaller institute, and did not occur so strongly in the main subfields of both institutes. Moreover, overall, we have found that the apparent increase in FCSm values was accounted for 75 percent by articles (which represent 91 percent of publications), while the other three document types accounted for 25 percent. As N increases, random changes in distribution of papers over years tend to cancel out. Note, however, that the distributional effect was considerable in the present study, although it covered nearly 3,800 papers. In early CWTS studies, fixed citation windows were used (e.g. three years), which allowed less scope for distributional effects than the present block indicators using variable citation windows (see Moed et al., 1995). However, fixed citation windows suffered from other drawbacks, such as a failure to cover publications (and their impact) from recent years.

It is evident that both JCSm and FCSm scores are determined by many factors. Especially when publication periods are compared, for example in trend studies, the actual values of JCSm and FCSm are often not very informative about the research performance of the group, nor of the development of impact over time in a certain subfield. JCSm and FCSm scores are norm scores that are tailored to patterns occurring in the publication sets of groups. Their main application is not solely as indicators, but as components in normalized scores, such as CPP/JCSm, CPP/FCSm, and JCSm/FCSm.

The same goes for the CPP score used as a “block” indicator. This value is also influenced by distribution and document type effects. However, the normalized indicators CPP/JCSm, CPP/FCSm and JCSm/FCSm are free from such effects and should therefore be consulted primarily in research performance assessments rather than their component parts. Here, every change in the distribution of papers across years is matched with a corresponding change in JCSm and FCSm values. Therefore, the comparison of the impact of a research group's impact with journal and subfield citation scores offers valid results whether or not block indicators are used.

The findings are of importance given the widespread use of block indicators in bibliometric studies, policy‐oriented reports (e.g. Science Watch, 2003) and elsewhere. This study may have contributed to an enhanced insight in factors that are of greater and lesser importance in determining those results.

Table I  Impact of target institutes

Table I

Impact of target institutes

Table II  Increase in FCSm by document type

Table II

Increase in FCSm by document type

Table III  FCSm as a function of publication shifts across subfields

Table III

FCSm as a function of publication shifts across subfields

Table IV  UU and WU subfields with increasing FCSm and decreasing JCSm/FCSm

Table IV

UU and WU subfields with increasing FCSm and decreasing JCSm/FCSm

Table V  JCSm/FCSm (world‐wide) over time for four types of journals

Table V

JCSm/FCSm (world‐wide) over time for four types of journals

Table VI  Journal coverage and article publication in all subfields

Table VI

Journal coverage and article publication in all subfields

Table VII  JCSm/FCSm over time of four types of journals for VAS articles in all subfields

Table VII

JCSm/FCSm over time of four types of journals for VAS articles in all subfields

Table VIII  Impact of articles of WU and UU in the Immunology subfield

Table VIII

Impact of articles of WU and UU in the Immunology subfield

Table IX  Distribution of articles over period and resulting FCSm scores

Table IX

Distribution of articles over period and resulting FCSm scores

Table X  Distribution of JCSm and JCSm/FCSm scores for journals in which WU publishes

Table X

Distribution of JCSm and JCSm/FCSm scores for journals in which WU publishes

References

Adair, J.G. and Vohra, N. (2003), “The explosion of knowledge, references, and citations: psychology's unique response to a crisis”, American Psychologist, Vol. 58 No. 1, pp. 1523.

Leydesdorff, L. (2003), “Can networks of journal‐journal citations be used as indicators of change in the social sciences?”, Journal of Documentation, Vol. 59 No. 1, pp. 84104.

Moed, H.F., De Bruin, R.E. and Van Leeuwen, T.N. (1995), “New bibliometric tools for the assessment of national research performance: database description, overview of indicators and first applications”, Scientometrics, Vol. 33 No. 3, pp. 381425.

Nederhof, A.J. and Van Wijk, E. (1999), “Profiling institutes: identifying high research performance and social relevance in the social and behavioral sciences”, Scientometrics, Vol. 44 No. 3, pp. 487506.

Nederhof, A.J., Van Leeuwen, T.N., Visser, M.S. and Van der Wurff, L.J. (1999), “Bibliometric profiles of academic Veterinary and Animal Sciences research in The Netherlands, 1989‐1998”, Veterinary and Animal Sciences, VSNU, Utrecht, pp. 5192.

Schubert, A. and Glänzel, W. (1983), “Statistical reliability of comparisons based on the citation impact of scientific publications”, Scientometrics, Vol. 5 No. 1, pp. 5974.

Schubert, A., Glänzel, W. and Braun, T. (1989), “Scientometric data files. A comprehensive set of indicators on 2,649 journals and 96 countries in all major science fields and subfields 1981‐1985”, Scientometrics, Vol. 16 No. 1, pp. 3478.

Science Watch (2003), “Twenty years of citation superstars”, Vol. 14 No. 5, pp. 12.

Van Raan, A.F.J. (1996), “Advanced bibliometric methods as quantitative core of peer review based evaluation and foresight exercises”, Scientometrics, Vol. 36 No. 3, pp. 396420.

Zitt, M., Ramanana‐Rahary, S. and Bassecoulard, E. (2003), “Correcting glasses help fair comparisons in international science landscape: country indicators as a function of ISI database delineation”, Scientometrics, Vol. 56 No. 2, pp. 25982.

Further Reading

Ingwersen, P., Larsen, B. and Noyons, E. (2001), “Mapping national research profiles in social science disciplines”, Journal of Documentation, Vol. 57 No. 6, pp. 71540.

Related articles