Critical rationalism and the search for standard (field-normalized) indicators in bibliometrics

Reading time: 5 minute
...

📝 Original Info

  • Title: Critical rationalism and the search for standard (field-normalized) indicators in bibliometrics
  • ArXiv ID: 1805.06153
  • Date: 2020-10-01
  • Authors: - Lutz Bornmann - Rüdiger Mutz - Wolfgang Glänzel —

📝 Abstract

Bibliometrics plays an increasingly important role in research evaluation. However, no gold standard exists for a set of reliable and valid (field-normalized) impact indicators in research evaluation. This opinion paper recommends that bibliometricians develop and analyze these impact indicators against the backdrop of Popper's critical rationalism. The studies critically investigating the indicators should publish the results in such a way that they can be included in meta-analyses. The results of meta-analyses give guidance on which indicators can then be part of a set of indicators used as standard in bibliometrics. The generation and continuous revision of the standard set could be handled by the International Society for Informetrics and Scientometrics (ISSI).

💡 Deep Analysis

Figure 1

📄 Full Content

Bibliometrics plays an increasingly important role in research evaluation (Bornmann, 2017). An evaluation report without bibliometrics is currently difficult to imagine. In the application of bibliometrics in evaluative practice, one can differentiate between professional and citizen bibliometrics (Leydesdorff, Wouters, & Bornmann, 2016). Citizen bibliometrics is based on simple indicators, such as the median and total number of citations. These indicators can possibly be used to compare similar units (concerning size and age) in the same or very similar fields. An important disadvantage of these indicators is, however, that they cannot be used in cross-field comparisons of units (e.g. research groups or institutions). The publication and citation cultures are so different in the fields that the median and total number of citations from different fields are not comparable.

In professional bibliometrics, established field-normalized indicators are applied in cross-field comparisons, which take into account that citations occur in different publication and citation cultures (Waltman, 2016). Normalized indicators can be used to assess whether a unit (e.g. an institute) shows above or below average performance in the corresponding fields.

Most of the studies in bibliometricsscientific studies and applied studies used to support research evaluationsare actually cross-field comparisons, because bibliometrics is especially suitable for greater units (e.g. institutes or countries) and these units publish as a rule in different fields. We argue in this opinion paper that bibliometrics should aim to develop standard approaches especially to field-normalization. Only with standard approaches, the results of different studies are comparable, reliable, and interpretable.

In bibliometrics, however, several variants of normalized indicators exist. For example, mean-based indicators compare the mean impact of the focal papers with the mean impact of papers published in the same field and publication year. Percentile-based indicators assign to every paper in a certain field and publication year a citation percentile, which can be used for cross-field comparisons. Some bibliometricians recommend mean-based citation indicators (e.g., the characteristic scores and scales method; see Glänzel & Schubert, 1988) and others deem percentiles the best alternative among the indicators (Bornmann, Leydesdorff, & Mutz, 2013). Further, different variants of mean-based and percentile-based indicators are currently in use (Todeschini & Baccini, 2016). No gold standards exist which can be recommended for the general application of impact indicators in research evaluation.

If it should be the aim of bibliometrics to develop indicators which gain general acceptance (among bibliometricians and beyond), it is essential to criticize the available indicators sharply. This is especially necessary for field-normalized indicators, because these indicators haveas a rulea more complicated design than simple indicators such as mean citation rates. We propose that indicators should be continuously investigated as to their reliability and validity against the backdrop of Popper’s critical rationalism (Gingras, 2016;Lindner, Torralba, & Khan, 2018;Ulrich, 2006): bibliometricians should try to create or find situations in which the indicators possibly do not show the desired behavior. For example, an indicator which is intended to reflect the quality of publications should be correlated with the assessments of peers (Bornmann & Marx, 2015). If the correlation turns out to be low, the validity of the indicator can be questioned. Situations should be created or found in empirical research in which the measurement with the new indicator can fail. An indicator is regarded as preliminarily valid if these situations could not be found or realized. Such an indicator should gain general acceptance in citizen and professional bibliometrics.

In bibliometrics, we frequently find inadequate investigations of the validity of indicators in the literature. Good examples are the developments of h index variants since the introduction of the h index in 2005 (Hirsch, 2005). Since then, about 50 variants have been introduced (Bornmann, Mutz, Hug, & Daniel, 2011). In many of the corresponding papers, one or several weaknesses of the h index are unrevealed, before the new variant is introduced (see, e.g., Tol, 2009). Then, it is demonstrated on the base of one or several examples that the new variant performs (significantly) better than the original index. However, these tests do not guarantee that the indicators have been proved successfully. The authors seem to be more interested in producing favorable results for the new indicators than their empirical validation.

For critical validation, situations should be sought in which the indicators probably fail and these situations should be specifically investigated. With the elimination of failed indicators, the set of indic

📸 Image Gallery

cover.png

Reference

This content is AI-processed based on open access ArXiv data.

Start searching

Enter keywords to search articles

↑↓
ESC
⌘K Shortcut