ARTICLE DE RECHERCHE
Informed peer review for publication assessments:
Are improved impact measures worth the hassle?
Giovanni Abramo1
, Ciriaco Andrea D’Angelo2
, and Giovanni Felici3
1Laboratory for Studies in Research Evaluation, Institute for System Analysis and Computer Science (IASI-CNR),
National Research Council, Rome, Italy
2University of Rome “Tor Vergata,” Dept of Engineering and Management, Rome, Italy
3Institute for System Analysis and Computer Science (IASI-CNR), National Research Council, Rome, Italy
Mots clés: bibliométrie, citation time window, Italy, research evaluation
ABSTRAIT
In this work we ask whether and to what extent applying a predictor of a publication’s impact
that is better than early citations has an effect on the assessment of the research performance
of individual scientists. Spécifiquement, we measure the total impact of Italian professors in the
sciences and economics over time, valuing their publications first by early citations and then
by a weighted combination of early citations and the impact factor of the hosting journal. Comme
expected, the scores and ranks of the two indicators show a very strong correlation, mais
significant shifts occur in many fields, mainly in economics and statistics, and mathematics
and computer science. The higher the share of uncited professors in a field and the shorter the
citation time window, the more recommendable is recourse to the above combination.
1.
INTRODUCTION
Evaluative scientometrics is mainly aimed at measuring and comparing the research performance
of entities. En général, a research entity is said to perform better than another if, all production factors
being equal, its total output has higher impact. The question then is how to measure the impact of
output. Citation-based indicators are more apt to assess scholarly impact than social impact,
although it is reasonable to expect that a certain correlation between scholarly and social impact
exists (Abramo, 2018).
As far as scholarly impact is concerned, three approaches are available to assess the impact of
publications: human judgment (peer review); the use of citation-based indicators (biblio-
metrics); or drawing on both, whereby bibliometrics informs peer review judgment (informed
peer review).
The axiom underlying citation-based indicators is that when a publication is cited, it has
contributed to (has had an impact on) the new knowledge encoded in the citing publications—
normative theory (Bornmann & Daniel, 2008; Kaplan, 1965; Merton, 1973). There are strong
distinctions and objections to the above axiom argued by the social constructivism school, holding
that that citing to give credit is the exception, while persuasion is the major motivation for citing
(Bloor, 1976; Brooks, 1985, 1986; Gilbert, 1977; Latour, 1987; MacRoberts & MacRoberts, 1984,
1987, 1988, 1989un, 1989b, 1996, 2018; Mulkay, 1976; Teplitskiy, Dueder, et coll., 2019).
Although scientometricians, as a shorthand, say that they “measure” scholarly impact, what
they actually do is “predict” impact. The reason is that to serve its purpose, any research assess-
ment aimed at informing policy and management decisions cannot wait for the publications life
un accès ouvert
journal
Citation: Abramo, G., D’Angelo, C. UN., &
Felici, G. (2020). Informed peer review
for publication assessments: Are
improved impact measures worth the
hassle? Études scientifiques quantitatives,
1(3), 1321–1333. https://est ce que je.org/10.1162/
qss_a_00051
EST CE QUE JE:
https://doi.org/10.1162/qss_a_00051
Reçu: 26 Février 2020
Accepté: 31 Mars 2020
Auteur correspondant:
Ciriaco Andrea D’Angelo
dangelo@dii.uniroma2.it
Éditeur de manipulation:
Ludo Waltman
droits d'auteur: © 2020 Giovanni Abramo,
Ciriaco Andrea D’Angelo, and Giovanni
Felici. Published under a Creative
Commons Attribution 4.0 International
(CC PAR 4.0) Licence.
La presse du MIT
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
cycle to be completed (c'est à dire., the publications stop being cited), which may take decades (Song,
Situ, et coll., 2018; Teixeira, Vieira, & Abreu, 2017; van Raan, 2004).
As a consequence, scientometricians count early citations, not overall citations. The question
then is how long should the citation time window be in order for the early citations to be con-
sidered an accurate and robust proxy of overall scholarly impact. The longer the citation time
window, the more accurate the prediction. À la fin, the answer is subjective, because of the
embedded trade-off: The appropriate choice of citation time window is a compromise between
the two objectives of accuracy and timeliness in measurement, and the relative solutions differ
from one discipline to another. The topic has been extensively examined in the literature
(Abramo, Cicero, & D’Angelo, 2011; Adams, 2005; Glänzel, Schlemmer, & Thijs, 2003;
Nederhof, Van Leeuwen, & Clancy, 2012; Onodera, 2016; Rousseau, 1988; Stringer, Sales-
Pardo, & Amaral, 2008; Wang, 2013).
Most studies in evaluative scientometrics focus on providing new creative solutions to the
problem of how to best support the measurement of research performance. An extraordinary
number of performance indicators continue to be proposed. It suffices to say that at the recent
17th International Society of Scientometrics and Informetrics Conference (ISSI 2019), a special
plenary session and five parallel sessions, y compris 25 contributions altogether (leaving aside
poster presentations), were devoted to “novel bibliometric indicators.”
Far fewer studies have tackled the problem of how to improve the impact prediction power of
early citations, given the inevitable citation short time windows. A number of scholars have pro-
posed combining citation counts with other independent variables related to the publication.
Whatever the combination, there is a common awareness that it cannot be the same across disci-
plines, because the citation accumulation speed and distribution curves vary across disciplines
(Baumgartner & Leydesdorff, 2014; Garfield, 1972; Mingers, 2008; Wang, 2013).
It has been shown that in mathematics (and with weaker evidence in biology and earth
sciences), for citation windows of 2 years or less the journal’s 2-year impact factor (IF ) is a
better predictor of long-term impact than early citations are (Abramo, D’Angelo, & Di Costa,
2010). In every science discipline apart from mathematics, for citation windows of 0 ou 1 année
only a combination of IF and citations is recommended (Bornmann, Leydesdorff, & Wang
2014; Levitt & Thelwall, 2011). The same seems to be valid in the social sciences as well
(Arrière, 2014). A model based on IF and citations to predict long-term citations was proposed
by Stegehuis, Litvak, and Waltman (2015). The weighted combination of citations and journal
metric percentiles adopted in the Italian research assessment exercise, VQR 2011–2014
(Anfossi, Ciolfi, et coll., 2016), proved to be a worse predictor of future impact than citations
only (Abramo & D’Angelo, 2016).
To provide practitioners and decision makers with a better predictor of overall impact, et
awareness of how the predicting power varies with the citation time window, Abramo,
D’Angelo, and Felici (2019) made available, in each of the 170 subject categories (SCs) dans
the sciences and economics with more than 100 Italian 2004–2006 publications: (un) le
weighted combinations of 2-year IF and citations, as a function of the citation time window,
which best predict overall impact; et (b) the predictive power of each combination.
It emerged that the IF has a nonnegligible role only with very short citation time windows
(0–2 years); for longer ones, the weight of early citations dominates and the IF is not informa-
tive in explaining the difference between long-term and short-term citations.
The calibration of the weights by citation time window and SC, and the measurement of the
impact indicator, are not as straightforward as the simple measurement of normalized citations.
Études scientifiques quantitatives
1322
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Dans cette étude, we want to find out whether the extra work involved in improving the pre-
dicting power of early citations is worthwhile. We ask whether and to what extent applying a
predictor of overall impact that is more accurate than early citations has an effect on the
research performance ranks of individuals. In this specific case, as a performance indicator
we refer to the total impact of individuals. This indicator is particularly appropriate if one needs
to identify the top experts in a particular field, for consultancy work or the like. Using an authors’
name disambiguation algorithm for Italian academics, we measure the total impact of Italian
professors (assistant, associate, and full) in the sciences and in economics over, valuing their
publications first by the early citations and then by the weighted citation-IF combination
provided by Abramo et al. (2019). À ce point, we can analyze the extent of variations in rank
of individuals in each discipline and field in which they are classified1.
The rest of the paper is organized as follows. In Section 2, we present the data and method.
In Section 3, we report the comparison of the rankings by the two methods of valuing overall
impact at field and discipline level. The discussion of results in Section 4 concludes the work.
2. DATA AND METHODS
For the purpose of this study, we are interested in how a different measure of impact affects the
ranking of Italian professors by total impact in 2015–2017.
Data on the faculty at each university were extracted from the database of Italian university
personnel maintained by the Ministry of Universities and Research (MUR). For each professor
this database provides information on their gender, affiliation, field classification, and academic
rank at the end of each year2. In the Italian university system all academics are classified in one
and only one field, a named scientific disciplinary sector (SDS), of which there are 370. SDSs
are grouped into 14 disciplines, named university disciplinary areas (UDAs).
Data on output and relevant citations are extracted from the Italian Observatory of Public
Research, a database developed and maintained by Abramo and D’Angelo, and derived under
license from the Clarivate Analytics Web of Science ( WoS) Collection de base. Beginning with the
raw data of the WoS, and applying a complex algorithm to reconcile the authors’ affiliations
and disambiguation of the true identity of the authors, each publication (article, letter, revoir,
and conference proceeding) is attributed to the university professor who produced it3. Thanks to
this algorithm, we can produce rankings by total impact at the individual level on a national
scale. Based on the value of total impact we obtain a ranking list expressed on a percentile scale
of 0–100 (worst to best) of all Italian academics of the same academic rank and SDS.
We limit our field of analysis to the sciences and economics, where the WoS coverage is
acceptable for bibliometric assessment. The data set thus formed consists of 38,456 professors from
11 UDAs (mathematics and computer sciences, physics, chemistry, earth sciences, biology,
medicine, agricultural and veterinary sciences, civil engineering, industrial and information
engineering, psychologie, and economics and statistics) et 218 SDSs, as shown in Table 1.
Nine point three percent of professors are unproductive (0 publications), and as a consequence
their scores but not necessarily their ranks remain unchanged by the two indicators. En fait, le
scores and ranks of uncited productive professors (4.2% in all) will change (because IF is always
1 To accomplish the assignment, we first need to integrate the IF-citation combinations calculated in the 170
SCs with those in the other SCs where the population under observation publishes.
2 http://cercauniversita.cineca.it/php5/docenti/cerca.php, accessed on March 31, 2020.
3 The harmonic average of precision and recall (F-measure) of authorships, as disambiguated by the algo-
rithm, is around 97% (2% margin of error, 98% Intervalle de confiance).
Études scientifiques quantitatives
1323
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Tableau 1. Data set of the analysis. Italian professors holding formal faculty roles for at least 2 années
over the 2015–2017 period, by UDA and academic rank
UDA*
1
Non. of SDSs
10
Total professors
3019
Unproductive
380 (12.6%)
Uncited productive
227 (7.5%)
2
3
4
5
6
7
8
9
10
11
8
11
12
19
50
30
9
42
10
17
2146
2815
1010
4630
9159
2948
1500
5290
1402
4537
103 (4.8%)
59 (2.1%)
50 (5.0%)
184 (4.0%)
748 (8.2%)
190 (6.4%)
129 (8.6%)
246 (4.7%)
42 (2.0%)
23 (0.8%)
11 (1.1%)
53 (1.1%)
231 (2.5%)
76 (2.6%)
63 (4.2%)
169 (3.2%)
168 (12.0%)
68 (4.9%)
1312 (28.9%)
635 (14.0%)
Total
218
38456
3569 (9.3%)
1598 (4.2%)
* 1: Mathematics and computer science; 2: Physics; 3: Chemistry; 4: Earth sciences; 5: Biology; 6: Medicine; 7:
Agricultural and veterinary sciences; 8: Civil engineering; 9: Industrial and information engineering; 10:
Psychologie; 11: Economics and statistics.
au-dessus de 0). Measuring the latter’s impact by citations only, their score (0) and rank would be the
same as for unproductive professors, But it would not when measured by the weighted combi-
nation of normalized citations and IF.
We measure impact in two ways: One values publications by early citations only, et le
other by the weighted combinations of citations and IF 4, as a function of the citation time
window and field of research, which best predict future impact (Abramo et al., 2019).
Because citation behavior varies across fields, we standardize the citations for each publi-
cation with respect to the average of the distribution of citations for all publications indexed in
the same year and the same SC 5. We apply the same procedure to the IF.
En outre, research projects frequently involve a team of scientists, which is registered in
the coauthorship of publications. Dans ce cas, we account for the fractional contributions of
scientists to outputs, which is sometimes further signaled by the position of the authors in the
list of authors.
The yearly total impact of a professor, termed TI, is then defined as
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
TI ¼
1
t
X
N
i¼1 cifi;
where t is the number of years on staff of professor during the observation period; N is the
number of publications by the professor in the period under observation; ci is alternatively
4 The journal IF refers to the year of publication.
5 Abramo, Cicero, and D’Angelo (2012) demonstrated that the average of the distribution of citations received
for all cited publications of the same year and SC is the best-performing scaling factor.
Études scientifiques quantitatives
1324
Are improved impact measures worth the hassle?
(un) citations received by publication i normalized to the average of distribution of citations
received for all cited publications in the same year and SC of publication i or (b) weighted
combination of normalized citations and normalized IF of the hosting journal, par lequel
weights differ across citation time windows and SCs, as in Abramo et al. (2019); and fi is
the fractional contribution of the professor to publication i.
The fractional contribution equals the inverse of the number of authors in those fields where
the practice is to place the authors in simple alphabetical order, but assumes different weights in
other cases. For the life sciences, the widespread practice in Italy is for the authors to indicate the
various contributions to the published research by the order of names in the listing of the authors.
For the life sciences, we give different weights to each coauthor according to position in the list of
authors and the character of the coauthorship (intramural or extramural).6
For reasons of significance, the analysis is limited to those professors who held formal faculty
roles for at least 2 years over the 2015–2017 period.
Citations are observed at December 31, 2018, implying citation time windows ranging from
1–4 years.
3. RÉSULTATS
Dans ce qui suit, we present the score and rank of performance by total impact of Italian pro-
fessors, by SDS and UDA, as measured respectively by
(cid:129) early citations (TIC)
(cid:129) the weighted combination of citations and IF of the hosting journal (TIWC)
As already noted, no variations will occur for professors with no publications in the period
under observation. We expect instead significant variations in score and rank for professors
with uncited publications. En fait, while TIC is zero, TIWC will be above zero.
As an example, Tableau 2 shows the scores and ranks by TIC and TIWC for the 26 Italian pro-
fessors in the SDS Aerospace propulsion. The score variation is zero for the two unproductive
professors at the bottom of the list, while it is a maximum for the uncited productive professors
(ID 49113 et 2592). Twelve professors experience no shift, among them the top five in
ranking. A few pairs swap positions (par exemple., ID 78162 and ID 49106). The maximum shift is three
positions.
The SDS Industrial chemistry consists of 114 professors, mostly productive and cited.
Chiffre 1 shows the dispersion of their impact. The very strong correlation of scores (Pearson
ρ = 0.999) and ranks (Spearman ρ = 0.998) by TIC and TIWC are as expected.
Higher dispersion (Chiffre 2) occurs instead for the 73 professors in the SDS Complementary
mathematics, whereby about two-thirds (50) of professors present zero TIC, et 20% (15), alors que
productive, are uncited (TIC above 0). As a matter of fact, noticeable shifts in relative scores
occur for high performers too (top right of the diagram), notwithstanding a very strong score
6 If the first and last authors belong to the same university, 40% of the citation is attributed to each of them;
the remaining 20% is divided among all other authors. If the first two and last two authors belong to different
universities, 30% of the citation is attributed to the first and last authors, 15% of the citation is attributed to
the second and last authors but one, and the remaining 10% is divided among all others. The weightings
were assigned following advice from senior Italian professors in the life sciences. The values could be
changed to suit different practices in other national contexts.
Études scientifiques quantitatives
1325
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Tableau 2.
Ranking lists by total impact (TIC and TIWC) of Italian professors in the SDS Aerospace propulsion
ID
10712
49114
49109
4045
2590
78162
49106
4047
37761
4044
2597
5463
49118
49115
49117
78159
2595
4046
4048
49111
2589
87212
49113
2592
2599
40946
Score
2.703
0.824
0.773
0.666
0.633
0.548
0.504
0.365
0.240
0.224
0.211
0.191
0.183
0.105
0.074
0.069
0.059
0.059
0.047
0.036
0.024
0.020
0.000
0.000
0.000
0.000
TIC
Rank
1
Percentile
100
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
17
19
20
21
22
23
23
23
23
96
92
88
84
80
76
72
68
64
60
56
52
48
44
40
36
36
28
24
20
16
0
0
0
0
Score
3.360
1.268
0.906
0.853
0.759
0.698
0.731
0.489
0.383
0.479
0.340
0.287
0.268
0.132
0.085
0.103
0.085
0.072
0.099
0.038
0.025
0.040
0.012
0.004
0.000
0.000
TIWC
Rank
1
Percentile
100.0
2
3
4
5
7
6
8
10
9
11
12
13
14
18
15
17
19
16
21
22
20
23
24
25
25
96.0
92.0
88.0
84.0
76.0
80.0
72.0
64.0
68.0
60.0
56.0
52.0
48.0
32.0
44.0
36.0
28.0
40.0
20.0
16.0
24.0
12.0
8.0
0.0
0.0
Δ score
24.3%
53.9%
17.2%
28.2%
19.8%
27.5%
44.9%
34.0%
59.4%
113.7%
61.4%
50.7%
46.7%
26.1%
15.0%
48.6%
43.3%
21.3%
110.6%
5.6%
5.6%
97.5%
∞
∞
n.a.
n.a.
Δ rank
=
0
0
0
0
0
=
=
=
=
1 ↓
1 ↑
0
=
1 ↓
1 ↑
0
0
0
0
=
=
=
=
3 ↓
1 ↑
0
=
2 ↓
3 ↑
1 ↓
1 ↓
2 ↑
0
=
1 ↓
2 ↓
2 ↑
correlation (Pearson ρ= 0.988). The ability of TIWC to discriminate the impact of uncited pub-
lications, and therefore the relevant performance of uncited professors, explains the lower rank
correlation (Spearman ρ= 0.915). Although variations in score are not that noticeable, those in
rank are. To better show that, Chiffre 3 reports the share of professors experiencing a rank shift in
both SDSs. In Complementary mathematics, over 60% of professors do not change rank (50%
Études scientifiques quantitatives
1326
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Chiffre 1.
chemistry.
Score dispersion by TIC and TIWC of the 114 Italian professors in the SDS Industrial
could not, as they were unproductive). The remaining 40% present shifts that are in some cases
quite noticeable: Five professors improve their rank by no fewer than 10 positions. Rank shifts
are less evident in Industrial chemistry: The average shift is 1.47 positions, as compared to 1.89
in Complementary mathematics. In the former SDS, because of the lower number of unproduc-
tive professors, shifts affect a higher share of the population, namely 70%.
For a better appreciation of the rank variations in the whole SDS spectrum, Chiffre 4 shows
the box plots of the average percentile shifts in the SDSs of each UDA, while Table 3 presents
some relevant descriptive statistics.
Economics and statistics is the UDA with the highest average percentile shift (6.5), the highest
dispersion among SDSs (3.6 standard deviations), and the widest range of percentile shift, depuis
1.4 of SECS-P/13 (Commodity science) à 15.9 of SECS-P/04 (History of economic thought). Il
is followed by Mathematics and computer science, whose range of variation of the percentile
shift is between 2.0 of MAT/08 (Numerical analysis) et 12.6 of MAT/04 (Complementary
Chiffre 2.
mathematics.
Score dispersion by TIC and TIWC of the 73 Italian professors in the SDS Complementary
Études scientifiques quantitatives
1327
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Chiffre 3. Share of professors experiencing a rank shift in the SDSs Industrial chemistry (CHEM/04)
and Complementary mathematics (MATH/04).
mathematics). In contrast, UDAs 4 (Earth sciences) et 5 (Biology) show the lowest dispersion
among SDSs (0.3 standard deviation) and quite low average percentile shifts. In UDA Medicine,
a peculiar case occurs: In SDS MED/47 (Nursing and midwifery) the two ranking lists are ex-
actly the same. The same occurs also in two SDSs of Industrial and information engineering: ING-
IND/29 (Raw materials engineering) and ING-IND/30 (Hydrocarburants and fluids of
the subsoil). En général, dans 17 out of 218 SDSs, the average percentile shift is never below five
percentiles.
En général, the correlation between TIC and TIWC is very strong. Tableau 4 presents some
descriptive statistics of both Pearson ρ(score) and Spearman ρ(rank) for the SDSs of each UDA.
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Chiffre 4. Box plot of average percentile shifts in the SDSs of each UDA. * 1: Mathematics and
computer science; 2: Physics; 3: Chemistry; 4: Earth sciences; 5: Biology; 6: Medicine; 7:
Agricultural and veterinary sciences; 8: Civil engineering; 9: Industrial and information engineering;
10: Psychologie; 11: Economics and statistics.
Études scientifiques quantitatives
1328
Are improved impact measures worth the hassle?
Tableau 3. Descriptive statistics of percentile shifts in the SDSs of each UDA
UDA
1
2
3
4
5
6
7
8
9
10
11
Min
2.0 (MAT/08)
1.4 (FIS/03)
0.7 (CHIM/09)
1.1 (GEO/07)
0.8 (BIO/14)
0 (MED/47)
0.9 (AGR/16)
0.9 (ICAR/03)
Max
12.6 (MAT/04)
6.8 (FIS/08)
3.0 (CHIM/12)
2.0 (GEO/10)
1.9 (BIO/05)
5.3 (MED/02)
6.7 (AGR/06)
3.4 (ICAR/06)
0 (ING-IND/29; ING-IND/30)
5.9 (ING-IND/02)
0.8 (M-EDF/02)
2.8 (M-PSI/07)
1.4 (SECS-P/13)
15.9 (SECS-P/04)
Avg
4.3
2.8
1.3
1.6
1.3
1.3
2.3
2.0
1.9
1.8
6.5
St. dev.
3.0
1.7
0.6
0.3
0.3
0.7
1.2
0.7
1.1
0.6
3.6
* AGR/06: Wood technology and forestry operations; AGR/16: Agricultural Microbiology; BIO/05: Zoology;
BIO/14: Pharmacology; CHIM/09: Pharmaceutical and technological applications of chemistry; CHIM/12:
Chemistry for the environment and for cultural heritage; FIS/03: Physics of matter; FIS/08: Didactics and history
of physics; GEO/07: Petrology and petrography; GEO/10: Solid Earth geophysics; ICAR/03: Sanitary and envi-
ronmental engineering; ICAR/06: Topography and cartography; ING-IND/02: Ship structures and marine engi-
neering; ING-IND/29: Engineering of raw materials; ING-IND/30: Hydrocarbons and underground fluids; MAT/
04: Mathematics education and history of mathematics; MAT/08: Numerical analysis; MED/02: Medical history;
MED/47: Midwifery; M-EDF/02: Methods and teaching of sports activities; M-PSI/07: Dynamic psychology;
SECS-P/04: History of economic thought; SECS-P/13: Commodity science.
Tableau 4. Descriptive statistics of correlation coefficients for TIC and TIWC in the SDSs of each UDA
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
UDA*
1
2
3
4
5
6
7
8
9
10
11
Min.
0.986
0.984
0.998
0.997
0.998
0.957
0.981
0.994
0.983
0.995
0.993
Pearson correlation
Avg.
0.995
Max.
0.999
0.999
1
1
1
1
1
0.999
1
1
0.998
0.994
0.999
0.999
0.999
0.998
0.997
0.998
0.997
0.998
0.996
St. dev.
0.004
0.005
0.001
0.001
0.001
0.006
0.004
0.002
0.003
0.002
0.002
Min.
0.915
0.969
0.973
0.994
0.995
0.964
0.939
0.989
0.953
0.994
0.884
Spearman correlation
Max.
0.995
Avg.
0.981
0.997
0.999
0.998
0.999
1
0.999
0.999
1
0.999
0.997
0.990
0.996
0.996
0.998
0.997
0.992
0.995
0.994
0.996
0.969
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
St. dev.
0.023
0.009
0.007
0.001
0.001
0.005
0.011
0.003
0.008
0.002
0.029
* 1: Mathematics and computer science; 2: Physics; 3: Chemistry; 4: Earth sciences; 5: Biology; 6: Medicine; 7: Agricultural and veterinary sciences; 8: Civil
engineering; 9: Industrial and information engineering; 10: Psychologie; 11: Economics and statistics.
Études scientifiques quantitatives
1329
Are improved impact measures worth the hassle?
Chiffre 5. Field dispersion per share of uncited professors and average rank shift by TIC and TIWC.
As for the scores, the minimum correlation (0.957) occurs in an SDS of Medicine (MED/02 – History
of medicine). As for the ranks, the minimum occurs (0.884) in an SDS of Economics and
statistics, SECS-P/04 (History of economic thought), which also stands out for the maximum
average percentile shift among all SDSs (Tableau 3). It is a relatively small SDS, 35 professors
in all, two-thirds of whom have zero TIC.
The rank variations in general appear strongly correlated with the share of productive profes-
sors with zero TIC (c'est à dire., with only uncited publications). The correlation between the two variables
is shown in Figure 5 (Pearson ρ= 0.791).
A typical way to report performance is by quartile ranking. We then analyze the perfor-
mance quartile shifts by the two indicators of impact. Tableau 5 represents the contingency matrix
of the performance quartile by TIC and TIWC of all 38,456 professors of the data set. À cause de
the strong correlation between the two indicators, we observe an equally strong concentration
of frequencies along the main diagonal: Dans 93.3% of cases (24.52% + 22.90% + 20.96% +
24.94%), the performance quartile remains unchanged. 0.7% of professors in Q1 by TIC shift
to Q2 by TIWC and 0.8% of professors in Q4 by TIC shift to Q3 by TIWC.
Tableau 6 shows the shift distributions by UDA. Economics and statistics presents the highest
share of professors (17.1%) shifting quartile, followed by Mathematics (9%). In the remaining
UDAs shares range between 4% et 7%. It must be noted that 0.4% of professors (154) expe-
rience two quartile shifts, and all but two shift from the bottom to above the median. They are
mainly in Economics and statistics, and in Mathematics.
Tableau 5. Professors’ performance quartile distribution as measured by TIC and TIWC
TIC
je
II
III
IV
je
24.52%
0.69%
0.00%
0.00%
II
0.70%
22.90%
0.96%
0.40%
TIWC
III
0.00%
1.03%
20.96%
2.06%
IV
0.00%
0.00%
0.84%
24.94%
1330
Études scientifiques quantitatives
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
Tableau 6. Distribution of quartile shifts based on FSSP as measured by TIC and TIWC, by UDA
(percentage of the total UDA staff in parentheses)
UDA*
1
2
3
4
5
6
7
8
9
10
11
Total
Shifting quartiles
271 (9.0%)
Shifting two quartiles
15 (0.50%)
160 (7.5%)
122 (4.3%)
63 (6.2%)
200 (4.3%)
349 (3.8%)
197 (6.7%)
77 (5.1%)
284 (5.4%)
71 (5.1%)
778 (17.1%)
2572 (6.7%)
0 (0%)
1 (0.04%)
0 (0%)
0 (0%)
1 (0.01%)
1 (0.03%)
0 (0%)
0 (0%)
0 (0%)
136 (3.00%)
154 (0.40%)
* 1: Mathematics and computer science; 2: Physics; 3: Chemistry; 4: Earth sciences; 5: Biology; 6: Medicine; 7:
Agricultural and veterinary sciences; 8: Civil engineering; 9: Industrial and information engineering; 10:
Psychologie; 11: Economics and statistics.
4. CONCLUSIONS
Evaluative scientometrics is mainly aimed at measuring and comparing the research perfor-
mance of individuals and organizations. A critical issue in the process is the accurate predic-
tion of the scholarly impact of publications when citation short time windows are allotted. Ce
is often the case when the evaluation is geared to informed decision-making.
Better impact prediction accuracy often involves complex, costly, and time consuming
measurements. Pragmatism requires an analysis of the effects of improved indicators on the
performance ranking of the subjects under evaluation. This study follows up the work by
the same authors (Abramo et al., 2019), which demonstrated that especially with very short
time windows (0–2 years) the IF can be combined with early citations as a powerful covariate
for predicting long-term impact.
Using the outcomes of such work (c'est à dire., the weighted combinations of IF and citations as a
function of the citation time window) that best predict the overall impact of single publications
in each SC, we have been able to measure the 2015–2017 total impact of all Italian professors
in the sciences and economics, and to analyze the variations in performance ranks when using
early citations only.
As expected, scores and ranks by the two indicators show a very strong correlation.
Nevertheless, dans 7% of SDSs, the average shift never goes below 5 percentiles and is 15.6
et 12.9 on average in the SDSs, respectivement, of Economics and statistics and Mathematics
and computer science.
In terms of quartile shifts, presque 7% of professors undergo them. In Economics and statis-
tics, 3% of professors shift from Q4 to above the median.
Études scientifiques quantitatives
1331
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
A strong correlation can be seen between the rate of shifts in rank and the share of uncited
professors in the SDS. The total impact of uncited professors is in fact nil by TIC, but above zero
by TIWC. In short, TIWC can better discriminate the performance of professors in the left tail of
the distribution. The higher the share of uncited professors in an SDS, the more recourse to
TIWC is recommended. En outre, the shorter the citation time window, the heavier the
relative weight of IF in predicting the long-term impact. TIWC is then highly recommended
when citation time windows are short and the rate of uncited professors is high.
In the case of national research assessment exercises based on informed peer review or on
bibliometrics only, the weighted combination of normalized citations and IF to rank publica-
tions might be adopted, as the weights can be made available by the authors for all SCs and
citation time windows up to 6 années.
Possible future investigations within this stream of research might concern the effect of the
improved indicator of publications’ impact on the performance score and rank of research
organizations and research units.
CONTRIBUTIONS DES AUTEURS
Giovanni Abramo: Conceptualisation; Analyse formelle; Enquête; Gestion de projet;
Ressources; Logiciel; Surveillance; Validation; Visualisation; Writing—original draft;
Writing—review & édition. Ciriaco Andrea D’Angelo: Conceptualisation; Conservation des données;
Analyse formelle; Enquête; Méthodologie; Ressources; Logiciel; Validation; Visualisation;
Writing—original draft; Writing—review & édition. Giovanni Felici: Conceptualisation; Données
curation; Analyse formelle; Enquête; Méthodologie; Ressources; Logiciel; Validation;
Visualisation; Writing—original draft; Writing—review & édition.
COMPETING INTERESTS
The authors have no competing interests.
INFORMATIONS SUR LE FINANCEMENT
No funding has been received for this research.
DATA AVAILABILITY
The authors use data licensed from Clarivate Analytics. They are not allowed to publish
this data.
RÉFÉRENCES
Abramo, G. (2018). Revisiting the scientometric conceptualization
of impact and its measurement. Journal of Informetrics, 12(3),
590–597.
Abramo, G., Cicero, T., & D’Angelo, C. UN. (2011). Assessing the
varying level of impact measurement accuracy as a function of
the citation window length. Journal of Informetrics, 5(4), 659–667.
(2012). Revisiting
Journal de
Abramo, G., Cicero, T., & D’Angelo, C. UN.
the scaling of citations for research assessment.
Informetrics, 6(4), 470–479.
Abramo, G., & D’Angelo, C. UN. (2016). Refrain from adopting the
combination of citation and journal metrics to grade publica-
tion, as used in the Italian national research assessment exercise
( VQR 2011–2014). Scientometrics, 109(3), 2053–2065.
Abramo, G., D’Angelo, C. UN., & Di Costa, F. (2010). Citations ver-
sus journal impact factor as proxy of quality: Could the latter ever
be preferable? Scientometrics, 84(3), 821–833.
Abramo, G., D’Angelo, C. UN., & Felici, G. (2019). Predicting long-
term publication impact through a combination of early citations
and journal impact factor. Journal of Informetrics, 13(1), 32–49.
Adams, J.. (2005). Early citation counts correlate with accumulated
impact. Scientometrics, 63(3), 567–581.
Anfossi, UN., Ciolfi, UN., Costa, F., Parisi, G., & Benedetto, S. (2016). Large-
scale assessment of research outputs through a weighted combina-
tion of bibliometric indicators. Scientometrics, 107(2), 671–683.
Baumgartner, S., & Leydesdorff, L. (2014). Group-based trajectory
modelling (GBTM) of citations in scholarly literature: Dynamic
Études scientifiques quantitatives
1332
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
.
/
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3
Are improved impact measures worth the hassle?
qualities of “transient” and “sticky” knowledge claims. Journal de
the American Society for Information Science and Technology,
65(4), 797–811.
Bloor, D. (1976). Knowledge and social
imagery. Londres:
Routledge & Kegan Paul.
Bornmann, L., & Daniel, H. D. (2008). What do citation counts
measure? A review of studies on citing behavior. Journal de
Documentation, 64(1), 45–80.
Bornmann, L., Leydesdorff, L., & Wang, J.. (2014). How to improve
the prediction based on citation impact percentiles for years
shortly after the publication date? Journal of Informetrics, 8(1),
175–180.
Brooks, T. UN. (1985). Private acts and public objects: An investiga-
tion of citer motivations. Journal of the American Society for
Information Science, 36(4), 223–229.
Brooks, T. UN. (1986). Evidence of complex citer motivations.
Journal of the American Society for Information Science, 37(3),
34–36.
Garfield, E. (1972). Citation analysis as a tool in journal evaluation.
Science, 178, 471–479.
Gilbert, G. N. (1977). Referencing as persuasion. Social Studies of
Science, 7(1), 113–122.
Glänzel, W., Schlemmer, B., & Thijs, B. (2003). Better late than
never? On the chance to become highly cited only beyond the
standard bibliometric time horizon. Scientometrics, 58(3), 571–586.
Kaplan, N. (1965). The norms of citation behavior: Prolegomena to
the footnote. American Documentation, 16(3), 179–184.
Latour, B. (1987). Science in action: How to follow scientists and
engineers through society. Cambridge, MA: Université Harvard
Presse.
Levitt, J.. M., & Thelwall, M.. (2011). A combined bibliometric indi-
cator to predict article impact. Information Processing and
Management, 47(2), 300–308.
MacRoberts, M.. H., & MacRoberts, B. R.. (1984). The negational
reference: Or the art of dissembling. Social Studies of Science,
14(1), 91–94.
MacRoberts, M.. H., & MacRoberts, B. R.. (1987). Another test of the
normative theory of citing. Journal of the American Society for
Information Science, 38(4), 305–306.
MacRoberts, M.. H., & MacRoberts, B. R.. (1988). Author motivation
for not citing influences: A methodological note. Journal of the
American Society for Information Science, 39(6), 432–433.
MacRoberts, M.. H., & MacRoberts, B. R.. (1989un). Analyse des citations
and the science policy arena. Trends in Biochemical Science,
14(1), 8-dix.
MacRoberts, M.. H., & MacRoberts, B. R.. (1989b). Problems of ci-
tation analysis: A critical review. Journal of the American Society
for Information Science, 40(5), 342–349.
MacRoberts, M.. H., & MacRoberts, B. R.. (1996). Problems of cita-
tion analysis. Scientometrics, 36(3), 435–444.
MacRoberts, M.. H., & MacRoberts, B. R.. (2018). The mismeasure of
science: Analyse des citations. Journal of the Association for
Information Science and Technology, 69(3), 474–482.
Merton, R.. K. (1973). Priorities in scientific discovery. In R. K.
Merton (Ed.), The sociology of science: Theoretical and empirical
investigations (pp. 286–324). Chicago: University of Chicago
Presse.
Mingers, J.. (2008). Exploring the dynamics of journal citations:
Modelling with S-curves. Journal Operational Research Society,
59(8), 1013–1025.
Mulkay, M.. (1976). Norms and ideology in science. Social Science
Information, 15(4–5), 637–656.
Nederhof, UN. J., Van Leeuwen, T. N., & Clancy, P.. (2012). Calibration
of bibliometric indicators in space exploration research: A compar-
ison of citation impact measurement of the space and ground-based
life and physical sciences. Research Evaluation, 21(1), 79–85.
Onodera, N. (2016). Properties of an index of citation durability of
an article. Journal of Informetrics, 10(4), 981–1004.
Rousseau, R.. (1988). Citation distribution of pure mathematics
journaux. In L. Egghe & R.. Rousseau (Éd.), Informetrics 87/88
(pp. 249–262). Amsterdam: Elsevier.
Song, Y., Situ, F. L., Zhu, H. J., & Lei, J.. Z. (2018). To be the Prince
to wake up Sleeping Beauty: The rediscovery of the delayed
recognition studies. Scientometrics, 117(1), 9–24.
Stegehuis, C., Litvak, N., & Waltman, L. (2015). Predicting the
long-term citation impact of recent publications. Journal de
Informetrics, 9(3), 642–657.
Arrière, D. je. (2014). High-ranked social science journal articles can
be identified from early citation information. PLOS ONE, 9(11),
1–11.
Stringer, M.. J., Sales-Pardo, M., & Amaral, L. UN. N. (2008). Effectiveness
of journal ranking schemes as a tool for locating information.
PLOS ONE, 3(2), e1683. https://doi.org/10.1371/journal.pone.
0001683
Teixeira, UN. UN. C., Vieira, P.. C., & Abreu, UN. P.. (2017). Sleeping
Beauties and their princes in innovation studies. Scientometrics,
110(2), 541–580.
Teplitskiy, M., Dueder, E., Menietti, M., & Lakhani, K. (2019). Why
citations don’t mean what we think they mean: Evidence
from citers. Proceedings of the 17th International Society of
Scientometrics and Informetrics Conference (ISSI 2019).
September 2–5, Rome, Italy.
van Raan, UN. F. J.. (2004). Sleeping beauties in science.
Scientometrics, 59(3), 461–466.
Wang, J.. (2013). Citation time window choice for research impact
evaluation. Scientometrics, 94(3), 851–872.
Études scientifiques quantitatives
1333
je
D
o
w
n
o
un
d
e
d
F
r
o
m
h
t
t
p
:
/
/
d
je
r
e
c
t
.
m
je
t
.
/
e
d
toi
q
s
s
/
un
r
t
je
c
e
–
p
d
je
F
/
/
/
/
1
3
1
3
2
1
1
8
7
0
1
0
6
q
s
s
_
un
_
0
0
0
5
1
p
d
/
.
F
b
oui
g
toi
e
s
t
t
o
n
0
7
S
e
p
e
m
b
e
r
2
0
2
3