RESEARCH ARTICLE

RESEARCH ARTICLE

Methodological considerations for identifying
questionable publishing in a national context: Der
case of Swedish Higher Education Institutions

Gustaf Nelhans1,2

and Theo Bodin3

1University of Borås, Schweden
2University of Southern Denmark, Denmark
3Karolinska Institutet, Schweden

Schlüsselwörter: national level, publication blacklists, publication ethics, questionable publishing

ABSTRAKT

The overall scope of this study is an attempt at a methodological framework for matching
publication lists at the national level against a combined set of blacklists for questionable
Veröffentlichung. Using the total verified set of publications from Swedish Higher Education
Institutions (HEI) as a case, we examined the number, distribution, and proportion of
publishing in questionable journals at the national level. Journal publication data was
extracted from the national SwePub database and matched against three curated blacklists
of questionable publishing. For the period 2012–2017, we identified 1,743 published
papers in blacklisted journals, equal to an average of 0.5–0.9% of the total publications
from Swedish HEIs. There was high variability between different HEI categories, mit
more established universities at the lower end of the spectrum, while university colleges
and new universities had a much higher proportion (∼2%). There was a general decreasing
trend during the study period (ρ = 0.83) for all categories of HEIs. The study presents a
methodology to identify questionable publishing in academia that could be applied to
other countries with similar infrastructure. Daher, it could serve as a starting point for the
development of a general framework for cross-national quantitative estimation of
questionable publishing.

1.

EINFÜHRUNG

There is an increasing focus on publications in scholarly research. Previously regarded as the
outcome of a successful research project, researchers are nowadays often expected to specify
how their results will be published already when planning research projects or applying for
funding. Publishing also plays a role in the evaluation of individual researchers’ performance
and the allocation of resources to and within academia. This is often popularized by the slogan
“publish or perish.” Over the last decade, there has also been a rise in open access (OA) jour-
nals financed through article processing charges (APCs), paid mainly by the author, instead of
the older system of subscription fees paid by the reader.

These two parallel trends have opened up a market for unscrupulous actors preying on un-
wary or desperate researchers. Dubious editorial activities include claiming to give peer review,
although clearly not doing so, as well as accepting any submitted manuscript as long as the APC
is paid. These journals also operate under false pretenses (z.B., using high-status entity names in

Keine offenen Zugänge

Tagebuch

Zitat: Nelhans, G., & Bodin, T.
(2020). Methodological considerations
for identifying questionable publishing
in a national context: The case of
Swedish Higher Education Institutions.
Quantitative Science Studies, 1(2),
505–524. https://doi.org/10.1162/
qss_a_00033

DOI:
https://doi.org/10.1162/qss_a_00033

Erhalten: 2 Juli 2019
Akzeptiert: 10 Februar 2020

Korrespondierender Autor:
Gustaf Nelhans
gustaf.nelhans@hb.se

Handling-Editor:
Ludo Waltman

Urheberrechte ©: © 2020 Gustaf Nelhans and
Theo Bodin. Published under a
Creative Commons Attribution 4.0
International (CC BY 4.0) Lizenz.

Die MIT-Presse

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

/

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

titles without having real connections, inventing fake journal impact metrics, or listing promi-
nent researchers as editors without their knowledge). Some studies have also raised the issue of
hijacked journals (Dadkhah & Maliszewski, 2015; Jalalian & Dadkhah, 2015; Shahri, Jazi, et al.,
2018) where the titles and ISSNs of both existing and discontinued journals have been picked up
by shady actors. The problem is perceived as so widespread that tools for distinguishing real from
hijacked journals have been developed (Asadi, Rahbar, et al., 2017).

The phenomenon is sometimes referred to as “predatory” publishing, but due to the intri-
cate activities of several actors as described above, a better term might be “questionable” or
“unethical” publishing practices, thus not blaming any single actor, but viewing it as a phe-
nomenon worthy of empirical study in itself. The reasons researchers chose to publish in such
journals are lack of awareness, speed, and ease of the publication process, and a chance to get
work rejected elsewhere published (Shaghaei et al., 2018).

The ethical aspects of questionable publishing have been conceptualized and discussed in
several publications and in a variety of research fields, such as nursing, biomedicine, econom-
ics, and agriculture (da Silva, 2017; Eriksson & Helgesson, 2017; Ferris & Winker, 2017;
Gasparyan, Yessirkepov, et al., 2016; Josephson & Michler, 2018; McLeod, Savage, &
Simkin, 2018).

Questionable publishing practices negatively affect legitimate journals, policymakers, Und
research funders and are seen by most stakeholders as an imminent threat to institutional
integrity, reputation, and public support. The issue is of great importance and merits an in-
vestigation, especially in countries like Sweden, where almost all universities are public and
most research is financed by taxpayers. Jedoch, academia is struggling to find ways to tackle
this issue.

zuerst, there is no consensus on how to identify and classify journals as predatory or not.
Zweitens, as many predatory journals are not indexed in established databases, the identifica-
tion of questionable published papers is difficult (Shen and Björk 2015). Even when agreed
criteria are developed, the authors found that it was difficult to evaluate single publishers
and journals based on these, since stated information did not always seem to match practical
considerations about how publishing actually was done, in which country they operated, oder
how the peer review process actually worked. Folglich, it has not been possible to esti-
mate the full scope of the problem.

Early attempts to categorize and collect blacklists of predatory journals have been met with
complaints and alleged hostility toward the creators and curators of such lists. A list curated by
an individual librarian at the University of Colorado, Jeffrey Beall, was publicly available but
discontinued in 2017, although archives exist online. Despite the merits and impact of Beall’s
blacklist, justified criticism has also been raised against it, especially regarding how some
criteria for inclusion on the list were employed. Auch, due to difficulties in maintaining the
list, its accuracy over time was debated. Jedoch, most attempts to assess or discuss the issue
of predatory publishing have frequently used Beall’s list as their only source, and thus it has
been highly influential.

Several actors have made efforts to fill the void after the discontinuation of Beall’s list. Der
US-based publisher of journal directories, Cabell Inc., recently launched a new blacklist of
journals, where they score journals against 66 publicly available quality criteria (Das &
Chatterjee, 2018; Hoffecker, 2018). The Ministry of Health and Medical Education in Iran also
curates a publicly available blacklist and the Directory of Open Access Journals (DOAJ) Profi-
vides a list that contains journals that are no longer indexed in the general DOAJ directory due

Quantitative Science Studies

506

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

.

/

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

to the stated ground of “suspected editorial misconduct by publisher.” Here, we propose that
using a combined set of blacklists overcomes some of the issues of lists being biased toward
specific research areas or interests. While a certain overlap is expected, we argue that the ad-
dition of more sources is needed, since this is a new area where no established practices for
defining and identifying sources have been developed.

In a wider context, there is a discussion regarding the role of standards and how institutional
knowledge is constructed in a specific context within science and technology studies. Es
should also be noted that there is an ongoing debate as to whether blacklisting of journals
fulfills any valuable purpose at all (Editorial, 2018; Strinzel, Severin, et al., 2019). Some argue
that this is an issue that cannot be weeded out as long as there is a demand from authors and
that the solution is to rank all journals. Indexing of legitimate journals, using so-called white-
lists, is another approach, where large journal directories such as Web of Science ( WoS),
Scopus, PubMed, and the DOAJ act as gatekeepers. Jedoch, populating a whitelist with
acceptable journals is arguably as difficult as identifying questionable journals, and it is well
known that suspected questionable journals manage to get listed in these directories
(Bohannon, 2013; Haider & Åström, 2017; Manca, Moher, et al., 2018). Zusätzlich,
noninclusion in a whitelist is not a clear sign that a journal is questionable, in the same
manner that noninclusion in a blacklist signals that the journal is of acceptable quality. Das
is because there could be many different reasons for not being included, such as the journal
being newly instigated or not known to the indexers. In a study that compares two blacklists
(Cabell’s and Beall’s) and two whitelists (Cabell’s and DOAJ), it was found that there are
overlaps in the contents of both blacklists and whitelists, and that inclusion criteria are vague
and biased (Strinzel et al., 2019). In this study we handled the latter by evaluating the criteria
used in the inclusion process for the blacklists, only using criteria that we found to be ade-
quate for the purposes of the study.

The closest studies to the present one are Eykens, Waffen, et al. (2018) and Eykens, Waffen,
et al. (2019). For some time, this group has used Beall’s lists in conjunction with WoS and
the DOAJ whitelist in order to screen registered publications in the performance-based re-
search funding system in Flanders, Belgien. Lately, they have added Cabell’s list to their
efforts. In their study, they reported 556 “potentially predatory open access” articles during
the years 2004 Und 2015 in the Flemish VABB-SHW publication system. Despite the short-
comings of blacklists, we believe that they are currently the best available method to identify
questionable research. Taken together they provide a breadth of inclusion criteria, yet focus
on the issue at hand.

1.1. Aims

This paper provides a methodological framework for evaluating the scope of national publi-
cations in questionable journals during 2012–2017. Sweden is used as a case for testing the
methodology since the authors are familiar with its HEI system. The aim can be articulated in
the following specific research questions:

1. What proportion of serial academic publishing can be matched to alleged questionable

publishing and are there any trends over time?

2. Are there differences in questionable publishing patterns between

A. different types of HEIs?
B.

research areas?

3. To what degree do different journal blacklists cover the same journal titles?

Quantitative Science Studies

507

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

.

/

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

2. MATERIALS AND METHODS

Hier, we will describe the methodological framework developed for matching suspected
questionable publishing at the national level. The choice of using Sweden as a case was de-
liberate, since the authors are familiar with its HEI landscape and have good contacts within
the National Library of Sweden, which hosts the SwePub publication database that was used
as base data. Interessant, it was found that third-party data such as CrossRef DOIs could not
be used, since many questionable publishers’ publications are not indexed in any way. To
produce solid research on these data, we developed a quite labor-intensive semiautomatic
Ansatz. daher, instead of making a comparative analysis at a shallow level, which might
contribute to fuelling preconceived notions when comparing countries, we opted for an in-
depth study to look at different kinds of higher education institutions. The Swedish HEIs are
divided between broader (comprehensive) established universities, specialized universities,
new universities, and university colleges (Hansson, Barriere, et al., 2019). This made it possi-
ble to try to understand if there are differences at this level instead. It is our intention to scale
up the methodology for other countries providing the availability of comparable publication
Daten.

2.1. Sources of Blacklisted Journals

2.1.1. Cabell’s blacklist
We received the blacklist curated by Cabell’s International Inc in August 2018. It contained
9,503 journals from 446 unique publishers. It included information on journal name, publisher,
Und (for some journals) ISSNs. Wie oben beschrieben, Cabell’s uses 66 different criteria to qualify the
inclusion of journals. During the process, we decided to benchmark the Cabell criteria against
subjective judgments of the “seriousness” of each criterion. The two authors ranked each cri-
terion based on a three-level scale, ranging from 1, “not serious,” to 3, “highly serious” criteria.
The rule for inclusion of a Cabell’s blacklisted journal was the following:

1. One or more 3s,
2. Two or more 2s, oder
3. Eins 2 and at least two 1s.

In all, 49 entries were identified as published in one of 14 journals that had articles matched
only in Cabell’s that were subjectively judged not serious. Six more journals were identified,
but since these were also matched in either the Iran or the DOAJ list, these articles were left in
the data set. The results are found in the supplementary material, Table XI.

2.1.2. The Iran MHME list

This is a curated blacklist maintained by the Ministry of Health and Medical Education in Iran
(http://blacklist.research.ac.ir/). It includes information on journal name, publisher, ISSN,
Journal website URL, and when the journal was added to the list. There is a field in the data-
base, called ”Status,” with two different inclusion criteria stated:
, Bedeutung
“invalid” and “fake,” respectively. Journals matching both criteria were included in the match-
ing process. The first set of criteria consisted of 12 journals that are known from the literature to
be hijacked (Jalalian & Dadkhah, 2015). In matching, publications from these journals were
manually inspected to identify if it was a publication in the hijacked version that was matched.
The complete list was downloaded on September 5, 2018 and contained 2,180 titles from 88
different publishers.

ﻧـﺎﻣـﻌـﺘـﺒـﺮ

ﺟـﻌـﻠـﯽ

Und

Quantitative Science Studies

508

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

/

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

2.1.3. DOAJ list

Seit 2014, the DOAJ has provided a Google spreadsheet workbook of journals “Added,”
“Removed,” and “Failed to submit a reapplication” of changes to the database. Wir haben uns entschieden
the most conservative approach, selecting only journals that have been removed from their
directory list because of “Suspected editorial misconduct by publisher” (https://blog.doaj.
org/2014/05/22/doaj-publishes-lists-of-journals-removed-and-added/), leaving journals out of
the matching exercise regarded as “not adhering to best practice,” (1,349 entries) Und
“Ceased publishing” and “Inactive” (259 Und 238 entries, jeweils). The complete list
was downloaded from their website (see URL above) on October 6, 2018 and contained
642 titles with associated ISSNs that were removed due to “Suspected editorial misconduct.”

2.2. Publication Databases

2.2.1. SwePub

SwePub is the centralized Swedish collection of published literature from all Swedish univer-
sities and a selection of public research organizations (Sīle, Pölönen, et al., 2018). It was de-
veloped by the National Library of Sweden. It collects data from the local publication
databases of its participants. Most Swedish research organizations, including HEIs, regularly
upload their publication lists to SwePub and maintain the database, including correcting errors
and managing duplicates.

2.2.2. Clarivate Web of Science list

The Swedish Research Council calculates the research output of Swedish HEIs annually based
on Clarivate WoS data. The data on full and fractionalized publications per HEI and year for
the time period were obtained with kind permission (Akt # 3.1-2018-6882).

2.3. Search Strategy

Publication data for all Swedish Higher Education Institutions during the years 2012 durch
2017 were extracted from SwePub on October 4, 2018, by using a SparQL query for extracting
journal data information. A total of 651,002 entries were returned in the query.

In SwePub we included all “journal publishing” categories. All but 19 entries were labeled
with “ref,” meaning that they had supposedly undergone peer review before being published.
After the full matching procedure described below, the search results included entries labeled
as editorials (n = 2), letters (n = 2), journal articles (n = 1,716) and review articles (n = 25). In
order to capture the scientific literature, we did not include “magazine publishing,” a category
which includes other types of periodical publishing, such as leisure magazines, Zeitungen,
and popular science magazines.

2.3.1. Data management

In order to reduce the risk of false positives in the blacklists due to journals with titles very
similar to already established journals (homonyms), we conducted manual searches in the
ISSN portal online (https://portal.issn.org). By establishing information such as country of pub-
lication and the organization responsible for the journal, and by browsing the journal’s web-
Seiten, we were able to identify the relationship between an article and the specific journal. Für
Beispiel, by identifying the article at the journal’s webpage, or by matching the DOI address
structure with other articles in the same journal, we were able to determine the relationship to
a high degree of certainty. We also matched journal titles and publishers with the SwePub data
by using DOIs to identify ISSNs where these were missing in the source data.

Quantitative Science Studies

509

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

.

/

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

Publication databases such as SwePub are prone to errors, as they are built on data origi-
nally submitted by researchers, which is then evaluated and corrected by librarians prior to
uploading the data to the national database (z.B., titles could have spelling errors or be short-
ened, and ISSN IDs could be written in different forms). Because of the size of the original data
set (almost 1 million entries), we generally did not correct the data, other than harmonizing the
ISSNs that were sometimes misrepresented by missing or different hyphen types.

Data was deduplicated at record ID level (_recordID).

Due to a known error in the SwePub database at the time of data extraction, some data for
Malmö University College was duplicated. These publications were manually removed from
the data set during analysis. Zusätzlich, the Swedish Agricultural University had not reported
publications correctly in 2017, and consequently had no data to match in that year.

Journal titles (_channel) und Verleger (_publishers), as well as ISSNs (as found in SwePub),
were matched against the blacklists in October (using the latest snapshots from the respective
blacklists (Cabell’s, latest entry 09 11 2018; IRAN MHME, datiert 08 05 2018, and DOAJ, latest
entry 10 01 2018).

At the suggestion of one of the reviewers, we conducted two additional tests to evaluate
that matched entries would not be found to be false positives: We first matched the set of
matched entries to the so-called Norwegian list (NSD; Oktober 29, 2019 NSDs register over
vitenskapelige publiseringskanaler – tidskrifter og serier.xlsx). Zusätzlich, we matched en-
tries containing a WoS ID, Document Object Identifier (DOI), or PubMed (PMID) ID. The re-
sults of each of these steps are reported below.

Matching publication records with blacklisted journals was carried out in a hierarchical

procedure.

Identical ISSNs in SwePub and blacklist(S).
Identical journal title AND publisher name in SwePub and blacklist(S).

1.
2.
3. Manual matching of entries with matching journal titles using the DOI string with the
same structure as a matched publication, and/or using other metadata information, solch
as article title, Namen, and affiliations, to strengthen the link between the SwePub entry
and the journal at hand.

After this procedure, 1,799 published works in blacklisted journals were identified.

Hereafter, a manual process took place of selecting only journals matching our evaluative
criteria for Cabell’s (49 entries omitted) as well as the evaluation against the NSD (0 entries
omitted) and WoS databases (7 entries omitted) for whitelists.

A total of 1,743 published works in blacklisted journals were identified. 1,548 were iden-
tified in step 1, 34 in step 2, Und 161 in step 3. The final data set of matched publications is
available in the supplementary material.

2.4. University Type Aggregation

Sweden has four distinct types of HEIs and results were aggregated according to this classifi-
cation (Hansson et al., 2019). The University of Gothenburg, Linköping University, Lund
Universität, Stockholm University, Umeå University, and Uppsala University are part of the
broader (comprehensive) established universities.

Quantitative Science Studies

510

l

D
Ö
w
N
Ö
A
D
e
D

F
R
Ö
M
H

T
T

P

:
/
/

D
ich
R
e
C
T
.

M

ich
T
.

/

e
D
u
Q
S
S
/
A
R
T
ich
C
e

P
D

l

F
/

/

/

/

1
2
5
0
5
1
8
8
5
7
8
9
Q
S
S
_
A
_
0
0
0
3
3
P
D

/

.

F

B
j
G
u
e
S
T

T

Ö
N
0
7
S
e
P
e
M
B
e
R
2
0
2
3

Identifying questionable publishing in a national context

Chalmers University of Technology, Karolinska Institutet, KTH Royal Institute of
Technologie, Luleå University of Technology, and SLU—Swedish University of Agricultural
Sciences—are specialized universities with a high proportion of research on specific topics
such as technology, medicine, or agriculture.

A third group, consisting of Karlstad University, Linnaeus University, Mid Sweden
Universität, and Örebro University, are previous university colleges that were upgraded to full
university status about 15 years ago and hence designated new universities.

Blekinge Institute of Technology, Dalarna University, the University of Borås, the University
of Gävle, Halmstad University, Jönköping University, University of Skövde, Kristianstad
Universität, University West, Malmö University, Mälardalen University, and Södertörn
University were all university colleges during the study period. Malmö University was up-
graded to full university status in 2018. In Sweden, university colleges were initially created
with the aim of providing both academic and professional training at the tertiary level, viel
like the former binary divide in the UK. This form roughly correlates with polytechnics in that
they do not have a general right to issue degrees at the postgraduate level, but must apply for
each right, and that there is a lower share of government funding issued for research.

2.5. Statistical Methods

Full counts were generally used for the presentation of data, since reporting of the number of
local authors and total numbers did not seem to be fully consistent. The difference was quite
konsistent, obwohl, with fractional numbers being 43–44% higher for all HEI types except for
comprehensive universities, where the difference was 35%. The latter was used, obwohl, für
comparing with WoS identified publications by the Swedish Research Council (supplementary
Material, Tables III and VI). In order to calculate fractionalized counts for each HEI, we divided
the count given in the fields for affiliations with that for authors (_numLocalCreator/
_creatorCount); das ist, we calculated ”affiliation_shares,” which is roughly comparable to
fractional publications for an HEI.

Proportions were calculated using standard descriptive statistics and the linear correlation
between the share of RQP and time was calculated using Pearson correlation coefficients.
Coword analysis and visualization were performed using VOSviewer v 1.6.8 (van Eck &
Waltman, 2009).

3. ERGEBNISSE

3.1. Suspected Questionable Publications (SQP) in Relation to Total Publication 2012–2017

In our analysis of the 5-year period 2012–2017, we identified 1,743 journal publications that
matched with the blacklists. The size of the institution was correlated with the total number of
publications and SQP (supplementary material, Table I). In order to put the numbers into con-
Text, we calculated the proportion of fractionalized SQP in relation to the total number of jour-
nal publications as found in SwePub. Figur 1 shows the proportion of SQP in relation to total
publishing during the years 2012–2017. Spearman’s correlation was run to determine the re-
lationship between SQP and time over six years. There was a strong, negative monotonic cor-
relation between SQP share and year (ρ = −0.83, n = 6, P < 0.05). The overall proportion over the period was 0.73% of all journal publications entered in SwePub. The proportion of SQP journal publishing was not evenly distributed between the Swedish universities. A difference can be discerned between comprehensive universities and special- ized universities (median share = 0.55, 0.43%) and new universities and university colleges Quantitative Science Studies 511 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context Figure 1. Share of SQP publishing (full counts) in Sweden 2012–2017. (both with a median share = 1.96, 2.13%) (Table 1; supplementary material, Tables II and IV). However, there were considerable differences between these four categories. For exam- ple, within the group of comprehensive universities, the rate of SQP was almost three times higher at the University of Gothenburg (0.97%) compared to Uppsala University (0.35%) (supplementary material, Table II). Most of the university colleges had a total rate above 1% SQP during the full period, and as many as four HEIs had an SQP ratio of 3% or higher for the period (Figure 2). An unflattering national record was set in 2013 by Kristianstad University when 7.5% of their total reported journal publishing was in blacklisted journals (supplementary material, Table II). As was mentioned above, the trend over time is declining. When different HEI types are distinguished, there is a clear difference in the share of SQP between new universities and uni- versity colleges, on the one hand, and comprehensive universities and specialized universities, on the other (Figure 3). Still, the declining trend is similar within each HEI type, at roughly half the share in the last year of the analysis as opposed to the year with the highest share. 3.2. Overlap with External Databases 3.2.1. The Norwegian NSD database An analysis based on whitelists was performed after the main analysis was done. Table 1. Proportion of in SQP publishing (mean values and range) among different organization categories (full counts) l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Total n 131,947 Total SQP n Median 0.55% 709 Mean 0.58% Range (min, yearly) 0.10% Range (max, yearly) 1.23% Range (min) 0.33% Range (max) 0.97% Comprehensive universities Specialized universities New universities University colleges 75,042 13,644 17,548 370 291 373 0.43% 0.85% 1.96% 2.25% 2.13% 2.44% 0.00% 0.45% 0.00% 0.73% 4.17% 4.64% 7.50% 0.73% 0.14% 2.20% 1.65% 3.41% 0.58% 4.97% 0.47% 0.92% Full set 238,181 1,743 0.77% 0.73% Quantitative Science Studies 512 Identifying questionable publishing in a national context Figure 2. Organization share of SQP journal publications (full counts) as a share of the total number of registered journal publications in SwePub 2012–2017. The cleaned ISSNs from the 1,743 papers were matched against the Norwegian NSD database. Table 2 shows the number of matched entries in any of the three blacklists used in the analysis. Seventy-three articles in 28 journals were matched as being at level 1 for NSD 2019. Of these, two were raised from level 0 in the 2019 edition of NSD, and one was newly introduced in 2019. Additionally, two journals were demoted to level 0 for the coming year (2020). Four of the journals were identified as SQP in two blacklists, while 24 were identified in only one. Six journals were indexed in the regular DOAJ index. Of these, two were each listed in the IRAN MHME list and the Cabell’s list, while three were also listed as “suspected editorial mis- conduct by publisher.” The reason for the inclusion in both the regular DOAJ and the list of removed journals is due to the publishing year of the matched entries. Three of these journals were entered into DOAJ in 2018 and 2019 after being suspended in 2017, which means that Figure 3. Share of SQP in different HEI types over time. Quantitative Science Studies 513 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context Table 2. Number of matched entries in the set found in the Norwegian NSD database, by blacklist Blacklist Cabell’s Iran MHME DOAJ Matched 11 16 5 Not Matched 18 12 24 they have been rehabilitated in the eyes of DOAJ, while it should be noted that one of these was also included in Cabell’s, albeit for a reason for blacklisting that could be debatable, namely that “authors are published several times in the same journal and/or issue.” There are clear discrepancies between the lists. One journal found at level 1 in NSD was also listed in the DOAJ general index since 2017. At the same time, it was also listed in Cabell’s blacklist with no fewer than seven criteria of malpractice, including “no physical or fake addresses to editorial offices or publisher,” “similar title as a legitimate journal,” and “prominent promises of unusual quick peer review.” Here, the decision about which criteria should prevail is difficult, but since there was no error in the matching process (i.e., no false positives due to entry error in SwePub), it was retained in the analysis. Ultimately, no entries were omitted after this analysis, but it shed light on some of the issues in combining blacklists and whitelists, that the same entry could actually appear in both. The full matching exercise may be found in the supplementary material, Table VII. 3.2.2. Clarivate WoS Of the 1,743 publications matched in the set, 1,313 had either a WoS ID (n = 60, 55 matched), DOI (n = 1,275, 65 matched), or PubMed ID (n = 112, 31 matched). A total of 91 WoS indexed entries were matched in the set. These represent 38 different journals. Five were included in two of the blacklists, while 33 were only matched in one. Table 3 shows the number of matched entries in any of the three blacklists used in the analysis. Additionally, each journal was manually searched for in the so-called Norwegian list (NSD) (https://dbh.nsd.uib.no/pub- liseringskanaler). If the title was either not listed at all, or marked with a zero, indicating that it was not judged as a valid peer-reviewed journal by the Norwegian authorities, it is marked as “matched” in the table. The number of matched journals in the blacklists were quite evenly distributed between the blacklists, with DOAJ and the NSD standing out with larger numbers. The calculation of Cohen’s Kappa coefficient for interrater reliability between the blacklists (using NSD as a fourth list) also showed the largest overlap between NSD and DOAJ for the set of journals that were identified in WoS (Table 4). Table 5 shows that 22 out of the 38 journals that were identified were found in the Emerging Sources Citation Index (ESCI), while 16 were found in one or two of the original Table 3. Number of matched entries in the set found in WoS, by blacklist Blacklist Cabell’s Iran MHME DOAJ NSD Matched 12 12 19 25 Not matched 26 26 19 13 514 Quantitative Science Studies l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context Table 4. Cohen’s Kappa for interreliability between blacklists for journals found in WoS NSD Cabell’s −0.179 DOAJ 0.474 −0.421 Iran −0.273 −0.096 −0.632 NSD Cabell’s DOAJ Iran WoS databases (SCI-e, SSCI). No matches were found in A&HCI. The ESCI database lists jour- nals that have not (yet) been selected for inclusion in Clarivate’s so-called Flagship Citation indexes, and are not included in the JCR, or assigned a JIF (Testa, n.d.). The full set of data is available in the supplementary material, Tables VIII and IX, where additional information, such as if the journal was demoted in 2019 in the NSD list and the blacklist it was matched in, as well as the corresponding WoS database the journal was identified in. It should also be noted that four of these journals were additionally identified in the list of journals removed from the Elsevier Scopus list between the years 2013 and 2016, based on “Publication con- cerns” (https://www.elsevier.com/solutions/scopus/how-scopus-works/content). l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 3.3. SQP in Relation to WoS Indexed Publishing 2012–2017 We compared SQP in relation to the number of fractionalized publications as calculated an- nually by the Swedish Research Council based on WoS data. Fractionalization was calculated by dividing the number of authors with the total number of authors (_numLocalCreator/ _creatorCount) in for the entry in the SwePub database. The percentages refer to the compar- ison between the number of SQP (fractionalized as calculated within SwePub data) with the fractionalized volume indicator calculated by the Swedish Research Council for each year. The comparison then reads: If the proportion of SQP is 1% then there is 1 SQP publication for every 100 WoS indexed output (fractionalized) from the organization. What is noteworthy is that the percentage of SQP/ WoS-indexed from comprehensive universities varies from 1% to up to 3% (Table 6; supplementary material, Tables III and V), and specialized universities have lower median (0.8%), but higher extremes (11.3%), while the percentage of SQP/ WoS- indexed from new universities and university colleges is about 6–7% with quite a few individ- ual years noted above 10% for SQPs. Publishing in WoS-indexed journals was quite rare by researchers at university colleges, which could explain the extreme interyear variability. 3.4. SQP in Different Research Topics A total of 638 different journal titles were identified in the 1,743 matched articles. From these, it was possible to create a set of 700 noun phrases. After removing the two most often Table 5. Source WoS database for journals matched in the dataset WoS database ESCI SCI-e SCI-e + SSCI SSCI Total Quantitative Science Studies n 22 13 1 2 38 515 Identifying questionable publishing in a national context Table 6. HEI category proportion of SQP journal publications (fractionalized) as a share of the number of WoS indexed journal articles calculated by the Swedish Research Council Comprehensive universities Specialized universities New universities University colleges Full set Median 1.04% 0.77% 5.96% 6.83% 1.49% Range (min, yearly) 0.24% Range (max, yearly) 2.96% 0.00% 1.38% 0.00% 0.96% 11.26% 15.56% 58.27% 1.91% occurring terms (“journal” and “international journal”), the largest segment the software was able to identify was a set of 219 terms that were connected to each other in a coword analysis. The result is shown in Figure 4. The most frequent topics were Nursing, Education, and Business. This does not reflect the relative research output of these areas in the Swedish con- text, which is dominated heavily by medicine and technology both in terms of number of pub- lications and financial resources (Hansson et al., 2019). 3.5. Blacklist Coverage and Overlap In Table 7 and Figure 5, a Venn diagram shows the overlap of matched entries in the three blacklists that were used in this project. In terms of matched entries of SQP after combining automatic publisher/name/ISSN-matching and manual additions there was a significant differ- ence. The actual overlap of matched entries (with Cohen’s Kappa coefficient in parenthesis) was Cabell’s–Iran MHME: 146 (−0.58), Cabell’s–DOAJ: 292 (0.05), Iran MHME–DOAJ: 23 (−0.50), Cabell’s–Iran MHME-DOAJ: 9. Cabell’s was by far the largest database and it also had the highest capture rate (904 of 1,743). However, only 3% of its journals had published l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Figure 4. Coword analysis of questionable publishing in Sweden based on the journal titles of the 1,743 matched articles. Quantitative Science Studies 516 Identifying questionable publishing in a national context Table 7. Number of matched publications of the 1,743 total entries Step 1 2 3 C+I+D ISSN Publisher + Title Manual Total Cabell’s 800 Iran MHME 669 33 71 904 4 104 777 DOAJ 472 4 42 518 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . Figure 5. Overlap between the matched documents in the three databases (full count). Cabell’s = Cabell’s blacklist, Iran MHME = Iran MHME, DOAJ = DOAJ list of removed journals based on “Suspected editorial misconduct.” work submitted by Swedish researchers. The Iran MHME and DOAJ lists were both smaller (777/1,743 and 518/1,743 respectively), but a larger proportion of their journal titles were matched (13% and 23% respectively). / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 4. DISCUSSION 4.1. Suspected Questionable Publishing Within Swedish Academia This study found that less than one in a hundred scientific publications from Swedish higher education institutions were in questionable journals. However, there were marked differences between broad-based established and technical universities on the one hand, and newer higher education institutions with a lower total research output on the other. University colleges in particular, but also new universities, had a substantial proportion of their output published in questionable journals. This can be explained by several factors, such as lack of resources allocated to newer HEIs, which makes it difficult to recruit and retain talent. Since research is also a smaller part of these HEIs’ mission (the primary one being education at bachelor level), it is possible that both ad- herence to scholarly norms and practice, as well as structures for formal and informal quality control are less well-developed at these institutions due to a lack of critical mass of active researchers in each subject area. Another reason for the differences might be differences in research profiles in different cat- egories of HEIs. Our results indicated that publishing in questionable journals is more likely to be found in applied research areas, as well as in newly professionalized and academicized areas. Both of these are predominantly found in new universities and university colleges, Quantitative Science Studies 517 Identifying questionable publishing in a national context where the share of teaching-intensive areas is more prominent. Nursing, Education, and Business stand out, probably reflecting the “demand” among researchers in these research areas. These areas have undergone an “academization” during recent decades, and it is pos- sible that their publishing cultures are less mature. Natural and medical science, which usually require more resources, are concentrated in larger HEIs. There is a decline in the matched share of SQP, at an overall level, from just below 1%, at the highest level, to almost half that share in 2017. The same level of decline was seen in all HEI types, although their respective rates were quite different. We can only speculate about the causes, but it might be due to heightened awareness and a general maturity in the publishing system. On the other hand, blacklists are created in retrospect, and some SQP might not have been identified and registered in blacklists yet. There is also a possi- bility that questionable publishers have adapted to increased awareness and stated criteria, making it harder to distinguish between clear offenders and just “poor publishing practices.” 4.2. Methodological Considerations for Matching Against Blacklists Using the proposed methodological framework for evaluating total publication output at a na- tional level provides for the opportunity to compare the proportion of SQP between countries and across regions. Furthermore, using the national publication database based on re- searchers’ reported publishing means that a larger set of publications are reported than if a third-party publication list, such as CrossRef, WoS, or Scopus, were used, since the curation would be done in a more standardized way instead of based on the coverage of the specific service chosen. At the same time, different practices of maintaining the local databases could have systematic effects on possibly reported questionable publishing. The general trends with- in each category suggest that differences do not seem to be attributable to different practices within the institutions that maintain the local databases. This study also showed that none of the blacklists used in this study were comprehensive and that very few publications could be found in all three lists. At the same time, a very high agreement between the lists would still not ensure that the study covers all aspects of SQP. An evaluation of the coverage of the list is inherently problematic if no ground truth is available. It is not the place here to delve into the Science & Technology Studies STS issue of inherent uncertainty of facts and that standards are developed in context (Bowker & Star, 1999). But since it cannot be expected that SQP is distributed similarly to the full publishing activities at the national level, these evaluations are tentative. Apart from the statistical analyses of cov- erage and overlap that have been done here, it can be argued that, at least up to a point, the inclusion of more sources of blacklists would mean that a wider selection of views could be heard, akin to the argument within qualitative studies that a good selection process becomes “saturated.” By combining a commercial provider of blacklists with a list based on evaluation by an independent nonprofit organization, as well as with a blacklist originating from a gov- ernment outside of the USA and Europe, there is at least the opportunity for a diverse and broad coverage of SQP journals. Additionally, by subjecting the results to an evaluation using the Norwegian NSD and Clarivate WoS as further resources for possible whitelisting of pub- lications, we argue that we have provided as good quality control as is possible at this point. This helped us to remove seven matches that were made due to entry error in SwePub, but it was less clear-cut that using whitelisting using these sources would help prune out errors. In discussing the results of the whitelisting exercises made using the Norwegian NSD list and Clarivate WoS we would like to add the following: While 4% and 5% of the matched results were found in each of the respective databases, upon detailed analysis, none of the Quantitative Science Studies 518 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context ”whitelisted” entries were actually found to be without suspicion. This has to do with entries being whitelisted at a specific point in time or that there was other evidence that rendered the whitelisting itself questionable. All but two of the journals that had matched articles that were found in the Norwegian NSD list were, at the same time, not included in the general DOAJ directory (“not whitelisted by DOAJ”) during the coverage of the study, together with being matched to at least one of the blacklists. The two journals listed despite being on the DOAJ directory are noted as being found questionable in additional sources and were retained for that reason. Additionally, one of these was recently dropped by its publisher. While WoS is generally used as a whitelist for credible publication outlets, when journals were compared to the coverage of the Norwegian list as well as DOAJ (whitelist), clear dis- crepancies were found. The fact that most of the WoS included journals that where blacklisted were found in the ESCI, which uses fewer rigid criteria for inclusion prompts for caution when using the full set of WoS databases to whitelist journals. All but five journals were either matched with more than one blacklist or not included in NSD during the coverage of the study. Among the five journals only matched in one blacklist, four were additionally not in- cluded in DOAJ coverage, while the fifth was the same journal that was dropped from its pub- lisher as described above. Since all journals found in WoS were also matched in at least one blacklist, as well as hav- ing additional criteria found during cross-examination of the results (such as not included in DOAJ, or listed as not scholarly “0” in NSD), none of these journals were omitted from the final results. With this in mind, we do not want to underplay the fact that using blacklists is error- prone and based on the judgments of evaluators following protocols for inclusion. Therefore, this highlights the problem with blacklisting as a method to identify questionable journals. 4.3. Results in Relation to Other Studies In a literature review consisting of all the 178 articles identified in WoS on the topics of fake, questionable, or predatory publishers or journals, we found that a large proportion of the stud- ies focused on the issue of questionable publishing from a normative perspective, trying to detect, identify, and distinguish it from proper scientific publishing, sometimes by distinguish- ing high-quality from low-quality publishing. One study, which aimed to quantify the extent of predatory publishing using Beall’s list, found that predatory journals had rapidly increased their publication volumes, reaching an estimated 420,000 articles in 2014, published by around 8,000 active journals (Shen & Björk, 2015). It has also been reported that researchers belonging to less developed areas are found to cite research in questionable journals at a higher level (Frandsen, 2017). This seems to be in line with the results obtained here, which seem to show that less developed research areas that have recently undergone professionalization and academization are more prone to be pub- lished in SQP journals. A number of studies focused on the issue at the national level in countries such as Brazil (Perlin, Imasato, & Borenstein, 2018), the Czech Republic (Mercier, Tardif, et al., 2018; Strielkowski, Gryshova, & Shcherbata, 2017; Vershinina, Tarasova, & Strielkowski, 2017), Iran (Erfanmanesh & Pourhossein, 2017), India (Mukherjee, 2018; Samal & Dehury, 2017; Quantitative Science Studies 519 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context Seethapathy, Kumar, & Hareesha, 2016), Turkey (Demir, 2018a, b; Onder & Erdil, 2017), and Kazakhstan (Yessirkepov, Nurmashev, & Anartayeva, 2015). However, these studies tackle questionable publishing at policy levels, which means that they focus on single countries or a single research area, or sometimes both. Additionally, al- most all studies rely exclusively on a single set of blacklists—Beall’s—as the sole source of identification that is both dated and lacks the means to match with accuracy (using ISSN). However, one recent study used fuzzy text matching (Strinzel et al., 2019). Although this is quite an elaborate technique, it might actually defeat its own purpose, as many questionable publishers choose titles that are so akin to already existing journals that false positives could occur from such matching. Finally, it is worth discussing if a share of SQP at 0.77% is low or high. Unfortunately, there are few studies that evaluate SQP at the national level. As noted above, the ECOOM group, which maintains the Flanders database in Belgium, reports lower numbers (Eykens, Guns, et al., 2019). In their recent PLOS study, covering five yearly reports on the issue, they identified 556 publications in potentially predatory blacklisted journals during the period 2003–2016 in the regional Flanders database. After evaluation by a panel of independent scholars that oversee the selection of publication outlets in the Flemish VABB-SHW data- base, 210 out of 73,694 published articles were identified as predatory OA publications. This would indicate a share of 0.28%, which would be significantly lower than our results. Furthermore, it was found that the numbers dropped significantly (by a factor of 10) after the year 2014, which coincides in time with the change in methodology when the independent panels started to evaluate the journal lists. We have also conducted tentative analyses of publishing data obtained from the Danish Bibliometric Research Indicator (BFI), Finland’s VIRTA Publication Information Service ( VIRTA), and the Norwegian Current Research Information System in Norway (CRIStin). Here, the share of SQP seems to be comparable between the SwePub data set used here and VIRTA and CRIStin, while it is significantly lower in BFI at the same level as in the VABB-SHW. Upon manual inspection, the BFI has purged all publications that are below level 1 in the Danish database (level 0 and pub- lications in publication channels that have not been evaluated), meaning that a large share of possible SQP has already been removed. This is in line with the VABB-SHV and seems to verify that the introduction of a screening process for “accepted publishing channels” is a viable way to significantly filter out questionable publishing from the evaluation in a performance-based funding system. In Norway and Finland, the same screening is in effect, since only publications at level 1 and above are used in the evaluation. It should also be noted that the Swedish Research Council is producing such a “Swedish list,” while there is also ongoing work on a collaborative “Nordic list.” Still, it does not alter the practices of researchers publishing in these journals, potentially padding publication lists with question- able publishing for use in individual evaluation for grant applications or applications for positions, although it might have a deterrent effect on bad publishing practices. 4.4. Limitations The handling of a number of blacklists and whitelists, as well as the original handling of pub- lication lists, is error prone. Specifically, the handling of deduplication, fractionalization, and incorrectly added information in the registration process, as well as systematic errors, such as one university failing to report one year and another university’s publications having been added twice in one year, meant that every step in the process had to be handled with care, and often reiterated after small adaptations to the matching process. It was also found, late in the analysis, when comparing the resulting set of 1,750 SQP with the WoS database, that Quantitative Science Studies 520 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context seven entries seemed to have been mislabeled upon registration. This could generally be at- tributed to a false ISSN being attached to a title or the DOI not matching the stated journal title and ISSN. Upon manual inspection, these entries were removed from the resulting set of 1,743 SQP entries in our analysis. All the blacklists used in this study identified current publications, which means that changes in questionable status might have occurred (e.g., it has been alleged that some ques- tionable publishers have taken over respectable journals and turned them into problematic outlets). However, we believe that the number of hijacked journals and journals transforming from legitimate to questionable over such a short time period could only marginally impact the results of this study. Studies have shown that authors who select journals based on spam emails have been un- able to halt the publication of manuscripts once they have been submitted (Oermann, Conklin, et al., 2016). Therefore, there could be a risk that authors are published in questionable jour- nals against their will. This paper, however, focuses on publications that have not only been published, but that also have been submitted to local publication databases, indicating that the researchers acknowledge these publications as legitimate. Although the results only apply to Swedish research, the methods are transferable to other countries. l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 4.5. Conclusions We have provided a methodology for identifying SQP by matching a national publication list to a combination of blacklists from different providers, omitting the outdated Beall’s list, and at the same time not just substituting it by one single list. We noted that no blacklist has near- complete coverage of questionable publication outlets (we manually cross-checked all black- list matches), which makes our study less reliant on a single blacklist provider, freely available, as well as subscription-based. Less than one in a hundred research papers reported by Swedish HEIs are published in questionable journals, and the problem does not seem to have increased in recent years. However, in relative terms, such publications are several times more common at newer HEIs, highlighting their vulnerability. We do not propose that our method should be used to publicly identify individual researchers, as our results indicate that blacklists are still in their infancy, to a certain degree are error prone, and probably only identify a certain fraction of all questionable publications, depending on the definition chosen. At the individual level, if made public, an SQP-matched publication list could have a stigmatizing effect and we find that the possibility of raising awareness at the individual level is not justified given that the risk of false positives is still quite high. When scaled to an aggregate level, errors, if found, could be ex- pected to level out. However, it is an easy-to-implement surveillance system that could be used by different stakeholders to identify problematic research environments or whole HEIs in need of support for tackling the issue of quality and ethics in publication. It should also be noted that the use of text-based analysis for identifying the subject structure of matched publications could help to make comparisons between countries without having to align na- tional subject classifications. Measures should be taken to raise awareness and enforce com- pliance with good publication practices. Information activities directed toward researchers, as well as general “mild" introduction of journal lists, such as the upcoming Nordic list and DOAJ for whitelisting, and preferably free lists, for blacklisting possible SQP journals needs to be developed and be maintained. There might also be a need to raise the issue in a larger context regarding research evaluation. We argue that there is a need to view the role of publishing as part of the full research cycle and not only as a means of presenting results at the end, and that Quantitative Science Studies 521 Identifying questionable publishing in a national context there is a need to critically question the role of publications in the merit system in light of the San Francisco Declaration on Research Assessment (DORA, https://sfdora.org/read/). Future elaborations would be to develop a comprehensive methodology using a similar methodological framework in collaboration with other research groups for comparing the in- cidence of questionable publishing in other European countries where a full set of HEI reported publications are available. Upon initial tests, comparable relationships between different types of HEI institution seems to be found in other countries (e.g., Danish, Norwegian, and Finnish data) that we have identified, but for obvious reasons, national data from Estonia and the Czech Republic that we have collected would need expert collaboration for trans- lation at the national level to be interpreted. There is a need to implement the methodology with some care, since every national publication database has its specifics. For instance, as noted in section 4.3, the BFI only includes entries that are found in publication channels at level 1 and above. A comparison might, therefore, show very different results if done without local knowledge of their construction. Furthermore, HEIs only stand for a certain part of the total academic publishing at the national level. A useful extension would be to include publishing from institutes, the private sector, and independent researchers. That would entail using other sources, such as Google Scholar and Microsoft Academic, or browsing known SQP publishers’ web sites (Sīle, 2019). ACKNOWLEDGMENTS The authors would like to thank two anonymous reviewers for constructive and useful feed- back on the submitted manuscript. We would also like to thank Peter Allebeck and Jonas Nordin for connecting the two authors with each other. We thank Henrik Aldberg, previously at the Swedish Research Council for supplying information publications in WoS journals, Camilla Lindelöw and Tuija Drake at the National Library of Sweden for guiding us in the peculiarities of SwePub, and Johan Eklund, SSLS, University of Borås, for help with implement- ing Cohen’s Kappa. Lastly, Cabell’s provided us with a copy of their blacklist, which made it possible for us to perform a large part of the matching. AUTHOR CONTRIBUTIONS Gustaf Nelhans: Conceptualization, Data curation, Methodology, Formal analysis, Investigation, Visualization Validation, Writing—original draft, Writing—review & editing. Theo Bodin: Conceptualization, Methodology, Formal analysis, Validation, Writing—original draft, Writing—review & editing. COMPETING INTERESTS The authors report no competing interests. Cabell’s Inc. had the right to read the manuscript before submission, but had no role in the design of the study, analysis or interpretation of data, or the writing of the manuscript. FUNDING INFORMATION No funding has been received for this research. DATA AVAILABILITY Supplementary data, as well as the raw matched data so that tables and figures in this article can be reproduced, are available at https://doi.org/10.5878/6dn9-yt13. Quantitative Science Studies 522 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context REFERENCES Asadi, A., Rahbar, N., Asadi, M., Asadi, F., & Paji, K. K. journals (2017). Online-based approaches to identify real and publishers from hijacked ones. Science and Engineering Ethics, 23(1), 305–308. https://doi.org/10.1007/s11948-015- 9747-9 Bohannon, J. (2013). Who’s afraid of peer review? Science, 342(6154), 60. https://doi.org/10.1126/science.342.6154.60 Bowker, G. C., & Star, S. L. (1999). Sorting things out: Classification and its consequences. Cambridge, MA: MIT Press. da Silva, J. A. T. (2017). Preprints: ethical hazard or academic lib- eration? Kome-an International Journal of Pure Communication Inquiry, 5(2), 73–80. https://doi.org/10.17646/kome.2017.26 Dadkhah, M., & Maliszewski, T. (2015). Hijacked journals—threats and challenges to countries’ scientific ranking. International Journal of Technology Enhanced Learning, 7(3), 281–288. https://doi.org/10.1504/ijtel.2015.072819 Das, S., & Chatterjee, S. S. (2018). Cabell’s blacklist: A new way to tackle predatory journals. Indian Journal of Psychological Medicine, 40(2), 197–198. https://doi.org/10.4103/ IJPSYM. IJPSYM_290_17 Demir, S. B. (2018a). Predatory journals: Who publishes in them and why? Journal of Informetrics, 12(4), 1296–1311. https://doi. org/10.1016/j.joi.2018.10.008 Demir, S. B. (2018b). Pros and cons of the new financial support policy for Turkish researchers. Scientometrics, 116(3), 2053–2068. https://doi.org/10.1007/s11192-018-2833-4 Editorial. (2018). Nature, 562(308). https://doi.org/10.1038/ d41586-018-07033-5 Erfanmanesh, M., & Pourhossein, R. (2017). Publishing in predatory open access journals: A case of Iran. Publishing Research Quarterly, 33(4), 433–444. https://doi.org/10.1007/s12109-017- 9547-y Eriksson, S., & Helgesson, G. (2017). The false academy: Predatory publishing in science and bioethics. Medicine Health Care and Philosophy, 20(2), 163–170. https://doi.org/10.1007/s11019- 016-9740-3 Eykens, J., Guns, R., Rahman, A. I. M. M., & Engels, T. C. E. (2019). Identifying publications in questionable journals in the context of performance-based research funding. PLOS ONE, 14(11), e0224541. https://doi.org/ https://doi.org/10.1371/journal. pone.0224541 Eykens, J., Guns, R., Rahman, A. I. M. M., Sīle, L., & Engels, T. C. E. (2018). Predatory open access journals: A review of past screenings within the Flemish performance based research funding system (2014–2018). Proceedings of the 23rd International Conference on Science and Technology Indicators, Leiden. Ferris, L. E., & Winker, M. A. (2017). Ethical issues in publishing in predatory journals. Biochemia Medica, 27(2), 279–284. https:// doi.org/10.11613/bm.2017.030 Frandsen, T. F. (2017). Are predatory journals undermining the credibility of science? A bibliometric analysis of citers. Scientometrics, 113(3), 1513–1528. https://doi.org/10.1007/ s11192-017-2520-x Gasparyan, A. Y., Yessirkepov, M., Voronov, A. A., Gorin, S. V., Koroleva, A. M., & Kitas, G. D. (2016). Statement on publica- tion ethics for editors and publishers. Journal of Korean Medical S c i e n c e , 3 1 ( 9 ) , 1 3 5 1 – 1 3 5 4 . h t t p s : / / d o i . o r g / 1 0 . 3 3 4 6 / jkms.2016.31.9.1351 Haider, J., & Åström, F. (2017). Dimensions of trust in scholarly communication: Problematizing peer review in the aftermath of John Bohannon’s “Sting” in science. Journal of the Association for Information Science and Technology, 68(2), 450–467. https://doi.org/10.1002/asi.23669 Hansson, G., Barriere, S. G., Gurell, J., Lindholm, M., Lundin, P., & Wikgren, M. (2019). The Swedish research barometer 2019: The Swedish research system in international comparison ( VR1915). Stockholm: Swedish Research Council. Hoffecker, L. (2018). Cabells Scholarly Analytics. Journal of the Medical Library Association, 106(2), 270–272. https://doi.org/ 10.5195/jmla.2018.403 Jalalian, M., & Dadkhah, M. (2015). The full story of 90 hijacked journals from August 2011 to June 2015. Geographica P a n n o n i c a , 1 9 ( 2 ) , 7 3 – 8 7 . h t t p s : / / d o i . o r g / 1 0 . 5 9 3 7 / GeoPan1502073J Josephson, A., & Michler, J. D. (2018). Viewpoint: Beasts of the field? Ethics in agricultural and applied economics. Food Policy, 79, 1–11. https://doi.org/10.1016/j.foodpol.2018.08.001 Manca, A., Moher, D., Cugusi, L., Dvir, Z., & Deriu, F. (2018). How predatory journals leak into PubMed. Canadian Medical Association Journal, 190(35), E1042–E1045. https://doi.org/ 10.1503/cmaj.180154 McLeod, A., Savage, A., & Simkin, M. G. (2018). The ethics of predatory journals. Journal of Business Ethics, 153(1), 121–131. https://doi.org/10.1007/s10551-016-3419-9 (2018). Mercier, E., Tardif, P. A., Moore, L., Le Sage, N., & Cameron, P. A. Invitations received from potential predatory publishers and fraudulent conferences: A 12-month early- career researcher experience. Postgraduate Medical Journal, 94(1108), 104–108. https://doi.org/10.1136/postgradmedj- 2017-135097 Mukherjee, B. (2018). Measuring scientific value of Indian journals beyond impact factor: a case with physics-chemistry-biology journals. Current Science, 114(8), 1613–1623. https://doi.org/ 10.18520/cs/v114/i08/1613-1623 Oermann, M. H., Conklin, J. L., Nicoll, L. H., Chinn, P. L., Ashton, K. S., Edie, A. H., … Budinger, S. C. (2016). Study of predatory open access nursing journals. Journal of Nursing Scholarship, 48(6), 624–632. Onder, C., & Erdil, S. E. (2017). Opportunities and opportunism: Publication outlet selection under pressure to increase research productivity. Research Evaluation, 26(2), 66–77. https://doi.org/ 10.1093/reseval/rvx006 Perlin, M. S., Imasato, T., & Borenstein, D. (2018). Is predatory pub- lishing a real threat? Evidence from a large database study. Scientometrics, 116(1), 255–273. Samal, J., & Dehury, R. K. (2017). The need and importance of in- corporating academic research results into the curricula of Ayurveda in India. Journal of Clinical and Diagnostic Research, 11(6), KA01–KA03. https://doi.org/10.7860/jcdr/2017/ 26742.10112 Seethapathy, G. S., Kumar, J. U. S., & Hareesha, A. S. (2016). India’s scientific publication in predatory journals: need for reg- ulating quality of Indian science and education. Current Science, 111(11), 1759–1764. https://doi.org/10.18520/cs/v111/i11/ 1759-1764 Shaghaei, N., Wien, C., Holck, J., Thiesen, A. L., Ellegaard, O., Vlachos, E., & Drachen, T. (2018). Being a deliberate prey of a predator: Researchers’ thoughts after having published in preda- tory journal. Liber Quarterly, 28(1). Shahri, M. A., Jazi, M. D., Borchardt, G., & Dadkhah, M. (2018). Detecting hijacked journals by using classification algorithms. Quantitative Science Studies 523 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d / . f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Identifying questionable publishing in a national context Science and Engineering Ethics, 24(2), 655–668. https://doi.org/ 10.1007/s11948-017-9914-2 Shen, C., & Björk, B.-C. (2015). “Predatory” open access: a longi- tudinal study of article volumes and market characteristics. BMC Medicine, 13(1), 230. Sīle, L. (2019). Entanglement of bibliographic database content and data collection practices: Rethinking data integration using findings from a European study. Procedia Computer Science, 146, 201–207. https://doi.org/10.1016/j.procs.2019.01.094 Sīle, L., Pölönen, J., Sivertsen, G., Guns, R., Engels, T. C. E., Arefiev, P., … Teitelbaum, R. (2018). Comprehensiveness of national bib- liographic databases for social sciences and humanities: Findings from a European survey. Research Evaluation, 27(4), 310–322. https://doi.org/10.1093/reseval/rvy016 Strielkowski, W., Gryshova, I., & Shcherbata, M. (2017). Predatory publishing and Beall’s list: lessons for the countries adapting novel research evaluation criteria. Science and Education, 8, 39–43. https://doi.org/10.24195/2414-4665-2017-8-5 Strinzel, M., Severin, A., Milzow, K., & Egger, M. (2019). “Blacklists” and “whitelists” to tackle predatory publishing: A cross-sectional comparison and thematic analysis. PeerJ Preprints. Testa, J. (n.d.). Journal selection process. Retrieved from https://clar- ivate.com/webofsciencegroup/essays/journal-selection-process/ van Eck, N., & Waltman, L. (2009). Software survey: VOSviewer, a computer program for bibliometric mapping. Scientometrics, 84(2), 523–538. Vershinina, S., Tarasova, O., & Strielkowski, W. (2017). Academic publishing, journal rankings, and scientific productivity. Terra Economicus, 15(4), 127–135. https://doi.org/10.23683/2073- 6606-2017-15-4-127-135 Yessirkepov, M., Nurmashev, B., & Anartayeva, M. (2015). A Scopus-based analysis of publication activity in Kazakhstan from 2010 to 2015: Positive trends, concerns, and possible solutions. Journal of Korean Medical Science, 30(12), 1915–1919. https:// doi.org/10.3346/jkms.2015.30.12.1915 l D o w n o a d e d f r o m h t t p : / / d i r e c t . m i t . / e d u q s s / a r t i c e - p d l f / / / / 1 2 5 0 5 1 8 8 5 7 8 9 q s s _ a _ 0 0 0 3 3 p d . / f b y g u e s t t o n 0 7 S e p e m b e r 2 0 2 3 Quantitative Science Studies 524RESEARCH ARTICLE image
RESEARCH ARTICLE image
RESEARCH ARTICLE image
RESEARCH ARTICLE image

PDF Herunterladen