- Research article
- Open Access
- Open Peer Review
Potential predatory and legitimate biomedical journals: can you tell the difference? A cross-sectional comparison
BMC Medicine volume 15, Article number: 28 (2017)
The Internet has transformed scholarly publishing, most notably, by the introduction of open access publishing. Recently, there has been a rise of online journals characterized as ‘predatory’, which actively solicit manuscripts and charge publications fees without providing robust peer review and editorial services. We carried out a cross-sectional comparison of characteristics of potential predatory, legitimate open access, and legitimate subscription-based biomedical journals.
On July 10, 2014, scholarly journals from each of the following groups were identified – potential predatory journals (source: Beall’s List), presumed legitimate, fully open access journals (source: PubMed Central), and presumed legitimate subscription-based (including hybrid) journals (source: Abridged Index Medicus). MEDLINE journal inclusion criteria were used to screen and identify biomedical journals from within the potential predatory journals group. One hundred journals from each group were randomly selected. Journal characteristics (e.g., website integrity, look and feel, editors and staff, editorial/peer review process, instructions to authors, publication model, copyright and licensing, journal location, and contact) were collected by one assessor and verified by a second. Summary statistics were calculated.
Ninety-three predatory journals, 99 open access, and 100 subscription-based journals were analyzed; exclusions were due to website unavailability. Many more predatory journals’ homepages contained spelling errors (61/93, 66%) and distorted or potentially unauthorized images (59/93, 63%) compared to open access journals (6/99, 6% and 5/99, 5%, respectively) and subscription-based journals (3/100, 3% and 1/100, 1%, respectively). Thirty-one (33%) predatory journals promoted a bogus impact metric – the Index Copernicus Value – versus three (3%) open access journals and no subscription-based journals. Nearly three quarters (n = 66, 73%) of predatory journals had editors or editorial board members whose affiliation with the journal was unverified versus two (2%) open access journals and one (1%) subscription-based journal in which this was the case. Predatory journals charge a considerably smaller publication fee (median $100 USD, IQR $63–$150) than open access journals ($1865 USD, IQR $800–$2205) and subscription-based hybrid journals ($3000 USD, IQR $2500–$3000).
We identified 13 evidence-based characteristics by which predatory journals may potentially be distinguished from presumed legitimate journals. These may be useful for authors who are assessing journals for possible submission or for others, such as universities evaluating candidates’ publications as part of the hiring process.
The Internet has transformed scholarly publishing. It has allowed for the digitalization of content and subsequent online experimentation by publishers, enabling print journals to host content online, and set the course for online open-access publishing. Nevertheless, an unwelcome consequence of the Internet age of publishing has been the rise of so-called predatory publishing.
In the traditional subscription model of publishing, journals typically require transfer of copyright from authors for articles they publish and their primary revenue stream is through fees charged to readers to access journal content, typically subscription fees or pay-per-article charges. Open access publishing, in contrast, typically allows for authors to retain copyright, and is combined with a license (often from Creative Commons), which enables free and immediate access to published content coupled with rights of reuse . Some open access journals  and many hybrid journals (i.e., those with some open access content and also with non-open access content)  use a business model that relies upon publication charges (often called article publication or processing charges, or APC) to the author or funder of the research to permit immediate and free access.
Predatory publishing is a relatively recent phenomenon that seems to be exploiting some key features of the open access publishing model. It is sustained by collecting APCs that are far less than those found in presumably legitimate open access journals and which are not always apparent to authors prior to article submission. Jeffrey Beall, a librarian at the University of Colorado in Denver, first sounded the alarm about ‘predatory journals’ and coined the term. He initiated and maintains a listing of journals and publishers that he deems to be potentially, possibly, or probably predatory, called Beall’s List  (content unavailable at the time of publishing). Their status is determined by a single person (Jeffrey Beall), against a set of evolving criteria (in its 3rd edition at the time of writing) that Beall has based largely on The Committee On Publication Ethics (COPE) Code of Conduct for Journal Editors and membership criteria of the Open Access Scholarly Publisher’s Association [5–7]. Others have suggested similar criteria for defining predatory journals [8, 9].
The phenomenon of predatory publishing is growing and opinions on its effects are divided. Critics say that it is extremely damaging to the scientific record and must be stopped [10, 11]. Others feel that, while problematic, predatory publishing is a transient state in publishing and will disappear or become obvious over time . A fundamental problem of predatory journals seems to be that they collect an APC from authors without offering concomitant scholarly peer review (although many claim to ) that is typical of legitimate journals . Additionally, they do not appear to provide typical publishing services such as quality control, licensing, indexing, and perpetual content preservation and may not even be fully open access. They tend to solicit manuscripts from authors through repeated email invitations (i.e., spam) boasting open access, rapid peer review, and praising potential authors as experts or opinion leaders . These invitations may seem attractive or an easy solution to inexperienced or early career researchers who need to publish in order to advance their career, or to those desperate to get a publication accepted after a number of rejections, or to those simply not paying attention. Predatory journals may also be a particular problem in emerging markets of scientific research, where researchers face the same pressure to publish, but lack the skills and awareness to discern legitimate journals from predatory ones.
Still, many researchers and potential authors are not aware of the problem of predatory journals and may not be able to detect a predatory journal or distinguish one from a legitimate journal. In order to assist readers, potential authors, and others in discerning legitimate journals from predatory journals, it would be useful to compare characteristics from both predatory and non-predatory journals to see how they differ.
In this study, we undertook a cross-sectional study comparing the characteristics of three types of biomedical journals, namely (1) potential predatory journals, (2) presumed legitimate, fully open access journals, and (3) presumed legitimate subscription-based biomedical journals that may have open access content (e.g., hybrid).
This was a cross-sectional study.
Journal identification and selection
We searched for journals on July 10, 2014. For feasibility, only journals with English-language websites were considered for inclusion and we set out to randomly select 100 journals within each comparison group. The following selection procedures were used to identify journals within each comparison group:
Potential predatory journals (‘Predatory’): We considered all journals named on Beall’s List of single publishers for potential inclusion. We applied the MEDLINE Journal Selection criteria : “[Journals] predominantly devoted to reporting original investigations in the biomedical and health sciences, including research in the basic sciences; clinical trials of therapeutic agents; effectiveness of diagnostic or therapeutic techniques; or studies relating to the behavioural, epidemiological, or educational aspects of medicine.” Three independent assessors (OM, DM, LS) carried out screening in duplicate. From the identified biomedical journals, a computer-generated random sample of 100 journals was selected for inclusion. Journals that were excluded during data extraction were not replaced.
Presumed legitimate fully open-access journals (‘Open Access’): A computer-generated, random sample of 95 journals from those listed on PubMed Central as being full, immediate open access, were included. In addition, five well-established open access journals were purposefully included: PLOS Medicine, PLOS One, PLOS Biology, BMC Medicine, and BMC Biology.
Presumed legitimate subscription-based journals (‘Subscription-based’): A computer-generated, random sample of 100 journals from those listed in the Abridged Index Medicus (AIM) was included. AIM was initiated in 1970 containing a selection of articles from 100 (now 119) English-language journals, as a source of relevant literature for practicing clinicians . AIM was used here since all journals in this group were initiated prior to the digital era and presumed to have a maintained a partially or fully subscription-based publishing model [confirmed by us].
For all journals, their names and URLs were automatically obtained during the journal selection process and collected in Microsoft Excel. Screening and data extraction were carried out in the online study management software, Distiller SR (Evidence Partners, Ottawa, Canada). Journals with non-functioning websites at the time of data extraction or verification were excluded and not replaced.
Data extraction process
Data were extracted by a single assessor (OM) between October 2014 and February 2015. An independent audit (done by LS) of a random 10% of the sample showed discrepancies in 34/56 items (61%) on at least one occasion. As such, we proceeded to verify the entire sample by a second assessor. Verification was carried out in April 2015 by one of eight assessors (RB, JC, JG, DM, JR, LS, BJS, LT) with experience and expertise on various aspects of biomedical publishing process. Any disagreements that arose during the verification process were resolved by third party arbitration (by LS or LT). It was not possible to fully blind assessors to study groups due to involvement in the journal selection process (OM, DM, LS).
Data extraction items
Items for which data were extracted were based on a combination of items from Beall’s criteria (version 2, December 2012) for determining predatory open-access publishers , the COPE Code of Conduct for Journal Publishers (http://publicationethics.org/resources/code-conduct), and the OASPA Membership criteria (http://oaspa.org/membership/membership-criteria/). Data for 56 items were extracted in the following nine categories: aims and scope, journal name and publisher, homepage integrity (look and feel), indexing and impact factor, editors and staff, editorial process and peer review, publication ethics and policies, publication model and copyright, and journal location and contact.
Data were descriptively summarized within each arm. Continuous data were summarized by medians and interquartile range (IQR); dichotomous data were summarized using proportions.
Ninety-three potential predatory journals, 99 open access journals, and 100 subscription-based journals were included in the analysis. The process of journal identification, inclusion, and exclusions within each study group is outlined in Fig. 1; 397 journals were identified as potential predatory journals. After de-duplication and screening for journals publishing biomedical content, 156 journals were identified, from which a random sample of 100 were chosen. Seven journals from the predatory group and one from the legitimate open access group were excluded during data extraction due to non-functional websites. No journal appeared in more than one study group.
There were four unanticipated journal exclusions during data extraction in the presumed legitimate open access and subscription-based groups for which randomly selected replacement journals were used. One journal was listed twice in the open access group and was deemed to be a magazine rather than a scientific journal. Two journals in the subscription-based journal group were deemed to be a magazine and a newsletter, respectively. The decision to exclude and replace these was made post-hoc, by agreement between LS and DM.
Homepage and general characteristics
About half of the predatory journals in our sample indicated interest in publishing non-biomedical topics (e.g., agriculture, geography, astronomy, nuclear physics) alongside biomedical topics in the stated scope of the journal and seemed to publish on a larger number of topics than non-predatory journals (Table 1). Predatory journals included pharmacology and toxicology (n = 59) in the scope of their journal four and a half times more often than open access journals (n = 13) and almost 30 times more than subscription-based journals (n = 2).
When we examined the similarity of the journal name to other existing journals (e.g., one or two words different on the first page of Google search results), we found that over half of predatory journals (n = 51, 55.84%) had names that were similar to an existing journal compared to only 17 open access journals (17.17%) and 22 subscription-based journals (22.00%) (Table 2). In all study groups, the journal name was well reflected in the website URL. For journals that named a country in the journal title, some journals named a different country in the journal contact information (11/21 (52.38%) predatory; 4/13 (30.77%) open access; 1/31 (3.23%) subscription-based) (Table 3). There was a high prevalence of predatory journals from low or low- to middle-income countries (LMICs) (48/64, 75.00%) compared to open access journals (18/92, 19.56%); none of the subscription-based journals listed LMIC addresses.
We assessed the integrity of the homepage by examining the content for errors (Table 4). Spelling and grammatical errors were more prevalent in predatory journals (n = 61, 65.59%) compared to in open access (n = 6, 6.06%) and subscription-based journals (n = 3, 3.00%). In addition, we found a higher frequency of distorted or potentially unauthorized image use (e.g., company logos such as Google, MEDLINE, COPE, Crossref) in predatory journals (n = 59, 63.44%) versus in open access (n = 5, 5.05%) and subscription-based journals (n = 1, 1%). Readers were the main target of language used on subscription-based journal webpages (n = 58, 58%) but less so in open access (n = 14, 14.14%) and predatory (n = 3, 3.23%) journals, where authors (predatory journals) or both authors and readers (open access journals) were the primary target.
Metrics and indexing
Most subscription-based journals indicated having a journal impact factor (assumed 2-year Thomson Reuters JIF unless otherwise indicated) (n = 80, median 4.275 (IQR 2.469–6.239)) compared to less than half of open access journals (n = 38, 1.750 (1.330–2.853)) and fewer predatory journals (n = 21, 2.958 (0.500–3.742)) (Table 5). More than half of predatory journals (n = 54, 58.06%) and subscription-based journals (n = 62, 62%) mentioned another journal-level metric, compared to only 16 (16.16%) open access journals. A metric called the Index Copernicus Value was the most common other metric mentioned in 31 predatory journals (33.33%) and in three open access journals (3.03%), followed by the 5-year impact factor (Thomson Reuters) mentioned in two open access journals (2.02%) and 27 subscription-based journals (27.00%), followed by the Scientific Journal Rankings (i.e., SCImago Journal Rank by Scopus) mentioned in seven predatory, six open access, and eight subscription-based journals. The top databases in which journals indicated being indexed were Google Scholar for predatory journals (n = 47, 50.54%), PubMed for open access journals (n = 85, 85.86%), and MEDLINE for subscription-based journals (n = 39, 39%). About half of predatory journals (n = 48, 51.61%) and 65 (65.65%) open access journals mention DOAJ (indexed in or applied for indexing). International Committee of Medical Journal Editors (ICMJE) was mentioned in some capacity in 16 predatory journals and about three quarters of non-predatory journals.
Editors and editorial process
Nearly a quarter (n = 22, 23.66%) of predatory journals, 17 (17.17%) open access journals, and 9 (9%) subscription-based journals did not name an editor-in-chief (EIC) (Table 6). Of those that did, 40 (56.33%) predatory, 71 (86.59%) open access, and 57 (62.64%) subscription-based journals provided an institutional affiliation for the named EIC. An editorial board listing individual members was provided in 60 (64.52%) predatory journals, 92 (92.93%) open access journals, and 72 (72%) subscription-based journals, each comprising a median of 23 (IQR 14–37), 32.5 (22–50), and 27.5 (16.5–62) board members, respectively. If editors, journal staff, or editorial board members were identified, we completed a subjective assessment of the validity of three arbitrary names and the likelihood of their association with the journal by performing a Google search of their name (in quotations) and searching any online profiles for affiliation with the journal. Details of this assessment can be found in Table 6. For journals with names of editors, staff, or board members available, 100% of names checked in subscription-based journals were found to be legitimate as well as in 95/98 (96.94%) open access journals. Only 24/90 (26.67%) named editors, staff, or board members were assessed as having a legitimate association with the journal among predatory journals. Almost 100% of non-predatory journals appear to use a manuscript submission system, whereas just over half of predatory journals use such a system; almost 70% of predatory journals request that authors send their manuscripts by email and 63% of those journals provide what appears to be a non-professional (e.g., Gmail, Yahoo) email address to do so. Almost all journals (95% predatory journals, 100% open access journals, 92% of subscription-based journals) indicate using peer review during publication consideration (Table 7).
Publication ethics and policies
We examined journals’ promotion and practices around publications ethics (Table 8). About three quarters (n = 77, 77.78%) of open access journals and about a third (n = 33, 33.00%) of subscription-based journals mentioned COPE somewhere on their website whereas only 13 predatory journals (13.98%) did. Few predatory journals had policies about retractions (n = 12, 12.90%), corrections/errata (n = 22, 23.66%), or plagiarism (n = 44, 47.31%) whereas more than half of all non-predatory journals had available policies for all three (retractions: n = 112, 56.28%; corrections/errata: n = 100, 50.25%; plagiarism: n = 199, 59.80%). Sixty-two subscription-based (62%), 56 open access (56.57%), and only 6 predatory (6.45%) journals suggested, recommended or required study registration. No predatory journals mentioned the Enhancing the Quality and Transparency of health Research (EQUATOR) Network, whereas about a quarter (49/195) of presumed legitimate journals did so.
Publication model, fees, and copyright
We assessed whether journals made any indication about accessibility, fees, and copyright (Table 9). Forty-two (42.00%) subscription-based journals indicated being partially open access in some capacity (e.g., hybrid or delayed access), with the remainder not mentioning open access. Almost all (n = 95, 95.00%) subscription-based journals indicated that there was a subscription charge. Eighty-three potential predatory (89.25%) and 94 open access (94.95%) journals claimed to be open access (presumed to be full, immediate open access as no qualification regarding partial or delayed access was stated). For the five (5.05%) open access journals that did not specifically indicate being open access, all had content that was free to access (we did not investigate this further). Subscription-based journals and open access journals seemed to collect revenue from a range of sources (Table 9), while predatory journals appeared to mainly collect revenues from APCs (n = 73, 78.49%) and to a lesser extent, subscription fees (n = 13, 13.98); in 14 predatory journals (15.05%), no sources of revenue (including an APC) could be found. Of journals listing an APC, the median fee (USD) was $100 ($63–$150) in predatory journals (n = 59), $1866 ($800–$2205) in open access journals (n = 70), and $3000 ($2500–$3000) in subscription-based hybrid journals (n = 44). Almost 90% of all journals indicated which party retained copyright of published work. Explicit statements that authors retained copyright were present in 68.09% (n = 64) of open access journals, 36.78% (n2 = 32) of the time in subscription-based journals, and in only 12% (n = 9) of predatory journals.
This study demonstrates that our sample of potential predatory journals is distinct in some key areas from presumed legitimate journals and provides evidence of how they differ. While criteria have been proposed previously to characterize potential predatory journals , measuring each journal against a long list of criteria is not practical for the average researcher. It can be time consuming and some criteria are not straightforward to apply, as we have learned during this study. For instance, whether or not the listed editors of a journal are real people or have real affiliations with a journal is quite subjective to assess. Another example pertains to preservation and permanent access to electronic journal content. We found that not all presumed legitimate journals made explicit statements about this; however, we know that in order to be indexed in MEDLINE, a journal must “Have an acceptable arrangement for permanent preservation of, and access to, the content” .
From our findings, we have developed a list of evidence-based, salient features of suspected predatory journals (Table 10) that are straightforward to assess; we describe them further below. We recognize that these criteria are likely not sensitive enough to detect all potentially illegitimate, predatory journals. However, we feel they are a good starting point.
Non-biomedical scope of interest
We found that predatory journals tend to indicate interest in publishing research that was both biomedical and non-biomedical (e.g., agriculture, geography, astrophysics) within their remit, presumably to avoid limiting submissions and increase potential revenues. While legitimate journals may do this periodically (we did not assess the scope of presumed legitimate biomedical journals), the topics usually have some relationship between them and represent a subgroup of a larger medical specialty (e.g., Law and Medicine). Authors should examine the scope and content (e.g., actual research) of the journals they intend to publish in to determine whether it is in line with what they plan to publish.
Spelling and grammar
The home page of a journal’s website may be a good initial indicator of their legitimacy. We found several homepage indicators that may be helpful in assessing a journal’s legitimacy and quality. The homepages of potential predatory journals’ websites contained at least 10 times more spelling and grammar errors than presumed legitimate journals. Such errors may be an artefact of foreign language translation into English, as the majority of predatory journals were based in countries where a non-English language is dominant. Further, legitimate publishers and journals may be more careful about such errors to maintain professionalism and a good reputation.
Fuzzy, distorted, or potentially unauthorized image
Potential predatory journals appeared to have images that were low-resolution (e.g., fuzzy around the edges) or distorted ‘knock-off’ versions of legitimate logos or images.
Language directed at authors
Another homepage check authors can do is to examine the actual written text to gauge the intended audience. We found that presumed legitimate journals appear to target readers with their language and content (e.g., highlighting new content), whereas potential predatory journals seem to target prospective authors by inviting submissions, promising rapid publication, and promoting different metrics (including the Index Copernicus Value).
Manuscript submission and editorial process/policies
Authors should be able to find information about what happens to their article after it is submitted. Potential predatory journals do not seem to provide much information about their operations compared to presumed legitimate journals. Furthermore, most potential predatory journals request that articles be submitted via email rather than a submission system (e.g., Editorial Manager, Scholar One), as presumed legitimate journals do. Typically, journals have requirements that must be met or checked by authors or the journal during submission (e.g., declaration of conflicts of interest, agreement that the manuscript adheres to authorship standards and other journal policies, plagiarism detection). When a manuscript is submitted via email, these checks are not automatic and may not ever occur. Authors should be cautious of publishing in journals that only take submissions via email and that do not appear to check manuscripts against journal policies as such journals are likely of low quality. In addition, the email address provided by a journal seems to be a good indicator of its legitimacy. Predatory journals seem to provide non-professional or non-academic email addresses such as from providers with non-secured servers like Gmail or Yahoo.
Very low APC and inappropriate copyright
Finally, authors should be cautious when the listed APC of a biomedical journal is under $150 USD. This is very low in comparison to presumed legitimate, fully open access biomedical journals for which the median APC is at least 18 times more. Hybrid subscription journals charge 30 times the amount of potential predatory journals to publish and make research openly accessible. It has been suggested that hybrid journals charge a higher fee in order to maintain their ‘prestige’ (e.g., journals can be more selective about their content based on who is willing to pay the high fee) . On the contrary, extremely low APCs may simply be a way for potential predatory journals to attract as many submissions as possible in order to generate revenue and presumably to build their content and reputation. Evidently, the APC varies widely across journals, perhaps more than any other characteristic we measured. Journal APCs are constantly evolving and increasing requirements by funders to make research open access may have a drastic impact on APCs as we know them over the coming years.
Researchers should be trained on author responsibilities, including how to make decision about where to publish their research. Ideally, authors should start with a validated or ‘white’ list of acceptable journals. In addition to considering the items listed in Table 10 in their decision-making, tools to guide authors through the journal selection process have started to emerge, such as ThinkCheckSubmit (http://thinkchecksubmit.org/). Recently, COPE, OASPA, DOAJ, and WAME produced principles of transparency against which, among other measures, DOAJ assesses journals in part, before they can be listed in the database (https://doaj.org/bestpractice). We also encourage researchers to examine all journals for quality and legitimacy using the characteristics in Table 10 when making a decision on where to submit their research. As the journal landscape changes, it is no longer sufficient for authors to make assumptions about the quality of journals based on arbitrary measures, such as perceived reputation, impact factor, or other metrics, particularly in an era where bogus metrics abound or legitimate ones are being imitated.
This study examined most of Beall’s criteria for identification of predatory publishers and journals together with items from the COPE and OASPA. While many of the characteristics we examined were useful to distinguish predatory journals from presumed legitimate journals, there were many that do not apply or that are not unique to predatory journals. For instance, defining criteria of predatory journals  suggest that no single individual is named as an editor and that such journals do not list an editorial board. We found that this was not the case in over two thirds of predatory journals and, in fact, a named EIC could not be identified for 26 (13.07%) of the presumed legitimate journals in our sample. Such non evidence-based criteria for defining journals may introduce confusion rather than clarity and distinction.
The existing designation of journals and publishers as predatory may be confusing for other reasons. For instance, more than one presumed-legitimate publisher has appeared on Beall’s list . In October 2015, Frontiers Media, a well-known Lausanne-based open access publisher, appeared on Beall’s List . Small, new, or under-resourced journals may appear to have the look and feel of a potential predatory journal because they do not have affiliations with large publishers or technologies (e.g., manuscript submission systems) or mature systems and the features of a legitimate journal. This is in line with our findings that journals from low-resourced (LMIC) countries were more often in the potentially predatory group of journals than either of the presumed-legitimate journal arms. However, this does not imply that they are necessarily predatory journals.
Another limitation is that the majority of the open access biomedical journals in our sample (95%) charged an APC, while generally many open access journals do not. May 2015 was the last time that the DOAJ provided complete information regarding APCs of journals that it indexes (fully open access, excluding delayed or partial open access). At that time, approximately 32% of journals charged an APC. At the time of writing this article, approximately 40% of medical journals in DOAJ appear to charge an APC. However, these figures do not account for the hybrid-subscription journals that have made accommodations in response to open access, many of which are included in our sample of subscription-based journals. For such journals, our data and that of others  show that their fees appear to be substantially higher than either potential predatory or fully open access journals.
In context of other research
To the best of our knowledge this is the first comparative study of predatory journal publishing and legitimate publishing models aimed at determining how they are different and similar. Previously, Shen and Björk  examined a sample of about 5% of journals listed on Beall’s List for a number of characteristics, including three that overlap with items for which we collected data: APC, country of publisher, and rapidity of (submission to) publishing . In a large part, for the characteristics examined, our findings within the predatory journal group are very similar. For example, Shen and Björk  found the average APC for single publisher journals to be $98 USD, which is very similar to our results ($100 USD). They also found that 42% of single predatory journal publishers were located in India, whereas our estimates were closer to 62%. Differences between their study and ours may exist because we focused on biomedical journals while they included all subject areas.
It was not possible to fully blind assessors to study groups since, given the expertise of team members, a minimum knowledge of non-predatory publishers was expected. In addition, we could only include items that could be assessed superficially rather than those requiring in-depth investigations for each journal. Many items can and should be investigated further.
Since some characteristics are likely purposely similar between journals (e.g., journals from all groups claim to be open access and indicate carrying out peer review) , and it was difficult to anticipate which, we did not carry out a logistic regression to determine whether characteristics were likely to be associated with predatory or presumed legitimate journals.
This research initiates the evidence-base illuminating the difference between major publishing models and, moreover, unique characteristics of potential predatory (or illegitimate) journals (Table 10).
The possibility that some journals are predatory is problematic for many stakeholders involved in research publication. Most researchers are not formally trained on publication skills and ethics, and as such may not be able to discern whether a journal is running legitimate operations or not. For early career researchers or for those who are unaware of the existence or characteristics of predatory journals, they can be difficult to distinguish from legitimate journals. However, this study indicates that predatory journals are offering at least 18-fold lower APCs than non-predatory journals, which may be attractive to uninformed authors and those with limited fiscal resources. Assuming that each journal publishes 100 articles annually, the revenues across all predatory journals would amount to at least a $USD 100 million dollar enterprise. This is a substantial amount of money being forfeited by authors, and potentially by funders and institutions, for publications that have not received legitimate professional editorial and publishing services, including indexing in databases.
Established researchers should beware of predatory journals as well. There are numerous anecdotes about researchers (even deceased researchers ) who have been put on a journal’s editorial board or named as an editor, who did not wish to be and who were unable to get their names delisted . Aside from this potentially compromising the reputation of an individual that finds him or herself on the board, their affiliation with a potential predatory journal may confer legitimacy to the journal that is not deserved and that has the potential to confuse a naïve reader or author. As our findings indicate, this phenomenon appears to be a clear feature of predatory journals.
In addition to the costs and potential fiscal waste on publication in predatory journals, these journals do not appear to be indexed in appropriate databases to enable future researchers and other readers to consistently identify and access the research published within them. The majority of predatory journals indicated being ‘indexed’ in Google Scholar, which is not an indexing database. Google does not search pre-selected journals (as is the case with databases such as Medline, Web of Science, and Scopus), rather it searches the Internet for scholarly content. Some potentially predatory journals indicate being indexed in well-known biomedical databases; however, we have not verified the truthfulness of these claims by checking the databases. Nonetheless, if legitimate clinical research is being published in predatory journals and cannot be discovered, this is wasteful , in particular when it may impact systematic reviews. Equally, if non-peer reviewed, low quality research in predatory journals is discovered and included in a systematic review, it may pollute the scientific record. In biomedicine, this may have detrimental outcomes on patient care.
What is contained (i.e., ‘published’) within potential predatory journals is still unclear. To date, there has not been a large-scale evaluation of the content of predatory journals to determine whether research is being published, what types of studies predominate, and whether or not data (if any) are legitimate. In addition, we have little understanding of who is publishing in predatory journals (i.e., experience of author, geographic location, etc.) and why. Presumably, the low APC is an attractive feature; however, whether or not authors are intentionally or unintentionally publishing within these journals is critical to understanding the publishing landscape and anticipate future potential directions and considerations.
The findings presented here can facilitate education on how to differentiate between presumed legitimate journals and potential predatory journals.
Abridged Index Medicus
article processing charge
CONsolidated Standards Of Reporting Trials
Committee On Publication Ethics
Directory Of Open Access Journals
Enhancing the QUAlity and Transparency of health Research
international standard serial number
journal impact factor
low- or middle-income country
Open Access Scholarly Publishers Association
Public Library Of Science
Preferred Reporting Items for Systematic reviews and Meta-Analyses
STAndards for Reporting Diagnostic accuracy
STrengthening the Reporting of OBservational studies in Epidemiology
United States Dollar
Bethesda Statement on Open Access Publishing. http://legacy.earlham.edu/~peters/fos/bethesda.htm. Accessed 31 Mar 2016.
Morrison H, Salhab J, Calve-Genest A, Horava T. Open access article processing charges: DOAJ survey May 2014. Publications. 2015;3:1–16.
Crotty D. Is it True that Most Open Access Journals Do Not Charge an APC? Sort of. It Depends. The Scholarly Kitchen on WordPress.com. The Scholarly Kitchen. 2015. http://scholarlykitchen.sspnet.org/2015/08/26/do-most-oa-journals-not-charge-an-apc-sort-of-it-depends/. 4 Apr 2016.
Beall J. Beall’s List: Potential, Possible, or Probable Predatory Scholarly Open-Access Publishers. 2015. http://scholarlyoa.com/publishers/. Accessed 7 Jan 2016.
Beall J. Criteria for Determining Predatory Open-Access Publishers. 1st ed. 2012. http://wp.me/p280Ch-g5. Accessed 1 Apr 2014.
Beall J. Criteria for Determining Predatory Open-Access Publishers. 2nd ed. 2012. http://scholarlyoa.com/2012/11/30/criteria-for-determining-predatory-open-access-publishers-2nd-edition/. Accessed 5 July 2016.
Beall J. Criteria for Determining Predatory Open-Access Publishers. 3rd ed. 2015. https://scholarlyoa.files.wordpress.com/2015/01/criteria-2015.pdf. Accessed 1 July 2016.
Dadkhah M, Bianciardi G. Ranking predatory journals: solve the problem instead of removing it! Adv Pharm Bull. 2016;6(1):1–4.
Glick M, Springer MD, Bohannon J, Bohannon J, Shen C, Björk B-C. Publish and perish. J Am Dent Assoc Elsevier. 2016;147(6):385–7.
Clark J, Smith R. Firm action needed on predatory journals. BMJ. 2015;350(jan16_1):h210.
Caplan AL, Bohannon J, Beall J, Sarwar U, Nicolaou M, Balon R, et al. The problem of publication-pollution denialism. Mayo Clin Proc. 2015;90(5):565–6.
Eisen M. Door-to-Door Subscription Scams: The Dark Side of The New York Times. It is NOT Junk. 2013. http://www.michaeleisen.org/blog/?p=1354. Accessed 22 Jun 2016
Moher D, Srivastava A. You are invited to submit…. BMC Med. 2015;13:180.
Bohannon J. Who’s afraid of peer review. Science. 2013;342:60–5.
Fact SheetMEDLINE® Journal Selection. U.S. National Library of Medicine. https://www.nlm.nih.gov/pubs/factsheets/jsel.html. Accessed Apr 2014.
Selection C. Abridged Index Medicus. N Engl J Med. 1970;282(4):220–1.
U.S. National Library of Medicine. MEDLINE Policy on Indexing Electronic Journals. https://www.nlm.nih.gov/bsd/policy/ejournals.html. Accessed 26 Jul 2016.
Van Noorden R. Open access: the true cost of science publishing. Nature. 2013;495(7442):426–9.
Butler D. Investigating journals: the dark side of publishing. Nature. 2013;495(7442):433–5.
Bloudoff-Indelicato M. Backlash after Frontiers journals added to list of questionable publishers. Nature. 2015;526(7575):613.
Wellcome Trust. Wellcome Trust and COAF Open Access Spend, 2014-15. Wellcome Trust blog. 2015. https://blog.wellcome.ac.uk/2016/03/23/wellcome-trust-and-coaf-open-access-spend-2014-15/. Accessed 21 Jun 2016
Shen C, Bjork B-C. ‘Predatory’ open access: a longitudinal study of article volumes and market characteristics. BMC Med. 2015;13:230.
Spears T. The Editor is Deceased: Fake Science Journals Hit New Low. Ottawa: Ottawa Citizen; 2016.
Kolata G. For Scientists, an Exploding World of Pseudo-Academia. NYTimes.com. 2016. http://www.nytimes.com/2013/04/08/health/for-scientists-an-exploding-world-of-pseudo-academia.html?pagewanted=all. Accessed July 2016.
Chan AW, Song F, Vickers A, Jefferson T, Dickersin K, Gøtzsche PC, et al. Increasing value and reducing waste: addressing inaccessible research. Lancet. 2014;383(9913):257–66.
No funding was received for this project.
Availability of data and materials
The screening and data extraction forms used, and the data generated, in this study are available from the authors on request.
DM and LS conceived of this project and drafted the protocol, with revisions by VB. RB, JC, JG, OM, DM, JR, LS, BJS, and LT were involved in the conduct of this project. LS and LT performed analysis of data. LS drafted the manuscript. All authors provided feedback on this manuscript and approved the final version for publication.
VB is the Chair of COPE and the Executive Director of the Australasian Open Access Strategy Group.
Consent for publication
Ethics approval and consent to participate
David Moher affirms that this manuscript is an honest, accurate, and transparent account of the study being reported, that no important aspects of the study have been omitted, and that any discrepancies from the study as planned (and, if relevant, registered) have been explained.