Tag Archives: scientific fraud

Scientific fraud: widespread and organized according to Northwestern University research + math fraud scandal

I have three stories about issues with science and mathematics: the research, the reporting, and the fraud.

Northwestern University and widespread scientific fraud

An August 4, 2025 article by Cathleen O’Grady for science.org describes a study into global networks instigating scientific fraud, Note: A link has been removed,

For years, sleuths who study scientific fraud have been sounding the alarm about the sheer size and sophistication of the industry that churns out fake publications. Now, an extensive investigation finds evidence of a range of bad actors profiting from fraud. The study, based on an analysis of thousands of publications and their authors and editors, shows paper mills are just part of a complex, interconnected system that includes publishers, journals, and brokers.

The paper, published today in the Proceedings of the National Academy of Sciences, paints an alarming picture. Northwestern University metascientist Reese Richardson and his colleagues identify networks of editors and authors colluding to publish shoddy or fraudulent papers, report that large organizations are placing batches of fake papers in journals, suggest brokers may serve as intermediaries between paper mills and intercepted journals, and find that the number of fake papers—though still relatively small—seems to be increasing at a rate far greater than the scientific literature generally.

The paper shows that misconduct “has become an industry,” says Anna Abalkina of the Free University of Berlin, who studies corruption in science and was not involved with the research. Richardson and colleagues hope their sweeping case will attract attention and spur change.

O’Grady’s August 4, 2025 article provides some fascinating detail, Note: Links have been removed,

They began their analysis by pinpointing corrupt editors. They focused their investigation on PLOS ONE, because the megajournal allows easy access to bulk metadata and publishes the names of the editors who have handled the thousands of papers it publishes each year, making it possible to detect anomalies without behind-the-scenes information. The researchers identified all the papers from the journal that had been retracted or received comments on PubPeer—a website that allows researchers to critique published work—and then identified each paper’s editors.

All told, 33 editors stood out as more frequently handling work that was later retracted or criticized than would be expected by chance. “Some of these were immense outliers,” Richardson says. For instance, of the 79 papers that one editor had handled at PLOS ONE, 49 have been retracted. Flagged editors handled 1.3% of papers published in the journal by 2024, but nearly one-third of all retracted papers.

The team also spotted that these editors worked on certain authors’ papers at a suspiciously high rate. These authors were often editors at PLOS [Public Library of Science] ONE themselves, and they often handled each other’s papers. It’s possible that some editors are being paid bribes, Richardson says, but “also possible that these are informal arrangements that are being made among colleagues.” The researchers detected similarly questionable editor behavior in 10 journals published by Hindawi, an open-access publisher that was shuttered because of rampant paper mill activity after Wiley acquired it. A spokesperson for Wiley told Science the publisher has made “significant investments to address research integrity issues.”

Renee Hoch, head of publication ethics at PLOS, said in an email to Science that the publisher has long been aware of networks like these, and will assess whether any of the editors implicated are still on the journal’s editorial board, opening investigations if they are. She emphasizes that the study focused on PLOS because of its readily accessible data: “Paper mills are truly an industry-wide problem.”

Researchers working on paper mills have long assumed that editors and authors have been colluding. The new findings are “killer evidence” for these suspicions, says Domingo Docampo, a bibliometrician at the University of Vigo (Spain). He adds that although the findings only show collusion at a limited number of journals, others are probably affected. Just last week, Retraction Watch reported that the publisher Frontiers had begun to retract 122 papers after discovering a network of editors and authors “who conducted peer review with undisclosed conflicts of interest,” according to a company statement. The network of 35 individuals has also published more than 4000 papers in journals from seven other publishers, the company said, which require further scrutiny. A Frontiers spokesperson said they planned to share information with the other affected publishers.

Richardson and his colleagues found that the problem goes far beyond networks of unscrupulous editors and authors scratching each other’s backs. They identified what appear to be coordinated efforts to arrange the publication of batches of dubious papers in multiple journals.

For the curious, there’s more in O’Grady’s August 4, 2025 article. An August 4, 2025 Northwestern University news release by Amanda Morris (received via email and available on EurekAlert) focuses on other aspects of the research,

From fabricated research to paid authorships and citations, organized scientific fraud is on the rise, according to a new Northwestern University study.

By combining large-scale data analysis of scientific literature with case studies, the researchers led a deep investigation into scientific fraud. Although concerns around scientific misconduct typically focus on lone individuals, the Northwestern study instead uncovered sophisticated global networks of individuals and entities, which systematically work together to undermine the integrity of academic publishing.

The problem is so widespread that the publication of fraudulent science is outpacing the growth rate of legitimate scientific publications. The authors argue these findings should serve as a wake-up call to the scientific community, which needs to act before the public loses confidence in the scientific process.

The study will be published during the week of August 4 the Proceedings of the National Academy of Sciences.

“Science must police itself better in order to preserve its integrity,” said Northwestern’s Luís A. N. Amaral, the study’s senior author. “If we do not create awareness around this problem, worse and worse behavior will become normalized. At some point, it will be too late, and scientific literature will become completely poisoned. Some people worry that talking about this issue is attacking science. But I strongly believe we are defending science from bad actors. We need to be aware of the seriousness of this problem and take measures to address it.”

An expert in complex social systems, Amaral is the Erastus Otis Haven Professor and professor of engineering sciences and applied mathematics at Northwestern’s McCormick School of Engineering. Reese Richardson, a postdoctoral fellow in Amaral’s laboratory, is the paper’s first author.

Extensive analysis

When people think about scientific fraud, they might remember news reports of retracted papers, falsified data or plagiarism. These reports typically center around the isolated actions of one individual, who takes shortcuts to get ahead in an increasingly competitive industry. But Amaral and his team uncovered a widespread underground network operating within the shadows and outside of the public’s awareness.

“These networks are essentially criminal organizations, acting together to fake the process of science,” Amaral said. “Millions of dollars are involved in these processes.”

To conduct the study, the researchers analyzed extensive datasets of retracted publications, editorial records and instances of image duplication. Most of the data came from major aggregators of scientific literature, including Web of Science (WoS), Elsevier’s Scopus, National Library of Medicine’s PubMed/MEDLINE and OpenAlex, which includes data from Microsoft Academic Graph, Crossref, ORCID, Unpaywall and other institutional repositories.

Richardson and his colleagues also collected lists of de-indexed journals, which are scholarly journals that have been removed from databases for failing to meet certain quality or ethical standards. The researchers also included data on retracted articles from Retraction Watch, article comments from PubPeer and metadata — such as editor names, submission dates and acceptance dates — from articles published in specific journals.

Buying a reputation

After analyzing the data, the team uncovered coordinated efforts involving “paper mills,” brokers and infiltrated journals. Functioning much like factories, paper mills churn out large numbers of manuscripts, which they then sell to academics who want to quickly publish new work. These manuscripts are mostly low quality — featuring fabricated data, manipulated or even stolen images, plagiarized content and sometimes nonsensical or physically impossible claims.

“More and more scientists are being caught up in paper mills,” Amaral said. “Not only can they buy papers, but they can buy citations. Then, they can appear like well-reputed scientists when they have barely conducted their own research at all.”

“Paper mills operate by a variety of different models,” Richardson added. “So, we have only just been able to scratch the surface of how they operate. But they sell basically anything that can be used to launder a reputation. They often sell authorship slots for hundreds or even thousands of dollars. A person might pay more money for the first author position or less money for a fourth author position. People also can pay to get papers they have written automatically accepted in a journal through a sham peer-review process.”

To identify more articles originating from paper mills, the Amaral group launched a parallel project that automatically scans published materials science and engineering papers. The team specifically looked for authors who misidentified instruments they used in their research. A paper with those results was accepted by the journal PLOS ONE.

Brokers, hijacking and collusion

Amaral, Richardson and their collaborators found fraudulent networks use several key strategies: (1) Groups of researchers collude to publish papers across multiple journals. When their activities are discovered, the papers are subsequently retracted; (2) brokers serve as intermediaries to enable mass publication of fraudulent papers in compromised journals; (3) fraudulent activities are concentrated in specific, vulnerable subfields; and (4) organized entities evade quality-control measures, such as journal de-indexing.

“Brokers connect all the different people behind the scenes,” Amaral said. “You need to find someone to write the paper. You need to find people willing to pay to be the authors. You need to find a journal where you can get it all published. And you need editors in that journal who will accept that paper.”

Sometimes these organizations go around established journals altogether, searching instead for defunct journals to hijack. When a legitimate journal stops publishing, for example, bad actors can take over its name or website. These actors surreptitiously assume the journal’s identity, lending credibility to its fraudulent publications, despite the actual publication being defunct.

“This happened to the journal HIV Nursing,” Richardson said. “It was formerly the journal of a professional nursing organization in the U.K., then it stopped publishing, and its online domain lapsed. An organization bought the domain name and started publishing thousands of papers on subjects completely unrelated to nursing, all indexed in Scopus.”

Fighting for science

To combat this growing threat to legitimate scientific publishing, Amaral and Richardson emphasize the need for a multi-prong approach. This approach includes enhanced scrutiny of editorial processes, improved methods for detecting fabricated research, a greater understanding of the networks facilitating this misconduct and a radical restructuring of the system of incentives in science.

Amaral and Richardson also underscore the importance of addressing these issues before artificial intelligence (AI) infiltrates scientific literature more than it already has.

“If we’re not prepared to deal with the fraud that’s already occurring, then we’re certainly not prepared to deal with what generative AI can do to scientific literature,” Richardson said. “We have no clue what’s going to end up in the literature, what’s going to be regarded as scientific fact and what’s going to be used to train future AI models, which then will be used to write more papers.”

“This study is probably the most depressing project I’ve been involved with in my entire life,” Amaral said. “Since I was a kid, I was excited about science. It’s distressing to see others engage in fraud and in misleading others. But if you believe that science is useful and important for humanity, then you have to fight for it.”

Here’s a link to and a citation for the paper,

The entities enabling scientific fraud at scale are large, resilient, and growing rapidly by Reese A. K. Richardson, Spencer S. Hong, Jennifer A. Byrne, Thomas Stoeger, and Luís A. Nunes Amaral. Proceedings of the National Academy of Sciences August 4, 2025 122 (32) e2420092122 DOI: https://doi.org/10.1073/pnas.2420092122

This paper is open access.

And now—math fraud

A September 19, 2025 news item on ScienceDaily features an investigation into fraudulent math research, Note: A link has been removed,

An international team of authors led by Ilka Agricola, professor of mathematics at the University of Marburg, Germany, has investigated fraudulent practices in the publication of research results in mathematics on behalf of the German Mathematical Society (DMV) and the International Mathematical Union (IMU), documenting systematic fraud over many years. The results of the study were recently published on the preprint server arxiv.org and in the Notices of the American Mathematical Society (AMS) and have since caused a stir among mathematicians.

Sanjana Gajbhiye’s September ??, 2025 article for earth.com delves further into the topic, Note: Links have been removed,

Quality lost to quantity

The findings show how the definition of research quality has shifted. Instead of focusing on content, originality, and insight, institutions and individuals are increasingly evaluated by commercial metrics. These include the number of publications, total citations, and the so-called impact factor of journals.

Such measures, calculated by private companies with little transparency, have gained outsized influence. Providers promote their databases globally, and universities use them to enhance prestige and compete internationally.

This environment rewards quantity over quality, pushing academics to publish more, even when contributions are marginal or flawed.

Fraudulent companies have seized this opportunity. They sell services that manipulate rankings, offering ghostwritten articles, fake peer reviews, and even bundles of citations. For individuals, this can mean better career prospects.

For universities, it can result in higher rankings, increased funding, and greater appeal to international students. The collateral damage is a growing pool of unread publications that add nothing to scientific understanding.

Fake mathematics success

The report documents striking examples that reveal how metrics can produce absurd outcomes. In 2019, Clarivate Inc., the market leader for citation data, ranked a Taiwanese university as having the most world-class mathematicians. The catch was startling: mathematics was not even offered at the institution.

Mathematic trust under threat

“‘Fake science‘ is not only annoying, it is a danger to science and society,” said IMU Secretary General Professor Christoph Sorger.

“Because you don’t know what is valid and what is not. Targeted disinformation undermines trust in science and also makes it difficult for us mathematicians to decide which results can be used as a basis for further research.”

This erosion of trust strikes at the heart of mathematics. Proofs rely on certainty, yet when fraudulent or hollow work appears in respected outlets, that certainty weakens.

Fixing trust in mathematics publishing

The commission’s work does not end with exposing the problem. It also outlines possible solutions for a healthier publication system. These recommendations emphasize the need to strengthen peer review, encourage collaboration among journals, and recenter the evaluation of research on quality rather than raw numbers.

Metrics are deeply tied to funding and prestige, so the shift won’t be simple, but it could reshape the landscape for future generations.

A September 20, 2025 Castle Journal blog posting provides more information,

The “Culture of Numbers” and its Consequences

The study, led by Professor Ilka Agricola of the University of Marburg, argues that the root cause of the problem is a “culture of numbers” that prioritizes commercial metrics over scientific content. Universities and research institutions have become increasingly reliant on commercial databases like Clarivate’s Journal Citation Reports (JCR) to evaluate researchers. These metrics, which are not transparent and are not vetted by the scientific community, have become the main currency for career progression, grants, and prestige.

 * “Megajournals”: The study highlights the rise of “megajournals,” which publish anything as long as the authors pay a fee. These journals now publish more articles per year than all reputable mathematics journals combined. The report cites a shocking example where a commercial database ranked a university in Taiwan as having the most world-class researchers in mathematics, despite the fact that the university does not even offer mathematics as a subject.

 * Paper Mills and Citation Cartels: The investigation found evidence of “paper mills,” which sell fabricated papers to researchers, and “citation cartels,” where academics agree to cite each other’s work to artificially inflate their metrics. These services are offered anonymously online, with prices for articles and citations ranging from hundreds to thousands of dollars. The report describes these networks as “criminal organizations” that have invaded the “ecosystem” of scientific publishing.

Ilka Agricola gave an interview to Retraction Watch, from the undated article, Note: Links have been removed,

A pair of papers posted to the arXiv addresses the issue of fraudulent publishing in math, particularly metrics gaming, and offers a list of recommendations to help detect and deal with that problem and other fraudulent activities. (The former was also published in the October AMS Notices; the latter will appear in the November issue.) “Fraudulent publishing undermines trust in science and scientific results and therefore fuels antiscience movements,” mathematician Ilka Agricola, lead author of both papers, told Retraction Watch. 

A professor of mathematics at Marburg University in Germany, Agricola was president of the German Mathematical Society in 2021-2022 and is chair of the Committee on Publishing of the International Mathematical Union. The new articles are the products of a working group of the IMU and the International Council of Industrial and Applied Mathematics. 

Retraction Watch: As you note in the new papers, Clarivate announced in 2023 it had excluded the entire field of math from its list of “Highly Cited Researchers,” or HCRs. What’s going on?

Agricola: The publication culture in math differs a bit from, say, experimental and life sciences. On average, mathematicians publish fewer papers with fewer authors than scientists in other fields. So, with the same absolute number of papers and citations, one can become a “highly-cited researcher” in math, but not in other fields. Thus, gaming the system is easier. 

The list of HCRs for mathematics became so screwed that Clarivate couldn’t pretend anymore that it had any value. This being said, Clarivate announced that they would look into new measuring tools, but didn’t come up with any alternative ideas in the meantime, nor did they contact any representatives of the international mathematical community. 

Retraction Watch:  Few people talk about fraudulent publishing in math. Why is that?

Agricola: For a long time, mathematicians thought that as long as they keep away from predatory journals or paper mills, the problem does not affect them. This turned out to be wrong. 

Retraction Watch: If you look at the number of papers that tripped Clear Skies’ Papermill Alarm in 2022 (we included a histogram in this article we wrote for The Conversation [link and excerpts follow]), math is pretty far down the list. Are there a lot of fake papers in math?

Agricola: It is probably fair to say that the problem is not as severe as in other fields like cancer research, but the community is smaller and the number of fake papers is growing at alarming speed. Predatory and low-quality mega-journals are trying hard to lure respected scientists into their parallel universe of fake science, thus trying to give themselves the impression of respectability. Thus, one of our goals is to raise awareness for the issue in the mathematical community!

Retraction Watch: You and your coauthors are mathematicians, and yet you argue against focusing on numbers like journal impact factors and publication and citation counts. Is that what’s driving all of this bad behavior?

Agricola: “When a measure becomes a target, it ceases to be a good measure.” This quote is from the British economist Charles Goodhart, and it also applies to bibliometrics measures. Of course, gaming these metrics has always existed, but some of us liked to believe that they would be roughly OK, with some error bar due to some cheating. Now, we realize the error bar is larger than the number one wants to measure. Perhaps one advantage of mathematicians is that they are not easily impressed by numbers, and we have the means to understand and analyze them — this is our job. And so, the conclusion is very clear: The correlation between bibliometrics and research quality is so low that we should not use bibliometrics. And I urge all colleagues to say so openly!

Retraction Watch: So how do we judge research quality if we shouldn’t use publication metrics?

Agricola: Read the actual publications instead of relying on bibliometrics! Plus, in mathematics, we are lucky to have two extremely well curated databases for math papers and journals, zbMath Open and MathReviews. If a journal is not included there, it’s either very interdisciplinary or one should get suspicious.

Retraction Watch: Is it possible for individual researchers to jump off the bibiometrics bandwagon without jeopardizing their careers?

Agricola: We need to fight for a change in culture, that’s for sure, and the path will be rash and hard. To young researchers, we should give the warning that being involved in predatory publishing can also just as well put their scientific integrity at risk. Remember the people who had to resign because of data falsification? 

I am providing citations (of a sort) to both papers and links to all three sites where both papers can be found and PDFs for both papers: Everything is open access.

Fraudulent Publishing in the Mathematical Sciences by Ilka Agricola, Lynn Heller, Wil Schilders, Moritz Schubotz, Peter Taylor, Luis Vega.

arXiv: https://arxiv.org/abs/2509.07257

AMS (American Mathematical Society) Notices October 2025: https://www.ams.org/journals/notices/202509/noti3217/noti3217.html?adat=October%202025&trk=3217&pdfissue=202509&pdffile=rnoti-p1038.pdf&cat=none&type=.html

Ilke Agricola’s Research Gate website: https://www.researchgate.net/profile/Ilka-Agricola (scroll down to see the listed papers)

PDF: https://www.ams.org/journals/notices/202509/rnoti-p1038.pdf

How to Fight Fraudulent Publishing in the Mathematical Sciences: Joint Recommendations of the IMU [International Mathematical Union] and the ICIAM [International Council for Industrial and Applied Mathematics] by Ilka Agricola, Lynn Heller, Wil Schilders, Moritz Schubotz, Peter Taylor, Luis Vega.

arXiv: https://arxiv.org/abs/2509.09877

AMS (American Mathematical Society) Notices November 2025: https://www.ams.org/journals/notices/202510/noti3266/noti3266.html?adat=November%202025&trk=3266&pdfissue=202510&pdffile=rnoti-p1179.pdf&cat=none&type=.html

Ilke Agricola’s Research Gate website: https://www.researchgate.net/profile/Ilka-Agricola (scroll down to see the listed papers)

PDF: https://www.ams.org/journals/notices/202510/rnoti-p1179.pdf

Fraud slows down research

Mentioned in the Retraction Watch/Agricola interview, this January 29, 2025 article by Frederik Joelving (contributing editor, Retraction Watch), Cyril Labbé, (professor of computer science, Université Grenoble Alpes [UGA]), Guillaume Cabanac, (professor of computer Science, Institut de Recherche en Informatique de Toulouse) is chilling, Note: Links have been removed,

Over the past decade, furtive commercial entities around the world have industrialized the production, sale and dissemination of bogus scholarly research, undermining the literature that everyone from doctors to engineers rely on to make decisions about human lives.

It is exceedingly difficult to get a handle on exactly how big the problem is. Around 55,000 scholarly papers have been retracted to date, for a variety of reasons, but scientists and companies who screen the scientific literature for telltale signs of fraud estimate that there are many more fake papers circulating – possibly as many as several hundred thousand. This fake research can confound legitimate researchers who must wade through dense equations, evidence, images and methodologies only to find that they were made up.

Even when the bogus papers are spotted – usually by amateur sleuths on their own time – academic journals are often slow to retract the papers, allowing the articles to taint what many consider sacrosanct: the vast global library of scholarly work that introduces new ideas, reviews other research and discusses findings.

These fake papers are slowing down research that has helped millions of people with lifesaving medicine and therapies from cancer to COVID-19. Analysts’ data shows that fields related to cancer and medicine are particularly hard hit, while areas like philosophy and art are less affected. Some scientists have abandoned their life’s work because they cannot keep pace given the number of fake papers they must bat down.

The problem reflects a worldwide commodification of science. Universities, and their research funders, have long used regular publication in academic journals as requirements for promotions and job security, spawning the mantra “publish or perish.”

But now, fraudsters have infiltrated the academic publishing industry to prioritize profits over scholarship. Equipped with technological prowess, agility and vast networks of corrupt researchers, they are churning out papers on everything from obscure genes to artificial intelligence in medicine.

These papers are absorbed into the worldwide library of research faster than they can be weeded out. About 119,000 scholarly journal articles and conference papers are published globally every week, or more than 6 million a year. Publishers estimate that, at most journals, about 2% of the papers submitted – but not necessarily published – are likely fake, although this number can be much higher at some publications.

… there is a bustling online underground economy for all things scholarly publishing. Authorship, citations, even academic journal editors, are up for sale. This fraud is so prevalent that it has its own name: paper mills, a phrase that harks back to “term-paper mills,” where students cheat by getting someone else to write a class paper for them.

The impact on publishers is profound. In high-profile cases, fake articles can hurt a journal’s bottom line. Important scientific indexes – databases of academic publications that many researchers rely on to do their work – may delist journals that publish too many compromised papers. There is growing criticism that legitimate publishers could do more to track and blacklist journals and authors who regularly publish fake papers that are sometimes little more than artificial intelligence-generated phrases strung together.

To better understand the scope, ramifications and potential solutions of this metastasizing assault on science, we – a contributing editor at Retraction Watch, a website that reports on retractions of scientific papers and related topics, and two computer scientists at France’s Université Toulouse III–Paul Sabatier and Université Grenoble Alpes who specialize in detecting bogus publications – spent six months investigating paper mills.

This included, by some of us at different times, trawling websites and social media posts, interviewing publishers, editors, research-integrity experts, scientists, doctors, sociologists and scientific sleuths engaged in the Sisyphean task of cleaning up the literature. It also involved, by some of us, screening scientific articles looking for signs of fakery.

What emerged is a deep-rooted crisis that has many researchers and policymakers calling for a new way for universities and many governments to evaluate and reward academics and health professionals across the globe.

Just as highly biased websites dressed up to look like objective reporting are gnawing away at evidence-based journalism and threatening elections, fake science is grinding down the knowledge base on which modern society rests.

The January 29, 2025 article highlights a number of problems including these,

To expedite the publication of one another’s work, some corrupt scientists form peer review rings. Paper mills may even create fake peer reviewers impersonating real scientists to ensure their manuscripts make it through to publication. Others bribe editors or plant agents on journal editorial boards.

María de los Ángeles Oviedo-García, a professor of marketing at the University of Seville in Spain, spends her spare time hunting for suspect peer reviews from all areas of science, hundreds of which she has flagged on PubPeer. ……

“One of the demanding fights for me is to keep faith in science,” says Oviedo-García, who tells her students to look up papers on PubPeer before relying on them too heavily. Her research has been slowed down, she adds, because she now feels compelled to look for peer review reports for studies she uses in her work. Often there aren’t any, because “very few journals publish those review reports,” Oviedo-García says.

An ‘absolutely huge’ problem

It is unclear when paper mills began to operate at scale. The earliest article retracted due to suspected involvement of such agencies was published in 2004, according to the Retraction Watch Database, which contains details about tens of thousands of retractions. (The database is operated by The Center for Scientific Integrity, the parent nonprofit of Retraction Watch.) Nor is it clear exactly how many low-quality, plagiarized or made-up articles paper mills have spawned.

“The threat of paper mills to scientific publishing and integrity has no parallel over my 30-year scientific career …. In the field of human gene science alone, the number of potentially fraudulent articles could exceed 100,000 original papers,” she [Jennifer Byrne, an Australian scientist] wrote to lawmakers, adding, “This estimate may seem shocking but is likely to be conservative.”

In one area of genetics research – the study of noncoding RNA in different types of cancer – “We’re talking about more than 50% of papers published are from mills,” Byrne said. “It’s like swimming in garbage.”

… in the global south, the publish-or-perish edict runs up against underdeveloped research infrastructures and education systems, leaving scientists in a bind. For a Ph.D., the Cairo physician who requested anonymity conducted an entire clinical trial single-handedly – from purchasing study medication to randomizing patients, collecting and analyzing data and paying article-processing fees. In wealthier nations, entire teams work on such studies, with the tab easily running into the hundreds of thousands of dollars.

“Research is quite challenging here,” the physician said. That’s why scientists “try to manipulate and find easier ways so they get the job done.”

Institutions, too, have gamed the system with an eye to international rankings. In 2011, the journal Science described how prolific researchers in the United States and Europe were offered hefty payments for listing Saudi universities as secondary affiliations on papers. And in 2023, the magazine, in collaboration with Retraction Watch, uncovered a massive self-citation ploy by a top-ranked dental school in India that forced undergraduate students to publish papers referencing faculty work.

According to the January 29, 2025 article, there is a root cause, Note: Links have been removed,

… unsavory schemes can be traced back to the introduction of performance-based metrics in academia, a development driven by the New Public Management movement that swept across the Western world in the 1980s, according to Canadian sociologist of science Yves Gingras of the Université du Québec à Montréal. When universities and public institutions adopted corporate management, scientific papers became “accounting units” used to evaluate and reward scientific productivity rather than “knowledge units” advancing our insight into the world around us, Gingras wrote.

This transformation led many researchers to compete on numbers instead of content, which made publication metrics poor measures of academic prowess. As Gingras has shown, the controversial French microbiologist Didier Raoult, who now has more than a dozen retractions to his name, has an h-index – a measure combining publication and citation numbers – that is twice as high as that of Albert Einstein – “proof that the index is absurd,” Gingras said.

Worse, a sort of scientific inflation, or “scientometric bubble,” has ensued, with each new publication representing an increasingly small increment in knowledge. “We publish more and more superficial papers, we publish papers that have to be corrected, and we push people to do fraud,” said Gingras.

In terms of career prospects of individual academics, too, the average value of a publication has plummeted, triggering a rise in the number of hyperprolific authors. One of the most notorious cases is Spanish chemist Rafael Luque, who in 2023 reportedly published a study every 37 hours.

There is some hope according to the January 29, 2025 article, Note: Links have been removed,

Stern [Bodo Stern, a former editor of the journal Cell and chief of Strategic Initiatives at Howard Hughes Medical Institute] isn’t the first scientist to bemoan the excessive focus on bibliometrics. “We need less research, better research, and research done for the right reasons,” wrote the late statistician Douglas G. Altman in a much-cited editorial from 1994. “Abandoning using the number of publications as a measure of ability would be a start.”

Nearly two decades later, a group of some 150 scientists and 75 science organizations released the San Francisco Declaration on Research Assessment, or DORA, discouraging the use of the journal impact factor and other measures as proxies for quality. The 2013 declaration has since been signed by more than 25,000 individuals and organizations in 165 countries.

Despite the declaration, metrics remain in wide use today, and scientists say there is a new sense of urgency.

Stern and his colleagues have tried to make improvements at their institution. Researchers who wish to renew their seven-year contract have long been required to write a short paragraph describing the importance of their major results. Since the end of 2023, they also have been asked to remove journal names from their applications.

That way, “you can never do what all reviewers do – I’ve done it – look at the bibliography and in just one second decide, ‘Oh, this person has been productive because they have published many papers and they’re published in the right journals,’” says Stern. “What matters is, did it really make a difference?”

Shifting the focus away from convenient performance metrics seems possible not just for wealthy private institutions like Howard Hughes Medical Institute, but also for large government funders. In Australia, for example, the National Health and Medical Research Council in 2022 launched the “top 10 in 10” policy, aiming, in part, to “value research quality rather than quantity of publications.”

Rather than providing their entire bibliography, the agency, which assesses thousands of grant applications every year, asked researchers to list no more than 10 publications from the past decade and explain the contribution each had made to science. …

Gingras, the Canadian sociologist, advocates giving scientists the time they need to produce work that matters, rather than a gushing stream of publications. He is a signatory to the Slow Science Manifesto: “Once you get slow science, I can predict that the number of corrigenda, the number of retractions, will go down,” he says.

At one point, Gingras was involved in evaluating a research organization whose mission was to improve workplace security. An employee presented his work. “He had a sentence I will never forget,” Gingras recalls. The employee began by saying, “‘You know, I’m proud of one thing: My h-index is zero.’ And it was brilliant.” The scientist had developed a technology that prevented fatal falls among construction workers. “He said, ‘That’s useful, and that’s my job.’ I said, ‘Bravo!’”

Sometimes, there’s a science reporting problem

A September 3, 2025 Universiteit van Amsterdam press release (also on EurekAlert) highlights a problem with science reporting and over confidence,

Science journalists aren’t particularly concerned about so-called “predatory journals”, confident that they have the skills and intuition needed to avoid reporting on problematic research. For many, a journal’s reputation and name-recognition are decisive factors in assessing the quality of scientific research – but this could be exacerbating existing imbalances in science and journalism. This perspective emerges from a new study, led by Dr Alice Fleerackers of the University of Amsterdam (UvA), and published on 2 September [2025] in Journalism Practice.

Predatory journals prioritise profit over editorial and publication standards. They often charge researchers publication fees but offer little to no real quality control, such as peer review. As a result, some journals publish almost everything submitted. ‘Predatory journals are not a harmless side effect of the academic publishing industry,’ says Fleerackers. ‘They are becoming increasingly common, raising concerns about the integrity of scientific publishing. They not only undermine the reliability of science but also jeopardise science journalism, as journalists can unknowingly report on weak or even flawed research.’

In the new study, Fleerackers – along with colleagues from Simon Fraser University (Canada) and San Francisco State University (US) – investigated how science journalists view predatory journals and what strategies they employ to ensure the reliability of the journals they report on. The researchers present a qualitative analysis of interviews with 23 health, science, and environmental journalists in Europe and North America.

Problematic, but only in theory

Some of the journalists interviewed were familiar with the phenomenon of predatory journals and acknowledged that they are theoretically problematic. However, most weren’t concerned that they might be using them in their own work. They acknowledged that these journals might be a problem for colleagues, but not for them.

Well-known, therefore reliable

Journalists in the study were confident they wouldn’t fall for a predatory journal because of their strong intuition, which they said allowed them to immediately distinguish high-quality from problematic research. Besides their intuition, they also relied on strategies for verifying the reliability of research that they had developed through years of experience. These strategies often centred on trust proxies – like the journal’s prestige, impact factor, and selectivity – as well as whether the journal claimed to conduct peer review.

Proofreading also played a role for some journalists: if an article contained grammatical or spelling errors, it could be a sign of low-quality research. Open access journals were also considered less reliable by several journalists. ‘But by far the most commonly used benchmark for reliability was the journal’s reputation,’ Fleerackers explains. ‘Some journalists avoid all journals they’re not familiar with and report only on research published in top journals like Science and Nature.’

Distortion in science news

According to Fleerackers, journalists’ focus on the reputation and prestige of journals has major consequences for the diversity of research in the news media. ‘Research from newer, lesser-known journals, and from journals in the Global South, for example, remains hidden from the public. Most journalists in our study didn’t realise that their selection strategies could perpetuate the existing imbalance in science news. I hope that our study can raise awareness of this among journalists.’


Here’s a link to and a citation for the paper.

“I’d Like to Think I’d Be Able to Spot One”: How Journalists Navigate Predatory Journals by Alice Fleerackers, Laura L. Moorhead & Juan Pablo Alperin. Journalism Practice 1–19. DOI: https://doi.org/10.1080/17512786.2025.2551984 Published online: 02 Sep 2025

Final comments

This has been a good wake up call for me. Bad apples, yes, but criminal networks? I had no idea. I will probably write more about this in my 2025 year post. In the meantime, This is a good reminder to exercise caution.