Tag Archives: UNIVERSITEIT VAN AMSTERDAM

Scientific fraud: widespread and organized according to Northwestern University research + math fraud scandal

I have three stories about issues with science and mathematics: the research, the reporting, and the fraud.

Northwestern University and widespread scientific fraud

An August 4, 2025 article by Cathleen O’Grady for science.org describes a study into global networks instigating scientific fraud, Note: A link has been removed,

For years, sleuths who study scientific fraud have been sounding the alarm about the sheer size and sophistication of the industry that churns out fake publications. Now, an extensive investigation finds evidence of a range of bad actors profiting from fraud. The study, based on an analysis of thousands of publications and their authors and editors, shows paper mills are just part of a complex, interconnected system that includes publishers, journals, and brokers.

The paper, published today in the Proceedings of the National Academy of Sciences, paints an alarming picture. Northwestern University metascientist Reese Richardson and his colleagues identify networks of editors and authors colluding to publish shoddy or fraudulent papers, report that large organizations are placing batches of fake papers in journals, suggest brokers may serve as intermediaries between paper mills and intercepted journals, and find that the number of fake papers—though still relatively small—seems to be increasing at a rate far greater than the scientific literature generally.

The paper shows that misconduct “has become an industry,” says Anna Abalkina of the Free University of Berlin, who studies corruption in science and was not involved with the research. Richardson and colleagues hope their sweeping case will attract attention and spur change.

O’Grady’s August 4, 2025 article provides some fascinating detail, Note: Links have been removed,

They began their analysis by pinpointing corrupt editors. They focused their investigation on PLOS ONE, because the megajournal allows easy access to bulk metadata and publishes the names of the editors who have handled the thousands of papers it publishes each year, making it possible to detect anomalies without behind-the-scenes information. The researchers identified all the papers from the journal that had been retracted or received comments on PubPeer—a website that allows researchers to critique published work—and then identified each paper’s editors.

All told, 33 editors stood out as more frequently handling work that was later retracted or criticized than would be expected by chance. “Some of these were immense outliers,” Richardson says. For instance, of the 79 papers that one editor had handled at PLOS ONE, 49 have been retracted. Flagged editors handled 1.3% of papers published in the journal by 2024, but nearly one-third of all retracted papers.

The team also spotted that these editors worked on certain authors’ papers at a suspiciously high rate. These authors were often editors at PLOS [Public Library of Science] ONE themselves, and they often handled each other’s papers. It’s possible that some editors are being paid bribes, Richardson says, but “also possible that these are informal arrangements that are being made among colleagues.” The researchers detected similarly questionable editor behavior in 10 journals published by Hindawi, an open-access publisher that was shuttered because of rampant paper mill activity after Wiley acquired it. A spokesperson for Wiley told Science the publisher has made “significant investments to address research integrity issues.”

Renee Hoch, head of publication ethics at PLOS, said in an email to Science that the publisher has long been aware of networks like these, and will assess whether any of the editors implicated are still on the journal’s editorial board, opening investigations if they are. She emphasizes that the study focused on PLOS because of its readily accessible data: “Paper mills are truly an industry-wide problem.”

Researchers working on paper mills have long assumed that editors and authors have been colluding. The new findings are “killer evidence” for these suspicions, says Domingo Docampo, a bibliometrician at the University of Vigo (Spain). He adds that although the findings only show collusion at a limited number of journals, others are probably affected. Just last week, Retraction Watch reported that the publisher Frontiers had begun to retract 122 papers after discovering a network of editors and authors “who conducted peer review with undisclosed conflicts of interest,” according to a company statement. The network of 35 individuals has also published more than 4000 papers in journals from seven other publishers, the company said, which require further scrutiny. A Frontiers spokesperson said they planned to share information with the other affected publishers.

Richardson and his colleagues found that the problem goes far beyond networks of unscrupulous editors and authors scratching each other’s backs. They identified what appear to be coordinated efforts to arrange the publication of batches of dubious papers in multiple journals.

For the curious, there’s more in O’Grady’s August 4, 2025 article. An August 4, 2025 Northwestern University news release by Amanda Morris (received via email and available on EurekAlert) focuses on other aspects of the research,

From fabricated research to paid authorships and citations, organized scientific fraud is on the rise, according to a new Northwestern University study.

By combining large-scale data analysis of scientific literature with case studies, the researchers led a deep investigation into scientific fraud. Although concerns around scientific misconduct typically focus on lone individuals, the Northwestern study instead uncovered sophisticated global networks of individuals and entities, which systematically work together to undermine the integrity of academic publishing.

The problem is so widespread that the publication of fraudulent science is outpacing the growth rate of legitimate scientific publications. The authors argue these findings should serve as a wake-up call to the scientific community, which needs to act before the public loses confidence in the scientific process.

The study will be published during the week of August 4 the Proceedings of the National Academy of Sciences.

“Science must police itself better in order to preserve its integrity,” said Northwestern’s Luís A. N. Amaral, the study’s senior author. “If we do not create awareness around this problem, worse and worse behavior will become normalized. At some point, it will be too late, and scientific literature will become completely poisoned. Some people worry that talking about this issue is attacking science. But I strongly believe we are defending science from bad actors. We need to be aware of the seriousness of this problem and take measures to address it.”

An expert in complex social systems, Amaral is the Erastus Otis Haven Professor and professor of engineering sciences and applied mathematics at Northwestern’s McCormick School of Engineering. Reese Richardson, a postdoctoral fellow in Amaral’s laboratory, is the paper’s first author.

Extensive analysis

When people think about scientific fraud, they might remember news reports of retracted papers, falsified data or plagiarism. These reports typically center around the isolated actions of one individual, who takes shortcuts to get ahead in an increasingly competitive industry. But Amaral and his team uncovered a widespread underground network operating within the shadows and outside of the public’s awareness.

“These networks are essentially criminal organizations, acting together to fake the process of science,” Amaral said. “Millions of dollars are involved in these processes.”

To conduct the study, the researchers analyzed extensive datasets of retracted publications, editorial records and instances of image duplication. Most of the data came from major aggregators of scientific literature, including Web of Science (WoS), Elsevier’s Scopus, National Library of Medicine’s PubMed/MEDLINE and OpenAlex, which includes data from Microsoft Academic Graph, Crossref, ORCID, Unpaywall and other institutional repositories.

Richardson and his colleagues also collected lists of de-indexed journals, which are scholarly journals that have been removed from databases for failing to meet certain quality or ethical standards. The researchers also included data on retracted articles from Retraction Watch, article comments from PubPeer and metadata — such as editor names, submission dates and acceptance dates — from articles published in specific journals.

Buying a reputation

After analyzing the data, the team uncovered coordinated efforts involving “paper mills,” brokers and infiltrated journals. Functioning much like factories, paper mills churn out large numbers of manuscripts, which they then sell to academics who want to quickly publish new work. These manuscripts are mostly low quality — featuring fabricated data, manipulated or even stolen images, plagiarized content and sometimes nonsensical or physically impossible claims.

“More and more scientists are being caught up in paper mills,” Amaral said. “Not only can they buy papers, but they can buy citations. Then, they can appear like well-reputed scientists when they have barely conducted their own research at all.”

“Paper mills operate by a variety of different models,” Richardson added. “So, we have only just been able to scratch the surface of how they operate. But they sell basically anything that can be used to launder a reputation. They often sell authorship slots for hundreds or even thousands of dollars. A person might pay more money for the first author position or less money for a fourth author position. People also can pay to get papers they have written automatically accepted in a journal through a sham peer-review process.”

To identify more articles originating from paper mills, the Amaral group launched a parallel project that automatically scans published materials science and engineering papers. The team specifically looked for authors who misidentified instruments they used in their research. A paper with those results was accepted by the journal PLOS ONE.

Brokers, hijacking and collusion

Amaral, Richardson and their collaborators found fraudulent networks use several key strategies: (1) Groups of researchers collude to publish papers across multiple journals. When their activities are discovered, the papers are subsequently retracted; (2) brokers serve as intermediaries to enable mass publication of fraudulent papers in compromised journals; (3) fraudulent activities are concentrated in specific, vulnerable subfields; and (4) organized entities evade quality-control measures, such as journal de-indexing.

“Brokers connect all the different people behind the scenes,” Amaral said. “You need to find someone to write the paper. You need to find people willing to pay to be the authors. You need to find a journal where you can get it all published. And you need editors in that journal who will accept that paper.”

Sometimes these organizations go around established journals altogether, searching instead for defunct journals to hijack. When a legitimate journal stops publishing, for example, bad actors can take over its name or website. These actors surreptitiously assume the journal’s identity, lending credibility to its fraudulent publications, despite the actual publication being defunct.

“This happened to the journal HIV Nursing,” Richardson said. “It was formerly the journal of a professional nursing organization in the U.K., then it stopped publishing, and its online domain lapsed. An organization bought the domain name and started publishing thousands of papers on subjects completely unrelated to nursing, all indexed in Scopus.”

Fighting for science

To combat this growing threat to legitimate scientific publishing, Amaral and Richardson emphasize the need for a multi-prong approach. This approach includes enhanced scrutiny of editorial processes, improved methods for detecting fabricated research, a greater understanding of the networks facilitating this misconduct and a radical restructuring of the system of incentives in science.

Amaral and Richardson also underscore the importance of addressing these issues before artificial intelligence (AI) infiltrates scientific literature more than it already has.

“If we’re not prepared to deal with the fraud that’s already occurring, then we’re certainly not prepared to deal with what generative AI can do to scientific literature,” Richardson said. “We have no clue what’s going to end up in the literature, what’s going to be regarded as scientific fact and what’s going to be used to train future AI models, which then will be used to write more papers.”

“This study is probably the most depressing project I’ve been involved with in my entire life,” Amaral said. “Since I was a kid, I was excited about science. It’s distressing to see others engage in fraud and in misleading others. But if you believe that science is useful and important for humanity, then you have to fight for it.”

Here’s a link to and a citation for the paper,

The entities enabling scientific fraud at scale are large, resilient, and growing rapidly by Reese A. K. Richardson, Spencer S. Hong, Jennifer A. Byrne, Thomas Stoeger, and Luís A. Nunes Amaral. Proceedings of the National Academy of Sciences August 4, 2025 122 (32) e2420092122 DOI: https://doi.org/10.1073/pnas.2420092122

This paper is open access.

And now—math fraud

A September 19, 2025 news item on ScienceDaily features an investigation into fraudulent math research, Note: A link has been removed,

An international team of authors led by Ilka Agricola, professor of mathematics at the University of Marburg, Germany, has investigated fraudulent practices in the publication of research results in mathematics on behalf of the German Mathematical Society (DMV) and the International Mathematical Union (IMU), documenting systematic fraud over many years. The results of the study were recently published on the preprint server arxiv.org and in the Notices of the American Mathematical Society (AMS) and have since caused a stir among mathematicians.

Sanjana Gajbhiye’s September ??, 2025 article for earth.com delves further into the topic, Note: Links have been removed,

Quality lost to quantity

The findings show how the definition of research quality has shifted. Instead of focusing on content, originality, and insight, institutions and individuals are increasingly evaluated by commercial metrics. These include the number of publications, total citations, and the so-called impact factor of journals.

Such measures, calculated by private companies with little transparency, have gained outsized influence. Providers promote their databases globally, and universities use them to enhance prestige and compete internationally.

This environment rewards quantity over quality, pushing academics to publish more, even when contributions are marginal or flawed.

Fraudulent companies have seized this opportunity. They sell services that manipulate rankings, offering ghostwritten articles, fake peer reviews, and even bundles of citations. For individuals, this can mean better career prospects.

For universities, it can result in higher rankings, increased funding, and greater appeal to international students. The collateral damage is a growing pool of unread publications that add nothing to scientific understanding.

Fake mathematics success

The report documents striking examples that reveal how metrics can produce absurd outcomes. In 2019, Clarivate Inc., the market leader for citation data, ranked a Taiwanese university as having the most world-class mathematicians. The catch was startling: mathematics was not even offered at the institution.

Mathematic trust under threat

“‘Fake science‘ is not only annoying, it is a danger to science and society,” said IMU Secretary General Professor Christoph Sorger.

“Because you don’t know what is valid and what is not. Targeted disinformation undermines trust in science and also makes it difficult for us mathematicians to decide which results can be used as a basis for further research.”

This erosion of trust strikes at the heart of mathematics. Proofs rely on certainty, yet when fraudulent or hollow work appears in respected outlets, that certainty weakens.

Fixing trust in mathematics publishing

The commission’s work does not end with exposing the problem. It also outlines possible solutions for a healthier publication system. These recommendations emphasize the need to strengthen peer review, encourage collaboration among journals, and recenter the evaluation of research on quality rather than raw numbers.

Metrics are deeply tied to funding and prestige, so the shift won’t be simple, but it could reshape the landscape for future generations.

A September 20, 2025 Castle Journal blog posting provides more information,

The “Culture of Numbers” and its Consequences

The study, led by Professor Ilka Agricola of the University of Marburg, argues that the root cause of the problem is a “culture of numbers” that prioritizes commercial metrics over scientific content. Universities and research institutions have become increasingly reliant on commercial databases like Clarivate’s Journal Citation Reports (JCR) to evaluate researchers. These metrics, which are not transparent and are not vetted by the scientific community, have become the main currency for career progression, grants, and prestige.

 * “Megajournals”: The study highlights the rise of “megajournals,” which publish anything as long as the authors pay a fee. These journals now publish more articles per year than all reputable mathematics journals combined. The report cites a shocking example where a commercial database ranked a university in Taiwan as having the most world-class researchers in mathematics, despite the fact that the university does not even offer mathematics as a subject.

 * Paper Mills and Citation Cartels: The investigation found evidence of “paper mills,” which sell fabricated papers to researchers, and “citation cartels,” where academics agree to cite each other’s work to artificially inflate their metrics. These services are offered anonymously online, with prices for articles and citations ranging from hundreds to thousands of dollars. The report describes these networks as “criminal organizations” that have invaded the “ecosystem” of scientific publishing.

Ilka Agricola gave an interview to Retraction Watch, from the undated article, Note: Links have been removed,

A pair of papers posted to the arXiv addresses the issue of fraudulent publishing in math, particularly metrics gaming, and offers a list of recommendations to help detect and deal with that problem and other fraudulent activities. (The former was also published in the October AMS Notices; the latter will appear in the November issue.) “Fraudulent publishing undermines trust in science and scientific results and therefore fuels antiscience movements,” mathematician Ilka Agricola, lead author of both papers, told Retraction Watch. 

A professor of mathematics at Marburg University in Germany, Agricola was president of the German Mathematical Society in 2021-2022 and is chair of the Committee on Publishing of the International Mathematical Union. The new articles are the products of a working group of the IMU and the International Council of Industrial and Applied Mathematics. 

Retraction Watch: As you note in the new papers, Clarivate announced in 2023 it had excluded the entire field of math from its list of “Highly Cited Researchers,” or HCRs. What’s going on?

Agricola: The publication culture in math differs a bit from, say, experimental and life sciences. On average, mathematicians publish fewer papers with fewer authors than scientists in other fields. So, with the same absolute number of papers and citations, one can become a “highly-cited researcher” in math, but not in other fields. Thus, gaming the system is easier. 

The list of HCRs for mathematics became so screwed that Clarivate couldn’t pretend anymore that it had any value. This being said, Clarivate announced that they would look into new measuring tools, but didn’t come up with any alternative ideas in the meantime, nor did they contact any representatives of the international mathematical community. 

Retraction Watch:  Few people talk about fraudulent publishing in math. Why is that?

Agricola: For a long time, mathematicians thought that as long as they keep away from predatory journals or paper mills, the problem does not affect them. This turned out to be wrong. 

Retraction Watch: If you look at the number of papers that tripped Clear Skies’ Papermill Alarm in 2022 (we included a histogram in this article we wrote for The Conversation [link and excerpts follow]), math is pretty far down the list. Are there a lot of fake papers in math?

Agricola: It is probably fair to say that the problem is not as severe as in other fields like cancer research, but the community is smaller and the number of fake papers is growing at alarming speed. Predatory and low-quality mega-journals are trying hard to lure respected scientists into their parallel universe of fake science, thus trying to give themselves the impression of respectability. Thus, one of our goals is to raise awareness for the issue in the mathematical community!

Retraction Watch: You and your coauthors are mathematicians, and yet you argue against focusing on numbers like journal impact factors and publication and citation counts. Is that what’s driving all of this bad behavior?

Agricola: “When a measure becomes a target, it ceases to be a good measure.” This quote is from the British economist Charles Goodhart, and it also applies to bibliometrics measures. Of course, gaming these metrics has always existed, but some of us liked to believe that they would be roughly OK, with some error bar due to some cheating. Now, we realize the error bar is larger than the number one wants to measure. Perhaps one advantage of mathematicians is that they are not easily impressed by numbers, and we have the means to understand and analyze them — this is our job. And so, the conclusion is very clear: The correlation between bibliometrics and research quality is so low that we should not use bibliometrics. And I urge all colleagues to say so openly!

Retraction Watch: So how do we judge research quality if we shouldn’t use publication metrics?

Agricola: Read the actual publications instead of relying on bibliometrics! Plus, in mathematics, we are lucky to have two extremely well curated databases for math papers and journals, zbMath Open and MathReviews. If a journal is not included there, it’s either very interdisciplinary or one should get suspicious.

Retraction Watch: Is it possible for individual researchers to jump off the bibiometrics bandwagon without jeopardizing their careers?

Agricola: We need to fight for a change in culture, that’s for sure, and the path will be rash and hard. To young researchers, we should give the warning that being involved in predatory publishing can also just as well put their scientific integrity at risk. Remember the people who had to resign because of data falsification? 

I am providing citations (of a sort) to both papers and links to all three sites where both papers can be found and PDFs for both papers: Everything is open access.

Fraudulent Publishing in the Mathematical Sciences by Ilka Agricola, Lynn Heller, Wil Schilders, Moritz Schubotz, Peter Taylor, Luis Vega.

arXiv: https://arxiv.org/abs/2509.07257

AMS (American Mathematical Society) Notices October 2025: https://www.ams.org/journals/notices/202509/noti3217/noti3217.html?adat=October%202025&trk=3217&pdfissue=202509&pdffile=rnoti-p1038.pdf&cat=none&type=.html

Ilke Agricola’s Research Gate website: https://www.researchgate.net/profile/Ilka-Agricola (scroll down to see the listed papers)

PDF: https://www.ams.org/journals/notices/202509/rnoti-p1038.pdf

How to Fight Fraudulent Publishing in the Mathematical Sciences: Joint Recommendations of the IMU [International Mathematical Union] and the ICIAM [International Council for Industrial and Applied Mathematics] by Ilka Agricola, Lynn Heller, Wil Schilders, Moritz Schubotz, Peter Taylor, Luis Vega.

arXiv: https://arxiv.org/abs/2509.09877

AMS (American Mathematical Society) Notices November 2025: https://www.ams.org/journals/notices/202510/noti3266/noti3266.html?adat=November%202025&trk=3266&pdfissue=202510&pdffile=rnoti-p1179.pdf&cat=none&type=.html

Ilke Agricola’s Research Gate website: https://www.researchgate.net/profile/Ilka-Agricola (scroll down to see the listed papers)

PDF: https://www.ams.org/journals/notices/202510/rnoti-p1179.pdf

Fraud slows down research

Mentioned in the Retraction Watch/Agricola interview, this January 29, 2025 article by Frederik Joelving (contributing editor, Retraction Watch), Cyril Labbé, (professor of computer science, Université Grenoble Alpes [UGA]), Guillaume Cabanac, (professor of computer Science, Institut de Recherche en Informatique de Toulouse) is chilling, Note: Links have been removed,

Over the past decade, furtive commercial entities around the world have industrialized the production, sale and dissemination of bogus scholarly research, undermining the literature that everyone from doctors to engineers rely on to make decisions about human lives.

It is exceedingly difficult to get a handle on exactly how big the problem is. Around 55,000 scholarly papers have been retracted to date, for a variety of reasons, but scientists and companies who screen the scientific literature for telltale signs of fraud estimate that there are many more fake papers circulating – possibly as many as several hundred thousand. This fake research can confound legitimate researchers who must wade through dense equations, evidence, images and methodologies only to find that they were made up.

Even when the bogus papers are spotted – usually by amateur sleuths on their own time – academic journals are often slow to retract the papers, allowing the articles to taint what many consider sacrosanct: the vast global library of scholarly work that introduces new ideas, reviews other research and discusses findings.

These fake papers are slowing down research that has helped millions of people with lifesaving medicine and therapies from cancer to COVID-19. Analysts’ data shows that fields related to cancer and medicine are particularly hard hit, while areas like philosophy and art are less affected. Some scientists have abandoned their life’s work because they cannot keep pace given the number of fake papers they must bat down.

The problem reflects a worldwide commodification of science. Universities, and their research funders, have long used regular publication in academic journals as requirements for promotions and job security, spawning the mantra “publish or perish.”

But now, fraudsters have infiltrated the academic publishing industry to prioritize profits over scholarship. Equipped with technological prowess, agility and vast networks of corrupt researchers, they are churning out papers on everything from obscure genes to artificial intelligence in medicine.

These papers are absorbed into the worldwide library of research faster than they can be weeded out. About 119,000 scholarly journal articles and conference papers are published globally every week, or more than 6 million a year. Publishers estimate that, at most journals, about 2% of the papers submitted – but not necessarily published – are likely fake, although this number can be much higher at some publications.

… there is a bustling online underground economy for all things scholarly publishing. Authorship, citations, even academic journal editors, are up for sale. This fraud is so prevalent that it has its own name: paper mills, a phrase that harks back to “term-paper mills,” where students cheat by getting someone else to write a class paper for them.

The impact on publishers is profound. In high-profile cases, fake articles can hurt a journal’s bottom line. Important scientific indexes – databases of academic publications that many researchers rely on to do their work – may delist journals that publish too many compromised papers. There is growing criticism that legitimate publishers could do more to track and blacklist journals and authors who regularly publish fake papers that are sometimes little more than artificial intelligence-generated phrases strung together.

To better understand the scope, ramifications and potential solutions of this metastasizing assault on science, we – a contributing editor at Retraction Watch, a website that reports on retractions of scientific papers and related topics, and two computer scientists at France’s Université Toulouse III–Paul Sabatier and Université Grenoble Alpes who specialize in detecting bogus publications – spent six months investigating paper mills.

This included, by some of us at different times, trawling websites and social media posts, interviewing publishers, editors, research-integrity experts, scientists, doctors, sociologists and scientific sleuths engaged in the Sisyphean task of cleaning up the literature. It also involved, by some of us, screening scientific articles looking for signs of fakery.

What emerged is a deep-rooted crisis that has many researchers and policymakers calling for a new way for universities and many governments to evaluate and reward academics and health professionals across the globe.

Just as highly biased websites dressed up to look like objective reporting are gnawing away at evidence-based journalism and threatening elections, fake science is grinding down the knowledge base on which modern society rests.

The January 29, 2025 article highlights a number of problems including these,

To expedite the publication of one another’s work, some corrupt scientists form peer review rings. Paper mills may even create fake peer reviewers impersonating real scientists to ensure their manuscripts make it through to publication. Others bribe editors or plant agents on journal editorial boards.

María de los Ángeles Oviedo-García, a professor of marketing at the University of Seville in Spain, spends her spare time hunting for suspect peer reviews from all areas of science, hundreds of which she has flagged on PubPeer. ……

“One of the demanding fights for me is to keep faith in science,” says Oviedo-García, who tells her students to look up papers on PubPeer before relying on them too heavily. Her research has been slowed down, she adds, because she now feels compelled to look for peer review reports for studies she uses in her work. Often there aren’t any, because “very few journals publish those review reports,” Oviedo-García says.

An ‘absolutely huge’ problem

It is unclear when paper mills began to operate at scale. The earliest article retracted due to suspected involvement of such agencies was published in 2004, according to the Retraction Watch Database, which contains details about tens of thousands of retractions. (The database is operated by The Center for Scientific Integrity, the parent nonprofit of Retraction Watch.) Nor is it clear exactly how many low-quality, plagiarized or made-up articles paper mills have spawned.

“The threat of paper mills to scientific publishing and integrity has no parallel over my 30-year scientific career …. In the field of human gene science alone, the number of potentially fraudulent articles could exceed 100,000 original papers,” she [Jennifer Byrne, an Australian scientist] wrote to lawmakers, adding, “This estimate may seem shocking but is likely to be conservative.”

In one area of genetics research – the study of noncoding RNA in different types of cancer – “We’re talking about more than 50% of papers published are from mills,” Byrne said. “It’s like swimming in garbage.”

… in the global south, the publish-or-perish edict runs up against underdeveloped research infrastructures and education systems, leaving scientists in a bind. For a Ph.D., the Cairo physician who requested anonymity conducted an entire clinical trial single-handedly – from purchasing study medication to randomizing patients, collecting and analyzing data and paying article-processing fees. In wealthier nations, entire teams work on such studies, with the tab easily running into the hundreds of thousands of dollars.

“Research is quite challenging here,” the physician said. That’s why scientists “try to manipulate and find easier ways so they get the job done.”

Institutions, too, have gamed the system with an eye to international rankings. In 2011, the journal Science described how prolific researchers in the United States and Europe were offered hefty payments for listing Saudi universities as secondary affiliations on papers. And in 2023, the magazine, in collaboration with Retraction Watch, uncovered a massive self-citation ploy by a top-ranked dental school in India that forced undergraduate students to publish papers referencing faculty work.

According to the January 29, 2025 article, there is a root cause, Note: Links have been removed,

… unsavory schemes can be traced back to the introduction of performance-based metrics in academia, a development driven by the New Public Management movement that swept across the Western world in the 1980s, according to Canadian sociologist of science Yves Gingras of the Université du Québec à Montréal. When universities and public institutions adopted corporate management, scientific papers became “accounting units” used to evaluate and reward scientific productivity rather than “knowledge units” advancing our insight into the world around us, Gingras wrote.

This transformation led many researchers to compete on numbers instead of content, which made publication metrics poor measures of academic prowess. As Gingras has shown, the controversial French microbiologist Didier Raoult, who now has more than a dozen retractions to his name, has an h-index – a measure combining publication and citation numbers – that is twice as high as that of Albert Einstein – “proof that the index is absurd,” Gingras said.

Worse, a sort of scientific inflation, or “scientometric bubble,” has ensued, with each new publication representing an increasingly small increment in knowledge. “We publish more and more superficial papers, we publish papers that have to be corrected, and we push people to do fraud,” said Gingras.

In terms of career prospects of individual academics, too, the average value of a publication has plummeted, triggering a rise in the number of hyperprolific authors. One of the most notorious cases is Spanish chemist Rafael Luque, who in 2023 reportedly published a study every 37 hours.

There is some hope according to the January 29, 2025 article, Note: Links have been removed,

Stern [Bodo Stern, a former editor of the journal Cell and chief of Strategic Initiatives at Howard Hughes Medical Institute] isn’t the first scientist to bemoan the excessive focus on bibliometrics. “We need less research, better research, and research done for the right reasons,” wrote the late statistician Douglas G. Altman in a much-cited editorial from 1994. “Abandoning using the number of publications as a measure of ability would be a start.”

Nearly two decades later, a group of some 150 scientists and 75 science organizations released the San Francisco Declaration on Research Assessment, or DORA, discouraging the use of the journal impact factor and other measures as proxies for quality. The 2013 declaration has since been signed by more than 25,000 individuals and organizations in 165 countries.

Despite the declaration, metrics remain in wide use today, and scientists say there is a new sense of urgency.

Stern and his colleagues have tried to make improvements at their institution. Researchers who wish to renew their seven-year contract have long been required to write a short paragraph describing the importance of their major results. Since the end of 2023, they also have been asked to remove journal names from their applications.

That way, “you can never do what all reviewers do – I’ve done it – look at the bibliography and in just one second decide, ‘Oh, this person has been productive because they have published many papers and they’re published in the right journals,’” says Stern. “What matters is, did it really make a difference?”

Shifting the focus away from convenient performance metrics seems possible not just for wealthy private institutions like Howard Hughes Medical Institute, but also for large government funders. In Australia, for example, the National Health and Medical Research Council in 2022 launched the “top 10 in 10” policy, aiming, in part, to “value research quality rather than quantity of publications.”

Rather than providing their entire bibliography, the agency, which assesses thousands of grant applications every year, asked researchers to list no more than 10 publications from the past decade and explain the contribution each had made to science. …

Gingras, the Canadian sociologist, advocates giving scientists the time they need to produce work that matters, rather than a gushing stream of publications. He is a signatory to the Slow Science Manifesto: “Once you get slow science, I can predict that the number of corrigenda, the number of retractions, will go down,” he says.

At one point, Gingras was involved in evaluating a research organization whose mission was to improve workplace security. An employee presented his work. “He had a sentence I will never forget,” Gingras recalls. The employee began by saying, “‘You know, I’m proud of one thing: My h-index is zero.’ And it was brilliant.” The scientist had developed a technology that prevented fatal falls among construction workers. “He said, ‘That’s useful, and that’s my job.’ I said, ‘Bravo!’”

Sometimes, there’s a science reporting problem

A September 3, 2025 Universiteit van Amsterdam press release (also on EurekAlert) highlights a problem with science reporting and over confidence,

Science journalists aren’t particularly concerned about so-called “predatory journals”, confident that they have the skills and intuition needed to avoid reporting on problematic research. For many, a journal’s reputation and name-recognition are decisive factors in assessing the quality of scientific research – but this could be exacerbating existing imbalances in science and journalism. This perspective emerges from a new study, led by Dr Alice Fleerackers of the University of Amsterdam (UvA), and published on 2 September [2025] in Journalism Practice.

Predatory journals prioritise profit over editorial and publication standards. They often charge researchers publication fees but offer little to no real quality control, such as peer review. As a result, some journals publish almost everything submitted. ‘Predatory journals are not a harmless side effect of the academic publishing industry,’ says Fleerackers. ‘They are becoming increasingly common, raising concerns about the integrity of scientific publishing. They not only undermine the reliability of science but also jeopardise science journalism, as journalists can unknowingly report on weak or even flawed research.’

In the new study, Fleerackers – along with colleagues from Simon Fraser University (Canada) and San Francisco State University (US) – investigated how science journalists view predatory journals and what strategies they employ to ensure the reliability of the journals they report on. The researchers present a qualitative analysis of interviews with 23 health, science, and environmental journalists in Europe and North America.

Problematic, but only in theory

Some of the journalists interviewed were familiar with the phenomenon of predatory journals and acknowledged that they are theoretically problematic. However, most weren’t concerned that they might be using them in their own work. They acknowledged that these journals might be a problem for colleagues, but not for them.

Well-known, therefore reliable

Journalists in the study were confident they wouldn’t fall for a predatory journal because of their strong intuition, which they said allowed them to immediately distinguish high-quality from problematic research. Besides their intuition, they also relied on strategies for verifying the reliability of research that they had developed through years of experience. These strategies often centred on trust proxies – like the journal’s prestige, impact factor, and selectivity – as well as whether the journal claimed to conduct peer review.

Proofreading also played a role for some journalists: if an article contained grammatical or spelling errors, it could be a sign of low-quality research. Open access journals were also considered less reliable by several journalists. ‘But by far the most commonly used benchmark for reliability was the journal’s reputation,’ Fleerackers explains. ‘Some journalists avoid all journals they’re not familiar with and report only on research published in top journals like Science and Nature.’

Distortion in science news

According to Fleerackers, journalists’ focus on the reputation and prestige of journals has major consequences for the diversity of research in the news media. ‘Research from newer, lesser-known journals, and from journals in the Global South, for example, remains hidden from the public. Most journalists in our study didn’t realise that their selection strategies could perpetuate the existing imbalance in science news. I hope that our study can raise awareness of this among journalists.’


Here’s a link to and a citation for the paper.

“I’d Like to Think I’d Be Able to Spot One”: How Journalists Navigate Predatory Journals by Alice Fleerackers, Laura L. Moorhead & Juan Pablo Alperin. Journalism Practice 1–19. DOI: https://doi.org/10.1080/17512786.2025.2551984 Published online: 02 Sep 2025

Final comments

This has been a good wake up call for me. Bad apples, yes, but criminal networks? I had no idea. I will probably write more about this in my 2025 year post. In the meantime, This is a good reminder to exercise caution.

When based on plastic materials, contemporary art can degrade quickly

There’s an intriguing April 1, 2016 article by Josh Fischman for Scientific American about a problem with artworks from the 20th century and later—plastic-based materials (Note: A link has been removed),

Conservators at museums and art galleries have a big worry. They believe there is a good chance the art they showcase now will not be fit to be seen in one hundred years, according to researchers in a project  called Nanorestart. Why? After 1940, artists began using plastic-based material that was a far cry from the oil-based paints used by classical painters. Plastic is also far more fragile, it turns out. Its chemical bonds readily break. And they cannot be restored using techniques historically relied upon by conservators.

So art conservation scientists have turned to nanotechnology for help.

Sadly, there isn’t any detail in Fischman’s article (*ETA June 17, 2016 article [for Fast Company] by Charlie Sorrel, which features some good pictures, a succinct summary of Fischman’s article and a literary reference [Kurt Vonnegut’s Bluebeard]I*) about how nanotechnology is playing or might play a role in this conservation effort. Further investigation into the two projects (NanoRestART and POPART) mentioned by Fischman didn’t provide much more detail about NanoRestART’s science aspect but POPART does provide some details.

NanoRestART

It’s probably too soon (this project isn’t even a year-old) to be getting much in the way of the nanoscience details but NanoRestART has big plans according to its website homepage,

The conservation of this diverse cultural heritage requires advanced solutions at the cutting edge of modern chemistry and material science in an entirely new scientific framework that will be developed within NANORESTART project.

The NANORESTART project will focus on the synthesis of novel poly-functional nanomaterials and on the development of highly innovative restoration techniques to address the conservation of a wide variety of materials mainly used by modern and contemporary artists.

In NANORESTART, enterprises and academic centers of excellence in the field of synthesis and characterization of nano- and advanced materials have joined forces with complementary conservation institutions and freelance restorers. This multidisciplinary approach will cover the development of different materials in response to real conservation needs, the testing of such materials, the assessment of their environmental impact, and their industrial scalability.

NanoRestART’s (NANOmaterials for the REStoration of works of ART) project page spells out their goals in the order in which they are being approached,

The ground-breaking nature of our research can be more easily outlined by focussing on specific issues. The main conservation challenges that will be addressed in the project are:

 

Conservation challenge 1Cleaning of contemporary painted and plastic surfaces (CC1)

Conservation challenge 2Stabilization of canvases and painted layers in contemporary art (CC2)

Conservation challenge 3Removal of unwanted modern materials (CC3)

Conservation challenge 4Enhanced protection of artworks in museums and outdoors (CC4)

The European Commission provides more information about the project on its CORDIS website’s NanoRestART webpage including the start and end dates for the project and the consortium members,

From 2015-06-01 to 2018-12-01, ongoing project

CHALMERS TEKNISKA HOEGSKOLA AB
Sweden
MIRABILE ANTONIO
France
NATIONALMUSEET
Denmark
CONSIGLIO NAZIONALE DELLE RICERCHE
Italy
UNIVERSITY COLLEGE CORK, NATIONAL UNIVERSITY OF IRELAND, CORK
Ireland
MBN NANOMATERIALIA SPA
Italy
KEMIJSKI INSTITUT
Slovenia
CHEVALIER AURELIA
France
UNIVERSIDADE FEDERAL DO RIO GRANDE DO SUL
Brazil
UNIVERSITA CA’ FOSCARI VENEZIA
Italy
AKZO NOBEL PULP AND PERFORMANCE CHEMICALS AB
Sweden
COMMISSARIAT A L ENERGIE ATOMIQUE ET AUX ENERGIES ALTERNATIVES
France
ARKEMA FRANCE SA
France
UNIVERSIDAD DE SANTIAGO DE COMPOSTELA
Spain
UNIVERSITY COLLEGE LONDON
United Kingdom
ZFB ZENTRUM FUR BUCHERHALTUNG GMBH
Germany
UNIVERSITAT DE BARCELONA
Spain
THE BOARD OF TRUSTEES OF THE TATE GALLERY
United Kingdom
ASSOCIAZIONE ITALIANA PER LA RICERCA INDUSTRIALE – AIRI
Italy
THE ART INSTITUTE OF CHICAGO
United States
MINISTERIO DE EDUCACION, CULTURA Y DEPORTE
Spain
STICHTING HET RIJKSMUSEUM
Netherlands
UNIVERSITEIT VAN AMSTERDAM
Netherlands
UNIVERSIDADE FEDERAL DO RIO DE JANEIRO
Brazil
ACCADEMIA DI BELLE ARTI DI BRERA
Italy

It was a bit surprising to see Brazil and the US as participants but The Art Institute of Chicago has done nanotechnology-enabled conservation in the past as per my March 24, 2014 posting about a Renoir painting. I’m not familiar with the Brazilian organization.

POPART

POPART (Preservation of Plastic Artefacts in museum collections) mentioned by Fischman was a European Commission project which ran from 2008 – 2012. Reports can be found on the CORDIS Popart webpage. The final report has some interesting bits (Note: I have added subheads in the [] square brackets),

To achieve a valid comparison of the various invasive and non-invasive techniques proposed for the identification and characterisation of plastics, a sample collection (SamCo) of plastics artefacts of about 100 standard and reference plastic objects was gathered. SamCo was made up of two kinds of reference materials: standards and objects. Each standard represents the reference material of a ‘pure’ plastic; while each object represents the reference of the same plastic as in the standards, but compounded with pigments, dyestuffs, fillers, anti oxidants, plasticizers etc.  Three partners ICN [Instituut Collectie Nederland], V&A [Victoria and Albert Museum] and Natmus [National Museet] collected different natural and synthetic plastics from the ICN reference collections of plastic objects, from flea markets, antique shops and from private collections and from their own collection to contribute to SamCo, the sample collection for identification by POPART partners. …

As a successive step, the collections of the following museums were surveyed:

-Victoria & Albert Museum (V&A), London, U.K.
-Stedelijk Museum, Amsterdam, The Netherlands
-Musée d’Art Moderne et d’Art Contemporaine (MAMAC) Nice, France
-Musée d’Art moderne, St. Etienne, France
-Musée Galliera, Paris, France

At the V&A approximately 200 objects were surveyed. Good or fair conservation conditions were found for about 85% of the objects, whereas the remaining 15% was in poor or even in unacceptable (3%) conditions. In particular, crazing and delamination of polyurethane faux leather and surface stickiness and darkening of plasticized PVC were observed. The situation at the Stedelijk Museum in Amsterdam was particularly favourable because a previous survey had been done in 1995 so that it was possible to make a comparison with the Popart survey in 2010. A total number of 40 objects, which comprised plastics early dating from the 1930’s until the newer plastics from the 1980’s, were considered and their actual conservation state compared with the 1995 records. Of the objects surveyed in 2010, it can be concluded that 21 remained in the same condition. 13 objects containing PA, PUR, PVC, PP or natural rubber changed due to chemical and physical degradation while works of art containing either PMMA or PS changed due to mechanical damages and incorrect artist’s technique (inappropriate adhesive) into a lesser condition. 6 works of art (containing either PA or PMMA or both) changed into a better condition due to restoration or replacements.  More than 230 objects have been examined in the 3 museums in France. A particular effort was devoted to the identification of the constituting plastics materials. Surveys have been undertaken without any sophisticated equipment, in order to work in museums everyday conditions. Plastics hidden by other materials or by paint layers were not or hardly accessible, it is why the final count of some plastics may be under estimated in the final results. Another outcome is that plastic identification has been made at a general level only, by trying to identify the polymer family each plastic belongs to. Lastly, evidence of chemical degradation processes that do not cause visible or perceptible damage have not been detected and could not be taken in account in the final results.

… The most damaged artefacts resulted constituted by cellulose acetate, cellulose nitrate and PVC.

[Polly (the doll)]

One of the main issues that is of interest for conservators and curators is to assess which kinds of plastics are most vulnerable to deterioration and to what extent they can deteriorate under the environmental conditions normally encountered in museums. Although one might expect that real time deterioration could be ascertained by a careful investigation of museum objects on display or in storage, real objects or artworks may not sampled due to ethical considerations. Therefore, reference objects were prepared by Natmus in the form of a doll (Polly) for simultaneous exposures in different environmental conditions. The doll comprised of 11 different plastics representative of types typically found in modern museum collections. The 16 identical dolls realized were exposed in different places, not only in normal exhibit conditions, but also in some selected extreme conditions to ascertain possible acceleration of the deterioration process. In most cases the environmental parameters were also measured. The dolls were periodically evaluated by visual inspection and in selected cases by instrumental analyses. 

In conclusion the experimental campaign carried out with Polly dolls can be viewed as a pilot study aimed at tackling the practical issues related to the monitoring of real three dimensional plastic artworks and the surrounding environment.

The overall exposure period (one year and half) was sufficient to observe initial changes in the more susceptible polymers, such as polyurethane ethers and esters, and polyamide, with detectable chromatic changes and surface effects. Conversely the other polymers were shown to be stable in the same conditions over this time period.

[Polly as an awareness raising tool]

Last but not least, the educational and communication benefits of an object like Polly facilitated the dissemination of the Popart Project to the public, and increased the awareness of issues associated with plastics in museum collections.

[Cleaning issues]

Mechanical cleaning has long been perceived as the least damaging technique to remove soiling from plastics. The results obtained from POPART suggest that the risks of introducing scratches or residues by mechanical cleaning are measurable. Some plastics were clearly more sensitive to mechanical damage than others. From the model plastics evaluated, HIPS was the most sensitive followed by HDPE, PVC, PMMA and CA. Scratches could not be measured on XPS due to its inhomogeneous surfaces. Plasticised PVC scratched easily, but appeared to repair itself because plasticiser migrated to surfaces and filled scratches.

Photo micrographs revealed that although all 22 cleaning materials evaluated in POPART scratched test plastics, some scratches were sufficiently shallow to be invisible to the naked eye. Duzzit and Scotch Brite sponges as well as all paper based products caused more scratching of surfaces than brushes and cloths. Some cleaning materials, notably Akapad yellow and white sponges, compressed air, latex and synthetic rubber sponges and goat hair brushes left residues on surfaces. These residues were only visible on glass-clear, transparent test plastics such as PMMA. HDPE and HIPS surfaces both had matte and roughened appearances after cleaning with dry-ice. XPS was completely destroyed by the treatment. No visible changes were present on PMMA and PVC.

Of the cleaning methods evaluated, only canned air, natural and synthetic feather duster left surfaces unchanged. Natural and synthetic feather duster, microfiber-, spectacle – and cotton cloths, cotton bud, sable hair brush and leather chamois showed good results when applied to clean model plastics.

Most mechanical cleaning materials induced static electricity after cleaning, causing immediate attraction of dust. It was also noticed that generally when adding an aqueous cleaning agent to a cleaning material, the area scratched was reduced. This implied that cleaning agents also functioned as lubricants. A similar effect was exhibited by white spirit and isopropanol.
Based on cleaning vectors, Judith Hofenk de Graaff detergent, distilled water and Dehypon LS45 were the least damaging cleaning agents for all model plastics evaluated. None of the aqueous cleaning agents caused visible changes when used in combination with the least damaging cleaning materials. Sable hair brush, synthetic feather duster and yellow Akapad sponge were unsuitable for applying aqueous cleaning agents. Polyvinyl acetate sponge swelled in contact with solvents and was only suitable for aqueous cleaning processes.

Based on cleaning vectors, white spirit was the least damaging solvent. Acetone and Surfynol 61 were the most damaging for all model plastics and cannot be recommended for cleaning plastics. Surfynol 61 dissolved polyvinyl acetate sponge and left a milky residue on surfaces, which was particularly apparent on clear PMMA surfaces. Surfynol 61 left residues on surfaces on evaporating and acetone evaporated too rapidly to lubricate cleaning materials thereby increasing scratching of surfaces.

Supercritical carbon dioxide induced discolouration and mechanical damage to the model plastics, particularly to XPS, CA and PMMA and should not be used for conservation cleaning of plastics.

Potential Impact:
Cultural heritage is recognised as an economical factor, the cost of decay of cultural heritage and the risk associated to some material in collection may be high. It is generally estimated that plastics, developed at great numbers since the 20th century’s interbellum, will not survive that long. This means that fewer generations will have access to lasting plastic art for study, contemplation and enjoyment. On the other hand will it normally be easier to reveal a contemporary object’s technological secrets because of better documentation and easier access to artists’ working methods, ideas and intentions. A first more or less world encompassing recognition of the problems involved with museum objects made wholly or in part of plastics was through the conference ‘Saving the twentieth century” held in Ottawa, Canada in 1991. This was followed later by ‘Modern Art, who cares’ in Amsterdam, The Netherlands in 1997, ‘Mortality Immortality? The Legacy of Modern Art’ in Los Angeles, USA in 1998 and, for example much more recent, ‘Plastics –Looking at the future and learning from the Past’ in London, UK in 2007. A growing professional interest in the care of plastics was clearly reflected in the creation of an ICOM-CC working group dedicated to modern materials in 1996, its name change to Modern Materials and Contemporary Art in 2002, and its growing membership from 60 at inception to over 200 at the 16th triennial conference in Lisbon, Portugal in 2011 and tentatively to over 300 as one of the aims put forward in the 2011-2014 programme of that ICOM-CC working group. …

[Intellectual property]

Another element pertaining to conservation of modern art is the copyright of artists that extends at least 50 years beyond their death. Both, damage, value and copyright may influence the way by which damage is measured through scientific analysis, more specifically through the application of invasive or non invasive techniques. Any selection of those will not only have an influence on the extent of observable damage, but also on the detail of information gathered and necessary to explain damage and to suggest conservation measures.

[How much is deteriorating?]

… it is obvious from surveys carried out in several museums in France, the UK and The Netherlands that from 15 to 35 % of what I would then call an average plastic material based collection is in a poor to unacceptable condition. However, some 75 % would require cleaning,

I hope to find out more about how nanotechnology is expected to be implemented in the conservation and preservation of plastic-based art. The NanoRestART project started in June 2015 and hopefully more information will be disseminated in the next year or so.

While it’s not directly related, there was some work with conservation of daguerreotypes (19th century photographic technique) and nanotechnology mentioned in my Nov. 17, 2015 posting which was a followup to my Jan. 10, 2015 posting about the project and the crisis precipitating it.

*ETA June 30, 2016: Here’s clip from a BBC programme, Science in Action broadcast on June 30, 2016 featuring a chat with some of the scientists involved in the NanoRestArt project (Note: This excerpt is from a longer programme and seemingly starts in the middle of a conversation,)