More scientific papers being written with help of ChatGPT—especially in computer science

Since its release in November of 2022, the use of ChatGPT and other large language models (LLMs) has proliferated throughout many disciplines, providing writing assistance for everything from speeches to contracts. So, it may not be surprising that some scientists might utilize ChatGPT to quicken the pace at which they publish their research.


There is little known about how the adoption of AI-generated content might affect the diversity, quality and reliability of research papers. And because these technologies are still new and constantly evolving, there is not yet a sure-fire way to detect the use of LLMs, and many institutions are still developing policies to curb their use.

To get a better grasp of how ChatGPT has been used in scientific writing over the last few years, a group of researchers recently conducted a study analyzing 1,121,912 scientific papers and preprints from arXivbioRxiv and Nature portfolio journals. The study, published in Nature Human Behaviour, used a new population-level framework based on word frequency shifts to estimate the increase of LLM-modified content between January 2020 and September 2024.

The study found that abstracts and introductions were most commonly affected, while methods and experiment sections showed less AI use, likely due to the summarization abilities of LLMs. A steady increase in the likely use of ChatGPT was observed across multiple topics of study, with the most dramatic being computer science—a notably AI-adjacent discipline.

The analysis showed likely LLM use in 22.5% of computer science abstracts and 19.5% of computer science introductions by September 2024. In November 2022, these numbers were only around 2.4% and similar across all article types at the time. LLM use was also relatively high in electrical engineering and systems science by 2024, at 18.0% for abstracts and 18.4% for introductions.

LLM usage was found to be much lower in areas like mathematics, with LLM usage at 7.7% for abstracts and 4.1% for introductions. The Nature portfolio of journals also showed a lower increase in AI use, with 8.9% for abstracts and 9.4% for introductions.

In addition to the field of study, the analysis was further stratified by author preprint frequency, paper length, and geographical region, in which the researchers found LLM modification to be more common in a few different cases. Authors who posted preprints more frequently were associated with more LLM usage in their papers, possibly due to increased pressure to put out more papers at a faster pace. Shorter papers—those less than 5,000 words—were also associated with more help from LLMs, along with those in more competitive research areas, like .

Detecting AI-generated text in non-English speaking geographical regions is trickier, and some bias has been pointed out in previous methods of AI detection against non-native English writers in . This study did show higher LLM usage in papers from China and Continental Europe, compared to North America and the UK, but much of this is likely for English-language assistance.

As the AI landscape rapidly evolves in the coming years, it has the potential to change how science is written and communicated, which in turn raises questions about transparency, originality and the future of scientific publishing.

The study authors point out many of the questions that should be answered as science continues to incorporate these technologies: "Our observations of the rise of generated or modified papers open many questions for future research. How do such papers compare in terms of accuracy, creativity or diversity? How do readers react to LLM-generated abstracts and introductions? How do citation patterns of LLM-generated papers compare with other papers in similar fields? How might the dominance of a limited number of for-profit organizations in the LLM industry affect the independence of scientific output?

"We hope our results and methodology inspire further studies of widespread LLM modified text and conversations about how to promote transparent, diverse and high-quality scientific publishing."

Written for you by our author Krystal Kasal, edited by Gaby Clark, and fact-checked and reviewed by Robert Eganthis article is the result of careful human work. 

More information: Weixin Liang et al, Quantifying large language model usage in scientific papers, Nature Human Behaviour (2025). DOI: 10.1038/s41562-025-02273-8

Journal information: Nature Human Behaviour , Nature , arXiv , bioRxiv

Comments

Pages

Archive

Show more

Popular posts from this blog

Top 10 Stocks Poised to Outperform in August 2025: AI, Clean Energy, and Tech Leaders to Watch

Top 10 ETF Picks for August 2025: Best Funds to Buy Now for Growth, Safety, and Diversification

Top August 2025 ETFs: In-Depth Performance Reviews

Top XRP ETFs to Watch in 2025: Dominating Holdings, Low Fees, and Performance Insights

7 Best AI Automated Trading Software for 2025

Top 10 Pharmaceutical Companies by Revenue in 2025 (July Edition)

Best Bitcoin ETFs of 2025: Should You Buy Today?

Top 10 Credit Cards for Hospital Bills in Singapore 2025

Did Steve Jobs Refuse Treatment for Pancreatic Cancer?

Top 10 Pharma Companies by Market Cap in 2025