AI Detected in 14% of Biomedical Abstracts in 2024
A study reveals that one in seven biomedical research abstracts published last year likely used AI assistance, raising concerns about transparency and impact on scientific integrity.
Around one in seven biomedical-research abstracts published last year was probably written with the help of artificial intelligence (AI), according to a massive analysis of the scholarly literature. More than 200,000 abstracts out of 1.5 million indexed in PubMed in 2024 contained words commonly suggested by large language models (LLMs).
The study was posted online as a preprint in June 2024, estimating that one in nine abstracts were written with AI assistance in the first half of last year. The updated analysis is published in Science Advances today.
“The total number of LLM-edited papers has continued to rise inexorably,” says Andrew Gray, a bibliometrics support officer at University College London. He thinks researchers haven’t fully grasped the scale on which these tools are being used to produce scholarly work. “Hopefully, this paper will help drive some attention to the problem,” he adds.
Many groups are trying to assess the impact of LLMs on scholarly output, but it is tricky because most users don’t disclose these practices. Efforts typically train models to identify differences between human-generated and LLM-generated text, and then apply those insights to assess the literature. However, it is not clear how such models distinguish between the two types of text, and training data sets don’t always represent up-to-date trends in LLM-generated writing.
Dmitry Kobak, a data scientist at the University of Tübingen in Germany, and his colleagues took a more open-ended approach. They searched abstracts for ‘excess words’ that started to appear more often than expected after November 2022, when ChatGPT became widely available. The team was inspired by work estimating ‘excess deaths’ during the COVID-19 pandemic.
The researchers found that 454 words appeared much more often in 2024 than in any other year since 2010. These were mostly ‘style’ words unrelated to the content of the research, and tended to be verbs and adjectives. Some were common, such as ‘findings’, ‘crucial’ and ‘potential’, whereas others were more unusual, including ‘delves’ and ‘showcasing’. Excess words that emerged in the second half of 2024 include ‘heighten’ and ‘hinder’, as well as superlatives such as ‘unparalleled’ and ‘invaluable’, says Kobak.
Shifts in the scientific lexicon happen over time, including dramatic changes that accompany major events, such as the COVID-19 pandemic that began in 2020. There were 190 excess words in 2021, and they tended to be nouns related to the content of the research, such as ‘mask’. But the lexical shift that has occurred since LLMs became popular has been even more pronounced, and mainly stylistic.
Geographic differences in the use of AI in research are also emerging. Some regions are more likely to adopt these tools, while others are more cautious. This can lead to variations in the quality and style of scientific writing, which may affect the global scientific community's ability to communicate effectively.
The implications of this trend are significant. If researchers rely heavily on AI to write abstracts, it could lead to a homogenization of scientific writing, making it harder to distinguish between different studies and potentially reducing the originality and creativity in scientific communication. It also raises ethical concerns about transparency and the integrity of the scientific process.
As AI continues to evolve and become more integrated into the research process, it is crucial for the scientific community to establish clear guidelines and standards for the use of these tools. This will help ensure that the benefits of AI are realized while minimizing the potential risks and maintaining the trust and integrity of scientific research.
Frequently Asked Questions
What percentage of biomedical abstracts used AI in 2024?
Around 14% of biomedical abstracts published in 2024 were likely written with the help of AI.
How do researchers detect AI-generated text in scientific papers?
Researchers search for ‘excess words’ that started appearing more frequently after the widespread availability of AI tools like ChatGPT.
What are some common excess words found in AI-generated abstracts?
Common excess words include ‘findings’, ‘crucial’, ‘potential’, ‘delves’, and ‘showcasing’.
What are the ethical concerns with using AI in scientific writing?
Ethical concerns include transparency, the integrity of the scientific process, and the potential homogenization of scientific writing.
How can the scientific community address the issues with AI in research?
The scientific community can establish clear guidelines and standards for the use of AI tools to ensure the benefits are realized while minimizing risks.