Blog icon

By  Jon Whittle Stefan Harrer 20 March 2025 4 min read

In February this year, Google announced it was launching “a new AI system for scientists”. It said this system was a collaborative tool designed to help scientists “in creating novel hypotheses and research plans”.

It’s too early to tell just how useful this particular tool will be to scientists. But what is clear is that artificial intelligence (AI) more generally is already transforming science.

Last year for example, computer scientists won the Nobel Prize for Chemistry for developing an AI model to predict the shape of every protein known to mankind. Chair of the Nobel Committee, Heiner Linke, described the AI system as the achievement of a “50-year-old dream” that solved a notoriously difficult problem eluding scientists since the 1970s.

But while AI is allowing scientists to make technological breakthroughs that are otherwise decades away or out of reach entirely, there’s also a darker side to the use of AI in science: scientific misconduct is on the rise.

AI makes it easy to fabricate research

Academic papers can be retracted if their data or findings are found to no longer valid. This can happen because of data fabrication, plagiarism or human error.

Paper retractions are increasing exponentially, passing 10,000 in 2023. These retracted papers were cited over 35,000 times.

One study found 8 per cent of Dutch scientists admitted to serious research fraud, double the rate previously reported. Biomedical paper retractions have quadrupled in the past 20 years, the majority due to misconduct.

AI has the potential to make this problem even worse.

For example, the availability and increasing capability of generative AI programs such as ChatGPT makes it easy to fabricate research.

This was clearly demonstrated by two researchers who used AI to generate 288 complete fake academic finance papers predicting stock returns.

While this was an experiment to show what’s possible, it’s not hard to imagine how the technology could be used to generate fictitious clinical trial data, modify gene editing experimental data to conceal adverse results or for other malicious purposes.

OpenAI, the makers of ChatGPT, just unveiled a new tool that can produce detailed reports. It's called Deep Research. The company says it uses reasoning to synthesize large amounts of complex data. They claim it can finish in-depth research in a fraction of the time it would take even the most knowledgeable human. That is a big sell. Let's hear from Jason, the Editor in Chief of ZDNet, to explain this.

Jason: So many of us have been impressed by what ChatGPT can do. Help us understand why they're introducing this new tool to the market now and what everything I just said really means practically.

Jason: Yeah, so this is the next stage of where generative AI is going. Generative AI includes tools like ChatGPT, Google Gemini, and some others out there. What this is, is an AI agent. Remember that term because you'll hear a lot of it in 2025. An AI agent takes tasks and goes and does something for you independently. So it's like an assistant, a virtual assistant. What we're seeing with AI agents like Deep Research is that they let you provide a research topic, and they go and do the research for you. This is typically research that might take you an hour or two. Maybe you ask for the history of the impact of tariffs on the U.S. economy and the world economy over the past 50 years. That might take you a while to dig up. What this will do is go and do it for you, typically in about 5 to 10 minutes. It will spit out a report that includes charts, images, and various data in ways that would be very difficult to do in the same amount of time.

Now, it has a few drawbacks. It has some challenges. It could hallucinate, as we've heard these models can do. However, OpenAI models are getting more and more accurate and less prone to hallucination. This is a powerful tool and one of the first examples of AI agents that almost anybody can look at and say, "I can see where that would be interesting for me."

Interviewer: I'm thinking about planning vacations, comparing and contrasting options in life. It's important to know it makes mistakes; it's not perfect. It also comes on the heels of Operator. What is that, and how might it mess up?

Jason: Yeah, so Operator is also an AI agent just released a couple of weeks ago. It's another one of these first stabs at an AI agent. This one is aimed at carrying out tasks for you more directly. You might say, "Order me a pair of AirPods Pro on Amazon," and it'll just go and do it, order it. Now, it's working off your browser, where you're logged in, your preferences, and your account. It could also order you an Uber, order food to be delivered, or take a recipe and order items to have them delivered. It is very early stages, but it is this AI agent that's going beyond just giving you information or an answer to a question and actually carrying out tasks for you. Again, this is the wave of AI that we're going to see in 2025.

Interviewer: Fascinating to get an early glimpse of what will be possible.

Fake references and fabricated data

There are already many reported cases of AI-generated papers passing peer-review and reaching publication – only to be retracted later on the grounds of undisclosed use of AI, some including serious flaws such as fake references and purposely fabricated data.

Some researchers are also using AI to review their peers’ work. Peer review of scientific papers is one of the fundamentals of scientific integrity. But it’s also incredibly time-consuming, with some scientists devoting hundreds of hours a year of unpaid labour. A Stanford-led study found that up to 17 per cent of peer reviews for top AI conferences were written at least in part by AI.

In the extreme case, AI may end up writing research papers, which are then reviewed by another AI.

This risk is worsening the already problematic trend of an exponential increase in scientific publishing, while the average amount of genuinely new and interesting material in each paper has been declining.

AI can also lead to unintentional fabrication of scientific results.

A well-known problem of generative AI systems is when they make up an answer rather than saying they don’t know. This is known as “hallucination”.

We don’t know the extent to which AI hallucinations end up as errors in scientific papers. But a recent study on computer programming found that 52 per cent of AI-generated answers to coding questions contained errors, and human oversight failed to correct them 39 per cent of the time.

AI is allowing scientists to make technological breakthroughs that are otherwise decades away or out of reach entirely. But it also comes with risks.

Maximising the benefits, minimising the risks

Despite these worrying developments, we shouldn’t get carried away and discourage or even chastise the use of AI by scientists.

AI offers significant benefits to science. Researchers have used specialised AI models to solve scientific problems for many years. And generative AI models such as ChatGPT offer the promise of general-purpose AI scientific assistants that can carry out a range of tasks, working collaboratively with the scientist.

These AI models can be powerful lab assistants. For example, researchers at CSIRO are already developing AI lab robots that scientists can speak with and instruct like a human assistant to automate repetitive tasks.

A disruptive new technology will always have benefits and drawbacks. The challenge of the science community is to put appropriate policies and guardrails in place to ensure we maximise the benefits and minimise the risks.

AI’s potential to change the world of science and to help science make the world a better place is already proven. We now have a choice.

Do we embrace AI by advocating for and developing an AI code of conduct that enforces ethical and responsible use of AI in science? Or do we take a backseat and let a relatively small number of rogue actors discredit our fields and make us miss the opportunity?

This article is republished from The Conversation under a Creative Commons license. Read the original article.

The Conversation