Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

Exam submissions by AI found to earn higher grades than real-life students

Experts say universities should focus on working out how to embrace the ‘new normal’ of AI in order to enhance education.

Nilima Marshall
Thursday 27 June 2024 04:34 EDT
AI-generated exam submissions evade detection in a real-world test (Chris Ison/PA)
AI-generated exam submissions evade detection in a real-world test (Chris Ison/PA) (PA Wire)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

Exam submissions generated by artificial intelligence (AI) can not only evade detection but also earn higher grades than those submitted by university students, a real-world test has shown.

The findings come as concerns mount about students submitting AI-generated work as their own, with questions being raised about the academic integrity of universities and other higher education institutions.

The data in our study shows it is very difficult to detect AI-generated answers

Peter Scarfe, Reading's School of Psychology and Clinical Language Sciences

It also shows even experienced markers could struggle to spot answers generated by AI, the University of Reading academics said.

Peter Scarfe, an associate professor at Reading’s School of Psychology and Clinical Language Sciences said the findings should serve as a “wake-up call” for educational institutions as AI tools such as ChatGPT become more advanced and widespread.

He said: “The data in our study shows it is very difficult to detect AI-generated answers.

“There has been quite a lot of talk about the use of so-called AI detectors, which are also another form of AI but (the scope here) is limited.”

For the study, published in the journal Plos One, Prof Scarfe and his team generated answers to exam questions using GPT-4 and submitted these on behalf of 33 fake students.

Exam markers at Reading’s School of Psychology and Clinical Language Sciences were unaware of the study.

Answers submitted for many undergraduate psychology modules went undetected in 94% of cases and, on average, got higher grades than real student submissions, Prof Scarfe said.

He said AI did particularly well in the first and second years of study but struggled more in the final year of study module.

Last year Russell Group universities, which include Oxford, Cambridge, Imperial College London and other top universities, pledged to allow ethical use of AI in teaching and assessments, with many others following suit.

We won't necessarily go back fully to hand-written exams, but global education sector will need to evolve in the face of AI

Prof Peter Scarfe

But Prof Scarfe said the education sector will need to constantly adapt and update guidance as generative AI continues to evolve and become more sophisticated.

He said universities should focus on working out how to embrace the “new normal” of AI in order to enhance education.

Prof Scarfe added that reverting back to in-person sit-down exam assessments, would “be a step backwards in many ways”.

He said: “Many institutions have moved away from traditional exams to make assessment more inclusive.

“Our research shows it is of international importance to understand how AI will affect the integrity of educational assessments.

“We won’t necessarily go back fully to hand-written exams, but the global education sector will need to evolve in the face of AI.”

We need to double down on our commitment to academic and research integrity

Co-author Professor Etienne Roesch

Study co-author Professor Etienne Roesch, of Reading’s School of Psychology and Clinical Language Sciences, added: “As a sector, we need to agree how we expect students to use and acknowledge the role of AI in their work.

“The same is true of the wider use of AI in other areas of life to prevent a crisis of trust across society.

“Our study highlights the responsibility we have as producers and consumers of information.

“We need to double down on our commitment to academic and research integrity.”

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in