Exam submissions by AI found to earn higher grades than real-life students
Experts say universities should focus on working out how to embrace the ‘new normal’ of AI in order to enhance education.
Your support helps us to tell the story
From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.
At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.
The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.
Your support makes all the difference.Exam submissions generated by artificial intelligence (AI) can not only evade detection but also earn higher grades than those submitted by university students, a real-world test has shown.
The findings come as concerns mount about students submitting AI-generated work as their own, with questions being raised about the academic integrity of universities and other higher education institutions.
It also shows even experienced markers could struggle to spot answers generated by AI, the University of Reading academics said.
Peter Scarfe, an associate professor at Reading’s School of Psychology and Clinical Language Sciences said the findings should serve as a “wake-up call” for educational institutions as AI tools such as ChatGPT become more advanced and widespread.
He said: “The data in our study shows it is very difficult to detect AI-generated answers.
“There has been quite a lot of talk about the use of so-called AI detectors, which are also another form of AI but (the scope here) is limited.”
For the study, published in the journal Plos One, Prof Scarfe and his team generated answers to exam questions using GPT-4 and submitted these on behalf of 33 fake students.
Exam markers at Reading’s School of Psychology and Clinical Language Sciences were unaware of the study.
Answers submitted for many undergraduate psychology modules went undetected in 94% of cases and, on average, got higher grades than real student submissions, Prof Scarfe said.
He said AI did particularly well in the first and second years of study but struggled more in the final year of study module.
Last year Russell Group universities, which include Oxford, Cambridge, Imperial College London and other top universities, pledged to allow ethical use of AI in teaching and assessments, with many others following suit.
But Prof Scarfe said the education sector will need to constantly adapt and update guidance as generative AI continues to evolve and become more sophisticated.
He said universities should focus on working out how to embrace the “new normal” of AI in order to enhance education.
Prof Scarfe added that reverting back to in-person sit-down exam assessments, would “be a step backwards in many ways”.
He said: “Many institutions have moved away from traditional exams to make assessment more inclusive.
“Our research shows it is of international importance to understand how AI will affect the integrity of educational assessments.
“We won’t necessarily go back fully to hand-written exams, but the global education sector will need to evolve in the face of AI.”
Study co-author Professor Etienne Roesch, of Reading’s School of Psychology and Clinical Language Sciences, added: “As a sector, we need to agree how we expect students to use and acknowledge the role of AI in their work.
“The same is true of the wider use of AI in other areas of life to prevent a crisis of trust across society.
“Our study highlights the responsibility we have as producers and consumers of information.
“We need to double down on our commitment to academic and research integrity.”
Subscribe to Independent Premium to bookmark this article
Want to bookmark your favourite articles and stories to read or reference later? Start your Independent Premium subscription today.