Study shows students using AI in exams earn higher grades – and can get away with it
The study has been called a ‘wake-up call’ for universities
Your support helps us to tell the story
From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.
At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.
The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.
Your support makes all the difference.Exams answered using AI can not only evade detection but also earn higher grades than those submitted by students, a real-world test has shown.
The study also showed even experienced markers could struggle to spot answers generated by AI.
The findings come as concerns mount about students submitting AI-generated work as their own, with questions being raised about the academic integrity of universities and other higher education institutions.
Peter Scarfe, an associate professor at Reading’s School of Psychology and Clinical Language Sciences said the findings should serve as a “wake-up call” for educational institutions as AI tools such as ChatGPT become more advanced and widespread.
He said: “The data in our study shows it is very difficult to detect AI-generated answers.
“There has been quite a lot of talk about the use of so-called AI detectors, which are also another form of AI but (the scope here) is limited.”
For the study, published in the journal Plos One, Prof Scarfe and his team generated answers to exam questions using GPT-4 and submitted these on behalf of 33 fake students.
Exam markers at Reading’s School of Psychology and Clinical Language Sciences were unaware of the study.
Answers submitted for many undergraduate psychology modules went undetected in 94% of cases and, on average, got higher grades than real student submissions, Prof Scarfe said.
He said AI did particularly well in the first and second years of study but struggled more in the final year of study module.
Last year Russell Group universities, which include Oxford, Cambridge, Imperial College London and other top universities, pledged to allow ethical use of AI in teaching and assessments, with many others following suit.
But Prof Scarfe said the education sector will need to constantly adapt and update guidance as generative AI continues to evolve and become more sophisticated.
He said universities should focus on working out how to embrace the “new normal” of AI in order to enhance education.
Prof Scarfe added that reverting back to in-person sit-down exam assessments, would “be a step backwards in many ways”.
He said: “Many institutions have moved away from traditional exams to make assessment more inclusive.
“Our research shows it is of international importance to understand how AI will affect the integrity of educational assessments.
“We won’t necessarily go back fully to hand-written exams, but the global education sector will need to evolve in the face of AI.”
Study co-author Professor Etienne Roesch, of Reading’s School of Psychology and Clinical Language Sciences, added: “As a sector, we need to agree how we expect students to use and acknowledge the role of AI in their work.
“The same is true of the wider use of AI in other areas of life to prevent a crisis of trust across society.
“Our study highlights the responsibility we have as producers and consumers of information.
“We need to double down on our commitment to academic and research integrity.”
Join our commenting forum
Join thought-provoking conversations, follow other Independent readers and see their replies
Comments