Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

‘Most exciting’ applications for AI lie in healthcare, says committee chairman

A House of Commons committee has published a report on the potential risks posed by advancements in the technology.

Storm Newton
Thursday 31 August 2023 03:10 EDT
Researchers are looking into how AI can be used to predict the damage long-term conditions such as diabetes can cause in a patient’s body (Joe Giddens/PA)
Researchers are looking into how AI can be used to predict the damage long-term conditions such as diabetes can cause in a patient’s body (Joe Giddens/PA) (PA Wire)

Your support helps us to tell the story

From reproductive rights to climate change to Big Tech, The Independent is on the ground when the story is developing. Whether it's investigating the financials of Elon Musk's pro-Trump PAC or producing our latest documentary, 'The A Word', which shines a light on the American women fighting for reproductive rights, we know how important it is to parse out the facts from the messaging.

At such a critical moment in US history, we need reporters on the ground. Your donation allows us to keep sending journalists to speak to both sides of the story.

The Independent is trusted by Americans across the entire political spectrum. And unlike many other quality news outlets, we choose not to lock Americans out of our reporting and analysis with paywalls. We believe quality journalism should be available to everyone, paid for by those who can afford it.

Your support makes all the difference.

Healthcare is where the “most exciting” opportunities for artificial intelligence (AI) lie, an MP has said, but is also an area where the technology’s major risks are illustrated.

Greg Clark, chairman of the Science, Innovation and Technology Committee (SITC), said the wider adoption of AI in healthcare would have a “positive impact”, but urged policy makers to “consider the risks to safety”.

He said: “If we’re to gain all the advantages, we have to anticipate the risks and put in place measures to safeguard against that.”

In the NHS, AI is currently being used to read X-rays – such as mammograms – helping doctors to speed up decision making and giving them more time to spend with patients, as well as helping medics to diagnose strokes faster.

Elsewhere, researchers are looking into how AI can be used to predict the damage long-term conditions such as diabetes can cause in a patient’s body.

The technology can also be utilised in drug discovery, giving researchers access to vast amounts of data to speed the process up.

Mr Clark told the PA news agency: “One of the things we were struck by is how medicine is becoming increasingly personalised.

“Quite a lot of drugs have failed clinical trials on a broad-spectrum basis, but now you can identify more closely who they are more suitable for. They can then be deployed straight away to help save lives.”

If you can consider that AI can be given the ability to develop new medicine, it also has the ability to develop toxins

Committee chairman Greg Clark

It comes as the SITC published an interim report on its inquiry into AI governance, which opened in October.

The document outlines 12 risks posed by the technology and provides a guide on how policy should be shaped to overcome them.

Mr Clark added: “If you can consider that AI can be given the ability to develop new medicine, it also has the ability to develop toxins.”

Among the risks highlighted by the SITC was AI’s potential to perpetuate “unacceptable” societal biases, as well as its potential to share personal information or generate material that “deliberately misrepresents someone’s behaviour, opinions or character”.

Debate is also ongoing about how AI would access large datasets, which are held by few organisations, or who would be liable if AI used by a third party were to cause harm.

“A lot of recommendations depend on the data on which AI is trained,” Mr Clark said. “If you’re conducting medical research on a particular sample or ethnic minority, then the data on which AI is trained may mean the recommendations are inaccurate.”

You need to drive the policy thinking at the same time as the tech development

Committee chairman Greg Clark

He added that no one risk included in the document was a priority and that they “all have to be addressed together”.

“It’s not the case if you just deal with one, or half of them, that everyone can relax. For example, on bias, that’s relevant in medical settings and others.”

The Government has backed research on how AI could benefit clinicians with almost £150 million in funding, while NHS England chief executive Amanda Pritchard promised in June that more uses of the technology were “on the horizon”.

In July, the National Institute for Health and Care Research (NIHR) released a collection of 10 studies on AI to demonstrate where it could be used in the future.

The pieces of research were published between 2020 and 2023, and explored how AI could be used to detect heart disease via “smart stethoscopes”, or predict the progression of diseases such as the bowel condition ulcerative colitis.

But the SITC has called for more urgency from the Government when it comes to developing policy on AI governance.

If the public lose confidence and are spooked by AI, then there will be a reaction standing in the way of some of the benefits

Committee chairman Greg Clark

“You need to drive the policy thinking at the same time as the tech development,” Mr Clark said.

“If the public lose confidence and are spooked by AI, then there will be a reaction standing in the way of some of the benefits.”

He added that the UK had done this in the past with the Warnock Report on fertility treatment, which was published in the 1980s at a time when in vitro fertilisation (IVF) was a controversial topic.

Mr Clark said “approaching it in an early and rigorous way” through the report would mean the subject “doesn’t have the contention and division it has in some other countries”.

He added: “You can instil public confidence in areas that are controversial.”

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in