ΑΙhub.org
 

AI is transforming medicine – but it can only work with proper sharing of data


by
23 January 2020



share this:
Anonymised data is crucial for AI to work. alphaspirit/Shutterstock

By Ara Darzi, Imperial College London

It is not often that one witnesses a transformational advance in medicine. But the application of artificial intelligence (AI) to improve the early detection of disease is exactly that.

I was a co-author of the paper recently published in Nature showing that an AI system developed by Google was better at spotting breast tumours than doctors. Now, researchers in the US have reported that AI-supported laser scanners are faster than doctors at detecting brain tumours. These are very exciting developments that will, ultimately, have a big impact on the accuracy, logistics and speed of diagnosis.

There are a multitude of similar projects underway that employ artificial intelligence in the early detection of macular degeneration, acute kidney failure, skin cancer, sepsis, Alzheimer’s disease and depression, among others.

Healthcare provides especially fertile territory for these advances because of the sheer volume and complexity of medical knowledge. No clinician, however smart, can hope to master it. The aim is not to replace the doctor (yet, at least) but to enhance their medical expertise.

Saving lives, and costs

As well as speeding up diagnosis and making it more accurate, potential savings of around US$400 billion in the US healthcare sector alone might be made from these developments, according to some estimates.

To achieve this, however, scientists need access to data. Data is as vital to artificial intelligence as coal was to the railways and oil to the motorcar. To fulfil the promise of AI in healthcare, medical data will need to be treated as precious to our health as drinking water, says Etta Pisano, chief research officer at the American College of Radiology, in a commentary published alongside our paper in Nature.

When medical data is a precious as water.Sebra/Shutterstock

It is a striking image: our global medical records may be as vital to our health as the water we drink and the air we breathe. And just as we seek to protect our drinking water from contamination and our air from pollution, we must protect our records from abuse. I know, as a researcher, that the risk is real when it comes to keeping patient records confidential. And, as a patient, I know the safety of our data is paramount.

For our research on AI in breast cancer detection, we studied over 29,000 mammograms from women in the UK and the US. All the records were anonymised and no woman could be identified – a standard procedure to protect patient confidentiality. But the research could not have gone ahead without access to this vital data.

We need to demonstrate why data sharing is a social benefit, if we are to generate the growth that these innovations could deliver. There is a moral imperative to improve care for others through research. The biggest challenge, however, is to devise a system of data governance that protects the interests of patients, provides access for researchers, distributes the fruits of success fairly and wins the confidence of the public.

As set out in its 2017 industrial strategy, the UK government has established an Artificial Intelligence Council to boost growth of AI in the UK and its ethical use, and a Centre for Data Ethics to establish a governance regime for data driven technologies. This is a welcome recognition of the central importance of adopting an ethical approach to the development of these technologies.

But we need to go further. Public trust demands more transparency and the creation of incentives to allow data to be shared – for example, by giving the NHS a share of any commercial benefits gained from use of NHS data. There also needs to be a health-specific data charter, with clear rules, norms and standards, setting out what can be done, what should be done and what may not be done. The government should offer an absolute guarantee that patient information will not be shared outside the health, care and research systems. Users of NHS data should be required to prove that they are using it to improve quality, safety and efficiency.

There are huge opportunities in these technologies to advance healthcare, benefit health systems and improve the outlook for millions of patients. But unless we establish clear rules from the outset, we risk sacrificing public trust, surrendering vital clinical gains and squandering the potential in the vast quantities of medical data we have spent decades accumulating.The Conversation

Ara Darzi, Director of the Institute of Global Health Innovation, Imperial College London

This article is republished from The Conversation under a Creative Commons license. Read the original article.




The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.
The Conversation is an independent source of news and views, sourced from the academic and research community and delivered direct to the public.




            AIhub is supported by:



Related posts :



We asked teachers about their experiences with AI in the classroom — here’s what they said

  05 Dec 2025
Researchers interviewed teachers from across Canada and asked them about their experiences with GenAI in the classroom.

Interview with Alice Xiang: Fair human-centric image dataset for ethical AI benchmarking

  04 Dec 2025
Find out more about this publicly-available, globally-diverse, consent-based human image dataset.

The Machine Ethics podcast: Fostering morality with Dr Oliver Bridge

Talking machine ethics, superintelligence, virtue ethics, AI alignment, fostering morality in humans and AI, and more.

Interview with Frida Hartman: Studying bias in AI-based recruitment tools

  02 Dec 2025
In the next in our series of interviews with ECAI2025 Doctoral Consortium participants, we caught up with Frida, a PhD student at the University of Helsinki.

Forthcoming machine learning and AI seminars: December 2025 edition

  01 Dec 2025
A list of free-to-attend AI-related seminars that are scheduled to take place between 1 December 2025 and 31 January 2026.
monthly digest

AIhub monthly digest: November 2025 – learning robust controllers, trust in multi-agent systems, and a new fairness evaluation dataset

  28 Nov 2025
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

EU proposal to delay parts of its AI Act signal a policy shift that prioritises big tech over fairness

  27 Nov 2025
The EC has proposed delaying parts of the act until 2027 following intense pressure from tech companies and the Trump administration.



 

AIhub is supported by:






 












©2025.05 - Association for the Understanding of Artificial Intelligence