ΑΙhub.org
 

AI in health care challenges us to define what better, people-centred care looks like


by
24 April 2023



share this:

cartoon of a doctor stood next to a large mobile phone
By Catherine Burns

From faster and more accurate disease diagnosis to models of using health care resources more efficiently, AI promises a new frontier of effective and efficient health care. If it’s done right, AI may allow for more people-centred care and for clinicians to spend more time with people, doing the work they enjoy most. But to achieve these aspirations, foundational work must occur in how we operate today and in defining what health care looks like in the future.

AI technologies are only as reliable as the data that drives them. To unlock the power of AI, it requires us to become better at sharing health data between primary care providers, specialists, hospitals, research universities, health companies and patients to develop reliable and accurate models. Without this data, AI technologies may make mistakes, generate inappropriate solutions and encourage inappropriate trust in their answers.

Our health data will also need to be better quality. Issues with noisy sensors, incomplete documentation and different data types must be solved. Health data will have to travel across individual health journeys through multiple providers to avoid reaching solutions that are limited in time and context. In some cases, AI solutions are being developed from clinical trial data. Clinical trial data sets are well known to exclude participants of certain ages, demographics or with multiple morbidities.

Our community and small hospitals can be a solution to this, and they need a louder voice in the health care conversation. More Canadians visit community hospitals than academic hospitals, so their data and experience must be part of the solution. Our small hospitals provide many services to our remote and often underserved communities. For this reason, the voices of those working in our remote communities are critically important at this time, where they are overworked and under-resourced. AI must be designed with a goal of promoting greater access and equity in health care. This means AI must be designed to support equity, be broadly inclusive and be designed to partner with our communities.

We need to understand what it means to have successful health care. Without understanding what a high-performance health-care system looks like, technologies will not be developed to align for effective solutions. We must define the right metrics to get the right results. Do we want to reduce the cost of surgery? Or do we want to reduce the likelihood of follow-up surgery years later? Those goals may have different solutions.

Similarly, do we believe strongly in growing towards a coordinated and shared health care vision? If we do, and I hope we do, AI must be people-centred and designed from an interprofessional lens. It means we must learn and teach each other more about practices of care, outcomes, technology, decision-making and quality of life.

AI learns from our data, so we must provide the proper foundation. Our next generation of AI designers will design their technologies for the problems we tell them are important. We need to define what those problems are and what success would mean.

Catherine Burns

Catherine Burns is the Chair in Human Factors in Health Care Systems and leads the University of Waterloo’s health initiatives. She is a professor in the Faculty of Engineering and an expert in human-centred approaches to the design and implementation of advanced health-care technologies.



tags: ,


University of Waterloo

            AIhub is supported by:



Subscribe to AIhub newsletter on substack



Related posts :

RWDS Big Questions: how do we balance innovation and regulation in the world of AI?

  06 Mar 2026
The panel explores the tensions, trade-offs and practical realities facing policymakers and data scientists alike.

Studying multiplicity: an interview with Prakhar Ganesh

  05 Mar 2026
What is multiplicity, and what implications does it have for fairness, privacy and interpretability in real-world systems?

Top AI ethics and policy issues of 2025 and what to expect in 2026

, and   04 Mar 2026
In the latest issue of AI Matters, a publication of ACM SIGAI, Larry Medsker summarised the year in AI ethics and policy, and looked ahead to 2026.

The greatest risk of AI in higher education isn’t cheating – it’s the erosion of learning itself

  03 Mar 2026
Will AI hollow out the pipeline of students, researchers and faculty that is the basis of today’s universities?

Forthcoming machine learning and AI seminars: March 2026 edition

  02 Mar 2026
A list of free-to-attend AI-related seminars that are scheduled to take place between 2 March and 30 April 2026.
monthly digest

AIhub monthly digest: February 2026 – collective decision making, multi-modal learning, and governing the rise of interactive AI

  27 Feb 2026
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

The Good Robot podcast: the role of designers in AI ethics with Tomasz Hollanek

  26 Feb 2026
In this episode, Tomasz argues that design is central to AI ethics and explores the role designers should play in shaping ethical AI systems.

Reinforcement learning applied to autonomous vehicles: an interview with Oliver Chang

  25 Feb 2026
In the third of our interviews with the 2026 AAAI Doctoral Consortium cohort, we hear from Oliver Chang.



AIhub is supported by:







Subscribe to AIhub newsletter on substack




 















©2026.02 - Association for the Understanding of Artificial Intelligence