ΑΙhub.org
 

Getting data right: governance for people and society


by
09 July 2020



share this:
Time Lapse Photography of Blue Lights. Photographer: Pixabay

By Carly Kind, Director of the Ada Lovelace Institute

Public scrutiny is critical for trust in, and democratic legitimacy for, the use of data-driven decision-making and algorithmic systems in our society.

We stand at the intersection of monumental and ongoing ruptures that will transform the data governance landscape. If they are to have a positive long-term influence it will be because we have heeded their lessons.

The Royal Society’s new publication, The UK data governance landscape, is a valuable resource published at a moment of immense uncertainty, as well as possibility, in the data governance ecosystem. Midway through 2020 we stand at the intersection of three monumental and ongoing ruptures: the coronavirus pandemic, which is accelerating the application of data-driven technologies (PDF) to health as well as policymaking; the Black Lives Matter movement, which is drawing long-overdue attention to the unequal distribution of the benefits of digital transformation as well the problem of bias in algorithmic systems (PDF); and the impending departure of the United Kingdom from the European Union, which is generating questions about the future of international data flows and the opportunities the UK faces to expand its leadership in artificial intelligence (AI). These three phenomena are already changing, and will transform, the data governance landscape. There is a pressing need to ensure the changes wrought benefit people and society.

The Ada Lovelace Institute was established by the Nuffield Foundation in 2018 as an independent research and deliberative body with a mission to ensure data and AI work for people and society. We are committed to addressing these challenges and realising the social value of data, our most important public good.

Ada Lovelace Institute Citizens’ Biometrics Council workshop surfacing public attitudes to biometric technologies © Ada Lovelace Institute. Photographer: Paul Wyatt.

Filling the data gaps

As decision-making becomes increasingly data-driven it is imperative that the right data is driving it: data that is comprehensive, accurate and representative. Yet gaps in critical datasets and missing data – what Professor David Hand calls ‘dark data’ – plague the data ecosystem.

Too often, those data gaps concern the most vulnerable in society: there is no data collected on vulnerable parties or witnesses who appear before the civil courts, for example, leading senior judges to declare the domain a ‘data desert’. There is an absence of public data on cost savings won through the roll out of Universal Credit and on rental costs throughout the UK. The recent Public Health England report (PDF) into disparities in deaths from COVID-19, which found that people from BAME backgrounds are up to twice as likely to die from the disease than White Britons, was unable to draw conclusions due to a lack of data to inform them. The openSAFELY study (PDF), which analysed 17 million patient records to identify factors associated with COVID-19 deaths, found that 26% of patient records did not contain ethnicity data.

Bad – or incomplete – data will lead to badly informed decisions. It may also make for bad algorithms. Biased and non-representative datasets cause the replication of bias in algorithmic systems, as has been evidenced in research in the United States into algorithmic systems used in healthcare and criminal sentencing, and in facial recognition systems. The economic fallout from the coronavirus crisis is likely to see increased government deployment of algorithmic decision-making systems; ensuring these are free from bias will be critical to prevent data-driven decision-making from entrenching, rather than rectifying, inequalities.

Making sure the right data is in the right hands

Too often, personal data is treated as exclusive, monetised by organisations and siloed in ways that inhibit research and development for public benefit. There is an imbalance in the data ecosystem which sees the private sector enjoy a monopoly on data generated by the public. Researchers have long called for social media firms to share their data, but governments too have found themselves begging at the feet of the large tech platforms for access to data to facilitate the pandemic response.

The present crisis has demonstrated how, in times of crisis, barriers to data sharing can be temporarily lowered and data shared to great benefit. As the crisis subsides and the democratic legitimacy for exceptional measures evaporates, questions emerge about how to constitute the political will, technological capability and legal safeguards necessary to continue to pursue socially beneficial data sharing and access arrangements. The Ada Lovelace Institute’s Rethinking Data programme aims to build the institutions – technical, legal, political – necessary to ensure that data is fairly distributed and equitably enjoyed, and that public institutions are able to realise the value of data generated by the public.

Ensuring our regulatory frameworks are fit for purpose

New and emerging concepts such as group and collective privacy (PDF) invite us to question whether additional regulatory frameworks are needed to complement existing laws – protecting privacy in a wider ecosystem where the privacy of groups, not just of individuals, is often compromised.

Other regulatory domains, such as competition law, are increasingly being shaped by the implications of our data-driven economy. The acquisitive practices of large technology companies are often criticised for being anti-competitive. However, existing competition law has struggled to articulate what it means for anti-competitive behaviour to exist in a data-enabled society (in which data-driven services rely on network effects, creating natural monopolies), where access to personal data, not price, is exchanged in return for access to platform services.

Particular types of data may also be under-protected by existing regulatory frameworks. The regulatory and policy framework for governing the use of biometric data in particular has been outpaced by advances in the technologies, such as facial recognition, that enable such data to be used by private companies or public bodies. We need to grapple with how the potential of new technologies change how certain data can be used, and to understand – through independent, impartial and evidence-led analysis – what the implications are for the regulatory landscape. On this issue the Ada Lovelace Institute has commissioned an independent review led by Matthew Ryder QC which will report in early 2021.

Much of the UK’s regulatory regime in the context of data has been led by European Union legislation, and the EU continues to be at the forefront of regulating to address harms in the digital economy. Beyond Brexit, we will need a new national vision for data regulation which articulates a positive vision for the protection and use of data for the public good while continuing to facilitate international data flows with the EU and beyond.

The next decade: challenges and opportunities

The next decade will be defined by how we respond to these three identified ruptures. If they are to have a positive long-term influence on the data governance landscape it will be because we will have heeded their lessons.

The Royal Society and British Academy’s 2017 report Data management and use: Governance in the 21st century called for a new interdisciplinary stewardship body that would be deeply connected to diverse communities, nationally focused and globally relevant. The Ada Lovelace Institute was created in response to this call. The evidence we are gathering for policy-making and regulatory reform includes scrutiny of social and technical infrastructures and systems, and – crucially – the perspectives of the those who will be most impacted: the public. This scrutiny will be critical for public trust and confidence in, and democratic legitimacy for, more widespread use of data-driven decision-making and algorithmic systems in our society.

You can download the Royal Society’s new publication The UK data governance landscape: explainer (PDF).

For another perspective on the UK’s data governance landscape and the Royal Society’s Data Governance Explainer, please see Using data for the public good: the roles of clear governance, good data and trustworthy institutions.

This article was originally published on The Royal Society blog.

About the author

Carly KindCarly Kind is Director of the Ada Lovelace Institute, an independent research institute and deliberative body with a mission to ensure data and AI work for people in society.




The Royal Society is a Fellowship of many of the world's most eminent scientists and is the oldest scientific academy in continuous existence.
The Royal Society is a Fellowship of many of the world's most eminent scientists and is the oldest scientific academy in continuous existence.




            AIhub is supported by:



Related posts :

monthly digest

AIhub monthly digest: January 2026 – moderating guardrails, humanoid soccer, and attending AAAI

  30 Jan 2026
Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.

The Machine Ethics podcast: 2025 wrap up with Lisa Talia Moretti & Ben Byford

Lisa and Ben chat about the prevalence of AI slop, the end of social media, Grok and explicit content generation, giving legislation more teeth, anthropomorphising reasoning models, and more.

Interview with Kate Larson: Talking multi-agent systems and collective decision-making

  27 Jan 2026
AIhub ambassador Liliane-Caroline Demers caught up with Kate Larson at IJCAI 2025 to find out more about her research.

#AAAI2026 social media round up: part 1

  23 Jan 2026
Find out what participants have been getting up to during the first few of days at the conference

Congratulations to the #AAAI2026 outstanding paper award winners

  22 Jan 2026
Find out who has won these prestigious awards at AAAI this year.

3 Questions: How AI could optimize the power grid

  21 Jan 2026
While the growing energy demands of AI are worrying, some techniques can also help make power grids cleaner and more efficient.

Interview with Xiang Fang: Multi-modal learning and embodied intelligence

  20 Jan 2026
In the first of our new series of interviews featuring the AAAI Doctoral Consortium participants, we hear from Xiang Fang.

An introduction to science communication at #AAAI2026

  19 Jan 2026
Find out more about our session on Wednesday 21 January.


AIhub is supported by:







 













©2026.01 - Association for the Understanding of Artificial Intelligence