ΑΙhub.org
 

Shortcuts to artificial intelligence – a tale


by
19 May 2020



share this:
shortcuts-to-ai

The current paradigm of artificial intelligence emerged as the result of a series of cultural innovations, some technical and some social. Among them are seemingly small design decisions, that led to a subtle reframing of some of the field’s original goals, and are now accepted as standard. They correspond to technical shortcuts, aimed at bypassing problems that were otherwise too complicated or too expensive to solve, while still delivering a viable version of AI.

Far from being a series of separate problems, recent cases of unexpected effects of AI are the consequences of those very choices that enabled the field to succeed, and this is why it will be difficult to solve them. Research at the University of Bristol has considered three of these choices, investigating their connection to some of today’s challenges in AI, including those relating to bias, value alignment, privacy and explainability.

1) Correlation vs causation
One important consequence of training statistical algorithms to emulate the decisions or behaviours of humans (e.g. recommending a book) is that we no longer value so highly the reason why the decision is made, so long as the action it generates is appropriate. Predictions count more than explanations, knowing ‘what’ counts more than knowing ‘why’, and ‘correlation trumps causation’.

2) Data from the wild
The second shortcut was summarised in a paper by Halevy, Norvig and Pereira which draws general lessons from the success stories of speech recognition and machine translation. It identifies the reason for those successes being the availability of large amounts of data, which had already been created for different purposes. Data gathered from the wild has been crucial in the design of object recognition systems, face recognition, and machine translation. The ubiquitous word embeddings that allow us to represent the meaning of words before we process them, are also all learned from data gathered from the wild.

3) Proxies and implicit feedback
Rather than asking users explicitly what they wanted the AI system to do, designers started making use of implicit feedback, which is another way of saying that they replaced unobservable quantities with cheaper proxies. Understanding the misalignment between a proxy and the intended target has become an important question for AI.

What has been accomplished by the AI research community over the past 20 years is remarkable, and much of this could not have been achieved at the time without taking “shortcuts”, including the three that have been summarised above. With the benefit of hindsight we can, however, reflect on how we introduced assumptions into our systems that are now generating problems, so that we can work on repairing and regulating the current version of AI. The same methods and principles that are perfectly innocuous in a certain domain, can become problematic only after being deployed in different domains. This is the space where we will need better informed regulation.

Read the research to find out more:
Shortcuts to Artificial Intelligence Cristianini, N. Machines We Trust. MIT Press (forthcoming)

Further papers that may be of interest:
Can Machines Read our Minds? Burr, C. & Cristianini, N. Minds & Machines (2019).

An Analysis of the Interaction Between Intelligent Software Agents and Human Users Burr, C., Cristianini, N. & Ladyman, J. Minds & Machines (2018) 28: 735.

This work is part of the ERC ThinkBIG project, Principal Investigator Nello Cristianini, University of Bristol.




Nello Cristianini is a Professor of Artificial Intelligence at the University of Bristol.
Nello Cristianini is a Professor of Artificial Intelligence at the University of Bristol.




            AIhub is supported by:


Related posts :



Dataset reveals how Reddit communities are adapting to AI

  25 Apr 2025
Researchers at Cornell Tech have released a dataset extracted from more than 300,000 public Reddit communities.

Interview with Eden Hartman: Investigating social choice problems

  24 Apr 2025
Find out more about research presented at AAAI 2025.

The Machine Ethics podcast: Co-design with Pinar Guvenc

This episode, Ben chats to Pinar Guvenc about co-design, whether AI ready for society and society is ready for AI, what design is, co-creation with AI as a stakeholder, bias in design, small language models, and more.

Why AI can’t take over creative writing

  22 Apr 2025
A large language model tries to generate what a random person who had produced the previous text would produce.

Interview with Amina Mević: Machine learning applied to semiconductor manufacturing

  17 Apr 2025
Find out how Amina is using machine learning to develop an explainable multi-output virtual metrology system.

Images of AI – between fiction and function

“The currently pervasive images of AI make us look somewhere, at the cost of somewhere else.”

Grace Wahba awarded the 2025 International Prize in Statistics

  16 Apr 2025
Her contributions laid the foundation for modern statistical techniques that power machine learning algorithms such as gradient boosting and neural networks.




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association