ΑΙhub.org
 

Bridging the gap between user expectations and AI capabilities: Introducing the AI-DEC design tool


by
06 August 2024



share this:

The challenge of AI integration in the workplace

Today, AI systems are increasingly integrated into everyday workplaces. However, as AI systems become more prevalent in everyday workplaces, their integration has not always been as smooth or successful as anticipated. A significant reason for this is the gap between user (worker) expectations and the actual capabilities of AI systems. This gap often leads to user dissatisfaction and poor adoption rates, highlighting the need for better design approaches to align user needs with AI functionalities.

Understanding dual perspectives

To address these design challenges, it’s crucial to understand dual perspectives: both the user’s and the AI system’s point of view. From the user’s viewpoint, designers need to comprehend their information and interaction needs, routines, and skills. From the AI system’s perspective, understanding its data inputs, capabilities, and limitations is essential. By merging these perspectives, we can create AI designs that are more user-centered and capable of delivering meaningful value.

The AI-DEC: A Card-based design tool for user-centered AI explanations


Our team developed the AI-DEC, a participatory design tool to bridge this gap. The AI-DEC is a card-based design method that enables users and AI systems to communicate their perspectives and collaboratively build AI explanations. Each card in the AI-DEC represents a design element with descriptions and example use cases, organized into four critical dimensions for explanation design: content, modality, frequency, and direction.

The four dimensions of the AI-DEC

  1. Content: This involves the main message that is transmitted.
  2. Modality: Refers to the communication medium used to transmit the message.
  3. Frequency: Entails the amount and duration of communication.
  4. Direction: Refers to the vertical and horizontal communication flow within an organization.

These dimensions are derived from a communication strategy model by Mohr and Nevin (Mohr & Nevin, 1990), which shows that effective communication strategies can enhance group outcomes such as coordination, satisfaction, commitment, and performance.

Practical application and effectiveness

To evaluate the AI-DEC’s effectiveness, we conducted co-design sessions with 16 workers from healthcare, finance, and management domains. These workers used AI systems for decision-making tasks like aiding in diagnoses, making financial loan decisions, and creating worker schedules. During the sessions, workers designed AI explanations tailored to their unique needs, resulting in diverse explanation designs across different domains.

Key themes and findings

During the co-design session, workers used the AI-DEC to design different types of explanations depending on their unique information and interaction needs. Let us present two different types as examples:
1. Adaptation Based on Worker Input: Healthcare workers designed AI explanations to enhance support based on worker feedback, tailoring explanations to different expertise levels.

2. Worker Acceptance of AI Systems: In the management domain, workers designed explanations to address resistance towards AI systems by providing detailed explanations during the introductory stage and more concise explanations in the active-use stage.

Implications for AI design

We believe the AI-DEC to be most effective when designers need to understand user needs and AI capabilities, prototype human-AI interactions, and collaborate with users and AI engineers. It facilitates communication between AI designers and users, resulting in prototypes that are more user-centered and context-specific.

Moreover, we envision the AI-DEC as a tool for UX designers to enhance their understanding of technical AI capabilities, enabling them to design more effective user experiences.

Future applications and takeaways

For various application contexts, the AI-DEC can be tailored to meet the needs of designers, users, or organizations, and used to design continuous interaction experiences, not just singular explanations. We envision the AI-DEC to enable effective user involvement in the AI design process, often a challenging aspect in many AI system design procedures today.

By bridging the gap between user expectations and AI capabilities, the AI-DEC holds promise for successful AI integration in various application settings. Its participatory approach ensures that both user and AI perspectives are aligned, leading to more effective and user-centered AI systems.

Read the work in full

The AI-DEC: A Card-based Design Method for User-centered AI Explanations, Christine P Lee, Min Kyung Lee, Bilge Mutlu.




Christine Lee is a fourth-year PhD student in the Department of Computer Sciences at the University of Wisconsin-Madison
Christine Lee is a fourth-year PhD student in the Department of Computer Sciences at the University of Wisconsin-Madison




            AIhub is supported by:


Related posts :



Generative AI is already being used in journalism – here’s how people feel about it

  21 Feb 2025
New report draws on three years of interviews and focus group research into generative AI and journalism

Charlotte Bunne on developing AI-based diagnostic tools

  20 Feb 2025
To advance modern medicine, EPFL researchers are developing AI-based diagnostic tools. Their goal is to predict the best treatment a patient should receive.

What’s coming up at #AAAI2025?

  19 Feb 2025
Find out what's on the programme at the 39th Annual AAAI Conference on Artificial Intelligence

An introduction to science communication at #AAAI2025

  18 Feb 2025
Find out more about our forthcoming training session at AAAI on 26 February 2025.

The Good Robot podcast: Critiquing tech through comedy with Laura Allcorn

  17 Feb 2025
Eleanor and Kerry chat to Laura Allcorn about how she pairs humour and entertainment with participatory public engagement to raise awareness of AI use cases

Interview with Kayla Boggess: Explainable AI for more accessible and understandable technologies

  14 Feb 2025
Hear from Doctoral Consortium participant Kayla about her work focussed on explanations for multi-agent reinforcement learning, and human-centric explanations.

The Machine Ethics podcast: Running faster with Enrico Panai

This episode, Ben chats to Enrico Panai about different aspects of AI ethics.

Diffusion model predicts 3D genomic structures

  12 Feb 2025
A new approach predicts how a specific DNA sequence will arrange itself in the cell nucleus.




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association