ΑΙhub.org
 

“I am here to assist you today” – how we respond to chatbots

by
05 July 2021



share this:
Speech bubble | NLP

Photo by Volodymyr Hryshchenko on Unsplash

As online users we are increasingly exposed to chatbots as one form of AI-enabled media technologies. Stand-alone chatbots are often used for product or service recommendations, for example when online shopping or making financial or health-related decisions. What is the persuasive potential of these chatbots? Carolin Ischen studies how our perceptions of AI and our experience with chatbots influence our attitudes and behaviour.

Carolin Ischen is a PhD candidate at the Persuasive Communication group and the Digital Communication Methods Lab of the Communication Science department at the University of Amsterdam. She studies the persuasive potential of AI-driven media, specifically chatbots. “Chatbots reshape today’s media environment and move the communication science field from computer-mediated communication into human-machine communication. I therefore study technology as communicators and not only tools for communication”, tells Ischen.

In her research Ischen focuses on chatbots that recommend products or services, like a recipe or insurance, and how our perceptions of and experience with these non-human assistants influence our behaviour. “People tend to respond to computers in comparable ways as they do to humans. I look at how people enjoy the interaction with the chatbot and how they judge the content and brand information that is offered to them”, Ischen adds.

Carolin Ischen
Carolin Ischen

Enjoyment is the key mechanism

In a first experiment Ischen and colleagues examined the effects of interacting with a stand-alone chatbot compared to more traditional interactive websites. With the use of tailormade websites and virtual assistants they randomly asked a representative sample of the Dutch population to interact either with the chatbot or with the interactive website to receive a recommendation for a health insurance. Afterwards participants had to complete a survey to assess the experience: How human- or machine-like did the perceive the assistant? How much did they enjoy the interaction? Would they purchase the brand that was recommended?

Ischen et al found that enjoyment is the key mechanism explaining the positive effect of chatbots versus websites: the interaction with a stand-alone chatbot resulted in more enjoyable user experiences, which subsequently translated into higher persuasive outcomes. But contrary to expectations, perceived anthropomorphism (i.e., attribution of human-like characteristics), seemed not to be particularly relevant in this comparison. “The mere presentation of a chatbot as the source of communication as done in this study was likely not sufficient to increase human-likeness.”

Chatbots and privacy concerns

Ischen also looks at privacy concerns relating to chatbot-interactions. “While such concerns are widely studied in an online (website-) context, research in the context of chatbot-interaction is still lacking.” Ischen investigated the extent to which chatbots with human-like cues are attributed human-like characteristics and how this affects privacy concerns, information disclosure and recommendation adherence. She finds that people have fewer privacy concerns and disclose more information to chatbots when they perceive chatbots as more human-like, often leading to a follow up of the recommendation.

Ischen calls the study of chatbots in consumer interaction important in light of consumer awareness and empowerment. “As consumers we need to be aware that what is being communicated to us by chatbots is not neutral or bias free. Are we as consumers informed enough how this technology works and how to use it safely? What kind of regulations are needed for advertisement practices driven by AI?”

Further reading

“I am here to assist you today”: The role of entity, interactivity and experiential perceptions in chatbot persuasion, Ischen, C., Araujo, T. B., van Noort, G., Voorveld, H. A. M., & Smit, E. G. (2020).
Privacy concerns in chatbot interactions, Ischen, C., Araujo, T., Voorveld, H., van Noort, G., & Smit, E. (2020).




University of Amsterdam




            AIhub is supported by:


Related posts :



The Machine Ethics Podcast: featuring Marc Steen

In this episode, Ben chats to Marc Steen about AI as tools, the ethics of business models, writing "Ethics for People Who Work in Tech", and more.
06 June 2023, by

On privacy and personalization in federated learning: a retrospective on the US/UK PETs challenge

Studying the use of differential privacy in personalized, cross-silo federated learning.
05 June 2023, by

VISION AI Open Day: Trustworthy AI

Watch the roundtable discussion on trustworthy AI, with a focus on generative models, from the AI Open Day held in Prague.
02 June 2023, by

PeSTo: an AI tool for predicting protein interactions

The model can predict the binding interfaces of proteins when they bind other proteins, nucleic acids, lipids, ions, and small molecules.
01 June 2023, by

Tetris reveals how people respond to an unfair AI algorithm

An experiment in which two people play a modified version of Tetris revealed that players who get fewer turns perceive the other player as less likeable, regardless of whether a person or an algorithm allocates the turns.
31 May 2023, by

AIhub monthly digest: May 2023 – mitigating biases, ICLR invited talks, and Eurovision fun

Welcome to our monthly digest, where you can catch up with AI research, events and news from the month past.
30 May 2023, by





©2021 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association