ΑΙhub.org
 

“I am here to assist you today” – how we respond to chatbots


by
05 July 2021



share this:
Speech bubble | NLP

Photo by Volodymyr Hryshchenko on Unsplash

As online users we are increasingly exposed to chatbots as one form of AI-enabled media technologies. Stand-alone chatbots are often used for product or service recommendations, for example when online shopping or making financial or health-related decisions. What is the persuasive potential of these chatbots? Carolin Ischen studies how our perceptions of AI and our experience with chatbots influence our attitudes and behaviour.

Carolin Ischen is a PhD candidate at the Persuasive Communication group and the Digital Communication Methods Lab of the Communication Science department at the University of Amsterdam. She studies the persuasive potential of AI-driven media, specifically chatbots. “Chatbots reshape today’s media environment and move the communication science field from computer-mediated communication into human-machine communication. I therefore study technology as communicators and not only tools for communication”, tells Ischen.

In her research Ischen focuses on chatbots that recommend products or services, like a recipe or insurance, and how our perceptions of and experience with these non-human assistants influence our behaviour. “People tend to respond to computers in comparable ways as they do to humans. I look at how people enjoy the interaction with the chatbot and how they judge the content and brand information that is offered to them”, Ischen adds.

Carolin Ischen
Carolin Ischen

Enjoyment is the key mechanism

In a first experiment Ischen and colleagues examined the effects of interacting with a stand-alone chatbot compared to more traditional interactive websites. With the use of tailormade websites and virtual assistants they randomly asked a representative sample of the Dutch population to interact either with the chatbot or with the interactive website to receive a recommendation for a health insurance. Afterwards participants had to complete a survey to assess the experience: How human- or machine-like did the perceive the assistant? How much did they enjoy the interaction? Would they purchase the brand that was recommended?

Ischen et al found that enjoyment is the key mechanism explaining the positive effect of chatbots versus websites: the interaction with a stand-alone chatbot resulted in more enjoyable user experiences, which subsequently translated into higher persuasive outcomes. But contrary to expectations, perceived anthropomorphism (i.e., attribution of human-like characteristics), seemed not to be particularly relevant in this comparison. “The mere presentation of a chatbot as the source of communication as done in this study was likely not sufficient to increase human-likeness.”

Chatbots and privacy concerns

Ischen also looks at privacy concerns relating to chatbot-interactions. “While such concerns are widely studied in an online (website-) context, research in the context of chatbot-interaction is still lacking.” Ischen investigated the extent to which chatbots with human-like cues are attributed human-like characteristics and how this affects privacy concerns, information disclosure and recommendation adherence. She finds that people have fewer privacy concerns and disclose more information to chatbots when they perceive chatbots as more human-like, often leading to a follow up of the recommendation.

Ischen calls the study of chatbots in consumer interaction important in light of consumer awareness and empowerment. “As consumers we need to be aware that what is being communicated to us by chatbots is not neutral or bias free. Are we as consumers informed enough how this technology works and how to use it safely? What kind of regulations are needed for advertisement practices driven by AI?”

Further reading

“I am here to assist you today”: The role of entity, interactivity and experiential perceptions in chatbot persuasion, Ischen, C., Araujo, T. B., van Noort, G., Voorveld, H. A. M., & Smit, E. G. (2020).
Privacy concerns in chatbot interactions, Ischen, C., Araujo, T., Voorveld, H., van Noort, G., & Smit, E. (2020).




University of Amsterdam




            AIhub is supported by:


Related posts :



Optimizing LLM test-time compute involves solving a meta-RL problem

  20 Jan 2025
By altering the LLM training objective, we can reuse existing data along with more test-time compute to train models to do better.

Generating a biomedical knowledge graph question answering dataset

  17 Jan 2025
Introducing PrimeKGQA - a scalable approach to dataset generation, harnessing the power of large language models.

The Machine Ethics podcast: 2024 in review with Karin Rudolph and Ben Byford

Karin Rudolph and Ben Byford talk about 2024 touching on the EU AI Act, agent-based AI and advertising, AI search and access to information, conflicting goals of many AI agents, and much more.

Playbook released with guidance on creating images of AI

  15 Jan 2025
Archival Images of AI project enables the creation of meaningful and compelling images of AI.

The Good Robot podcast: Lithium extraction in the Atacama with Sebastián Lehuedé

  13 Jan 2025
Eleanor and Kerry chat to Sebastián Lehuedé about data activism, the effects of lithium extraction, and the importance of reflexive research ethics.

Interview with Erica Kimei: Using ML for studying greenhouse gas emissions from livestock

  10 Jan 2025
Find out about work that brings together agriculture, environmental science, and advanced data analytics.

TELL: Explaining neural networks using logic

  09 Jan 2025
Alessio and colleagues have developed a neural network that can be directly transformed into logic.




AIhub is supported by:






©2024 - Association for the Understanding of Artificial Intelligence


 












©2021 - ROBOTS Association