Virtual assistants, so-called chatbots, have become an integral part of many company websites and are playing an increasingly important corporate role. A study by TU Dresden published in the Journal of Management Information Systems has investigated whether errors made by chatbots lead to aggressive behavior among their users and what influence users' perceived humanness of the virtual assistants has on their reactions.
Chatbots are designed to make it easier for internet users to find the information they need quickly by responding directly to questions and requests. However, the reactions to chatbots have proven to be not always positive, says Professor Alfred Benedikt Brendel from TU Dresden.
"If a chatbot gives incorrect or confusing answers, this can trigger aggression from users towards the digital conversation partner," explains the holder of the Chair of Information Systems, in particular Intelligent Systems and Services.
In the worst case, aggression towards the virtual assistant, including verbal abuse, can also have other repercussions—or example increasing users' negative attitude towards the provider or the website itself.
In their study, the international research team led by Alfred Brendel hypothesized that the design of chatbots has an influence on how users react to unsatisfactory answers. If a chatbot is provided with human attributes, it can be assumed that aggressive behavior occurs less frequently than with a neutrally designed chatbot.
"In our experiments, some of the participants used a human chatbot that was equipped with a name, gender and picture. It answered questions in a very friendly manner and reinforced its messages with appropriate emojis." On the other hand, the neutral chatbot, with which another part of the study participants interacted, did not contain any such design elements.
The results of the study show first and foremost that the human chatbot generally increases user satisfaction. This, in turn, also reduces the occurrence of frustration. However, if a chatbot gives unsatisfactory answers, this leads to frustration and aggression even towards a chatbot with human attributes, which contradicts the researchers' original assumption. Overall, around 10% of users show aggressive behavior towards the virtual assistants.
In contrast to a neutral chatbot, however, the intensity of aggressive behavior is reduced when faced with a more human chatbot. Users were less likely to use offensive language when interacting with a human chatbot, for example. The results have far-reaching consequences, particularly in practice, explains Alfred Brendel.
"I would advise software developers to be cautious in their approach to human-like design and think carefully about the positive and negative effects that additional human-like design elements such as gender, age or certain names can have."
More information: Alfred Benedikt Brendel et al, The Paradoxical Role of Humanness in Aggression Toward Conversational Agents, Journal of Management Information Systems (2023). DOI: 10.1080/07421222.2023.2229127
Citation: The paradoxical role of 'humanness' in aggression toward conversational agents (2024, April 10) retrieved 10 April 2024 from https://techxplore.com/news/2024-04-paradoxical-role-humanness-aggression-conversational.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.