TAMPA 鈥 As online retailers increasingly use artificial intelligence chatbots to streamline customer service tasks and replace their human counterparts, new research from the 新澳门六合彩内幕信息 examines how emotion-expressing AI chatbots can impact customer service.
AI chatbots that show positive feelings 鈥 such as adding an 鈥淚 am excited to do so!鈥 or a few exclamation marks 鈥 do not necessarily translate into positive reactions or contribute to higher customer satisfaction, the study revealed.
鈥淭he reason is that people do not expect chatbots to have feelings. People don鈥檛 react to chatbots the same way as they react to humans,鈥 said Denny Yin, an associate professor in the School of Information Systems and Management in the Muma College of Business, who co-authored the study.
The article was published online on Dec. 2 in Information Systems Research, a premier journal included among the top 24 business journals compiled by the University of Texas at Dallas.
The research is particularly relevant with the rise of emotional AI 鈥 the branch of AI that deals with processing and replicating human emotions 鈥 and the recent claim from a Google engineer that an unreleased AI chatbot was 鈥渟entient.鈥
That claim and the resulting backlash led to the employee鈥檚 firing for violating employment and data security policies. In addition, a national debate ensued around what 鈥渟entient鈥 means and whether Google鈥檚 chatbot has consciousness or feelings.
In order to bring real-world applications to the debate, the study explored how customers make sense of and react to positive emotions from an AI chatbot when compared to a human customer service agent.
Yin said impactful research is needed in emotional AI technologies, because the industry is projected to grow to $100 billion by 2024 and $200 billion by 2026, according to market research from Global Industry Analysts and Reports and Data.
The researchers conducted three experiments using emotion-capable chatbots in a customer service scenario.
In the first test, participants interacted with either an AI or human customer service agent to resolve a hypothetical service issue. Half the participants chatted with an agent where positive adjectives and exclamation marks were added into its responses, such as 鈥淚 am delighted to handle your request today!鈥
The remainder of the group chatted with an agent without any emotions, such as 鈥淚 am handling your request today.鈥 Participants rated the customer service quality and satisfaction on a seven-point scale.
The experiments revealed some surprising findings:
- While positive emotions from a human are beneficial and increase customer satisfaction, the same emotions from an AI chatbot are not as effective.
- Customers鈥 expectations play an important role, as customers don鈥檛 expect an AI chatbot to be able to feel emotion. Too much chatbot positivity can be a turn-off for consumers and lead to negative reactions.
鈥淭o our knowledge, this research is the very first to explore the ramifications of equipping service AI agents with emotional expression capabilities,鈥 Yin said.
The work further expands the understanding of customers鈥 reactions to emotional AIs as well as gives companies guidance on how and when retailers should use emotion-capable AI service agents.
Researchers cautioned that companies should understand the expectations of customers exposed to AI-powered services before haphazardly equipping AIs with emotion-expressing capabilities.
Other major takeaways from the study include:
- Good feelings can spread from a human agent鈥檚 positive emotional expressions to a customer, and the same can occur for an AI-powered chatbot. But sometimes those good feelings are cancelled out by a negative reaction to not expecting chatbots to display emotions.
- Expectations play an important role on whether an AI chatbot expressing positive emotions is a positive or negative experience. Different types of customers will react to the same AI-expressed emotions differently.
Servion Global Solutions predicts that by 2025, AI will drive 95% of all customer interactions, including live phone calls and online conversations, according to Finance Digest.
鈥淥ur work reveals the unique impact of expressed positive emotion by an AI agent on customers鈥 service evaluations along with the underlying mechanisms and a boundary condition, thus opening up exciting research opportunities in the area of human-AI interaction,鈥 Yin said.
Researchers conducted the experiments at the Muma College of Business at 新澳门六合彩内幕信息and at the Scheller College of Business at the Georgia Institute of Technology through in-person and online lab sessions from 2019 to 2020.
Aside from Yin, the article鈥檚 co-authors include Elizabeth Han, McGill University, and Han Zhang, Georgia Institute of Technology.