Caring For Others Is a Key Driver in Getting People to Use Chatbots for Mental Health
A new study from North Carolina State University and Syracuse University assessed what would motivate people to use chatbots for mental health services in the wake of a mass shooting. The researchers found that users’ desire to help others with mental health problems was a more powerful driver than seeking help for their own problems.
“We saw a sharp increase in mass shootings in the U.S. in recent years, and that can cause increases in the need for mental health services,” says Yang Cheng, first author of the study and an assistant professor of communication at NC State. “And automated online chatbots are an increasingly common tool for providing mental health services – such as providing information or an online version of talk therapy. But there has been little work done on the use of chatbots to provide mental health services in the wake of a mass shooting. We wanted to begin exploring this area, and started with an assessment of what variables would encourage people to use chatbots under those circumstances.”
The researchers conducted a survey of 1,114 U.S. adults who had used chatbots to seek mental health services at some point prior to the study. Study participants were given a scenario in which there had been a mass shooting, and were then asked a series of questions pertaining to the use of chatbots to seek mental health services in the wake of the shooting. The survey was nationally representative and the researchers controlled for whether study participants had personal experiences with mass shootings.
The researchers found a number of variables that were important in driving people to chatbots to address their own mental health needs. For example, people liked the fact that chatbots were fast and easy to access, and they thought chatbots would be good sources of information. The study also found that people felt it was important for chatbots to be humanlike, because they would want the chatbots to provide emotional support.
But researchers were surprised to learn that a bigger reason for people to use chatbots was to help other people who were struggling with mental health issues.
“We found that the motivation of helping others was twice as powerful as the motivation of helping yourself,” Cheng says.
Helping others, in this context, would include talking to a chatbot in order to help a loved one experiencing mental illness from getting worse; finding ways to encourage the loved one to access the chatbot services; or to demonstrate to the loved one that the services are easy to use.
“Our study offers detailed insights into what is driving people to access mental health information on chatbot platforms after a disaster, as well as how they are using that information,” Cheng says. “Among other applications, these findings should be valuable for the programmers and mental healthcare providers who are responsible for developing and deploying these chatbots.”
The paper, “AI-Powered Mental Health Chatbots: Examining Users’ Motivations, Active Communicative Action, and Engagement after Mass-Shooting Disasters,” is published in the Journal of Contingencies and Crisis Management. The paper was co-authored by Hua Jiang of Syracuse. The work was done with support from a CUSE seed grant from Syracuse University.
Note to Editors: The study abstract follows.
“AI-Powered Mental Health Chatbots: Examining Users’ Motivations, Active Communicative Action, and Engagement after Mass-Shooting Disasters”
Authors: Yang Cheng, North Carolina State University; Hua Jiang, Syracuse University
Published: Sept. 29, Journal of Contingencies and Crisis Management
Abstract: Chatbots are gaining their popularity in society and have triggered heated discussions in academia as well. Currently, few studies explored the applications of AI-powered mental health chatbots in a mass-shooting disaster context. Via integrating literature from multi-disciplines such as crisis management, mental health, and digital communication, this quantitative study intends to contribute to close this gap and explore the associations between perceived gratifications and protection motivations of using mental health chatbot services, active communicative action, and online and offline engagement behaviors of solving mental health problems after disasters. This study surveyed 1,114 U.S. participants who ever used chatbot services from top healthcare companies. Implications of the results enhance theoretical discussions on how artificial intelligence has shaped individuals’ motivations, communicative action, and engagement behavior to treat mental health problems. This study also benefits professionals who want to learn more about chatbots for mental healthcare, crisis management, and customer engagement.
This post was originally published in NC State News.