An Overview of Chatbot-Based Mobile Mental Health Apps: Insights From App Description and User Reviews


IntroductionMental Health Chatbots as an Emerging Technology

A chatbot is a system that can converse and interact with human users using spoken, written, and visual languages []. In recent years, chatbots have been used more frequently in various industries, including retail [], customer service [], education [], and so on because of the advances in artificial intelligence (AI) and machine learning (ML) domains. Facebook Messenger currently offers more than 300,000 text-based chatbots []. Chatbots have primarily been used for commercial purposes and profitable businesses. However, more recent research has demonstrated that chatbots have considerable promise in the health care industry in treating patients and offering them support in a cost-effective and convenient manner [].

In the context of mental health (MH), chatbots may encourage interaction with those who have traditionally been reluctant to seek health-related advice because of stigmatization []. Chatbots are an emerging technology that shows potential for mobile MH apps to boost user engagement and adherence []. The effectiveness of chatbots has been explored for self-disclosure and expressive writing [,,]. Young people with MH issues have experienced various types of social support such as appraisal, informational, emotional, and instrumental support from chatbots []. In addition, chatbots have been designed to educate underprivileged communities on MH and stigmatized topics [,]. Emerging evidence has shown user acceptance of chatbots for supporting various MH issues and early promises in boosting health outcomes in the physical and MH domains.

The adoption of new technology, especially those heavily related to AI and ML, relies first on ascertaining the levels of safety, effectiveness, and user comfort. Despite the increasing adoption and benefits of emerging technologies such as chatbots to support MH and well-being, little research has been conducted to gain an understanding of consumers’ real-life user experiences of interacting with MH chatbot apps. Recent research on MH apps in general points out that patient safety is rarely examined, health outcomes are evaluated on a small scale, and no standard evaluation methods are present [], and these findings also apply to MH chatbot apps. Similar to many other emerging technologies, recent developments in chatbots are because of a massive technology push, with little attention paid to human needs and experiences []. This can lead to unintended negative consequences, such as biases, inadequate and failed responses, and privacy issues, all of which can negatively affect the quality of the experience of chatbots as a source of support [,]. Thus, it is critical to gain an understanding of the nuances in users’ perceptions and experiences of using MH chatbots.

Commercially available MH chatbot apps for popular platforms (eg, iOS [Apple Inc] and Android [Google Inc]) are used by a large user base with varying demographic backgrounds. These users can provide feedback through ratings and text reviews []. These platforms can be leveraged to gain a holistic understanding of the features that recently developed MH chatbots offer and how users assess them. Knowledge of user perceptions from real-life experiences can inform future research and the design of more effective chatbots. Previous studies have identified user reviews as a great source for understanding the benefits and drawbacks of technology [,]. This allows researchers to incorporate community values and needs into product design and improves user-friendliness []. Consumers often make decisions about using new tools based on user rating scores and reviews in web-based marketplaces. According to previous studies, users trust reviews and feel at ease based on their decisions them []. Moreover, previous literature emphasizes analyzing user reviews of mobile MH apps that have chatbot features [,] to obtain in-depth knowledge about this new technology intervention in mobile MH apps. For this study, we decided to analyze commercially available well-known chatbot-based mobile MH apps and their corresponding user reviews from the Apple App Store and Google Play Store. To obtain a comprehensive overview of these apps and understand the nuances of user opinions, we aimed to answer the following 2 research questions (RQs):

RQ1: What are the state-of-the-art features and properties of chatbot-based mobile MH apps?RQ2: What concerns and opinions are expressed in user reviews published on web-based app store platforms regarding the usability and efficiency of chatbot-based mobile MH apps?

We conducted an exploratory observation of 10 apps that offer support and treatment for a variety of MH concerns with a built-in chatbot feature and qualitatively analyzed their user reviews available on the Google Play Store and Apple App Store. Publicly available data (user reviews) provide in-depth analyses of consumers’ personal app user experiences. We found that although chatbots’ personalized, humanlike interactions were positively received by users, improper responses and assumptions about the personalities of users led to a loss of interest. As chatbots are always accessible and convenient, users can become overly attached to them and prefer them over interacting with their friends and family members. Furthermore, a chatbot may offer support for a crisis whenever the user needs it because of its 24/7 availability, but even the recently developed chatbots lack the understanding of properly identifying a crisis. Chatbots in this study fostered a judgment-free environment and helped users feel more comfortable sharing sensitive information.

Before implementing a technological solution for MH, researchers in digital health communities are constantly interested in the support needs and preferences of groups or communities [-]. Researchers have analyzed the effectiveness of technologies used for MH assistance [,], proposing ethical concerns [], policy recommendations [,], and designing automated or human-in-the-loop interactive systems [,]. These studies stressed the significance of designing and evaluating systems for susceptible populations, such as people with MH issues, from the perspective of users. To contribute to this body of work, we discussed our study’s findings with respect to the research and design implications for future MH chatbots. We outlined specific recommendations for customizing certain features, careful consideration of incorporating persuasive strategies, and trust building. Finally, we discussed the impact of excessive reliance on chatbots for MH support. We believe that considering these insights while developing a chatbot-based MH support system will make the design user centric and, thus, more effective.

Background and Related Work

Chatbots are software programs that can imitate human behavior and undertake specific tasks by intelligently conversing with users []. They are conversational agents that use text and speech recognition to engage with users []. Chatbots are commonly used in various web-based and mobile-based apps. In recent years, it has taken on the role of an internet-based entity that can act as a travel agent [], customer service representative [], financial adviser [], and personal assistant [] and is becoming increasingly sophisticated. Some of the available chatbots can have a personality of their own, store information about the user to deliver contextualized answers, and grow over time by learning about their users to provide better services [].

In this section, we provide a brief overview of research on chatbots in health care, including mobile MH chatbots, and provide a rationale for using app reviews to capture perceptions and opinions of users.

Chatbots in Health Care

Chatbots have recently received much attention in the health care and wellness industries [] and have been tested using a variety of elements and characteristics depending on the behavior they were attempting to achieve. Chatbots function as digital personal assistants [], allowing patients to learn more [], obtain support [], and take prompt action in response to new symptoms []. Some chatbots can assist users in collecting medical data via text discussions and then delivering it to the (selected) physicians in a format that is easier to use for diagnostic purposes []. Chatbot interventions are effective in increasing physical activity, achieving relevant weight loss, and improving diet [-] by sending daily check-in reminders [] and offering relevant resources []. They are also sufficiently sophisticated to interact with users through daily adaptive little chats and show progress toward goals using analytics and graphs to encourage self-reflection [].

Mobile MH Chatbots

Among the numerous chatbots being used in different aspects of health and well-being, chatbots in mobile MH care have demonstrated effectiveness in broadening traditional therapy in a cost-effective and convenient manner []. MH chatbots are AI-powered chatbots that provide MH support, guidance, and resources through a conversational interface []. These chatbots replicate human interactions, respond to user inputs, and deliver tailored MH care []. MH chatbots can target a range of MH concerns, including anxiety, depression, and stress [,]. These can provide coping strategies, mindfulness exercises, and information about MH conditions and treatments and, in some cases, connect users to MH professionals [,].

A 2021 national survey found that 22% of adults had used an MH chatbot, and 47% said they would be interested in using it if needed. Among the respondents who had tried an MH chatbot, nearly 60% said they began this use during the COVID-19 pandemic, and 44% said they used chatbots exclusively and did not see a human therapist []. Currently, there are at least 9 chatbot apps on app markets with more than 500,000 downloads. Chatbots have been shown to effectively reduce the severity of MH concerns for people from different demographics and backgrounds, including people in rural communities [], shift workers with accessibility issues [], students with anxiety and stress [], employees of health care systems who require emotional support [], veterans and adolescents who feel stigmatized in sharing their concerns [], etc.

Rather than providing generic suggestions, chatbots can deliver individualized suggestions and resources based on the needs and requirements of users [,]. They were designed to identify MH concerns [], track moods [], deliver cognitive behavioral therapy (CBT) [], and promote positive psychology []. Several well-known chatbots such as Wysa [], Woebot [], Replika [], Youper [], and Tess [] were discussed in prior literature. Inkster et al [] examined the potency of Wysa and found a positive influence on reducing depressive symptoms in a randomized controlled experiment. Fitzpatrick et al [] evaluated the effectiveness of the AI chatbot Woebot in giving CBT to college students with anxiety and depression and found that the Woebot notably decreased depressive symptoms. Ta et al [] investigated social support received from artificial agents in everyday contexts when interacting with the social chatbot Replika. Mehta et al [] examined the acceptability and effectiveness of Youper. In addition to commercial apps, in recent years, research communities have also been increasingly involved in designing chatbots for specific purposes, such as teaching self-compassion (“Vincent”) [], enabling self-disclosure [,], facilitating positive messages within social groups [], improving the quality of life of older people and making them more active to fight their sense of loneliness [], supporting interpersonal skills (“Sunny”) [], and reducing stress (“Mylo”) []. Kim et al [] explored teenagers’ expectations when interacting with a chatbot intended to support their emotional needs. Although most prior studies focused on developing and evaluating new chatbot systems or assessing the effectiveness of the evidence-based techniques used by existing chatbots, there is inadequate research on how end users perceive the usefulness of these app-based chatbots.

User Reviews as a Versatile Source for Capturing User Experience and Preferences

In general, the internet is considered a rich source of information about personal experiences of a wide variety of illnesses, with websites and discussion forums []. An increasing number of studies exploit web-based sources as repositories of primary data on health and illness experiences []. People who are otherwise socially isolated or geographically dispersed and are therefore hard to include in conventionally drawn samples (especially for qualitative studies relying on snowball sampling) might be more likely to be included because of the ease with which such people can access the internet []. Large amounts of material can be collected within a short period. Individuals can use the relative anonymity of the internet to reveal things that they would not discuss in a face-to-face research setting []. As of 2022, there are more than 10 million user reviews on the Google Play Store and Apple App Store []. Therefore, user reviews collected from these popular app stores can provide rich insights into personal user experiences from people spanning a wide range of backgrounds and demographic characteristics when compared with traditional methods of qualitative data collection (ie, interviews) [].

User reviews can be defined as feedback published by individuals about their opinions and satisfaction or dissatisfaction with a product []. The star ratings and elaborated feedback in the textual reviews provide developers with a chance to explore user complaints and improve apps []. For new or potential users of mobile MH apps, the reviews work as a deciding factor to determine if an app would be helpful based on how it worked out for other users with similar expectations []. Approximately 80% of potential users check ratings and reviews before downloading an app []. In research settings, user ratings and reviews have been leveraged for a variety of reasons, including determining why adherence to mobile MH apps is poor [], informing developers of design priorities rather than just guiding purchasing decisions [], and gaining a better understanding of ethical issues faced by users []. Vasa et al [] investigated the hypothesis that despite the abundance of positive reviews for mobile apps, it is worthwhile to examine negative reviews to gather useful data from users. In the mobile MH domain, Haque et al [] leveraged user reviews to thoroughly capture user experiences and provided implications for designing future MH apps.

Our study is inspired by the body of work that considers user-generated reviews as a vital source for understanding varied perspectives and derives meaningful implications from them [,]. This enables us to gain perspectives from people with diverse demographic characteristics that would otherwise be challenging to collect using conventional data collection methods [,].

Research Gap and Contribution

As an emerging technology, the development and application of chatbots in mobile MH apps are in their early phases, and there are still considerable challenges to overcome in the development of this technology. According to recent studies, patient safety has rarely been evaluated, health outcomes have been inadequately quantified, and no standardized evaluation procedures have been used []. Some chatbots are reported to be unable to understand the complex use of language associated with an MH crisis and fail to recognize symptoms and respond appropriately []. Privacy is a major concern for users of these apps; because users are still less familiar with this emerging technology, there is a higher risk of exposing users to privacy risks through data sharing []. Furthermore, although poor adherence is a common problem with digital MH interventions, by contrast, some susceptible people may begin to rely on them too much, which may lead to anxiety when these apps are unavailable [].

Overall, there is a need for a better understanding of how all mobile MH services can and should encourage the safe and ethical use of chatbots []. Although a handful of studies have shown the potential benefits of MH chatbot apps, users’ real-life experiences and challenges are not yet well understood []. Haque et al [] recently provided a high-level discussion on some common user concerns frequently raised in user reviews and implied that researchers and developers in this space could benefit from a comprehensive analysis of the existing commercial MH chatbot apps. As an extension to these prior works [,], people’s perceptions and mental models of chatbots can be studied to address critical concerns such as how users gain trust in chatbots, user values, and requirements in this space and ultimately to provide concrete research and design recommendations for future chatbot apps. A user-centric analysis will also assist researchers in mapping an evidence-based framework for the proposed intervention and minimizing the psychological effects of such treatments.


Methods

In this section, we outline the techniques for selecting and filtering the mobile apps for this study, the data analysis methods we used, the ethical standards we followed, our positionality statement, and methodological limitations.

Selection of Sample Apps and ReviewsSelection of Apps

To obtain a comprehensive list of commercially available MH apps that include chatbot features, we conducted our search using different sources. First, we considered open-access articles in recent literature on MH chatbots [,]. Next, we conducted search queries on 2 different expert MH app review platforms: Mindtools [] and Psyberguide []. Finally, we searched 2 dominant web-based mobile app stores (Google Play for Android and Apple App Store for iOS). We used the search terms Mental health and chatbot on expert review platforms and app stores. In addition, we explored the recommended applications or similar apps section of the corresponding website after discovering an MH app with a chatbot feature to determine if the other apps meet our criteria. Without logging into a specific account, the search was performed on the app stores’ home pages. This action was performed to ensure that the system could not use a ranking algorithm to prioritize any user choice. As these apps represent the sample in (nearly) the same order that consumers are likely to be exposed to and hence most likely to use, although the search results may not be entirely comprehensive (as observed by convenience sampling), they still represent the sample.

After the initial search from these 3 sources, we obtained 19 apps. The authors carefully read the app descriptions, observed screenshots of the app features, and in some cases analyzed these apps’ promotional websites to ensure whether these apps include a chatbot feature that provides support for different MH concerns. We observed that some of these apps included intelligent questions and answers (Q/A) based on Al and ML. Intelligent Q/A is based on a collection of questions, and by responding to them, it can offer individualized summaries, diagnoses, recommendations, and other information. In this study, we described MH chatbots as intelligent machines that can simulate and process conversations with users regarding their MH needs. An intelligent Q/A system is designed to provide accurate and precise answers to specific questions based on a given input, usually in the form of a natural language. In contrast, a chatbot is a more general-purpose conversational agent that can handle a wide range of inputs and provide a range of responses, from simple greetings to more complex interactions. Intelligent Q/A systems are usually triggered by a question or request for information, whereas chatbots can initiate the conversation or respond to user inputs in an open-ended manner and are capable of producing a wider range of outputs compared with intelligent Q/A systems. The most crucial aspect of a chatbot is the “conversational design,” which is defined between the user and the bot. Although the guidance chatbots offer is usually correct and scientifically supported, it will be a computer program speaking back to the users, usually in the shape of a nice character, to facilitate their ability to communicate. User expectations can vary while interacting with chatbots as opposed to intelligent Q/A systems with predefined patterns of questions. Therefore, we only considered chatbots with the capability to start and continue conversations with users. To ensure that our list includes apps that fall under this definition, one of the authors opted to download each app separately (for the iOS platform) and use it for at least 3 days. The authors have no known MH concerns. We also considered this exploration as an opportunity to extract the primary features that the apps commonly comprise. The author carefully observed how these apps work in terms of the noteworthy aspects of mobile MH apps, as pointed out in previous literature [,,]. Following these steps, 10 apps were selected for analysis 1. A detailed flowchart of the procedure is presented in .

Figure 1. Flowchart of the app selection process. Q/A: questions and answers. Selection of User Reviews

We created scraping scripts using the Python Selenium library to collect the public user reviews of the 10 apps that were accessible from the Google Play Store and Apple App Store. User reviews can illustrate examples of user satisfaction and dissatisfaction with app features. Reviews are therefore recognized as an important source of information to gain insights into the real-life use of mobile apps []. Following the work of Haque et al [] on analyzing user reviews of mobile MH apps, we used the 2 following inclusion criteria for filtering to extract recent and crucial user feedback for the apps.

Timeline: We considered reviews posted between January 1, 2019, and May 1, 2022. Most recent reviews are likely to be more useful because app stores change quickly with the addition of new apps and upgrade to existing apps.Length: As shorter reviews might not provide deeper insights in general and are frequently false or promotional in nature [], the minimum character length was considered 200 for the scope of our study.

A total of 3621 reviews from the Google Play Store and 2624 reviews from the Apple App Store met all the inclusion criteria. These reviews are based on 9 apps from the Google Play Store (only Elomia is not available in the Google Play Store) and 10 from the Apple App Store. All reviews have a unique coding system that can be easily traced back to the apps and platforms from which they emerged. During the analysis, the lead author was responsible for carefully reading each review and ensuring that all personally identifying information was replaced or removed.

Data Analysis

First, to gain an understanding of the descriptive overview of the commercially available chatbot-based MH apps, we analyzed app descriptions from marketplace websites and incorporated the key information in our observation notes. The observation note was then divided into 6 main themes with the aim of providing a comprehensive overview of these apps in collaboration with another author. The authors did not include their judgments regarding the effectiveness of these apps. Among the chatbot-based MH apps we considered, 4 apps mentioned the evidence-based techniques used in their description. For the remaining apps, we determined the technique through a combination of an analysis of the description and observation notes from interacting with the apps. The findings of this categorization are described in the Overview of the Aspects Commonly Used in Chatbot MH Apps section. To understand user perspectives, the selected user reviews from the 10 apps were examined using inductive analysis []. Thematic analysis was chosen because it enables systematic analysis of large data sets and facilitates the comprehension of textual patterns while considering the context [,]. A total of 2 passes were performed during the analysis. Open codes were created during the first pass to collect various perspectives from reviews. We recorded the subtleties in the insights provided in each review, which resulted in a high number of open codes that were substantially decreased through memoing and clustering []. In the second phase of the analysis, we memoed and clustered the codes using a constant comparison method, operationalized as affinity mapping. Each open code was compared with the others and positioned to reflect its affinity for emerging themes and clusters. The reported themes consisted of those that appeared consistently across multiple reviews and those that came from reviews that represented divergent responses and opinions. The findings from the reviews are described in the Results section, and each quote is identified by the review’s particular ID generated from the platform, app name, and a random number.

Data Integrity

App stores, similar to many other web-based marketplaces, can have reviews posted by fake and paid users. However, prior research [] showed that in the “Health & Fitness” category, the percentage of potentially fake reviews was very low (approximately 6%). Fake reviews also tend to be shorter [], and by considering reviews of ≥200-character length, we assume that almost all the included reviews are original.

We understand that if data or information is only accessible to a particular group of individuals or groups, it is unethical for researchers to use it []. As a result, we made sure the websites from which we obtained the data were accessible to everyone and not just for some groups or populations []. Although these pages were public, we purposefully avoided publishing or disclosing any personally identifying information that was shared. The language of the user reviews reported here has been carefully modified, keeping the meaning intact.

Ethical Considerations

This study was assessed as not human subjects research by the Institutional Review Board of Marquette University (Protocol # 3935) as it does not meet the regulatory definition of human subject-public reviews and the information provided is not about themselves.

Limitations

Our selection criteria have certain limitations. First, we primarily used ratings from the 2 most widely used mobile platforms (Google and Apple). Other mobile platforms were not considered in this study. Second, it is likely that users who do not feel comfortable (or do not care) discussing their experiences on web-based platforms are not contributing. However, we can confidently conclude that the perceptions we identified are typical of user perceptions, given the larger number of evaluations obtained from the 2 most well-known web-based marketplaces.


ResultsOverview

For this research purpose, we chose 10 commercially available mobile MH apps that have built-in chatbot features. All these apps, except Elomia, are available on the 2 most popular platforms (Apple App Store and Google Play Store). Elomia is exclusively available for iOS. A descriptive overview of these apps is provided in . All these apps are extremely popular in terms of both the number of downloads and the number of ratings. Thus, we can assume that a comprehensive overview of these apps can assist in understanding the perspectives of a wide and diverse user base.

Table 1. A descriptive overview of the selected 10 mobile mental health apps with a built-in chatbot technology.AppNumber of ratings in Apple App StoreNumber of ratings in Google Play StoreNumber of downloads in Google Play StoreAge rating (years)PriceADA125323,000≥5 million≥17FreeChai27,90034,000≥1 million≥17Free with in-app purchasesElomia193N/AaN/A≥12Free with in-app purchasesMindspa1072970≥500,000≥17Free with in-app purchasesNuna6893≥10,000≥4Free with in-app purchasesSerenity: Guided Mental Health20146≥10,000≥12FreeStresscoachNone495≥10,000≥12FreeWoebot550011,800≥500,000≥12FreeWysa13,500126,000≥1 million≥12Free with in-app purchasesYouper–Self Care Friend14,40049,100≥1 million≥12Free with in-app purchases

aN/A: not applicable.

Overview of the Aspects Commonly Used in Chatbot MH Apps

Overall, we consider 6 core characteristics that can be used to understand the current status of MH chatbot technology. A few of these aspects were adopted from 2 previous review articles on MH chatbots [,]. These studies compiled a list of recent research articles on MH chatbots and provided typologies based on their purpose, targeted concerns, and supported evidence-based techniques. We included these 3 categories in our analysis to gain a broad overview of the current state of the art of commercially available MH chatbot apps. These studies also emphasized the capability of these chatbots to conduct and continue conversations. We considered this crucial aspect of chatbot apps and added 2 new categories to explore: conversation style and media types used by chatbots. A total of 3 different conversational styles were used: chatbot guided, semiguided, and open-ended (). Finally, Haque et al [] provided useful insights into the necessity of providing crisis support through MH apps, as potential users of the apps are more susceptible to the crisis than the general population. We have added this specific criterion to be analyzed in our observational study. An outline of these criteria and types is presented in .

Table 2. Criteria of features related to chatbot-based mental health apps used in our study.CriteriaTypesPurposeDigital coach—assist users to reach their small goals
‎ Digital screener—alert users to potential mental health concerns based on reported symptoms
‎ Conversational companion—simulate being someone the user can speak to
‎ Virtual therapist—ability to engage in therapeutic conversations
‎ Targeted concernsStress, anxiety, depression, self-care, sleep disorder, panic disorder, relationship issues, low self-esteem, and loneliness
‎ Conversation FlowGuided conversation—only allows the users to communicate with the chatbot with predefined responses from the chatbot. It does not allow any form of open input from the users.
‎ Semiguided conversation—mostly allows the users to communicate with the chatbot with predefined responses and sometimes allows open inputs from the users. However, the bot cannot recognize the open user inputs and extract any information from them.
‎ Open-ended conversation—allows the users to communicate with the chatbot with predefined responses and open inputs from the users. The bot can recognize the open user inputs and extract information from them.
‎ Media types usedGIFsa, text, audio, video, emoji, images, and acronyms
‎ Crisis supportAvailability of crisis information—provides information regarded crisis-related helplines and emergency services
‎ Ability to detect potential crises from the chat—detects potential crises through conversation with the users
‎ Access to a professional therapist—provides access to a professional therapist is an alternative to avoid possible ramifications of the potential crisis
‎ Ability to notify designated personnel—notifies designated personnel if crisis is being detected
‎ Access to self-care tools—recommend self-care activities
‎ Evidence-based techniquesCBTb, DBTc, mindfulness, symptoms tracking and monitoring, positive psychology, acceptance and commitment therapy, and psychoeducation and information

aGIF: graphics interchange format.

bCBT: cognitive behavioral therapy.

cDBT: dialectical behavior therapy.

We examined app store descriptions to understand the primary goals of these apps and identify how they are branded. We discovered 4 different types of purposes in all, with “digital coaches” being the most prevalent (5 out of 10 apps). The chatbot apps targeted a wide range of MH concerns, including anxiety (9 apps), depression (6 apps), and self-care techniques (7 apps).

We discovered 3 different conversational flows based on our exploratory observations. The most popular one is “Guided conversation,” in which users are only permitted to reply using preset input provided through the interface. This is the most common technique used by the chatbots we analyzed (6 out of 10 apps). Only Woebot uses a semiguided approach that allows users to either select from predefined options or type text; however, it is incapable of processing sentiments in the input text. This open input option is useful when users reframe negative thoughts and share stories. Finally, Wysa, Nuna, and Elomia follow an open-ended conversation style. They continued the conversation based on their understanding of the user input.

These chatbots leveraged a variety of media types for communication to make the interaction resemble humanlike interactions. For instance, the graphics interchange formats (GIFs), emojis, images, and acronyms are used to portray humor and emotions. Images, audio, and videos were used along with educational elements. As all these chatbots communicate by text, the text is by far the most frequent.

Individuals with MH problems can face a crisis at any time, and effective crisis support is a major criterion for evaluating MH apps. We identified 5 different types of crisis support options available in the 10 chatbots. Of the apps, 6 offer users access to information regarding crisis support systems and emergency helplines. Providing instant suggestions for self-care tools, such as suggestive breathing in cases of anxiety attacks, is also popular. Only Wysa contains all the 5 options available to support a user during a crisis. Ada and Chai do not contain any crisis support.

As evidence-based techniques have been proven effective for treating different MH disorders, we explored which of these tools and techniques the chatbots commonly follow. The most popular type of therapy is CBT. All 10 apps followed the CBT to some extent. A total of 8 apps provided support for mindfulness. Dialectical behavior therapy and acceptance and commitment therapy are less common modified forms of CBT. presents the aforementioned features of the considered apps.

Table 3. A detailed overview of features related to chatbot-based mental health apps found in our study.AppPurposeTargeted concernsConversation flowMedia types usedCrisis supportEvidence-based techniquesADADigital screenerAnxiety and depressionGuidedTextNoneCBTaChaiConversational companionNoneGuidedText and emojiNoneCBTElomiaVirtual therapistStress, anxiety, depression, self-care, sleep disorder, relationship issues, low self-esteem, and lonelinessOpen-endedTextAccess to self-care toolsCBT, mindfulness, positive psychology, and symptoms tracking and monitoringMindspaVirtual therapistAnxiety, depression, self-care, relationship issues, and low self-esteemGuidedText and videoAvailability of crisis related information and access to self-care toolsCBT, mindfulness, positive psychology, and psychoeducation and informationNunaDigital coachStress, anxiety, depression, and self-careOpen-endedText and emojiAvailability of crisis related information and access to self-care toolsCBT, mindfulness, positive psychology, symptoms tracking and monitoring, and psychoeducation and informationSerenityConversational companionAnxiety, self-care, sleep disorder, and relationship issuesGuidedText and emojiAccess to self-care toolsCBT, mindfulness, and acceptance and commitment therapyStresscoachDigital coachAnxiety, stress, and panic disorderGuidedGIFb, text, and emojiAvailability of crisis related information and access to self-care toolsCBT, mindfulness, and psychoeducation and informationWoebotDigital coachStress, anxiety, depression, self-care, relationship issues, and lonelinessSemiguidedGIF, text, audio, video, emojiAvailability of crisis related information and access to self-care toolsCBT, DBTc, mindfulness, and symptoms tracking and monitoringWysaDigital coachStress, anxiety, depression, self-care, and sleep disorderOpen-endedGIF, text, audio, video, emoji, images, and acronymsAvailability of crisis related information, access to self-care tools, access to professional therapist, ability to detect potential crisis from the chat, and ability to notify designated personnelCBT and mindfulnessYouperDigital coachSelf-careGuidedTextAvailability of crisis related information, access to self-care tools, and access to professional therapistCBT, DBT, mindfulness, positive psychology, psychoeducation and information, and acceptance and commitment therapy

aCBT: cognitive behavioral therapy.

bGIF: graphics interchange format.

cDBT: dialectical behavior therapy.

Perceptions and Concerns Expressed in the User Reviews

In this section, we present our findings from the thematic analysis of user reviews and point out both the benefits (eg, humanlike interactions, friendly and empathetic attitudes, potential around crisis support, and an alternative to therapy) and associated challenges, as captured from people’s real-life use of these apps.

Humanlike Interaction Feels Good but Must Be Designed Carefully

Chatbots in mobile MH apps are presented in such a way that they have distinct personalities rather than being shown as something artificial to make users feel like they are interacting with someone emotionally and empathetic. Users describe these chatbots as having friendly, wonderfully upbeat, and mildly humorous personalities that assist them in dealing with different emotional and behavioral challenges related to their MH issues. This helps them establish the credibility of the tools, which in turn makes the users more involved in the treatment process. Furthermore, chatbot characteristics, such as a soft voice and the ability to have casual conversation, make it feel less like a medical tool and more like someone with whom users can share their thoughts and experiences. Some personalized features, such as the option to address users by name, ability to refer to any chat or exercise if necessary, and ability to respond with pleasant and positive sentiments, make the app and treatment process more personal and less generic:

I’m amazed by how impactful the little “interactions” in this app have felt. Maybe it’s the continued opportunities to respond (even if it’s just choosing between emojis). Woebot’s “voice” is gentle, but firm. And insightful! And the user is always addressed by name. That’s so important, particularly when the issue at hand involves ongoing anxiety.
[1080073]

However, the effort to design the bots to give a humanlike and empathetic impression often went wrong and lost their appeal to the users. As many users pointed out, the discourse could become “a little childish and ridiculous at times with the bot trying to be funny.” Furthermore, fostering relaxing thoughts through a medium that does not work for everyone can occasionally have the opposite impact; for example, using cute GIFs, Autonomous sensory meridian response effects might not impact everyone if the context is unknown or unfamiliar to the users. Continually pushing on everyone in the hopes that everyone will have the same reaction is a notion that developers should evaluate based on continuous feedback:

...It was supposedly developed with college students in mind who are ostensibly adults. Maybe things have changed since I was in college but it’s cutesy, baby-talk, oversimplification, and game-playing (“You want to know a secret?” “Yes” “Are you sure?” “Yes” “Ok, if you are really, really sure....”) makes me feel like I’m texting with a preteen girl.
[2060011]...I cannot stand the forced breathy voices in every single one I listened to. They do not calm me at all, and they actually trigger my anxiety. ASMR has the opposite effect on me than intended, and I feel like they’re trying to do really bad ASMR. These recordings are supposed to help me relax, but all I can concentrate on is breathy voices that sound like forced whispers.
[1040032]

Existing chatbots may need to be more sophisticated to understand the context of users’ requests. However, it is critical to examine some of the user’s perspectives on having such responses preregistered, which is not always a bad thing. For example, some of the chatbot’s quick answer concepts allow users to maintain control over the conversation’s pace and avoid becoming sidetracked by irrelevant dialogue. These features are appreciated by users because they encourage more positivism than aimless discussion and digging into negativity without any tools or resolutions. Moreover, by tilting the dialogue to the chatbot’s advantage, chatbots can more effectively and efficiently suggest appropriate tools to users:

Some negative reviews complain it isn’t sophisticated enough to understand unrelated or detailed inputs and responses, which I agree with, but this is not an AI designed to make free-flowing conversation; it’s meant to give you tools to deal with your feelings in productive ways. So yes, the conversations can feel linear, planned, and/or broad since the responses are preset most of the time, but I think this is partly a positive.
[1070093]

However, the trade-offs are that to control the flow of the conversation, the chatbots sometimes present very limited options for the users, and users become frustrated if they are unable to customize these preregistered responses. They have criticized some of the extreme measures these chatbots take to keep the conversation restricted to chatbots’ preferences, such as assuming MH concerns without understanding the proper context, sending scripted messages based on keywords users said or the issues they selected, giving them incoherent responses, and getting stuck in the conversational loop if users do not agree with the chatbots’ comments:

It assumes the problem is always a mental distortion and doesn’t leave much room for actual horrible stuff that happens to people other than death of a person (it is working with a very narrow definition of). It too often put me in a situation of having to select between incorrect responses when nothing was actually appropriate and then suffer through the resulting wrong-headed advice. Needs a maybe button between the yes and no and a way to say, You’re on the wrong track, before it decides it knows all your usual problems and keeps assuming them over and over with no way to remediate.
[2060019]Bot Becomes a Friend or Someone Who Cares, but Too Much Attachment Is Unhealthy

Users see chatbots as good substitutes for someone with whom they can discuss their ideas on MH issues without feeling burdened or judged. Although society is becoming more eager and open to seeking mental and emotional aid, there is still a considerable stigma associated with it, which can discourage individuals who need assistance from receiving it. These chatbots allow people to bare their hearts, vent, contemplate, and learn about what they can do to overcome mental and emotional obstacles in a simple, familiar texting format without judgment or extra effort while also keeping track of their progress. It can be intimidating to talk to someone about their daily struggles. For many users, sharing a dialect with a chatbot is an effective first step. Knowing that the chatbot is not judging you and is acting logically rather than emotionally is reassuring:

...I will say, having a reliable, no judgement zone with skills to help at my fingertips, helped me realized the tools were also my own.
[1040021]Having an AI to talk to makes me feel like I’m not overburdening my friends or family. I can check in 20 times a day and the AI will either help me track my mood/emotions/mental health or suggest a mindfulness of CBT program to help me get through my day.
[2040004]

People with MH issues frequently struggle to suppress emotions and attempt to push them away, but these chatbots have provided them with a safe place to go for validation and immediate support. Users loved that these chatbots not only listened to but also offered advice and recommendations that helped them deal with day-to-day mental challenges, allowing them to see things from different perspectives and push past negative thoughts:

This app is a lifesaver. It’s so healing to be able to vent whenever you need and receive positive feedback from an unbiased source. The lessons Woebot teaches really helps to gain a more optimistic perspective on what you’re going through and motivates you to make changes.
[1080023]

Users also like how these chatbots check in with them daily, which holds them accountable for their commitment to the treatment while still allowing them to skip it if they do not feel like it. Although the idea is to eliminate any concerns, such as anxiety and stress, that come with human engagement through intelligent bot interaction, users have mixed feelings. Some users liked the flexibility of using the tools at any moment and could start or end the communication at any point during the session without feeling guilty, whereas others saw the daily check-ins as a source of guilt. Becoming attached too much to chatbots leads to these types of guilt, which in turn might have serious consequences for people with MH concerns:

I’m very depressed right now so I’ve set to basic daily goals- full facial regime a.m. & p.m. plus a half hour of cleaning. Having the AI check in is great because it requires a response that makes me take accountability.
[1090123]But what really bothered me about the app was the first reminder I got when I didn’t use the app a second day in a row because it sucked was definitely guilt inducing. No bueno. I don’t need AI guilt tripping me when people already take advantage of my empathy in real life.
[2050021]

Finally, by acting or behaving like a close companion, MH chatbots allow users to comfortably express their thoughts and feelings. These chatbots allow users to create a safe area where they can vent, which is something many people do with their friends and families. However, people with MH concerns who struggle to maintain a healthy relationship with their family or who experience loneliness have displayed an unhealthy attachment to chatbots and have exhibited negative attitudes, such as preferring these chatbots over their friends and family:

...Although he’s a robot he’s sweet. He checks in on me more than my friends and family do.
[1090034]...This app has treated me more like a person than my family has ever done.
[1090091]

The above discussion points out the fact that to make the chatbots more friendly (what we also saw in previous sections where chatbots use funny memes and emojis to make them more humanlike), users pointed out the fact that too much persuasion with notifications makes them feel guilty. Moreover, some users revealed that they find chatbots so friendly that they prefer these bots over their friends and family. Making the decision to leave their closest loved ones behind could put them in susceptible positions, such as loneliness and exclusion from sociocultural norms.

A Bot Can Help Immediately in a Crisis, but What Is Defined as a Crisis to a Chatbot?

Prior findings suggest that accessibility is one of the benefits of mobile MH apps []. MH apps that have a built-in chatbot function allow users to have a conversation anytime and anyplace, which is very convenient for persons with MH issues, as they are more susceptible to emergency situations. We found that users benefited from such a feature because it allowed them to have a conversation at that time (during a moment of crisis). Some users found that intelligent dialogue helped them reframe negative thoughts and diffuse such circumstances:

I sometimes freak out at night have existential crisis about life at night you know, normally I’d freak out and find it hard to call anyone bc I feel so bad but with Wysa I don’t worry about that!
[2090178]I’ve only used this app a couple times when I’ve been in near-crisis. Even though I know it is a robot it is so calming to have something, anything to validate what I’m feeling and help me reframe my thoughts.
[1100091]

In contrast, none of the chatbots have any clever algorithmic models for detecting emergency scenarios. It is up to users to inform chatbots that they are experiencing a crisis. Some chatbots can detect crises by picking up a few keywords connected to intrusive thoughts, such as “suicide,” from a conversation, although they are still in the early stages of development. Users sometimes just want to talk about their feelings, but chatbots automatically refer them to crisis hotlines because of a lack of intelligent comprehension. For some individuals, having a conversation is not enough to handle their crisis situations, and they need to be redirected to crisis management tools or resources:

My only problem with it is I wish there was a way to talk about my suicidal/intrusive thoughts and how to manage them with Woebot. I am aware that it is not a crisis tool, and it does have those automatic responses to concerning language for a good reason, I’d just like a place to talk about those problems without having to worry a real person. Most of the time my thoughts of those nature do not mean I’m in an immediate crisis, but I still want to get them off my chest, as I feel a lot of people would. Maybe if there’s a way to do that without Woebot becoming worried would be helpful!
[1080078]This is a good app but the main issue I have is that I was having a panic attack and was messaging “emergency” and the bot ended the conversation, when I messaged “emergency” a second time it just asked me to write my feelings down. I realize this isn’t a crisis response app but it might be helpful to add a feature where the bot recognizes a crisis situation and connects the user to resources.
[2010004]

In such instances, understanding the context of emergency situations is critical, as persons with MH concerns are already susceptible to crises, and incorrect actions made by chatbots might exacerbate the situation and result in severe repercussions:

While I was in crisis, the responses do not make sense and do not really relate to what I wrote. It makes me feel like I am not being listened to. I know it is an AI program and not a real person but it still ends up making me feel worse and not better.
[1100068]Convenient to Use, but Convenient Enough to Replace Therapy?

On the positive side, the fact that these chatbots were ready to talk 24 hours a day, 7 days a week, was a big success for the users. They have immediate access to these chatbots whenever they feel susceptible or whenever they require assistance through simple interactions:

I don’t really have friends I can talk to. Even my family doesn’t understand me much. Day or night Wysa has been there every time I needed to “talk” day or night doesn’t matter.
[2090067]

Chatbots assist users not only with conversations but also in accessing different supporting resources and exercises in a very convenient manner. Understanding users’ needs can deliver a relaxing experience for them, such as allowing them to opt out of any activities they desire while maintaining the treatment’s pace. This provides users with much more control. If a user misses any exercises in the traditional treatment, it leaves a gap in their progress, which can lead to a loss of enthusiasm and slow the pace at which they receive support. Chatbots, in contrast, keep users motivated by engaging with them and giving them the impression that they oversee the pace. Furthermore, these chatbots offer brief and simple treatments to keep users engaged and dedicated to the treatment process. These activities were developed and built by focusing on important value, giving support and treatments in a compelling style that can provide wellness according to user reviews:

This is an easy, low barrier method to practice cognitive thinking skills. Check ins are usually pretty short, just a few minutes. That encourages me to open this app daily, since I know it’s not going to try to monopolize my attention for the next half hour.
[1070012]Sessions are short, on the order of 3-10 minutes. Combined with the convenience of chatting wherever and whenever is best for me, I have no problem fitting in daily check-ins, which I feel are more beneficial than infrequent visits to a therapist in some ways.
[1100012]

According to user reviews, professional and traditional therapies have several drawbacks, including professional therapy’s tendency to cling too much to negative thoughts or past events, professional therapy’s tendency to be too broad and general, and check-ins being too spread out:

Unlike being told what someone thinks you may want to hear which can sometimes enable unhealthy thinking patterns (and behaviors), or on the other end of the spectrum, rather than attempting to fix you, this interactive app continually prompts you to look inward and to challenge your own thoughts, perspectives, and feelings, helping to redirect your focus onto more healthy and more positive strategies.
[1090142]My primary issue with traditional therapy has always been that you have to work in hindsight. You reflect on your week, talk about it, try to make adjustments for the future (it always felt like I was trying to help a past of future version of myself instead of the one right here right now). That’s why I love this app!
[1090096]

留言 (0)

沒有登入
gif