Abstract
Background: Health professionals face significant psychological burdens including burnout, anxiety, and depression. These can negatively impact their well-being and patient care. Traditional psychological health interventions often encounter limitations such as a lack of accessibility and privacy. Artificial intelligence (AI) chatbots are being explored as potential solutions to these challenges, offering available and immediate support. Therefore, it is necessary to systematically evaluate the characteristics and effectiveness of AI chatbots designed specifically for health professionals.
Objective: This scoping review aims to evaluate the existing literature on the use of AI chatbots for psychological health support among health professionals.
Methods: Following Arksey and O’Malley’s framework, a comprehensive literature search was conducted across eight databases, covering studies published before 2024, including backward and forward citation tracking and manual searching from the included studies. Studies were screened for relevance based on inclusion and exclusion criteria, among 2465 studies retrieved, 10 studies met the criteria for review.
Results: Among the 10 studies, six chatbots were delivered via mobile platforms, and four via web-based platforms, all enabling one-on-one interactions. Natural language processing algorithms were used in six studies and cognitive behavioral therapy techniques were applied to psychological health in four studies. Usability was evaluated in six studies through participant feedback and engagement metrics. Improvements in anxiety, depression, and burnout were observed in four studies, although one reported an increase in depressive symptoms.
Conclusions: AI chatbots show potential tools to support the psychological health of health professionals by offering personalized and accessible interventions. Nonetheless, further research is required to establish standardized protocols and validate the effectiveness of these interventions. Future studies should focus on refining chatbot designs and assessing their impact on diverse health professionals.
doi:10.2196/67682
Keywords
Introduction
Health professionals are on the front line of patient care, enduring enormous responsibilities and constant pressure. Their work often occurs in an environment in which they must make life-and-death decisions, making them particularly vulnerable to psychological burnout [
]. According to recent studies, more than 50% of health professionals worldwide experience symptoms of burnout, such as fatigue, cynicism, and decreased effectiveness, leading to low job satisfaction [ ], high turnover [ ], and low-quality patient care [ ]. This phenomenon not only causes an individual crisis but also has broader implications for health care systems [ , ]. Since the COVID-19 pandemic, psychological health issues such as anxiety, depression, and burnout among health professionals have become more prevalent because of long working hours, high-stress working environments, and exposure to traumatic events [ ]. These issues directly affect health professionals’ well-being and work performance and can ultimately negatively impact the quality of health care [ ]. Therefore, supporting the psychological health of health professionals and preventing burnout are urgent challenges.While traditional face-to-face counseling has been used to prevent psychological illness among health professionals, this method presents various limitations including temporal and spatial constraints, limited availability of mental health professionals, accessibility barriers, and anonymity concerns [
- ]. Health professionals face significant challenges in accessing psychological care owing to demanding work schedules and high work intensity [ ]. Additionally, they may feel psychologically burdened when seeking help in situations in which anonymity is not guaranteed [ ]. To address these challenges, artificial intelligence (AI) chatbots have emerged as a potential solution, offering a new way to deliver psychological health interventions that can compensate for the limitations of traditional methods [ , ]. AI chatbots have the potential to be accessible anytime and anywhere while maintaining user anonymity, which may provide an alternative approach to psychological health support [ ].Recently, rapid advances in AI and natural language processing (NLP) have opened new possibilities for psychological health interventions, with AI chatbots gaining particular attention in the field of psychological health. Studies have demonstrated the potential of AI chatbots—AI-driven conversational agents—in providing mental health support [
] while emphasizing the importance of their safe and explainable implementation [ ]. AI chatbots analyze users’ conversation histories and data to provide advice and support tailored to their specific needs. This can help health professionals better manage psychological health issues, such as burnout, and enable timely interventions tailored to individual situations [ ]. Previous research has shown that AI chatbots can improve coping skills through natural conversations with users [ ], enhance the user experience, facilitate effective intervention delivery [ , ], and provide interventions tailored to specific needs [ , ]. This conversational approach enhances the effectiveness of interventions by fostering positive user engagement, as people are more likely to empathize with and respond to tools that feel engaging and relatable [ , , ].Despite the potential of AI chatbots for enhancing psychological health, there remains a lack of research on their effectiveness for health professionals. Much of the existing research has focused on using AI chatbots in patient care, whereas relatively little research exists on the use of AI chatbots to support health professionals. This research gap highlights the need to systematically investigate the characteristics and effectiveness of AI chatbots that target health professionals. Therefore, this study aimed to fill this gap and analyze the characteristics and effectiveness of AI chatbot interventions targeting health professionals to contribute to future research and practice.
Methods
Study Design
We conducted a scoping review to identify the available evidence regarding AI chatbot interventions for health professionals’ psychological health. A scoping review maps the relevant literature in the field and is useful for identifying gaps in knowledge [
]. This review was conducted in accordance with the scoping review framework of Arksey and O’Malley. The study adopted the following five steps: (1) identifying the research question; (2) identifying relevant studies; (3) study selection; (4) charting the data; and (5) collating, summarizing, and reporting the results. The five stages provided a structured approach to assist in understanding the current evidence by examining the literature. There was no rating of quality performed on the included studies, as scoping studies do not require a quality evaluation [ ]. Therefore, a scoping review is suitable for providing a broader perspective on the current research and mapping the relevant literature on AI chatbot interventions for health professionals’ psychological health.Identifying the Research Question and Relevant Studies
The research question was, “What is the status of the current body of knowledge regarding AI chatbots for the psychological health of health professionals?” According to Arksey and O’Malley, the search strategy involves identifying relevant studies, including published and unpublished works, with an emphasis on being as comprehensive as possible [
]. A literature search was performed between May and August 2024 from the following eight databases: seven global search engines in English (Cochrane Library, CINAHL, Embase, ProQuest Dissertations & Theses Global, PsyclNFO, PubMed, and Web of Science) and one Korean search engine in Korean (Research Information Sharing Service). Various combinations of search terms ( ) were used to either broaden or narrow the search, depending on the results in a specific database. The reference lists of the retrieved studies were searched manually. Finally, gray literature, relevant letters, and editorials were searched.medical team OR medical staff* OR healthcare worker* OR healthcare provider* OR health personnel* OR health professional* OR doctor* OR surgeon* OR physician* OR nurse* OR midwife* OR Licensed Practical Nurse* OR licensed vocational nurse* OR physician assistant OR nurse practitioner OR clinical nurse specialist* OR nurse clinician* OR registered nurse* OR resident* OR therapist* OR pharmacist* OR nurse assistant* OR nurse aid* AND burnout OR burn-out* OR depress* OR depression OR stress* OR exhaust* OR depersonalization OR personal accomplishment OR anxiety* AND chatbot* OR artificial intelligence OR conversational agent* OR dialogue system* OR machine learning OR natural language process*
Study Selection
Overview
Studies published before May 2024 were selected for this review. The studies retrieved from the literature search were reviewed by three researchers using the following inclusion and exclusion criteria.
Inclusion Criteria
(1) The purposes of the studies included the development or implementation of AI chatbot interventions. (2) Participants of the studies included licensed health professionals such as physicians or surgeons, dentists, medical doctors, registered nurses, licensed practice or vocational nurses, pharmacists, and allied health professionals such as respiratory and physical therapists. Studies that included nonhealth professionals in addition to health professionals as part of their participants were included if the main purpose of the AI chatbot was to alleviate the psychological distress of health professionals. (3) The intention of the AI chatbots was to improve the psychological health of health professionals, focusing on outcomes such as stress, burnout, depression, and anxiety. (4) The studies were written in English or Korean.
Exclusion Criteria
Studies in which we could not retrieve the full text were excluded.
A total of 2465 studies were retrieved from the initial database search (
). After removing duplicates, 1766 studies remained. The titles and abstracts were reviewed, and 1750 studies were excluded, leaving 16 studies for a full-text review. Six studies completely met the inclusion criteria after full-text review. Then, by reviewing the reference lists and manually searching, five additional studies were obtained, of which four studies met the criteria for the scoping review. Finally, 10 studies were included in this scoping review.Throughout the review process, two researchers (GB and JH) took the lead in data extraction, and all researchers reviewed the content. Any uncertainties and disagreements were resolved through weekly meetings. The literature review software Covidence (Veritas Health Innovation Ltd), a study-managing tool for collaborating researchers [
], was used to manage the retrieved studies and to screen for redundant studies.
Charting the Data
The data are presented in tables, summarized, and synthesized qualitatively according to the purpose and research questions of the review.
Collating, Summarizing, and Reporting the Results
The final stage of a scoping review involves collating, summarizing, and reporting the results [
]. Data extraction was performed by two researchers (GB and JH) and checked for accuracy by all the researchers. Uncertainties and disagreements were resolved through discussions. The data were extracted using a common data extraction format.Results
General Characteristics
presents the characteristics of the analyzed studies. Ten studies published between 2018 and 2024 were retrieved. The studies were published worldwide: in Canada [ , ], Japan [ ], Malawi [ ], Netherlands [ ], Singapore [ , ], Spain [ ], and the United States [ , ]. Four studies examined diverse health professionals, including doctors, nurses, nurse assistants, clinical officers, pharmacists, lab technicians, and other allied health professionals [ , , , ]. Two studies specifically targeted doctors and nursing staff [ , ] such as nurse practitioners, registered nurses, licensed practical nurses, and nursing assistants, while one study exclusively covered chatbots for nurses [ ]. Three studies identified their participants as health professionals, without specifying the specific types [ , , ]. Of the ten studies, eight studies focused solely on health professionals [ , - , , ]. Two studies included both health professionals and their families [ , ].
The study designs were diverse. Of the 10 studies, three were development studies [
, , ], one was a design verification study [ ], one was a pilot study [ ], one was a randomized controlled trial [ ], one was a mixed-method feasibility study [ ], one was an intervention study [ ], one was a cross-sectional study protocol study [ ], and one was a qualitative study [ ]. Three studies evaluated the development and feasibility of AI chatbots for psychological health support [ , , ].Author (year)/ country | Study design | Participants | Source of participants | Sample size (M:F ratio) | Age (years), mean (SD) | Publication type |
Anmella et al [ | ](2023)/SpainDevelopment study | Primary care and health care professionals; medical doctors (GPs | and specialists from diverse disciplines), nurses, and nurse assistants
|
|
| Journal study |
Chaudhry and Islam [ | ](2021)/United StatesDesign verification study | Health professionals in their workplaces during the COVID-19 pandemic; RNs | , NPs , LPN , CNAs , PAs , RT , PTs , students or intern nursesLocal medical services departments and health professionals’ mailing lists | 22 (9.1:90.9) | 34.98 (8.55) | Journal study |
Chang et al [ | ] (2024)/SingaporePilot exercise (service evaluation study) | Health professionals (types unspecified) | National tertiary health care cluster | 527 (N/A) | N/A | Journal study |
Jackson-Triche et al [ | ] (2023)/United StatesDevelopment and usability study | Health professionals including faculty, staff, and trainees across various disciplines within the health care system | UCSF | Over 26,000 (N/A) | N/A | Journal study |
Kleinau et al [ | ] (2024)/MalawiRandomized controlled trial | Health professionals (doctors, nurses, clinical officers, medical assistants, pharmacists, lab technicians, physiotherapy technicians) | Public and private health care facilities within Blantyre and Lilongwe districts | 1584 (N/A) Completed: 511 (E=296, C=215) | N/A | Journal study |
Kroon [ | ](2018)/NetherlandsMixed method feasibility study (one group pre-posttest design) | Health professionals (radiation expert, social worker, nurse in training, training officer, pedagogue, health professional disability care HRD | adviserEmail using personal and iThrive connections. | 12 (8.3:91.7) (interview participant=9) | Half of the participants were aged 18 to 25 years | Master’s thesis |
Matsumoto et al [ | ](2021)/JapanIntervention study | Nurses | Two hospitals in Tokyo | 70 (N/A) | N/A | Journal study |
Noble et al [ | ](2022)/CanadaCross-sectional study protocol | Health professionals and their families | The Canadian provinces of Alberta and Nova Scotia | 2200 (N/A) | N/A | Journal study |
Yoon et al [ | ] (2021)/SingaporeQualitative study | Frontline health professionals (doctors and nurses) | Singapore general hospital and community care facilities (large scale institutional isolation units) | 42 (24:76) | 29.6 (3.9) | Journal study |
Zamani [ | ](2022)/CanadaDevelopment study | Health professionals and their families | The Canadian provinces of Alberta and Nova Scotia | N/A | N/A | Master’s thesis |
aM:F: male:female.
bGP: general practitioner.
cPCMHSP: Primary Care Mental Health Support Program.
dRN: registered nurse.
eNP: nurse practitioner.
fLPN: licensed practical nurse.
gCNA: certified nurse assistant.
hPA: physician assistant.
iRT: respiratory therapist.
jPT: physical therapist.
kN/A: not available.
lUCSF: University of California, San Francisco
mHRD: human resource development.
AI Chatbot Characteristics
shows the characteristics of the AI chatbots developed or implemented in the retrieved studies. The types of AI chatbot algorithms were diverse; NLP-based AI was the most commonly used algorithm, appearing in six studies [ , , , , , ]; rule-based AI, an algorithmic method that does not learn from data or create rules on its own, was used in two studies [ , ]; and learning-based AI that learns from data was adopted in two studies [ , ]. AI chatbots were delivered through mobile apps in six studies [ , , , , , ], and web-based platforms via web links in four studies [ , , , ]. The interaction type in all studies was 1:1 interaction: Among them, one study included 1:1 interaction alongside group chat functionality [ ], and another study included 1:1 interaction with optional human coach support [ ]. In seven studies, the chatbot actively initiated the conversation [ , , - , , ], while in three studies, the conversation was passively initiated [ , , ].
Participant completion of the AI chatbot interventions was reported in four studies [
, , , ]; among these, three studies provided the percentage of participants who completed the intervention: 14.3% [ ], 67.7% [ ], and 73% [ ] while one study reported the employee access rate of the intervention as 10.88% [ ]. Of the 10 studies, six evaluated the AI chatbot’s usability [ , , , , , ]. Four studies relied on participant feedback: 54.6% rated it as useful [ ], the average satisfaction score was 4.07 out of five [ ], more than 80% reported it as helpful [ ], and 91% of the users found the chatbot easy to use [ ]. One study used user engagement indicators to identify high subjective usability (acceptability, usability, and satisfaction) but low objective engagement (completion, adherence, compliance, and engagement) [ ]. Another study assessed chatbot performance metrics, reporting an intent detection accuracy of 99.1% and an entity extraction accuracy of 95.4% for the Mira Chatbot [ ]. Chatbot usability was not evaluated in the remaining four studies [ , , , ].Author (year)/ Country | Chatbot name | Purposes of chatbot | AI chatbot type | AI algorithm | Mode of delivery | Interaction type | Initiates conversation | Completion related | Chatbot usability evaluation |
Anmella et al [ | ] (2023)/SpainVickybot | To screen anxiety, depression, and burnout with interventions | Digital decision support platform (PRESTO | )
| Mobile app | 1:1 interaction | Active | 27% completed second self-assessment |
|
Chaudhry and Islam [ | ] (2021)/United StatesiChatBot (AI O) PeerChat (AI X) | To provide stress management support during COVID-19 | Intelligent chatbot |
| Mobile app | iChatBot: 1:1 PeerChat: group chat | Passive (on-demand queries) | Not reported |
|
Chang et al [ | ] (2024)/SingaporeWysa | To evaluate Wysa for health professionals’ support | Conversational chatbot |
| Mobile app | 1:1 interaction with optional human coach support | Passive | Not reported |
|
Jackson-Triche et al [ | ]/United StatesUCSF | CopeTo provide behavioral health support via chatbot | An algorithm-based, automated, web-based AI conversational tool |
| Web-based interface | 1:1 interaction | Active | 10.88% of employees accessed the chatbot |
|
Kleinau et al [ | ] (2024)/MalawiViki (Vitalk’s virtual mental health care assistant) | To assess chatbot impact on mental health in Malawi | Automated conversational agent |
| Mobile app | 1:1 interaction | Active | Of 1584 participants, 511 completed |
|
Kroon [ | ] (2018)/NetherlandsiThrive | To help health professionals manage inner critics and enhance self-compassion | Preprogrammed conversations and interaction |
| Web-based interface | 1:1 interaction | Passive (users initiate interaction with the chatbot) | 12 of 14 participants included in the final analysis |
|
Matsumoto et al [ | ] (2021)/JapanCB | app (a chatbot app)To support mental health self-care for busy professionals, especially nurses | Digital decision support platform |
| Mobile app, VR | 1:1 interaction | Active | Not reported |
|
Noble et al [ | ](2022)/CanadaMIRA | To provide tailored mental health information and services for health professionals and their families | Hybrid NLP | and decision tree AI chatbot
| Web-based platform | 1:1 interaction | Active | Not reported |
|
Yoon et al [ | ](2021)/SingaporemHealth | appTo support frontline health professionals’s psychosocial well-being during COVID-19 | N/A |
| Mobile app | 1:1 interaction | Active | Not reported |
|
Zamani [ | ] (2022)/CanadaMIRA | To offer strategic mental health resources for health professionals and their families | Hybrid NLP and decision tree AI chatbot |
| Web-based platform | 1:1 interaction | Active | Not reported |
|
aAI: artificial intelligence.
bPRESTO: primary care digital support tool in mental health.
cUEI: user engagement indicator.
dUCSF cope: University of California, San Francisco faculty, staff, and trainee coping and resiliency program.
eNot available.
fCB: cognitive behavioral.
gVR: virtual reality
hMIRA: mental health intelligent information resource assistant.
iNLP: natural language processing.
jmHealth: mobile health.
AI Chatbot Intervention Characteristics and Outcomes
lists the contents and outcomes of the AI chatbot interventions. The primary component is the screening function for mental health conditions such as anxiety, depression, and stress, which was included in seven studies [ - , , ]. Four studies included cognitive behavioral therapy techniques [ , , , ]. Additionally, four studies implemented stress management techniques, such as breathing exercises and meditation [ , , , ]. Three studies incorporated problem-solving strategies [ , , ] and psychoeducation [ , , ]. Two studies used behavioral activation [ , ], cognitive restructuring [ , ], and gratitude training [ , ]. One study featured storytelling [ ]. In addition, there was a function to support portal sites connected to community resources in three studies [ , , ].
Author (year)/ country | Chatbot contents | Additional features | Intervention doses | Main outcomes |
Anmella et al [ | ](2023)/Spain
|
|
|
|
Chaudhry and Islam [ | ] (2021)/United States
|
|
|
|
Chang et al [ | ](2024)/Singapore
|
|
|
|
Jackson-Triche et al [ | ] (2023)/United States
|
|
|
|
Kleinau et al [ | ](2024)/Malawi
|
|
|
|
Kroon [ | ](2018)/Netherlands
|
|
|
|
Matsumoto et al [ | ](2021)/Japan
|
|
|
|
Noble et al [ | ](2022)/Canada
|
|
|
|
Yoon et al [ | ](2021)/Singapore
|
|
|
|
Zamani [ | ](2022)/Canada
|
|
|
|
aAI: artificial intelligence.
bCBT: cognitive behavior therapy.
cDBT: dialectical behavioral therapy.
dACT: acceptance and commitment therapy.
eN/A: not available.
fGAD-7: general anxiety disorder-7.
gPHQ: patient health questionnaire.
hUCSF cope: University of California, San Francisco faculty, staff, and trainee coping and resiliency program
iOLBI: Oldenburg burnout inventory.
jUCLA: University of California, Los Angeles loneliness scale.
kRS-14: 14-item resilience scale.
lProQOL: Professional Quality of Life Scale.
mSCS: Self-Compassion Scale
nSAT: structured association technique.
oVR: virtual reality.
pVR-CB: virtual reality-chatbot.
Across the studies, a variety of additional features were incorporated. Seven studies included AI chatbot functionalities for crisis management and support [
, , , , , ]. Specific features encompassed emergency contact information for crises [ , , , ], peer support groups and in-app forums for social support [ ], expert support through in-app counseling, and options for connecting with human coaches [ , , ]. Six studies incorporated reminders to encourage chatbot use [ , , , , , ]. Gamification elements, such as point-based rewards to encourage participation, were featured in four studies [ , - ]. Peer support functionalities were available in three studies [ , , ]. Two studies included mood-tracking features for health and lifestyle management [ , ], while another two studies provided an audio recording option [ , ]. Additionally, one study integrated devices for tracking vital signs [ ] and provided an immersive virtual reality experience [ ]. A total of 4 out of 10 studies explicitly stated the duration of chatbot use or intervention dose ([ ]: 4 wk; [ ]: 8 wk; [ ]: 2 wk; [ ]: 1 mo) while it was not specified in six studies. The dose varied among users in two studies [ , ].The research in four studies evaluated psychological health outcomes [
- , ], whereas, in six studies, the focus was primarily on the development and description of user perceptions of the AI chatbot without providing specific outcome measures [ , , , , , ]. The most commonly measured psychological health outcomes were anxiety and depression, which were examined in three studies [ , , ]. Burnout was investigated in two studies [ , ]. Compassion fatigue [ ], loneliness [ ], resilience [ ], and self-compassion [ ] were also evaluated. Anxiety was measured in three studies: one reported a significant reduction of scores [ ] while the other two reported no significant differences in anxiety scores [ , ]. Burnout was assessed in two studies, and each reported a significant decrease in its scores [ , ]. Depressive symptoms were measured in two studies: one reported a significant reduction of scores [ ] while the other did not [ ]. Other psychological variables were measured in one study each; of those, the variables that showed significant change in scores were health counseling needs [ ] and resilience [ ].Discussion
Principal Results
The studies retrieved for this scoping review were published during the past seven years, between 2018 and 2024, and reflect the rapidly evolving trend of AI chatbot technology for supporting psychological health. The global distribution of studies across countries indicates a widespread interest in using AI for psychological health support and health professionals’ psychological health.
Our review revealed that chatbots were used by a broad spectrum of health professionals, highlighting their diverse applications. This diversity underscores the potential for universally designed chatbots to effectively address psychological health needs across different professional settings. However, prior research has shown that nurses, who have direct and continuous patient contact, experience greater mental health distress compared to other health professionals [
]. Therefore, future studies should consider incorporating role-specific options into chatbot designs to create tailored interventions for diverse professional groups.The diverse research methods used in the retrieved studies signify the nascent stage of research on AI chatbots for psychological health support, emphasizing technology refinement and user experience comprehension. Such a variety of research approaches is crucial for validating the efficacy and safety of AI chatbot technology, enabling iterative improvements and a deeper understanding of how health professionals interact with AI chatbots [
]. However, it is noteworthy that only 4 out of 10 studies measured psychological health outcomes. This limitation underscores the significant gap between technological developments and clinical validation. As the field matures, it becomes imperative to quantitatively assess the impact of AI chatbots on psychological health indicators. Further studies measuring outcome variables to demonstrate the clinical effectiveness of AI chatbots are needed.The studies reviewed here demonstrate the variety of AI algorithms used in chatbot design, including NLP-based-, rule-, and learning-based approaches. NLP-based AI chatbots which provide personalized responses by analyzing user inputs [
] are the most commonly used AI algorithms. Evidence has shown that an NLP-based AI chatbot, which simulates natural conversations, is successful at delivering tailored interventions [ ]. Rule-based AI algorithms, adopted in 2 studies out of the 10 retrieved studies, are effective in providing fixed and predetermined responses; however, they lack the adaptability required to provide more dynamic and engaging psychological interventions [ ]. Previous research has shown that NLP technology, which interprets psychological health cues from user-generated texts, significantly improves the effectiveness of interventions [ , ], a finding consistent with our review [ , , , , , ]. Moreover, the development of deep learning techniques has led to the widespread use of general-purpose pretrained language models in various NLP applications [ ]. This suggests that NLP-based algorithms are particularly suitable for delivering personalized psychological support in real-time interactions. Machine learning and deep learning algorithms show promise in enhancing AI chatbots’ effectiveness by continuously learning through user interactions. These algorithms can provide accurate and personalized interventions over time [ ]. Future studies should investigate how these advanced algorithms can be used to offer more precise psychological support tailored to individual user experiences.User engagement is a critical factor in the success of mental health interventions; however, several studies did not report intervention completion rates [
, , , ]. These issues might become more pronounced in real-world settings, where factors such as interface complexity, the chatbot’s ability to provide meaningful and personalized feedback, and technical issues impact the overall user satisfaction and perceived effectiveness of the intervention [ ]. This underscores the need for further exploration of how chatbot interactions can be designed to maintain user engagement, possibly through improved user experiences and personalization techniques [ , ]. From a different perspective, unlike traditional interventions, which are typically designed to be completed from start to finish, chatbot interactions often do not follow this linear structure. Users may terminate interactions once their immediate concerns are addressed, reflecting the unique flexibility of chatbots [ ]. For instance, if a chatbot provides critical information for crisis management and the user responds affirmatively, the user is likely to exit the chat after their need is met. This immediate resolution highlights the efficiency of chatbot interventions in addressing user concerns promptly, even if the interaction appears incomplete. Recognizing this distinction is essential for accurately interpreting dropout rates and evaluating chatbot effectiveness.The AI chatbots reviewed here comprised various therapeutic approaches. The incorporation of established psychological treatment techniques such as cognitive behavioral therapy, mindfulness, and dialectical behavior therapy into AI chatbots represents a novel approach to delivering psychological health interventions. This approach is considered novel because it uses AI to deliver evidence-based psychological techniques 24/7, serve multiple users simultaneously, tailor interventions based on individual responses, and combine various therapeutic modalities on a single platform.
This multidisciplinary therapeutic approach has the potential to address health professionals’ diverse psychological needs [
]. AI chatbot interventions are characterized by a digital and responsive design, which allows for integrating practical tools such as stress management techniques, mood tracking, and personalized goal setting [ ]. These features align with the immediate and personalized support often required by health professionals. However, while personalization is a key advantage of AI chatbots, a need also exists for some level of standardization to ensure the interventions’ quality and effectiveness [ ]. This standardization should focus on establishing evidence-based core components and assessment methods while still allowing for personalized delivery [ ]. Future research should focus on identifying the most effective components of these chatbots and tailoring them to the specific psychological health needs of health professionals while maintaining a balance between standardization and personalization.In this study, a temporary increase in depression scores was observed in one study [
], whereas previous research has reported AI chatbots’ positive effects in reducing anxiety, depression, and work-related burnout among health professionals [ ]. These contrasting findings may reflect the challenges faced by such professionals, who frequently experience high levels of stress and emotional demands [ ]. This suggests the complexity of psychological interventions and highlights the importance of careful monitoring to determine the appropriate duration and intensity of chatbot use.Some studies in our review demonstrated the potential of AI chatbots to act as first-line responders in identifying and managing mental health crises. Prior research has also shown the potential benefits of AI chatbots not only in improving mental and emotional well-being but also in promoting behavioral changes [
]. While these features may only benefit a subset of users, their impact on preventing escalation to severe mental health conditions or emergencies is invaluable. Future research should prioritize the integration and evaluation of these features to enhance the safety and efficacy of chatbot interventions.Although AI chatbots address the unique time constraints and stress among health professionals, balancing automated responses with genuine therapeutic interactions remains a significant challenge. Personalization technologies that provide tailored features to AI chatbots have emerged as key factors in their effectiveness, with studies reporting increased user engagement and satisfaction through customized responses and immediate support [
]. Consequently, the development of tailored intervention strategies that consider the individual needs and preferences of health professionals is essential to maximize AI chatbots’ effectiveness. These strategies should integrate ongoing user feedback and dynamically adapt interventions to maintain long-term user engagement and effectiveness. Future research should focus on larger sample sizes, extended follow-up periods, and diverse health care professional groups of health professionals. Such comprehensive studies will play a crucial role in developing evidence-based guidelines for implementing AI chatbots in psychological health and in advancing the sophistication of personalization algorithms.Limitations
This scoping review has several limitations. First, the limited number of studies adopting diverse study designs and using various AI algorithms complicated making direct comparisons among the retrieved studies. Second, the findings’ generalizability is restricted because many studies were published in specific regions, particularly high-income countries. Additionally, the short duration of many interventions and limited follow-up periods make it difficult to evaluate the long-term effectiveness and sustainability of AI chatbots. Future research should address these limitations by conducting more rigorous trials, standardizing the outcome measures, and exploring the long-term applications of AI chatbots across diverse populations and settings.
Conclusions
This scoping review explored the current state of AI chatbots aimed at supporting the psychological health of health professionals. Although the reviewed studies demonstrated AI chatbots’ potential to reduce stress, anxiety, depression, and burnout, research in this area remains in its early stages. The diversity of study design, AI algorithms, therapeutic approaches, and outcome measures highlights this field’s innovative but fragmented nature.
Despite promising results, particularly with NLP-based chatbots, a significant need exists for more rigorous and standardized studies to fully evaluate their clinical efficacy. Challenges such as usability issues and limited generalizability of the findings must be addressed to enhance the long-term application and effectiveness of AI chatbots in real-world settings. Future research should focus on refining chatbot designs, expanding the research to diverse populations, and conducting long-term studies to clarify the role of AI chatbots in supporting the psychological health of health professionals. Thus, AI chatbots could become a valuable solution for addressing the psychological health challenges faced by health professionals worldwide.
Acknowledgments
This work was supported by the National Research Foundation of Korea (NRF) grant funded by the Korea government (MSIT) (RS-2024-00351194).
Authors' Contributions
CC conceptualized the study, acquired funding, and oversaw project administration. GB and JH were responsible for data curation and methodology development. CC, GB, and JH conducted the formal analysis. GB created the visualizations. CC, GB, and JH collaboratively wrote the original draft, and all authors contributed to reviewing and editing the manuscript. All authors reviewed and approved the final manuscript.
Conflicts of Interest
None declared.
Checklist 1
PRISMA-ScR checklist. The checklist has been completed in accordance with the journal's guidelines and includes all the items required for a systematic scoping review.
PDF File, 98 KBReferences
- Moss M, Good VS, Gozal D, Kleinpell R, Sessler CN. An official Critical Care Societies Collaborative statement: burnout syndrome in critical care health care professionals: a call for action. Am J Crit Care. Jul 2016;25(4):368-376. [CrossRef] [Medline]
- Aronsson G, Theorell T, Grape T, et al. A systematic review including meta-analysis of work environment and burnout symptoms. BMC Public Health. Mar 16, 2017;17(1):264. [CrossRef] [Medline]
- Maslach C, Schaufeli WB, Leiter MP. Job burnout. Annu Rev Psychol. 2001;52(1):397-422. [CrossRef] [Medline]
- Bakker AB, Costa PL. Chronic job burnout and daily functioning: a theoretical analysis. Burn Res. Dec 2014;1(3):112-119. [CrossRef]
- Smiley RA, Lauer P, Bienemy C, et al. The 2017 National Nursing Workforce survey. J Nurs Regul. Oct 2018;9(3):S1-S88. [CrossRef]
- Gamble A. Artificial intelligence and mobile apps for mental healthcare: a social informatics perspective. AJIM. Jun 2, 2020;72(4):509-523. [CrossRef]
- Noble JM, Zamani A, Gharaat M, et al. Developing, implementing, and evaluating an artificial intelligence-guided mental health resource navigation chatbot for health care workers and their families during and following the COVID-19 pandemic: protocol for a cross-sectional study. JMIR Res Protoc. Jul 25, 2022;11(7):e33717. [CrossRef] [Medline]
- Fortuna KL, Naslund JA, LaCroix JM, et al. Digital peer support mental health interventions for people with a lived experience of a serious mental illness: systematic review. JMIR Ment Health. Apr 3, 2020;7(4):e16460. [CrossRef] [Medline]
- Jacob C, Sanchez-Vazquez A, Ivory C. Social, organizational, and technological factors impacting clinicians’ adoption of mobile health tools: systematic literature review. JMIR Mhealth Uhealth. Feb 20, 2020;8(2):e15935. [CrossRef] [Medline]
- Grover S, Dua D, Shouan A, Nehra R, Avasthi A. Perceived stress and barriers to seeking help from mental health professionals among trainee doctors at a tertiary care centre in North India. Asian J Psychiatr. Jan 2019;39:143-149. [CrossRef] [Medline]
- Chaudhry BM, Debi HR. User perceptions and experiences of an AI-driven conversational agent for mental health support. Mhealth. 2024;10:22. [CrossRef] [Medline]
- Sarkar S, Gaur M, Chen LK, Garg M, Srivastava B. A review of the explainability and safety of conversational agents for mental health to identify avenues for improvement. Front Artif Intell. 2023;6:1229805. [CrossRef] [Medline]
- Yoon S, Goh H, Nadarajan GD, et al. Perceptions of mobile health apps and features to support psychosocial well-being among frontline health care workers involved in the COVID-19 pandemic response: qualitative study. J Med Internet Res. May 31, 2021;23(5):e26282. [CrossRef] [Medline]
- Omarov B, Zhumanov Z, Gumar A, Kuntunova L. Artificial intelligence enabled mobile chatbot psychologist using AIML and cognitive behavioral therapy. IJACSA. 2023;14(6):137-146. [CrossRef]
- Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. Feb 2005;8(1):19-32. [CrossRef]
- Levac D, Colquhoun H, O’Brien KK. Scoping studies: advancing the methodology. Implement Sci. Sep 20, 2010;5(69):69. [CrossRef] [Medline]
- Babineau J. Product review: Covidence (systematic review software). J Can Health Libr Assoc. 2014;35(2):68. [CrossRef]
- Zamani A. Developing a mental health virtual assistance (chatbot) for healthcare workers and their families. University of Alberta; 2022. [CrossRef]
- Matsumoto A, Kamita T, Tawaratsumida Y, et al. Combined use of virtual reality and a chatbot reduces emotional stress more than using them separately. J Univers Comput Sci. 2021;27(12):1371-1389. [CrossRef]
- Kleinau E, Lamba T, Jaskiewicz W, et al. Effectiveness of a chatbot in improving the mental wellbeing of health workers in Malawi during the COVID-19 pandemic: a randomized, controlled trial. PLoS ONE. 2024;19(5):e0303370. [CrossRef] [Medline]
- Kroon YHE. The appreciation of the ithrive chatbot and its impact on self-compassion and compassion fatigue in healthcare workers: a mixed method feasibility study. University of Twente; 2018. URL: https://purl.utwente.nl/essays/75151
- Chang CL, Sinha C, Roy M, Wong JCM. AI-led mental health support (Wysa) for health care workers during COVID-19: service evaluation. JMIR Form Res. Apr 19, 2024;8:e51858. [CrossRef] [Medline]
- Jackson-Triche M, Vetal D, Turner EM, Dahiya P, Mangurian C. Meeting the behavioral health needs of health care workers during COVID-19 by leveraging chatbot technology: development and usability study. J Med Internet Res. Jun 8, 2023;25:e40635. [CrossRef] [Medline]
- Lai J, Ma S, Wang Y, et al. Factors associated with mental health outcomes among health care workers exposed to coronavirus disease 2019. JAMA Netw Open. Mar 2, 2020;3(3):e203976. [CrossRef] [Medline]
- Laranjo L, Dunn AG, Tong HL, et al. Conversational agents in healthcare: a systematic review. J Am Med Inform Assoc. Sep 1, 2018;25(9):1248-1258. [CrossRef] [Medline]
- Chaudhry B, Islam A. Design validation of a workplace stress management mobile app for healthcare workers during COVID-19 and beyond. Springer; 2021. Presented at: Proceedings of EAI MobiQuitous 2021. [CrossRef]
- Anmella G, Sanabra M, Primé-Tous M, et al. Vickybot, a chatbot for anxiety-depressive symptoms and work-related burnout in primary care and health care professionals: development, feasibility, and potential effectiveness studies. J Med Internet Res. Apr 3, 2023;25:e43293. [CrossRef] [Medline]
- Lim SM, Shiau CWC, Cheng LJ, Lau Y. Chatbot-delivered psychotherapy for adults with depressive and anxiety symptoms: a systematic review and meta-regression. Behav Ther. Mar 2022;53(2):334-347. [CrossRef] [Medline]
- Inkster B, Sarda S, Subramanian V. An empathy-driven, conversational artificial intelligence agent (Wysa) for digital mental well-being: real-world data evaluation mixed-methods study. JMIR Mhealth Uhealth. Nov 23, 2018;6(11):e12106. [CrossRef] [Medline]
- Denecke K, Vaaheesan S, Arulnathan A. A mental health chatbot for regulating emotions (SERMO)—concept and usability test. IEEE Trans Emerg Topics Comput. 2020;9(3):1170-1182. [CrossRef]
- Vaidyam AN, Wisniewski H, Halamka JD, Kashavan MS, Torous JB. Chatbots and conversational agents in mental health: a review of the psychiatric landscape. Can J Psychiatry. Jul 2019;64(7):456-464. [CrossRef] [Medline]
- Thunström AO, Carlsen HK, Ali L, Larson T, Hellström A, Steingrimsson S. Usability comparison among healthy participants of an anthropomorphic digital human and a text-based chatbot as a responder to questions on mental health: randomized controlled trial. JMIR Hum Factors. Apr 29, 2024;11:e54581. [CrossRef] [Medline]
- Ouali Y, Hudelot C, Tami M. An overview of deep semi-supervised learning. arXiv. Preprint posted online on Jun 9, 2020. URL: https://arxiv.org/abs/2006.05278
- Chakraborty C, Bhattacharya M, Pal S, Lee SS. From machine learning to deep learning: advances of the recent data-driven paradigm shift in medicine and healthcare. Curr Res Biotechnol. 2024;7:100164. [CrossRef]
- Agnihotri A, Bhattacharya S. Chatbots’ effectiveness in service recovery. Int J Inf Manage. Jun 2024;76:102679. [CrossRef]
- Xu Y, Niu N, Zhao Z. Dissecting the mixed effects of human-customer service chatbot interaction on customer satisfaction: an explanation from temporal and conversational cues. J Retail Consum Serv. Sep 2023;74:103417. [CrossRef]
- Abd-Alrazaq AA, Alajlani M, Ali N, Denecke K, Bewick BM, Househ M. Perceptions and opinions of patients about mental health chatbots: scoping review. J Med Internet Res. Jan 13, 2021;23(1):e17828. [CrossRef] [Medline]
- Ashfaq M, Yun J, Yu S, Loureiro SMC. I, Chatbot: modeling the determinants of users’ satisfaction and continuance intention of AI-powered service agents. Telematics and Informatics. Nov 2020;54:101473. [CrossRef]
- Balcombe L. AI chatbots in digital mental health. Informatics. 2023;10(4):82. [CrossRef]
- Aggarwal A, Tam CC, Wu D, Li X, Qiao S. Artificial intelligence-based chatbots for promoting health behavioral changes: systematic review. J Med Internet Res. Feb 24, 2023;25:e40789. [CrossRef] [Medline]
- Abd-Alrazaq AA, Rababeh A, Alajlani M, Bewick BM, Househ M. Effectiveness and safety of using chatbots to improve mental health: systematic review and meta-analysis. J Med Internet Res. Jul 13, 2020;22(7):e16021. [CrossRef] [Medline]
- Tamminga SJ, Emal LM, Boschman JS, et al. Individual-level interventions for reducing occupational stress in healthcare workers. Cochrane Database Syst Rev. May 12, 2023;5(5):CD002892. [CrossRef] [Medline]
- Bragazzi NL, Crapanzano A, Converti M, Zerbetto R, Khamisy-Farah R. The impact of generative conversational artificial intelligence on the lesbian, gay, bisexual, transgender, and queer community: scoping review. J Med Internet Res. Dec 6, 2023;25:e52091. [CrossRef] [Medline]
Abbreviations
AI: artificial intelligence |
NLP: natural language processing |
Edited by Andre Kushniruk; submitted 17.10.24; peer-reviewed by Christian Pankhurst, Eva-Marie Turner; final revised version received 02.02.25; accepted 14.02.25; published 19.03.25.
Copyright© Gumhee Baek, Chiyoung Cha, Jin-Hui Han. Originally published in JMIR Human Factors (https://humanfactors.jmir.org), 19.3.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Human Factors, is properly cited. The complete bibliographic information, a link to the original publication on https://humanfactors.jmir.org, as well as this copyright and license information must be included.