Abstract
Conversational agents have gained their ground in our daily life and various domains including healthcare. Chronic condition self-management is one of the promising healthcare areas in which conversational agents demonstrate significant potential to contribute to alleviating healthcare burdens from chronic conditions. This survey paper introduces and outlines types of conversational agents, their generic architecture and workflow, the implemented technologies, and their application to chronic condition self-management.
Keywords: Conversational agents, intelligent agents architecture, chronic conditions, self-management
I. Introduction
Chronic conditions (e.g., diabetes, hypertension, and cancer), are long-term illnesses that last one year or longer and require ongoing medical attention [1]. 60% of Americans have at least one chronic condition and 40% have two or more [2]. To illustrate, over 30 million people, approximately 10% of Americans, have diabetes, and two-thirds of them also have hypertension [3]. Chronic conditions are expensive for both clinical and economic reasons. They are the leading causes of death and disability globally and a leading driver of 4.1 trillion in annual healthcare costs in the United States alone [2]. As life expectancy globally rises and the prevalence of associated risks (e.g., unhealthy diet and smoking) are increasing, the burden of chronic conditions gets heavier.
Self-management represents a promising strategy for successful healthcare and positive health outcomes for people with chronic conditions [4]. Self-management refers to the set of relevant skills of individuals to manage and monitor their symptoms, treatment, physical and psychosocial consequences and lifestyle inherent in living with chronic conditions [5]. Self-management prevents the development of secondary morbidity and mortality, helps people with chronic conditions act in concert with providers, and take actions (e.g., self-monitoring) [6], [7]. Self-monitoring, a process of self-management, in turn helps them aware of patients’ symptoms, bodily sensations, and cognitive processes, as well as create the potential to provide information for action [8]. However, people with chronic conditions are often confronted with a range of challenges in managing their conditions routinely such as difficulties in tracking accurate health indicators (e.g., weight, blood sugar, and blood pressure) [9], difficulty in negotiating self-care commitment, and vertical communication healthcare providers and patients [10].
Conversational agents, or dialogue systems or chatbots, have been developed and used in the field of health among many of their applications (e.g., commerce, helpdesk, and websites). Conversational agents are computer systems that communicate with users in natural language in multiple communication channels notably by text or speech [11], [12]. Self-management of chronic conditions is one of the popular applications of these systems via two-way communication and tailored responses that mirror therapeutic processes (e.g., cognitive behavioral therapy) by emulating human-to-human interactions. Through the interactions, the agents engage users in self-management by promoting goal-setting, self-monitoring, provision of feedback, alleviation of negative emotions, and so on. The agents also can provide rich contextual information about the users’ health status [12]. The underlying motivation is to promote preventive measures and continuous monitoring with the help of technologies so that costs for emergency care and hospital admissions can be avoided [13].
Due to the advancement in computer technologies, particularly natural language processing (NLP) and artificial intelligence (AI), conversational agents have the significant potential to play an increasingly important role in healthcare. In particular, in the context of chronic conditions, it has a range of potential to improve patients’ self-management. This opportunity also comes with some challenges regarding the design of interfaces and the application of personal touches for communicating with different users [14], [15]. Thus, this paper aims to broadly outline the concepts, building blocks, underlying techniques, and applications for chronic condition self-management.
II. Conversational Agents
A. Definition
Conversational agents refer to dialog systems that engage with users in the natural language [14], [16], [17]. Dialog systems indicate any human-computer interaction systems that use natural language to understand and generate content to communicate with the user [16]. In other words, a premise of conversational agents is engagement, thus the capability of handling natural language and the contents should not be constrained to a predetermined set of words or limited sentence structure [14], [16]. In this sense, conversational agents are not necessarily the same as chatbots. In literature, the term conversational agents are often synonymously used with chatbots. Chatbots are, however, software components that are a type of text-based and goal-oriented conservational agent [16]–[18]. Chatbots are often designed to simulate conversations with the user or perform a particular task and respond to the conversation in human language with predefined replies and are mainly based on pattern matching [19], [20]. On the other hand, conversational agents are more context-aware, based on theoretically motivated techniques that enable conversation, and use more advanced technologies such as deep learning and natural language understanding [16], [18]. In that regard, conversational agents are said a pragmatic usage of computational linguistics techniques [17].
B. Types of Conversational Agents
Conversational agents can be classified by different criteria such as their communication mode, capability, embodiment, methods of response generation, and domains. Accounting for that natural language can be either written or spoken, the mode of communication of conversational agents can be differentiated by text, voice, or combined. To facilitate engagement, conversational agents are required to process the input and provide relevant output (e.g., advice or feedback) in a form of text, speech, graphics, haptics, gesture, handwriting recognizer, etc [21]. Text-based agents work through text messages or web platforms [22]. The design interfaces of this type of agent aim to support text messages. For example, chatbots like ELIZA [23] developed by Joseph Weizenbaum in 1966, simulate conversations similar to those in social media and service platforms (e.g., banks, bookings, and e-commerce), using text communication mode [18]. Speech or voice-based agents communicate with users through voice. Assistants like Amazon Alexa, Apple Siri, Microsoft Cortana, and Google Assistant Now are more commonplace examples. This type of agent integrates speech technology to recognize human speech [14]. The inputs from users trigger specific information retrieval or function, and then the agents respond with synthesized speeches [14]. Speech technology in voice-based agents is to support a range of purposes, including conducting small talk or expressing emotions [24], supporting learning [25], and fostering communication with others [26]. Some studies address that voice-based agents can better assist children and the elderly in a more natural form of interaction [27], [28]. Conversational agents with multimodal communication models use text, voice, non-verbal communication (e.g., hand gestures and facial expressions), and body stance in combination [29]. This type of agent encompasses physically- (e.g., social robots) and graphically embodied (e.g., virtual agents with a human-like body on a computer screen or avatars) agents [29].
Conversational agents also can be classified into two categories based on their capabilities [14]. Conversational agents’ main capabilities are often constrained to communication only and do not execute actions. ELIZA, Mitsuku [30], and Cleverbot [31] are examples. ELIZA is a basic person-centered or Rogerian psychotherapist chatbot. Its users type questions and ELIZA answers in the input box. ELIZA simulates conversations based on the pattern matching directives, MAD-Slip, that were provided in separate scripts and used rules to respond with to user inputs [32]. Mitsuku is a social chatbot that tells stories and plays games with its users while simulating a young female from Leeds [33]. Other agents, on the other hand, are capable of taking specific actions either in the physical or virtual world such as turning on lights or booking a flight [17]. Alexa [34] and Bluebot [35] are examples of those with capabilities of conducting actions.
The conversational agents also vary by their forms of representation: none embodiment, virtual (static or interactive), and physical embodiment. Those agents with no embodiment lack any form of representation and have a natural language interface only [36]. Embodied conversational agents have representation that demonstrates the same or similar properties as humans in face-to-face conversation, including the ability to communicate through both verbal and nonverbal means such as gestures, facial expressions, and body posture [29]. A physical-based embodied agent has a physical body such as a social robot or mobile robot. Embodied conversational agents encompass both virtual, graphical, and physical agents [17]. These forms of embodiment were found to influence users’ perception of a conversational agent and offer a range of design options [36], [37]. For example, an overly realistic representation of the agent negatively influences the users’ willingness to disclose personal or sensitive information [38], [39]. Users perceive a conversational agent with matching ethnicity of virtual representation as more enjoyable, sociable, and useful [40]. These embodiments of conversational agents also found to create a personal connection with a user through verbal and non-verbal interactions [14].
How to generate a response is another criterion to categorize conversational agents into two groups: rule-based and generative-based [41]. Rule-based agents are provided with pre-determined rules and a database of responses, rather than generating their own answers. ELIZA is an example of a rule-based agent that uses pattern matching. Generative-based agents use machine learning algorithms to generate responses based on the data provided and keep on learning and improving based on learning models. These types of agents with AI technologies are also called AI-based or intelligent agents.
The agents can also be differenced on the basis of domain: general or domain-specific. In the general or open domain, conversational agents are to support conversation in various topics and disciplines including entertainment [40]. Examples include Botta, an Arabic dialect education chatbot [42], AliMe [43], SOGO, a semi-automatic negotiation dialog system [44], and ChatGPT [45]. Goal-oriented conversational agents focus on particular topics in a specific domain to assist target users with completing a particular task, such as healthcare, education, and business [41], [46]. For example, in the medical domain, conversational agents can be applied to socially support patients with medical conditions or learners in the learning process with answers to specific health-related queries on particular health conditions such as HIV/DIDS, children’s health, and mental health [41].
III. System Architecture and Enabling Technologies
The rise and maturity of technologies, such as speech recognition and NLP have enabled the omnipresence of conversational agents. Thus, this section outlines a high-level genetic architectural overview, a foundational set of building structures [47], which can be used for developing a range of applications and related design technologies.
A. Architecture
The specification of conversational agent architecture varies by type and domain, but the generic diagram suggested by Kahn [47] comprises layers of presentation, business, service, data, and utility. The presentation layer encapsulates the various user interface components that are implemented and displayed for the users. The users interact with the agent via this layer using communication channels such as email, SMS, text messaging, voice, or IoT devices [20]. The core responsibility of this layer is supporting multi-channel and multi-platform to make the agent scalable and extendable enough to work with different channels and platforms (e.g., iOS or Android) [20]. To support multiple channels and platforms, user interfaces may contain any interface components that enable easy communication with the current and new platforms. The presentation layer then interacts with the underlying business layer components to access the functional capabilities of the agent [47].
The business layer is responsible for data processing, data formatting, and dialog management [47]. The data processing involves multiple steps that have to be performed in the correct order. Besides, the data from the service layer also need to be converted into real-world business entities (e.g., orders and products). The data processing components often gather data from services and identify common entity usage patterns and implement them as separate components. Data formatting is to convert the data into the required format that is in line with UI requirements in the presentation layer and capabilities. All the data from the service layer components needs processing to be structured as per the requirements of an agent. This process is needed to make a conversational agent extendable and scalable enough for any existing and new platforms and channels [47]. Dialog management is responsible for managing dialog with the user. This component is one of the core and unique components of a conversational agent solution that determines the user experience. For a flawless experience for the user, dialog management should be able to manage the context of dialog, the user’s preferences, and profiles [47]. With the assistance of specific AI/NLP services, Natural Language Understanding (NLU) enables the agent to understand and respond to the user by converting an abstract statement into a natural language surface utterance [20].
The service layer is responsible for integrating with external party services (e.g., business functionality and middleware) for sharing the internal and external data for further usage [47]. The components of the service layers can be deployed on the same tier or a separate tier, depending on the performance and requirements of the implementation. The service layer can consist of NLP services, data access services, and external service interfaces. An NLP service has NLU as a fundamental service component that can determine the success or failure of an agent. The degree of efficiency of an NLP service relies on its machine learning capability which is based on the number of and advancement of AI algorithms. Data access services serve as adaptors to provide access to a set of services that convert the data from those services into a format that the other components of an agent understand. An agent may have this layer separately or on the same layer, depending on the requirements of the implementation. External service interfaces manage the integration of a different set of external services (e.g., social media, CRM, and support center). The specific services may vary by the functionalities provided by an agent [47].
The data layer concerns access to local data collection, which is provided to other layers via the service layer [47]. These storage systems serve as the basis of many services and components of an agent. For instance, the history of communication with users, analysis of collected data, and performance of machine learning techniques rely on this layer. In case of an intelligent agent based on AI may consist of Knowledge Base and user interaction history in this layer [20]. The knowledge Base represents the domain-specific source content for training an agent [20].
The utility layer provides common services such as security, logging, and configuration. This layer includes configuration and security. These are not considered to be a functional part of an agent but play a significant role in the overall operations of any system implementation. The components of this layer affect the success of an agent’s feature and functionality by determining the scalability and repeatability of characteristics of an agent, and security management. Fig. 1 portrays an architectural view of the generic conversational agent solution and its ecosystem.
Fig. 1.

A generic architectural overview of conversational agents (redrawn and adapted from Kahn [47])
B. Workflow and Related Techniques
A common workflow for both rule-and AI-based agents is pre-processing, processing, and generation [46] as shown in Fig 2. The figure depicts the main three steps and associated components of a conversational agent. These steps involve different design approaches that may employ various design methods and techniques that are introduced below.
Fig. 2.

Workflow and associated components of conversational agent (redrawn and adapted from Maroengsit et al.,[41])
1). Pre-processing step
Once a user launches his/her interaction with an conversational agent via its user interface, the pre-processing step revolves around NLP to collect external data and format data for other processes of the agent. This step involves design methods such as pattern matching, parsing, Artificial Intelligence Markup Language (AIML), and statistical techniques like Term Frequency-Inverse Document Frequency (TF-IDF) and word2vec:
Pattern matching technique classifies the user input as “pattern” and produces a suitable response stored in “template” [46]. Although the complexity of algorithms of this technique varies, it basically uses “pattern”-“template” pairs which are hand-craft. This technique is the most commonly used approach which is used in both early and modern conversational agent design [46]. A disadvantage of the technique is limited capabilities because manual building restricts the scaling, making the agent’s responses predictable and repetitive [46].
AIML is to create conversational flow in an agent. AIML is made up of data objects, which consist of two units: topics and category units [46]. The topic has a name attribute ad a set of categories related to the topic. A category is a rule which comprises two elements “pattern” and “template” [48]. A user’s input is matched using “pattern” and “template” is used to generate a response using [48]. AIML is a powerful design tool but requires NLP programming too.
Parsing is a method to extract meaningful information from a text input by parsing that text input into a set of words that can be stored and manipulated. The extracted word or keyword is matched against the documents in the corpus to find an appropriate response [23]. Semantic parsing, which is a more advanced form of parsing, converts the user’s input to a machine-understanding representation of its meaning [23].
Chatscript is an open-source authoring tool that combines a natural language engine and dialog management system for interactive conversation [46]. This rule-based engine uses rules dialog flow script creates. Machine learning tools can also be used to improve dialog flow [46].
Domain ontologies are also used to replace a manually built knowledge base. Domain ontologies allow an agent to explore the concept nodes of an ontology to establish the relationship between concepts [49] or enable to yield standardization and sharing [50].
TF-IDF is a statistical technique to quantify the importance or relevance of text representations (e.g., words, lemmas) in a document among a corpus [51]. It is often used as a weighting factor in information retrieval, text mining, and user modeling. The weights of text representations are measured by calculating word frequencies (TF) and by multiplying TF with inverse document frequency [51]. TF-IDF is a basic way to convert a document of words into numbers but it can be used to obtain a vector of words frequencies [52].
Word2vec is a group of models for NLP that turns every word in text into a unique vector called “word embeddings” based on the context in which they appear in text [53]. This NLP technique can be used in conversational agents to improve the ability of the agent to understand the meanings of user inputs and to generate more appropriate responses. An example of using word2vec in a conversational agent is using pre-trained word2vec embeddings to initialize the word representation weights for individual utterance [54].
2). Processing step
The processing step mainly relies on NLU to collect and manage conversations based on pre-processed user input data [41]. Intent classification and name entity recognition (NER) can be integrated into NLU service of a conversational agent [20]. When an agent has machine learning components, it can consist of NLP, NLU, and decision engine.
NER is an NLP technique to extract and classify mentions of designators from text such as biological species, temporal expressions, and names. In conversational agents, NER can be a component of NLU [55]. NLU maps an input text into a pre-defined semantic slot according to scenarios. In this process, NLU typically uses two types of presentations: the utterance level such as the user’s intent category; and the world-level information extraction such as named entity recognition and slot filling [20].
Intent classification is the process of automatically categorizing users’ intent by analyzing the user input into intents such as purchase, unsubscribe, and demo requests [56].
Decision engine decides what response to generate based on information in the knowledge base. The engine learns domain-specific knowledge from the source data to generate the response [20]. Many recent intelligent agents employ neural networks in NLP, such as language modeling [57], paraphrase detection [58], and word embedding extraction [59]. The use of machine learning has risen in the conversational agent development [46].
Especially artificial neural network (ANN) models and their variants, such as convolutional neural networks (CNNs), recurrent neural networks (RNNs), sequence-to-sequence neural networks (Seq2Seq), and long short-term memory networks (LSTMs), are one of the learning algorithms within machine learning [46]. The presence of a learning algorithm in the ANN approach distinguishes intelligent AI-based agents from rule-based ones [46]. ANN-based agents can use both retrieval- and generative-based approaches for response generations. The use of deep learning neural networks, which are capable of learning unsupervised from unstructured or unlabeled data, has dominated intelligent conversational modeling, using especially RNNs and LSTMs [46]. Recently NLU with transformer model also have proven to offer benefits as a building block of conversational agents [60].
An RNN is a recursive ANN. RNNs save the output of a layer and feeding this saved output to the new input to predict the next output. Due to their ability to remember the previous computation, RNNs useful in understanding the semantical meaning and conversational context based on the previous words in the sentence. This ability to handle the sequential nature of natural language changed the landscape of conversational agent development as the understanding of conversational context is essential to understand the user inputs and generating more contextually appropriate responses [61].
Seq2Seq models are designed based on RNN architecture and have two RNNs, which are an encoder and a decoder. The encoder encodes the <status> or input sentence. The decoder decodes the <status> and generates the desired responses [61]. These two neural networks enable the model to process variable length input sentences. In conversational agents, this model is widely used to generate responses by converting between input <status> and output <response> [62].
LSTMs are a special type of RNN, which is designed to resolve the RNN model’s long-term dependency issue [63]. To hold the previous information for a long period of time, memory cells and gates are introduced in LSTMs. The memory cells store information in, write new information, and read information from them, and the gates are used to control the flow of information [64]. The capabilities of LSTM show higher performance in learning from experiences, classification, and processing and prediction of time series as compared to other available RNNs, hidden Markov models, and other sequence learning models. These abilities make LSTMs useful in designing conversational agents [64].
Transformer proposed by Vaswani et al., [65] is a new architecture that is based on encoder and encoder that are built on self-attention mechanism, and parallelizable to understand speech sequences [66]. This model utilizes pretrained models that are trained for the patterns and features of a sizable dataset, can be adjusted for other tasks with a minimal amount of additional training data [15], [17], [21]. A transformer model has ability to processes many words at the same time which overcomes slow training time of LSTM and RNNs [65]. Examples of those models include BERT (Bidrectional Encoder Representations from Transformers), RoBERTa (Robustly Optimized BERT Pretraining Approach), and GPT (Generative Pre-trained Transformer). In order to learn the semantics of words and contexts, BERT, developed by researchers [67] at Google in 2018, is an open-source machine learning framework for NLP. BERT consists of Transformer encoder layers and is pre-trained on two tasks: language modeling and next sentence prediction that enable BERT to learn latent representations of words and sentences in context [67]. After pretraining, the model can be quickly and easily adjusted to suit various tasks with fewer resources on smaller datasets to optimize its performance on specific tasks such as sequence-to-sequence-based language generation (e.g., question-answering and conversational response generations) [67]. RoBERTa is built on BERT by Yinhan et al [68]. This language model optimized key hyperparameters of BERT and trained with a larger training data size [68]. RoBERTa has the same architecture as BERT but is trained with bigger batches over more data and longer sequences, removed the next sentence prediction objective, and uses dynamic masking [68]. GPT is another family of language models by OpenAI that uses several blocks of the transformer [69]. A GPT model is generally trained on an enormous corpus of publicly available datasets and generates human-like texts based on a given input [69]. The latest version is GPT-4 as of March 2023, which can accept images as well [69]. Conversational agents with transformer models hold the potential to understand context around a word over a longer distance in a more effective way, so that the agent delivers a better response.
3). Generation step
In this last step, an agent responds to the user whether the response is either to provide information to the user or to ask a question to gain more information from the user [41]. This step evolves around natural language generation (NLG), converting an abstract statement into a natural language surface utterance [20]. To construct a response and render the output, dialog manager first uses response models (e.g. generative and retrieval-base), and then returns a response based on the priority [70]. Retrieval-based agents search a user utterance in a database and return the best matched response [20]. Agents used generative approaches typically employ machine learning models aforementioned to synthesize new responses. For instance, an LSTM-based agent incorporates the encoder-decoder to incorporate the question information, semantic slot values and dialog act type to generate correct responses [20].
IV. Conversational Agents for Chronic Condition Self-Management
Healthcare is an area where conversational agents have gained their ground along with business. Besides, conversational agents show the potential to play an increasingly more significant role, such as assisting healthcare providers and patients with consultation and supporting health behavior changes in their own living environment [17]. Self-management supports, such as continuous feedback and reinforcement, from conversational agents present the potential to reduce the health system burden and to empower people with the skills, knowledge, and confidence needed to take an active role in managing their health conditions [71]. This section examines how conversational agents have been discussed for self-management.
A. Characteristics of Chronic Conditions
Conversational agents that are applied for chronic conditions are domain specific that focus on specific conditions and tasks. Psychological conditions, such as depression and anxiety [72]–[74] are the most commonly addressed target condition in literature [14], [75], [76]. Other conditions include obesity [77], [78], diabetes [79], [80], cancer [81], asthma [82], [83], neuropsychiatric disorders [74], [84], cardiovascular diseases [85]–[87], neurological disorders [88], chronic obstructive pulmonary disease (COPD) [83], arthritis [89], hypertension [90], and hearing impairment [91]. These agents are designed to assist with maintaining a healthy lifestyle [77], [92]–[95], managing symptoms [72], [96], participating in social activities [97], problem-solving [98], working with healthcare professionals [99], [100], managing medications [99], and a combination of these. (see Table I for examples)
Table I.
Examples of Conversational Agents Attributes for Chronic Conditions Self-Management
| Chronic Condition | Example Agent | Target Area | Theory or Evidence/Evaluation | Communication Mode | Techniques |
|---|---|---|---|---|---|
| Alzheimer | N/A By Griol et al | Disease monitoring | N/A Ease of use |
Multi-mode (voice & visual) |
AI-based: NN, ML, ASR, NLU, NLG, TTS |
| Asthma | mASMAA by Rhee et al [82] | Disease monitoring and treatment adherence | N/A User satisfaction |
Text | AI-based: NLP |
| Breast Cancer | Vik By Chaix et al. [81] | Education, &Treatment adherence | N/A User satisfaction |
Multi-mode (text & visual) |
AI-based: NLP, ML |
| Cardio-vascular | Personal Health Management Assistant by Ferguson et al. [85] | Symptom monitoring | N/A Feasibility |
Multi-mode (text & voice) |
AI-based: SR, NLP |
| COPD | Avachat by Easton et al. [83] | Info seeking, problem solving, & motivation | N/A User Acceptance |
Multi-mode (text, voice & video) |
AI-based: SR, NLP |
| Depress-ion | N/A By Ly et al. [94] | Symptom monitoring | Cognitive behavior therapy / Perceived Usefulness | Multi-mode (voice & video) |
AI-based: Emotion algorithms, ML, NL |
| Diabetes | Healthy Coping by Cheng et al. [79] | Symptom monitoring | N/A User satisfaction & Ease of use |
Voice | AI-based: SR, NLU, ML |
| Heart failure | CARDIAC by Galescu et al. [86] | Monitoring & Self-care | N/A Perceived Usefulness |
Voice | AI-based: NLU, SR |
| Parkinson | Harlie by Ireland et al. [88] | Education & Disease monitoring | N/A Ease of use |
Multi-mode (voice & text) |
AI-based: NN, ML, ASR, NLU, NLG, TTS |
| Psycholo-gical disorders | Pocket Skills by Schroeder et al. [96] | Goal-setting, skill practice & self-tracking | Dialectical behavior theory / Symptom reduction | Multi-mode (text, voice & video) |
N/A |
| HIV /AIDS | N/A by van Heerden et al. [100] | Disease monitoring | CDC guidelines for HIV counseling and testing / Sentiment toward the agent | Multi-mode (text, voice, & muti-media) |
N/A |
Notes: AIML: artificial intelligence markup language; ASR: automatic speech recognition; API: application programming interface; DL: deep learning; NLG: natural language generation; NLP: natural language processing. NLU: natural language understanding; NN: neural network; ML: machine learning; SR: speech recognition; STT: speech-to-text; TTS: text-to-speech.
Conversational agents for self-management are discussed as suitable, personalized, and affordable solutions to react to challenges of self-management and slow down individual disease deterioration. Studies report that conversational agents have the ability to simulate a therapeutic process such as cognitive behavioral therapy or motivational interviewing [101] which goes beyond traditional healthcare sites and reaches personal settings [102], [103]. This could promote a range of self-management skills, including goal setting, provision of feedback, education, behavioral changes, self-care, illness monitoring, and personalized treatment [5]. Some agents exclusively focus on particular self-management skills such as monitoring [86], [91] and support [104], [105]. Some agents support multiple self-management skills [79], [87].
B. Acceptance and Efficacy
Despite limited, evidence suggests conversational agents are useful and effective for engaging users, and acceptable to those living with chronic conditions [75], [76], [106]. Conversational agents were found to be effective in a range of aspects, including treatment outcomes, adherence to self-care, and health literacy. To illustrate, the agents were found to help alleviate negative emotions or symptoms of mental problems by forming an alliance or rapport with patients through conversations and showing empathy [12], [107]. The agents are also capable of supporting natural health routine tasks when in-person health care is not available [108], [109]. Furthermore, the agents can support the understanding of the health status of the users through rich linguistic information using conversational syntax and semantics [12]. Its embodiment and design also affect users’ attitudes toward conversational agents. For example, text-based chatbots are found to engage users about sensitive or stigmatized topics which alleviates negative symptoms or emotions [12], [110]. Voice-based conversational agents were found to facilitate users with low literacy or with visual [111], intellectual [112], linguistic, motor, or cognitive disability [113]. AI-based agents also demonstrated positive feedback for chronic condition self-management regarding its helpfulness, satisfaction, and ease of use [76].
Several studies reported the efficacy of conversational agents in a range of aspects [72], [73], [114]–[117]. To illustrate, the engagement was found beneficial to treatment outcomes [118], [119]. The effects of utilizing conversational agents in treating chronic conditions are not superior to active intervention with human psychological therapy [98], [120]. However, in some controlled studies such as studies by Freeman et al., [108] and Pinto et al., [109], the agents were found relatively effective, showing small-to-large effects on chronic condition intervention. Study participants also reported satisfaction with accessibility, availability [94], [115], interactivity [95], agents’ ability to form a relationship and show empathy [72], [94], [97] and learning from input [72], customizability of its gender and appearance of embodied conversational agents [94], and the option to tailor the session to the users’ own needs [94].
There are, however, challenges in using conversational agents to provide interventions. The noticeable reasons include repetitive content [72], [94], [97], [106], [116] and limitations in understanding user inputs or generating responses appropriately [96], [98] and tailoring content and response to the users’ needs [117]. Concerns for privacy invasion and reduction in independence were also reported [75].
C. Attributes of Conversational Agents
A range of types of conversational agents have been developed and designed for chronic condition management. Besides, various technologies are utilized in those agents to make them efficient and useful in managing their target conditions. To illustrate, the conversational agents are developed as chatbots [72], [92]–[96], [99], [106], [122], voice-based [79], [79], [87], [88], [91], [105], [123], and embodied as robots or virtual human-like agents [101]. Their dialogue management strategies are mostly either finite-state or frame-based, letting either the users or the systems initiate the conversation [72], [94]–[98]. These agents have mobile, web, text messaging, and robot interfaces. Some conversational agents have multiple interfaces such as smartphones and the web. The conversational agents found in the literature utilized various techniques including speech recognition, NLP, text-to-speech, speech-to-text, AI techniques including machine learning and deep learning.
Some of the conversational agents for chronic condition self-management are developed based on theories of behavior change or contained evidence-based content. To illustrate, an agent developed by Elmasri and Maeder [92] for substance use was developed based on the alcohol use disorders identification test, which is a method of screening for unhealthy alcohol consumption or any alcohol use disorder. Another chatbot to assist with substance use developed by Kazemi and colleagues [93] was based on ecological momentary interventions (EMI) and the transtheoretical model of change. EMI moves to intervention from clinical settings to an individual’s daily life [124]. The transtheoretical model construes health behavior changes are an intentional process that takes multiple stages [125]. A chatbot for depression management developed by Gaffney and colleagues [106] was also based upon a behavioral theory, i.e., perceptual control theory. This theory posits human behavior varies in a way of keeping their perception of the world [126]. The task-oriented agent developed by Watson and colleagues [127] employed behavioral and social cognitive theory to promote exercise behavior change to manage overweight. Several chatbots were based on cognitive behavioral theory. For example, the chatbots developed by Fitzpatrick et al., [72] and Ly et al., [94] both for mental health management used cognitive behavioral theory (CBT). Another chatbot developed by Stein and colleagues also employed the same theory for diabetes management. CBT indicates a range of psychological treatments grounded on the theory that our thoughts, emotions, sensations, and behavior are all linked, therefore, our thoughts and actions affect the way we feel [128]. Another chatbot developed by Schroeder and colleagues [96] adopted dialectical behavior theory which is the modified CBT to apply CBT to broader mental health conditions by developing healthy ways to cope and regulate emotions and stress [129]. Some other chatbots utilized established evidence. For example, van Heerden and colleagues utilized CDC guidelines for HIV counseling in non-medical settings to assist symptom and healthy lifestyle management and collaboratively work with healthcare providers [100]. Wang and collaborators used the PubMed medical information database for their conversational agents for healthy lifestyle management and social activity participation [97].
D. Evaluation Measures
Conversational agents that target chronic condition management were assessed in three dimensions: technical performance, user experience, and health-related aspects [130]. Technical performance was assessed with accuracy [80], [90], [116], [131], specificity [80], [116], [132], [133], and task completion rates [80]. To illustrate, Azzini and colleagues [90] evaluated the task completion rate and the accuracy of their voice-activated agent. Rehman and colleagues [80] calculated Cohen’s D to assess task completion and the accuracy of their agent for diabetic conditions. Bickmore et al [133] and Philip et al [134] assessed the specificity of their agents that were developed for depressive disorder.
As for user experience, commonly evaluated features found in the literature are helpfulness, satisfaction, and ease of use [130]. According to literature reviews [75], [130], the majority of studies included in their reviews reported high to moderate to high positive user experiences. However, some studies reported dissatisfaction. For instance, in Baptista et al.’s study, some participants reported that the embodied agent for diabetes coaching was annoying and boring whereas the majority of other participants reported the agent is helpful and friendly [135]. Schroeder et al.’s participants also addressed the difficulty in engaging with the agent [96].
Regarding health-related measures, the most common measures found in the literature are symptom relief [30] and improvement in self-management skills. These studies used methods like quasi-experiments and randomized controlled trials.
One issue with these evaluations of conversational agents in healthcare is inconsistency and invalidated questionnaires [15], [75]. Inconsistency found in the reporting of design methods includes the number of study arms, method of assignment, allocation ratio, outcome measurement, and attrition [15]. To illustrate, the studies that tested the engagement of patients using highly variable metrics for engagement from 30-minute sessions over 2 weeks [114], a single session [98], [136], daily sessions for 2 weeks [72], [94], to unlimited access for 30 days [115]. To improve the consistency in reporting and the validity, Laranjo et al [15] suggest using the followings: Consolidated Standards of Reporting Trials of electronic and mobile health applications and online telehealth [137], the Transparent Reporting of Evaluations with Nonrandomized Designs statement [138], and the Standards for Reporting Diagnostic accuracy studies [139].
V. Conclusion
Despite mostly being preliminary and limited, the findings in the literature suggest conversational agents show advantages for self-management of chronic conditions. The overall acceptance of conversational agents for the self-management of chronic conditions was found promising. Particularly the recent advances in NLP and AI present multiple benefits for healthcare providers and those living with chronic conditions by supporting them with behavior changes and completing specific tasks for managing their conditions. These features of conversational agents make scalable, less costly medical support solutions for chronic condition management. However, there are limitations; firstly, more reliable and comparable evidence is needed to determine the efficacy of conversational agents for chronic conditions due to inconsistent reporting of technical implementation details, users’ attitudes towards agents, and medical efficacy.
This survey paper is similar to others in terms of outlining the definition, architecture, and underlying NLP techniques of conversational agents. But this survey paper extended prior similar work by examining the applications of conversational agents for chronic condition self-management including agent examples of agents, medical context, and theoretical framework applied in the development of those agents with specific purposes.
Acknowledgment
This work is supported by the National Institutes of Health (NIH) Exploratory/Developmental Research Grant 1R21NR019707-01.
References
- [1].Centers for Disease Control and Prevention, “About Chronic Diseases.” https://www.cdc.gov/chronicdisease/about/index.htm (accessed Mar. 27 2023).
- [2].Centers for Disease Control and Prevention, “Chronic Diseases in America.” https://www.cdc.gov/chronicdisease/resources/infographic/chronic-diseases.htm (accessed Mar. 27 2023).
- [3].National Center for Chronic Disease Prevention and Health Promotion, “National diabetes statistics report, 2017,” CS279910-A, 2017. (accessed Mar. 27 2023).
- [4].Rijken M, Jones M, Heijmans M, and Dixon A, “Supporting self-management,” in Caring for People with Chronic Conditions. A Health System Perspective., Nolte E and McKee M, Eds. Berkshire, UK: Open University Press, 2008, pp. 116–142. [Google Scholar]
- [5].Barlow J, Wright C, Sheasby J, Turner A, and Hainsworth J, “Self-management approaches for people with chronic conditions: a review,” Patient Educ. Couns, vol. 48, no. 2, pp. 177–187, 2002. [DOI] [PubMed] [Google Scholar]
- [6].Lorig K, “Self-management of chronic illness: a model for the future,” Gener. J. Am. Soc. Aging, vol. 17, no. 3, pp. 11–14, 1993. [Google Scholar]
- [7].Grady PA and Gough LL, “Self-management: a comprehensive approach to management of chronic conditions,” Am. J. Public Health, vol. 104, no. 8, pp. e25–e31, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [8].Wilde MH and Garvin S, “A concept analysis of self‐monitoring,” J. Adv. Nurs, vol. 57, no. 3, pp. 339–350, 2007. [DOI] [PubMed] [Google Scholar]
- [9].Susannah F and Maeve D, “Tracking for health.” https://www.pewinternet.org/2013/01/28/tracking-for-health/. (accessed Mar. 21, 2023).
- [10].Fort MP, Alvarado-Molina N, Peña L, Montano CM, Murrillo S, and Martínez H, “Barriers and facilitating factors for disease self-management: a qualitative analysis of perceptions of patients receiving care for type 2 diabetes and/or hypertension in San José, Costa Rica and Tuxtla Gutiérrez, Mexico,” BMC Fam. Pract, vol. 14, pp. 1–9, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [11].Jurafsky D and Martin JH, Speech and Language Processing: An Introduction to Natural Language Processing, Computational Linguistics, and Speech Recognition. 2nd ed. Upper Saddle River, New Jersey: Prentice Hall; 2009 [Google Scholar]
- [12].Guy I, “Searching by talking: Analysis of voice queries on mobile web search,” in the 39th International ACM SIGIR conference on Research and Development in Information Retrieval, Pisa, Italy, Jul. 2016, pp. 35–44. [Google Scholar]
- [13].Agnihothri S, Cui L, Delasay M, and Rajan B, “The value of mHealth for managing chronic conditions,” Health Care Manag. Sci, vol. 23, pp. 185–202, 2020. [DOI] [PubMed] [Google Scholar]
- [14].Montenegro JLZ, da Costa CA, and da Rosa Righi R, “Survey of conversational agents in health,” Expert Syst. Appl, vol. 129, pp. 56–67, 2019. [Google Scholar]
- [15].Laranjo L et al. , “Conversational agents in healthcare: a systematic review,” J. Am. Med. Inform. Assoc, vol. 25, no. 9, pp. 1248–1258, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [16].Vishnoi L, “Conversational agent: A more assertive form of chatbots,” Towards Data Science. https://towardsdatascience.com/conversational-agent-a-more-assertive-form-of-chatbots-de6f1c8da8dd (accessed Mar. 21, 2023). [Google Scholar]
- [17].Allouch M, Azaria A, and Azoulay R, “Conversational agents: Goals, technologies, vision and challenges,” Sensors, vol. 21, no. 24, p. 8448, 2021. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [18].Radziwill NM and Benton MC, “Evaluating quality of chatbots and intelligent conversational agents,” in Computers and Society, Apr 2017. [Online] Available: ArXiv170404579. [Google Scholar]
- [19].Masche J and Le N-T, “A review of technologies for conversational systems,” presented at the Advanced Computational Methods for Knowledge Engineering: Proceedings of the 5th International Conference on Computer Science, Applied Mathematics and Applications, ICCSAMA 2017, Berlin, Germany, Jul. 2017, pp. 212–225. [Google Scholar]
- [20].Som B and Nandi S, “Survey of textbased chatbot in perspective of recent technologies,” presented at the Computational Intelligence, Communications, and Business Analytics: Second International Conference, CICBA 2018, Kalyani, India, July 27–28, 2018, Revised Selected Papers, Part II, 2019, vol. 1031, p. 84. [Google Scholar]
- [21].Gudivada VN, Rao D, and Raghavan VV, “Big data driven natural language processing research and applications,” in Handbook of Statistics, vol. 33, Oxford, UK: Elsevier, 2015, pp. 203–238. [Google Scholar]
- [22].Alesanco Á, Sancho J, Gilaberte Y, Abarca E, and García J, “Bots in messaging platforms, a new paradigm in healthcare delivery: application to custom prescription in dermatology,” presented at the EMBEC & NBC 2017: Joint Conference of the European Medical and Biological Engineering Conference (EMBEC) and the Nordic-Baltic Conference on Biomedical Engineering and Medical Physics (NBC), Tampere, Finland, June 2017, 2018, pp. 185–188. [Google Scholar]
- [23].Weizenbaum J, “ELIZA—a computer program for the study of natural language communication between man and machine,” Commun. ACM, vol. 9, no. 1, pp. 36–45, 1966. [Google Scholar]
- [24].Ball G and Breese J, “Emotion and personality in a conversational agent,” Embodied Conversational Agents, vol. 189, 2000. [Google Scholar]
- [25].Garg R and Sengupta S, “Conversational technologies for in-home learning: using co-design to understand children’s and parents’ perspectives,” presented at the Proceedings of the 2020 CHI conference on human factors in computing systems, Honolulu, HI, Apr. 2020, pp. 1–13. [Google Scholar]
- [26].Storer KM, Judge TK, and Branham SM, “‘All in the same boat’: Tradeoffs of voice assistant ownership for mixed-visual-ability families,” presented at the Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, Apr. 2020, pp. 1–14. [Google Scholar]
- [27].Ring L, Bickmore T, and Pedrelli P, “Real-time tailoring of depression counseling by conversational agent,” Iproceedings, vol. 2, no. 1, p. e6065, 2016. [Google Scholar]
- [28].Garg R et al. , “The last decade of HCI research on children and voice-based conversational agents,” presented at the Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, New Orleans, LA, Apr. 2022, pp. 1–19. [Google Scholar]
- [29].Cassell J, Sullivan J, Prevost S, and Churchill E, Eds., Embodied Conversational Agents. Cambridge, Massachussets: The MIT Press, 2000. [Google Scholar]
- [30].Higashinaka R et al. , “Towards an open-domain conversational system fully based on natural language processing,” presented at the Proceedings of COLING 2014, the 25th International Conference on Computational Linguistics: Technical Papers, Dublin, Ireland, Aug. 2014, pp. 928–939. [Google Scholar]
- [31].Gehl RW, “Teaching to the Turing test with Cleverbot,” Transform. J. Incl. Scholarsh. Pedagogy, vol. 24, no. 1–2, pp. 56–66, 2013. [Google Scholar]
- [32].Bassett C, “The computational therapeutic: exploring Weizenbaum’s ELIZA as a history of the present,” AI Soc, vol. 34, pp. 803–812, 2019. [Google Scholar]
- [33].Croes EA and Antheunis ML, “Can we be friends with Mitsuku? A longitudinal study on the process of relationship formation between humans and a social chatbot,” J. Soc. Pers. Relatsh, vol. 38, no. 1, pp. 279–300, 2021. [Google Scholar]
- [34].Lopatovska I et al. , “Talk to me: Exploring user interactions with the Amazon Alexa,” J. Librariansh. Inf. Sci, vol. 51, no. 4, pp. 984–997, 2019. [Google Scholar]
- [35].Patil A, Munson J, Wood D, and Cole A, “Bluebot: Asset tracking via robotic location crawling,” presented at the ICPS’05. Proceedings. International Conference on Pervasive Services, Santorini, Greec, Jul. 2005, pp. 117–126. [Google Scholar]
- [36].Diederich S, Brendel AB, and Kolbe LM, “On conversational agents in information systems research: analyzing the past to guide future work,” presented at the 14th International Conference on Wirtschaftsinformatik, Siegen, Germany, Feb 2019. pp. 1550–1564. [Google Scholar]
- [37].Sah YJ and Peng W, “Effects of visual and linguistic anthropomorphic cues on social perception, self-awareness, and information disclosure in a health website,” Comput. Hum. Behav, vol. 45, pp. 392–401, 2015. [Google Scholar]
- [38].Schuetzler RM, Grimes GM, Giboney JS, and Nunamaker JF Jr, “The influence of conversational agents on socially desirable responding,” Decision Support Systems, vol. 114, pp. 94–102, 2018. [Google Scholar]
- [39].Schroeder J and Schroeder M, “Trusting in machines: How mode of interaction affects willingness to share personal information with machines,” presented at the the 51st Hawaii International Conference on System Sciences, 2018. [Online]. Available: https://scholarspace.manoa.hawaii.edu/server/api/core/bitstreams/ed6ac49a-82f5-4a34-9a56-c479784a4f0a/content [Google Scholar]
- [40].Qiu L and Benbasat I, “A study of demographic embodiments of product recommendation agents in electronic commerce.,” Int. J. Hum.-Comput. Stud, vol. 68, no. 10, pp. 669–688, 2010, doi: 10.1016/j.ijhcs.2010.05.005. [DOI] [Google Scholar]
- [41].Maroengsit W, Piyakulpinyo T, Phonyiam K, Pongnumkul S, Chaovalit P, and Theeramunkong T, “A survey on evaluation methods for chatbots,” presented at the 2019 7th International conference on information and education technology, Aizu, Japan, Mar 2019, pp. 111–119. [Google Scholar]
- [42].Ali D and Habash N, “Botta: An arabic dialect chatbot,” presented at COLING 2016, the 26th International Conference on Computational Linguistics: System Demonstrations, Osaka, Japan, Dec. 2016, pp. 208–212. [Google Scholar]
- [43].Qiu M et al. , “AliMe chat: A sequence to sequence and rerank based chatbot engine,” presented at the the 55th Annual Meeting of the Association for Computational Linguistics, Vancouver, Canada, Jul 2017, pp. 498–503. [Google Scholar]
- [44].Zhao R, Romero O, and Rudnicky A, “SOGO: A social intelligent negotiation dialogue system,” presented at IVA ‘18: International Conference on Intelligent Virtual Agents, Sydney, Australia, Nov. 2018, pp. 239–246. doi: 10.1145/3267851.3267880. [DOI] [Google Scholar]
- [45].Karagiannis G, “Unlocking the full potential of customer service with ChatGPT,” 2023. https://www.departmentofproduct.com/blog/unlocking-the-full-potential-of-customer-service-with-chatgpt/ (accessed March 27, 2023).
- [46].Hussain S, Ameri O, and Ababneh N, “A survey on conversational agents/chatbots classification and design techniques,” in Web, Artificial Intelligence and Network Applications: Proceedings of the Workshops of the 33rd International Conference on Advanced Information Networking and Applications. 2019, pp. 946–956. [Google Scholar]
- [47].Khan R, “Standardized architecture for conversational agents aka chatbots,” Int. J. Comput. Trends Technol, vol. 50, no. 2, pp. 114–121, 2017. [Google Scholar]
- [48].Shawar BA and Atwell E, “Chatbots: are they really useful?,” J. Lang. Technol. Comput. Linguist, vol. 22, no. 1, pp. 29–49, 2007. [Google Scholar]
- [49].Al-Zubaide H and Issa AA, “Ontbot: Ontology based chatbot,” presented at the 4th International Symposium on Innovations in Information and Communications Technology, Amman, Joran, Nov. 2011. pp. 7–12. [Google Scholar]
- [50].Amith MT, Cui L, Roberts K, and Tao C, “Towards an ontology-based medication conversational agent for PrEP and PEP,” presented at the Proceedings of the conference. Association for Computational Linguistics. Meeting, Online, Jul. 2020, pp. 31–40. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [51].Anand R and Jeffrey David U, Mining of Massive Datasets. Cambridge, UK: Cambridge University Press, 2012. [Google Scholar]
- [52].Joigneau A, “Utterances classifier for chatbots’ intents,” M.S. thesis, KTH Royal Institute of Technology, School of Electrical Engineering and Computer Science, Stockholm, Sweden, 2018. [Google Scholar]
- [53].Tamir M, “A short introduction to using word2vec for text classification,” 2016. https://www.linkedin.com/pulse/short-introduction-using-word2vec-text-classification-mike/ (access March 21, 2023).
- [54].Ahmadvand A, Choi JI, and Agichtein E, “Contextual dialogue act classification for open-domain conversational agents,” in the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, 2019, pp. 1273–1276. [Google Scholar]
- [55].Nadeau D and Sekine S, “A survey of named entity recognition and classification,” Lingvisticae Investig, vol. 30, no. 1, pp. 3–26, 2007. [Google Scholar]
- [56].Schuurmans J and Frasincar F, “Intent classification for dialogue utterances,” IEEE Intell. Syst, vol. 35, no. 1, pp. 82–88, 2019. [Google Scholar]
- [57].Bengio Y, Ducharme R, and Vincent P, “A neural probabilistic language model,” Adv. Neural Inf. Process. Syst, vol. 13, 2000. [Online]. Available: https://proceedings.neurips.cc/paper/2000/file/728f206c2a01bf572b5940d7d9a8fa4c-Paper.pdf [Google Scholar]
- [58].Socher R, Huang E, Pennin J, Manning CD, and Ng A, “Dynamic pooling and unfolding recursive autoencoders for paraphrase detection,” Adv. Neural Inf. Process. Syst, vol. 24, 2011. [Online]. Available: https://papers.nips.cc/paper_files/paper/2011/file/3335881e06d4d23091389226225e17c7-Paper.pdf [Google Scholar]
- [59].Mikolov T, Sutskever I, Chen K, Corrado GS, and Dean J, “Distributed representations of words and phrases and their compositionality,” Adv. Neural Inf. Process. Syst, vol. 26, 2013. [Online]. Available: https://proceedings.neurips.cc/paper/2013/file/9aa42b31882ec039965f3c4923ce901b-Paper.pdf [Google Scholar]
- [60].Golovanov S, Tselousov A, Kurbanov R, and Nikolenko SI, “Lost in conversation: A conversational agent based on the transformer and transfer learning,” in The NeurIPS’18 Competition: From Machine Learning to Intelligent Conversations, 2020, pp. 295–315. [Google Scholar]
- [61].Cahn J, “CHATBOT: Architecture, design, & development,” M.S. thesis. Dep. Computer and Information Science, Univ. of Pennsylvania, Philadelphia, PA, 2017. [Google Scholar]
- [62].Cho K et al. , “Learning phrase representations using RNN encoder-decoder for statistical machine translation,” in Computation and Language, Sep. 2014. [Online]. Available: ArXiv14061078, 2014. [Google Scholar]
- [63].Ramesh K, Ravishankaran S, Joshi A, and Chandrasekaran K, “A survey of design techniques for conversational agents,” presented at the Information, Communication and Computing Technology: Second International Conference, ICICCT 2017, New Delhi, India, May 13, 2017, pp. 336–350. [Google Scholar]
- [64].Sak H, Senior AW, and Beaufays F, “Long short-term memory recurrent neural network architectures for large scale acoustic modeling,” in Neural and Evolutionary Computing, Feb. 2014. [Online]. Available: arXiv:1402.1128 [Google Scholar]
- [65].Vaswani A et al. , “Attention is all you need,” Adv. Neural Inf. Process. Syst, vol. 30, 2017. [Online]. Available: https://arxiv.org/abs/1706.03762 [Google Scholar]
- [66].Srivastava A, “What Are Transformers In NLP And It’s Advantages,” knoldus. 2023. https://blog.knoldus.com/what-are-transformers-in-nlp-and-its-advantages/ (accessed Mar. 21, 2023). [Google Scholar]
- [67].Devlin J, Chang M-W, Lee K, and Toutanova K, “Bert: Pre-training of deep bidirectional transformers for language understanding,” in Computation and Language, May 2019. [Online]. Available: ArXiv181004805, 2018. [Google Scholar]
- [68].Liu Y et al. , “Roberta: A robustly optimized bert pretraining approach,” in Computation and Language, Jul. 2019. [Online]. Available: ArXiv190711692, 2019. [Google Scholar]
- [69].OpenAI, “GPT-4 Technical Report.” [Online]. Available: https://cdn.openai.com/papers/gpt-4.pdf (accessed Mar. 23, 2023).
- [70].Nuruzzaman M and Hussain OK, “A survey on chatbot implementation in customer service industry through deep neural networks,” presented at the 2018 IEEE 15th International Conference on e-Business Engineering (ICEBE), Xi’an, China, Oct. 2018, pp. 54–61. [Google Scholar]
- [71].Chatterjee S, Davies MJ, Heller S, Speight J, Snoek FJ, and Khunti K, “Diabetes structured self-management education programmes: a narrative review and current innovations,” Lancet Diabetes Endocrinol, vol. 6, no. 2, pp. 130–142, 2018. [DOI] [PubMed] [Google Scholar]
- [72].Fitzpatrick KK, Darcy A, and Vierhile M, “Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): A randomized controlled trial,” JMIR Ment. Health, vol. 4, no. 2, p. e7785, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [73].Fulmer R, Joerin A, Gentile B, Lakerink L, and Rauws M, “Using psychological artificial intelligence (Tess) to relieve symptoms of depression and anxiety: randomized controlled trial,” JMIR Ment. Health, vol. 5, no. 4, p. e9782, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [74].Griol D and Callejas Z, “Mobile conversational agents for context-aware care applications,” Cogn. Comput, vol. 8, pp. 336–356, 2016. [Google Scholar]
- [75].Griffin AC et al. , “Conversational agents for chronic disease self-management: A systematic review,” in the AMIA Annual Symposium Proceedings, 2020, vol. 2020, pp. 504–514. [PMC free article] [PubMed] [Google Scholar]
- [76].Schachner T, Keller R, and v Wangenheim F, “Artificial intelligence-based conversational agents for chronic conditions: systematic literature review,” J. Med. Internet Res, vol. 22, no. 9, p. e20701, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [77].l’Allemand D, Shih C, Heldt K, Buchter D, Brogle B, and Ruegger D, “Design and interim evaluation of a smartphone app for overweight adolescents using a behavioural health intervention platform,” Obes Rev, vol. 102, 2018. [Google Scholar]
- [78].Kowatsch T et al. , “Text-based healthcare chatbots supporting patient and health professional teams: Preliminary results of a randomized controlled trial on childhood obesity,” presented at Persuasive Embodied Agents for Behavior Change (PEACH 2017) Workshop, co-located with the 17th International Conference on Intelligent Virtual Agents (IVA 2017), Stockholm, Sweden, Aug 2017. Available: https://www.research-collection.ethz.ch/handle/20.500.11850/348306 [Google Scholar]
- [79].Cheng A, Raghavaraju V, Kanugo J, Handrianto YP, and Shang Y, “Development and evaluation of a healthy coping voice interface application using the Google home for elderly patients with type 2 diabetes,” presented at the 2018 15th IEEE Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, LA, Jan. 2018, pp. 1–5. [Google Scholar]
- [80].Rehman UU, Chang DJ, Jung Y, Akhtar U, Razzaq MA, and Lee S, “Medical instructed real-time assistant for patient with glaucoma and diabetic conditions,” Appl. Sci, vol. 10, no. 7, p. 2216, 2020. [Google Scholar]
- [81].Chaix B et al. , “When chatbots meet patients: one-year prospective study of conversations between patients with breast cancer and a chatbot,” JMIR Cancer, vol. 5, no. 1, p. e12856, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [82].Rhee H, Allen J, Mammen J, and Swift M, “Mobile phone-based asthma self-management aid for adolescents (mASMAA): a feasibility study,” Patient Prefer. Adherence, vol. 8, pp. 63–72, 2014. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [83].Easton K et al. , “A virtual agent to support individuals living with physical and mental comorbidities: co-design and acceptability testing,” J. Med. Internet Res, vol. 21, no. 5, p. e12996, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [84].Khodabakhsh A, Kuscuoglu S, and Demiroglu C, “Detection of Alzheimer’s disease using prosodic cues in conversational speech,” presented at the 2014 22nd signal processing and communications applications conference (SIU), Trabzon, Turkey, Apr. 2014, pp. 1003–1006. [Google Scholar]
- [85].Ferguson G, Quinn J, Horwitz C, Swift M, Allen J, and Galescu L, “Towards a personal health management assistant,” J. Biomed. Inform, vol. 43, no. 5, pp. S13–S16, 2010. [DOI] [PubMed] [Google Scholar]
- [86].Galescu L, Allen J, Ferguson G, Quinn J, and Swift M, “Speech recognition in a dialog system for patient health monitoring,” presented at the 2009 IEEE International Conference on Bioinformatics and Biomedicine Workshop, Washington, DC, Nov. 2009, pp. 302–307. [Google Scholar]
- [87].Lobo J, Ferreira L, and Ferreira AJ, “CARMIE: a conversational medication assistant for heart failure,” Int. J. E-Health Med. Commun. IJEHMC, vol. 8, no. 4, pp. 21–37, 2017. [Google Scholar]
- [88].Ireland D et al. , “Hello Harlie: enabling speech monitoring through chat-bot conversations,” Stud. Health Technol. Inform, vol. 227, pp. 55–60, 2016. [PubMed] [Google Scholar]
- [89].Rose-Davis B, Van Woensel W, Stringer E, Abidi S, and Abidi SSR, “Using an artificial intelligence-based argument theory to generate automated patient education dialogues for families of children with juvenile idiopathic arthritis,” in MEDINFO 2019: Health and Wellbeing e-Networks for All, IOS Press, 2019, pp. 1337–1341. [DOI] [PubMed] [Google Scholar]
- [90].Azzini I et al. , “Automated spoken dialog system for home care and data acquisition from chronic patients,” in The New Navigators: from Professionals to Patients, IOS Press, 2003, pp. 146–151. [PubMed] [Google Scholar]
- [91].Ooster J, Moreta PNP, Bach J-H, Holube I, and Meyer BT, “‘ Computer, test my hearing’: Accurate speech audiometry with smart speakers.,” presented at the INTERSPEECH, Graz, Austria, Sep. 2019, pp. 4095–4099. [Google Scholar]
- [92].Elmasri D and Maeder A, “A conversational agent for an online mental health intervention,” presented at the Brain Informatics and Health: International Conference, BIH 2016, Omaha, NE, Oct. 2016, pp. 243–251. [Google Scholar]
- [93].Kazemi DM, Cochran AR, Kelly JF, Cornelius JB, and Belk C, “Integrating mHealth mobile applications to reduce high risk drinking among underage students,” Health Educ. J, vol. 73, no. 3, pp. 262–273, 2014. [Google Scholar]
- [94].Ly KH, Ly A-M, and Andersson G, “A fully automated conversational agent for promoting mental well-being: A pilot RCT using mixed methods,” Internet Interv, vol. 10, pp. 39–46, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [95].Stein N and Brooks K, “A fully automated conversational artificial intelligence for weight loss: Longitudinal observational study among overweight and obese adults,” JMIR Diabetes, vol. 2, no. 2, p. e8590, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [96].Schroeder J et al. , “Pocket skills: A conversational mobile web app to support dialectical behavioral therapy,” In the 2018 CHI Conference on Human Factors in Computing Systems, 2018, pp. 1–15. [Google Scholar]
- [97].Wang H, Zhang Q, Ip M, and Lau JTF, “Social media–based conversational agents for health management and interventions,” Computer, vol. 51, no. 8, pp. 26–33, 2018. [Google Scholar]
- [98].Gaffney H, Mansell W, Edwards R, and Wright J, “Manage your life online (MYLO): A pilot trial of a conversational computer-based intervention for problem solving in a student sample,” Behav. Cogn. Psychother, vol. 42, no. 6, pp. 731–746, 2014. [DOI] [PubMed] [Google Scholar]
- [99].Tsiourti C, Joly E, Wings C, Moussa MB, and Wac K, “Virtual assistive companions for older adults: qualitative field study and design implications,” in the 8th International Conference on Pervasive Computing Technologies for Healthcare, 2014, pp. 57–64. [Google Scholar]
- [100].van Heerden A, Ntinga X, and Vilakazi K, “The potential of conversational agents to provide a rapid HIV counseling and testing services,” presented at the 2017 International Conference on the Frontiers and Advances in Data Science (FADS), Salzburg, Austria, Oct. 2017, pp. 80–85. [Google Scholar]
- [101].Bickmore T, Gruber A, and Picard R, “Establishing the computer–patient working alliance in automated health behavior change interventions,” Patient Educ. Couns, vol. 59, no. 1, pp. 21–30, 2005. [DOI] [PubMed] [Google Scholar]
- [102].Bodenheimer T, Lorig K, Holman H, and Grumbach K, “Patient self-management of chronic disease in primary care,” JAMA, vol. 288, no. 19, pp. 2469–2475, 2002. [DOI] [PubMed] [Google Scholar]
- [103].Lenferink A et al. , “Self‐management interventions including action plans for exacerbations versus usual care in patients with chronic obstructive pulmonary disease,” Cochrane Database Syst. Rev, vol. 8, no. 8, p. CD011682, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [104].Amith M et al. , “Examining potential usability and health beliefs among young adults using a conversational agent for HPV vaccine counseling,” in AMIA Summits Transl. Sci. Proc, vol. 2020, pp. 43–52, 2020. [PMC free article] [PubMed] [Google Scholar]
- [105].Boyd M and Wilson N, “Just ask Siri? A pilot study comparing smartphone digital assistants and laptop Google searches for smoking cessation advice,” PLoS One, vol. 13, no. 3, p. e0194811, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [106].Gaffney H, Mansell W, and Tai S, “Conversational agents in the treatment of mental health problems: mixed-method systematic review,” JMIR Ment. Health, vol. 6, no. 10, p. e14166, 2019. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [107].Horvath AO and Luborsky L, “The role of the therapeutic alliance in psychotherapy,” J. Consult. Clin. Psychol, vol. 61, no. 4, pp. 561–573, 1993. [DOI] [PubMed] [Google Scholar]
- [108].Miner G, Delen D, Elder J, Fast A, Hill T, and Nisbet R, Practical Text Mining and Statistical Analysis for Non-structured Text Data Applications. Waltham, MA: Elsevier, 2012. [Google Scholar]
- [109].Sezgin E, Huang Y, Ramtekkar U, and Lin S, “Readiness for voice assistants to support healthcare delivery during a health crisis and pandemic,” NPJ Digit. Med, vol. 3, no. 1, p. 122, 2020. Available: https://www.nature.com/articles/s41746-020-00332-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- [110].Klein J, Moon Y, and Picard R, “This computer responds to user frustration: Theory, design, results,” Interacting with Computers, vol. 14, no. 2, pp. 119–140, 2002. [Google Scholar]
- [111].Barata M, Galih Salman A, Faahakhododo I, and Kanigoro B, “Android based voice assistant for blind people,” Libr. Hi Tech News, vol. 35, no. 6, pp. 9–11, 2018. [Google Scholar]
- [112].Balasuriya SS, Sitbon L, Bayor AA, Hoogstrate M, and Brereton M, “Use of voice activated interfaces by people with intellectual disability,” in the 30th Australian Conference on Computer-Human Interaction, 2018, pp. 102–112. [Google Scholar]
- [113].Masina F et al. , “Investigating the accessibility of voice assistants with impaired users: mixed methods study,” J. Med. Internet Res, vol. 22, no. 9, p. e18431, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [114].Freeman D et al. , “Automated psychological therapy using immersive virtual reality for treatment of fear of heights: a single-blind, parallel-group, randomised controlled trial,” Lancet Psychiatry, vol. 5, no. 8, pp. 625–632, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [115].Gardiner PM et al. , “Engaging women with an embodied conversational agent to deliver mindfulness and lifestyle recommendations: A feasibility randomized control trial,” Patient Educ. Couns, vol. 100, no. 9, pp. 1720–1729, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [116].Inkster B, Sarda S, and Subramanian V, “An empathy-driven, conversational artificial intelligence agent (Wysa) for digital mental well-being: real-world data evaluation mixed-methods study,” JMIR MHealth UHealth, vol. 6, no. 11, p. e12106, 2018. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [117].Pinto MD, Greenblatt AM, Hickman RL, Rice HM, Thomas TL, and Clochesy JM, “Assessing the critical parameters of eSMART-MH: a promising avatar-based digital therapeutic intervention to reduce depressive symptoms,” Perspect. Psychiatr. Care, vol. 52, no. 3, pp. 157–168, 2016. [DOI] [PubMed] [Google Scholar]
- [118].Mead N and Bower P, “Patient-centred consultations and outcomes in primary care: a review of the literature,” Patient Educ. Couns, vol. 48, no. 1, pp. 51–61, 2002. [DOI] [PubMed] [Google Scholar]
- [119].Martin DJ, Garske JP, and Davis MK, “Relation of the therapeutic alliance with outcome and other variables: a meta-analytic review.,” J. Consult. Clin. Psychol, vol. 68, no. 3, p. 438, 2000. [PubMed] [Google Scholar]
- [120].Burton C et al. , “Pilot randomised controlled trial of Help4Mood, an embodied virtual agent-based system to support treatment of depression,” J. Telemed. Telecare, vol. 22, no. 6, pp. 348–355, 2016. [DOI] [PubMed] [Google Scholar]
- [121].Pinto MD, Hickman RL Jr, Clochesy J, and Buchner M, “Avatar-based depression self-management technology: Promising approach to improve depressive symptoms among young adults,” Appl. Nurs. Res, vol. 26, no. 1, pp. 45–48, 2013. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [122].Baskar J and Lindgren H, “Human-agent dialogues on health topics-an evaluation study,” presented at the Highlights of Practical Applications of Agents, Multi-Agent Systems, and Sustainability-The PAAMS Collection: International Workshops of PAAMS 2015, Salamanca, Spain, Jun. 2015, pp. 28–39. [Google Scholar]
- [123].Kadariya D, Venkataramanan R, Yip HY, Kalra M, Thirunarayanan K, and Sheth A, “kBot: knowledge-enabled personalized chatbot for asthma self-management,” presented at the 2019 IEEE International Conference on Smart Computing (SMARTCOMP), Washington, DC, Jun. 2019, pp. 138–143. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [124].Myin-Germeys I, Klippel A, Steinhart H, and Reininghaus U, “Ecological momentary interventions in psychiatry,” Curr. Opin. Psychiatry, vol. 29, no. 4, pp. 258–263, 2016. [DOI] [PubMed] [Google Scholar]
- [125].Prochaska JO, “Transtheoretical model of behavior change,” Encycl. Behav. Med, vol. 12, no. 1, pp. 2266–2270, 2020. [Google Scholar]
- [126].Powers WT and Powers WT, Behavior: The Control of Perception. New Canaan, Conn: Benchmark Publications, 1973. [Google Scholar]
- [127].Watson A, Bickmore T, Cange A, Kulshreshtha A, and Kvedar J, “An internet-based virtual coach to promote physical activity adherence in overweight adults: randomized controlled trial,” J. Med. Internet Res, vol. 14, no. 1, p. e1629, 2012. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [128].Kingdon DG and Turkington D, Cognitive-Behavioral Therapy of Schizophrenia. New York: NY: Routledge, 2022. [Google Scholar]
- [129].May JM, Richardi TM, and Barth KS, “Dialectical behavior therapy as treatment for borderline personality disorder,” Ment. Health Clin, vol. 6, no. 2, pp. 62–67, 2016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [130].Bin Sawad A et al. , “A systematic review on healthcare artificial intelligent conversational agents for chronic conditions,” Sensors, vol. 22, no. 7, p. 2625, 2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [131].Shamekhi A and Bickmore T, “Breathe deep: A breath-sensitive interactive meditation coach,” in the Proceedings of the 12th EAI International Conference on Pervasive Computing Technologies for Healthcare, 2018, pp. 108–117. [Google Scholar]
- [132].Piau A, Crissey R, Brechemier D, Balardy L, and Nourhashemi F, “A smartphone Chatbot application to optimize monitoring of older patients with cancer,” Int. J. Med. Inf, vol. 128, pp. 18–23, 2019. [DOI] [PubMed] [Google Scholar]
- [133].Bickmore TW, Mitchell SE, Jack BW, Paasche-Orlow MK, Pfeifer LM, and O’Donnell J, “Response to a relational agent by hospital patients with depressive symptoms,” Interact. Comput, vol. 22, no. 4, pp. 289–298, 2010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [134].Philip P et al. , “Virtual human as a new diagnostic tool, a proof of concept study in the field of major depressive disorders,” Sci. Rep, vol. 7, no. 1, pp. 1–7, 2017. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [135].Baptista S, Wadley G, Bird D, Oldenburg B, and Speight J, “The my diabetes coach research group acceptability of an embodied conversational agent for type 2 diabetes self-management education and support via a smartphone app: Mixed methods study,” JMIR MHealth UHealth, vol. 8, p. e17038, 2020. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [136].Bird T, Mansell W, Wright J, Gaffney H, and Tai S, “Manage your life online: a web-based randomized controlled trial evaluating the effectiveness of a problem-solving intervention in a student sample,” Behav. Cogn. Psychother, vol. 46, no. 5, pp. 570–582, 2018. [DOI] [PubMed] [Google Scholar]
- [137].Eysenbach G and Consort-EHEALTH Group, “CONSORT-EHEALTH: Improving and standardizing evaluation reports of Web-based and mobile health interventions,” J. Med. Internet Res, vol. 13, no. 4, p. e1923, 2011. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [138].Des Jarlais DC, Lyles C, Crepaz N, and Trend Group, “Improving the reporting quality of nonrandomized evaluations of behavioral and public health interventions: The TREND statement,” Am. J. Public Health, vol. 94, no. 3, pp. 361–366, 2004. [DOI] [PMC free article] [PubMed] [Google Scholar]
- [139].Bossuyt PM et al. , “STARD 2015: An updated list of essential items for reporting diagnostic accuracy studies,” Clin. Chem, vol. 61, no. 12, pp. 1446–1452, 2015. [DOI] [PubMed] [Google Scholar]
