Abstract
Objectives
We conducted an implementation planning process during the pilot phase of a pragmatic trial, which tests an intervention guided by artificial intelligence (AI) analytics sourced from noninvasive monitoring data in heart failure patients (LINK-HF2).
Materials and methods
A mixed-method analysis was conducted at 2 pilot sites. Interviews were conducted with 12 of 27 enrolled patients and with 13 participating clinicians. iPARIHS constructs were used for interview construction to identify workflow, communication patterns, and clinician’s beliefs. Interviews were transcribed and analyzed using inductive coding protocols to identify key themes. Behavioral response data from the AI-generated notifications were collected.
Results
Clinicians responded to notifications within 24 hours in 95% of instances, with 26.7% resulting in clinical action. Four implementation themes emerged: (1) High anticipatory expectations for reliable patient communications, reduced patient burden, and less proactive provider monitoring. (2) The AI notifications required a differential and tailored balance of trust and action advice related to role. (3) Clinic experience with other home-based programs influenced utilization. (4) Responding to notifications involved significant effort, including electronic health record (EHR) review, patient contact, and consultation with other clinicians.
Discussion
Clinician’s use of AI data is a function of beliefs regarding the trustworthiness and usefulness of the data, the degree of autonomy in professional roles, and the cognitive effort involved.
Conclusion
The implementation planning analysis guided development of strategies that addressed communication technology, patient education, and EHR integration to reduce clinician and patient burden in the subsequent main randomized phase of the trial. Our results provide important insights into the unique implications of implementing AI analytics into clinical workflow.
Keywords: artificial intelligence, heart failure, home monitoring, computerized decision support, implementation
Background and significance
Heart failure (HF) is a common medical condition with prevalence increasing with age.1 HF is one of the most frequent discharge diagnoses in the United States.2 The morbidity related to HF remains very high; the rate of readmission for HF is approximately 20% at 30 days and 30% at 90 days after discharge. Readmission in HF patients is related to adverse outcomes including increased mortality.3,4 Reduction of HF readmissions is, therefore, a strategic goal for both civilian and Veterans Health Administration hospitals.
Clinicians and patients are typically unaware of the physiologic changes of early HF exacerbation. Once overt symptoms and signs of HF exacerbation are present, it usually is too late to prevent hospitalization.5 This may, in-part, explain why monitoring of parameters such as body weight or blood pressure has not resulted in reduction of hospitalizations in previous clinical investigations.6,7 Importantly, detection of HF exacerbation at its early stages could provide an opportunity for therapeutic intervention that would, in turn, decrease the risk of hospital admission.5
In an earlier study, we tested a noninvasive remote patient monitoring system consisting of a wearable disposable multisensory patch (sensor) placed on the patient chest that recorded continuous physiological data. We showed that artificial intelligence (AI) methods based on machine learning analysis of the continuous data streams accurately predicted HF rehospitalization approximately 1 week before the hospitalization took place. In addition, we demonstrated high patient adherence rate with the use of the sensor.8 The next step would be to determine whether early, standardized intervention based on this AI predictive algorithm could result in improved outcomes.
Use of AI in healthcare is gaining more focus across academia and industry.9–11 Despite rising popularity, however, few real-world results of integration of AI analytics into clinical care have been published. Several authors have suggested that the lack of information and guidance on implementation is a limiting factor.12–14 In addition, systematic reviews on the use of AI in clinical care also suggest that implementation in clinical practice is now the most important barrier to AI adoption.13,15,16 Investigations on the barriers and facilitators for effective implementation of AI tools are a current research imperative. Implementation planning research explores factors that might impact actual use in a real setting17 which can be incorporated into a real-world pragmatic trial of an intervention’s effectiveness.
Continuous wearable monitoring analytics predict heart failure decompensation (LINK-HF2) is a pragmatic trial consisting of 2 phases. The nonrandomized pilot phase that explores issues related to implementation and the main randomized phase that implements the intervention in a real-world setting (Figure 1). This report presents results of the implementation planning conducted during the pilot phase, which focused on provider and patient beliefs, workflow, possible barriers and facilitators, and organizational factors that could impact the fidelity of the clinical intervention.17,18 Specific attention was paid to the various issues of implementing AI in clinical settings,9,13 including a multi-level approach,19 tailoring strategies by context,20–22 and incorporating internal formative evaluation metrics into the pragmatic trial itself.23
Figure 1.
Study design of the LINK-HF2 pragmatic trial.
Our specific objectives were to: (1) Characterize the workflow, environmental constraints, and communication channels relevant to implementation planning of an ambulatory monitoring program at the clinic and team level; (2) characterize provider perceptions of using AI data generated from an ambulatory monitoring system relevant to implementation planning in a real clinical setting; and (3) design generalizable and measurable implementation metrics relevant to implementation planning in a real clinical setting for a future pragmatic trial intervention.
Materials and methods
Setting and patient population
This work was conducted at 2 VA sites located in different geographical areas, with different patient populations and clinicians. The study received Institutional Review Board approval and Research and Development Committee approval at all institutions. All subjects provided written informed consent for study participation. Both sites had cardiac care services. Table 1 describes the clinics and settings.
Table 1.
Description of pilot settings.
Setting | Patients (n) | Clinicians (n) | Description |
---|---|---|---|
Salt Lake City, UT VA Cardiac Care |
17 | 9 | Academic urban hospital with 160 HF inpatient visits/year. Cardiac care service sees 763 unique HF patients/year. |
Gainesville, FL VA Cardiac Care |
10 | 5 | Academic urban hospital with 190 HF inpatient visits/year. Cardiac care service sees 717 unique HF patients per year. |
Description of design
LINK-HF2 is a pragmatic controlled trial funded by the VA’s Health Services Research and Development Service (HSR&D). It consists of 2 phases, a nonrandomized implementation planning pilot phase and the main randomized phase (Figure 1).
In the nonrandomized pilot phase, all patients had the study sensor were monitored remotely and all notifications were transferred to the clinical team. The design of the pilot work was both qualitative (interviews and observations) and quantitative (behavioral data) using a cohort design that included patients discharged after HF hospitalization as potential participants.
In the subsequent main randomized phase, all patients will wear the sensor, but will be randomly assigned to a clinician-monitoring group, where the clinicians receive notifications from the AI analytics, versus a control group, where the AI notifications are not communicated to the clinical team.
Description of intervention and study procedures
Study participants are enrolled at the time of discharge from a HF hospitalization when they are trained on how to activate and apply the disposable sensor and pair it with a provided cell phone. Patients are expected to wear the sensor 24 hours a day for up to 90 days and are instructed to change the sensor when the battery is depleted. In addition, patients complete a short version of the Kansas City Cardiomyopathy Questionnaire (KCCQ-12) and a visual analog scale at the time of enrollment, at 30 days and at 90 days after enrollment. No constraints are placed on subjects’ activities.24,25
Continuous data were automatically uploaded from the sensor to the cell phone and then to a cloud-based server for analytical processing (PhysIQ portal). The system includes a cloud-based analytics platform that implements a machine learning method called similarity-based modeling that builds a personalized model of physiology behavior based on the patient’s initial data.26 Physiological data obtained include heart rate, respiratory rate, and activity level. Windowed evaluations of relative change from baseline in the behavior of the monitored physiological data are used to generate notifications that trigger the intervention workflow we studied (Appendix SA). When changes that correlate with impending HF exacerbation are identified, a clinical notification is generated, messaged to site coordinators by email and forwarded to clinical team through patient’s electronic health record (EHR). Clinicians follow a standardized response algorithm aimed on altering therapy in patients with suspected incipient HF decompensation (Figure 2).
Figure 2.
Algorithmic decision protocol for response to notifications.
Stakeholder needs assessment in the pilot phase
Clinicians
Clinicians were interviewed 1-on-1 via secure video technology using a semistructured approach which focused on workflow, organizational structure, communication processes and technology, and general attitudes toward the intervention (Appendix SB). The initial 13 interviews were audiorecorded and transcribed. In addition, 5 observations of clinicians (2 RNs, 1 PharmD at first site and 2 RNs at the second site) preparing for patient calls post notification were conducted to identify information needs, EHR use, team communication, and patterns of preparation. All members of the clinic staff in each setting agreed to be interviewed and observed. The team members ranged in age from 29 to 63 years, averaged 13.3 years of experience in the VA, and 21.4 years of clinical experience.
Patients
Twelve patients were interviewed by phone (due to COVID restrictions) using a structured interview form with specific questions (Appendix SC). The phone calls were not recorded, but patient answers were recorded manually. The patients ranged in age from 61 to 82 years were all male, and had used the monitoring system for 2 weeks on average. The interviewers had advanced training in human factors engineering and biomedical informatics.
Qualitative analysis
Promoting Action on Research Implementation in Health Services (i-PARIHS) was the specific theoretical framework for implementation used in this study and includes 3 broad areas of focus: (1) Innovation (characteristics of the intervention), (2) Recipients (beliefs, attitudes, experience of participants), and (3) Context (culture, leadership, and policies). Each focus has subheadings which are discussed in the Results section.27Facilitation is viewed as the underlying mechanism of action and is defined as “the construct that activates implementation through assessing and responding to characteristics of the innovation and the recipients (both as individuals and in teams) within their contextual setting.”27 The overall goal of the planning process is to identify key strategies and tools that facilitators could use to tailor the AI intervention appropriately, that can contribute to an i-PARIHS facilitation planning tool, and that can be measured during the RCT portion of the pragmatic trial.28,29
A hybrid analytical approach was used to qualitatively analyze the interview transcripts. The process involved an initial open-coding, inductive review followed by deductive categorization using theoretical constructs.30 The analysis was conducted in 3 stages. Individual members of the research team conducted an initial review of transcripts to identify concepts and constructs of interest using NVivo (@Lumivero, 2018). After group discussion and refinement of codes, the team conducted a theoretical recoding and reanalysis based on i-PARIHS factors. Subcategories were rereviewed by 2 investigators to ensure consistency. The extracted quotations were again reviewed by 3 investigators (1 MD and 2 PhD cognitive psychologists) in order to agree on overarching themes. Table 2 describes the i-PARIHS constructs.
Table 2.
i-PARIHS theoretical constructs and description.
Constructs | Description |
---|---|
Innovation | The degree to which the intervention “fits” the context, the amount of familiarity that clinicians have with components of the intervention, trialability, and perceived relative advantage. |
Recipients | Attitudes, roles, power structures, communication networks, goals, time, resources, and perceived support. Includes local organizational structures and technology. |
Context | Three-level construct; (1) local (clinic) culture, organizational leadership, evaluation and feedback processes, workflow metrics, and learning environment; (2) organizational priorities, absorptive capacity, experience with similar changes, learning networks; (3) national priorities, regulatory requirements, incentives, and policy drivers. |
Penetration | The degree to which the intervention is integrated as a practice within a service setting and its subsystems. Similar to the concept of “breadth” of adoption. |
Sustainability | Defined as the extent to which a newly implemented treatment is maintained or institutionalized within a service setting’s ongoing, stable operations. It could also emphasize the integration of a given program within an organization’s culture through policies and practices as well. |
Rigor and reproducibility
We addressed qualitative coding rigor by referencing the Standards for Reporting Qualitative Research recommendations31 and addressing recommended constructs for establishing reliability and validity in qualitative research.32
Behavioral and adoption data
Data collected from the PhysIQ portal covered the 90-day pilot study period and 27 enrolled patients. All notifications were tracked automatically and the specific information collected included: (1) The number of notifications per patient and the number per event; (2) Response time for the study coordinators to acknowledge; (3) Response time to clinician response; (4) Type of response; (5) Timing of response; (6) Review of effort, information needs, and patient interaction; (7) Communication patterns between clinicians; and (8) Patient response.
Designing implementation evaluation metrics
An important aim of the pilot work was to propose measurable implementation metrics to be used by future facilitators. We used the RE-AIM (Reach, Effectiveness, Adoption, Implementation and Maintenance) framework for this purpose. RE-AIM is a planning and evaluation model widely used in public health, clinical implementation studies and in informatics research consisting of 5 dimensions—reach, effectiveness, adoption, implementation, and maintenance.23,33,34 RE-AIM has been updated to include more attention to contextual issues35 and to include implementation planning.36 Development of the metrics consisted of multiple and iterative discussions within the research team.
Results
The results are presented in 3 sections: (1) Stakeholder needs assessment findings which include the provider and patient interviews, (2) behavioral results which include provider responses to notifications, notification data, and the observation data, and (3) proposed final RE-AIM implementation metrics.
Clinician stakeholder needs assessment
The results reflect the content coding of the transcripts of provider and patient interviews using the i-PARIHS constructs. Table 3 outlines broader categories of identified stakeholder needs with example quotations.
Table 3.
Qualitative coding examples of i-PARIHS constructs from clinicians and staff.
Innovation | Example quotations |
---|---|
|
|
|
|
|
|
| |
Recipient | Quotations |
| |
|
|
|
|
|
|
| |
Context | Quotations |
| |
|
|
|
“It’s a self-consult consulting service. In other words, we get a list that’s generated of who’s been admitted. Look through that list. That’s heart failure… Dr. xx (Attending) and the fellow, the cardiology fellow will go take up manage to impatient with those patients, and then we’ll put them into clinic within you know, seven to 14 days after discharge.”(RN) |
|
“… if it’s a difficult adjustment, or like there’s a lot of moving pieces that need to be addressed, then I’ll usually consult with the pharmacist or the nurse practitioner or the attending. I don’t necessarily cosign them to the note unless they request to be or unless I need a response from them. Or if it’s something that’s very tenuous situation that that I want to make sure that I have their signature on there to back me up.” (RN) |
|
“And then it's really kind of tricky, because CPRS does not have a good way for us to like run a list or do data mining or anything like that. So, we have to be really careful on these patients, not to lose them to follow up….” (RN) |
Patient stakeholder experience interview results
In total, 27 patients were enrolled in the pilot phase of the study and 12 of the patients were interviewed. All interviewed patients reported satisfaction with the usability of the equipment and the effectiveness of clinical support. Eight patients reported some initial difficulty in getting used to the equipment but reported it was “easy” and did not interfere with their general quality of life, social activities, or movement. Qualitative responses noted that they “felt safer” knowing they were being monitored. Three commented on some negative feelings of “being measured all the time” or to always being reminded of “how fat I am.” Ten expressed beliefs that the remote monitoring system would minimize trips to the clinic and to the hospital, which many veterans reported experiencing as a burden.
Behavioral and adoption data
Over a 90-day study period, 27 patients were enrolled, all were White males, with a mean age of 72.6 ± 8.6 years. Of these, 15 (55.6%) had HF with reduced ejection fraction, and 12 (44.4%) had HF with preserved ejection fraction. A total of 135 notifications were generated, including 38 repeat notifications for the same event. Of the 27 patients, 17 had at least 1 notification generated (63%) resulting in a median of 5 notifications per patient. Of the notifications that were responded to, clinicians responded to > 95% within 24 hours. In 26.7% of the notifications, clinicians intensified the patient’s medication regimen, scheduled a follow-up clinic appointment or referred the patient to the emergency department. In 40.7% of the cases, clinicians monitored the patients over the following 72-hour period. For 32.6% of the notifications, clinicians did not achieve patient contact. In 28.2%, multiple notifications were generated during 72 hours following the initial notification; clinicians identified these as repeated notifications and did not pursue contact. In 4.4% of the notifications, contact efforts were unsuccessful (Table 4). Of the 27 patients, 7 were hospitalized and 2 died.
Table 4.
RE-AIM based tracking metrics.
Construct | Description/definition | Metrics | Data source |
---|---|---|---|
Reach
|
Who is the target patient population for the LINK intervention? How many are agreeing to enroll? How long do they stay in the program? Which clinicians are involved and what are their roles? |
Patients:
Clinicians:
|
Research coordinator tracking |
Implementation | How is the notification understood by clinicians and patients? How can the notification be integrated into the workflow and communication strategies? What is the length of time to respond to notifications? Is the documentation chain complete? What technologies are involved in documenting the notification, communicating the notification, and integrating the notification into practice? |
|
Team OpenClinicaa |
System response times
|
VA CPRS | ||
Responses to notification (count):
|
|
||
Effectiveness | Is the monitoring platform working? How well do the patches work? (# that fail, days of nonrecordings). |
|
|
Adoption | Individual provider level | Qualitative data on satisfaction and reports of burden | Interviews at culmination of study |
Adoption | Clinic level | Descriptive data on resources use, enrollment, and dropout rates | Interviews at culmination of study |
Maintenance | Program level | Annual and monthly reports of patients’ enrollment and dropout rates, data on provider resources, and cost of program | VA CPRS annual and monthly reports |
OpenClinica is the study database linked to the PhysIQ platform.
The clinicians’ behavioral responses varied. All clinicians used the same general approach to respond to notification: (1) Read the notification and determine when to call the patient; (2) review in the patient’s EHR medications, labs, and recent appointment notes; (3) call the patient and assess the patient’s clinical status, asking specifically about weight gain, blood pressure, dizziness, dyspnea, edema, and current medication profile. The order of the questions varied. Only in 1 of the 5 observations did the clinician review data on the remote monitoring portal. Clinicians waited to make calls depending on their time burden and whether or not they knew the patient. Some patients had several notifications in a row, which resulted in some clinicians responding the following day. Observed chart review lasted 3-15 min and calls lasted between 12 and 22 min. There was a variation between clinicians on how much they engaged the patients in problem-solving. The template used in the progress notes to document notification responses did not pull in clinical data, which created extra effort. The template did provide some decision support instructions.
Implementation implications: The sensitivity, frequency, and information display of the notifications was modified based on this initial experience. The major reasons were: (1) Notifications were sent out repeatedly for the same event, resulting in clinician notification fatigue and complaints; (2) the information provided to clinicians on the web-based portal was expanded in order to assist the clinician with interpretation and to encourage portal use; and (3) the decision options were expanded from the original 3 options to 4, in order to include an “Increased monitoring” (or “watch and wait”) option based on clinicians’ suggestions.
Thematic description and recommended strategies
The results of data collection across interviews, behavior, and system performance were integrated in order to design initial implementation strategies. The thematic results and their implementation implications are as follows:
Theme 1: Anticipatory expectations for reliable patient communications, reduced patient burden, and less proactive provider monitoring were high : Clinicians expected that the system would replace the need for patients to supply information proactively, making the information flow more reliable and less burdensome on patients. A noninvasive intervention was expected to minimize harm and maximize access. Observations of the early patient-provider calls indicated a desire for both the patient and the clinician to identify “causes” resulting in a notification.
Implementation implications: Both the patient and the clinician expected that if they receive a notification, a cause could be identified. More information could be included with the notification and the EHR could be programmed to “pull in” critical patient history that could be reviewed “at a glance” as part of a note template.
Theme 2: The degree of trust required for the notification is associated with both the action implications and the role. RNs need to trust that the MDs are in agreement regarding the action decisions and would “have their back.” In other words, RNs will revert to “script” and trust the real-life patient reports more than the notification if they are not sure of their support.
The other component of trust for RNs is that the system is accurately “catching” patients with problems, thereby minimizing their constant monitoring work.
However, MDs tended to assume that the notification would provide more actionable information that goes beyond symptoms, but is based on the AI findings. Without specific decision support, MDs might then also simply go back to relying on patient symptoms.
Implementation implications: The notification might be perceived as more trustworthy and therefore more actionable by RNs if clinic leadership explicitly embedded it into current scope of practice protocols for RNs. For MDs, the notification may require both more information from the portal regarding the basis of the notification as well as specific action recommendations.
Theme 3: Prior clinic experience with other remote monitoring local programs influenced utilization : Sites appear to have significant and varied experience in home monitoring and other complex monitoring programs that include remote service, telecommunication tools, and experienced staff. The result is variability in both the monitoring components and the information/communication tools used.
Implementation implications: Key program workflow components that need to be addressed include: (1) Identification of high-risk patients; (2) assignment of responsibility for tracking; (3) assignment of responsibility for making calls; (4) specifying data exchanged (EHR, patient, monitoring devices), (5) documentation of content in notes; and (6) definition of rules for who else is informed. Communication flow should be assessed before implementation and the tools need to be standardized across the clinic.
Theme 4: Notification response involved significant effort, including EHR review, patient contact, and consultation with other clinicians : The conversation usually involved more than talking with the patient, but also reviewing several other data sources in EHR as well as other provider notes and plans. This requires a mental model of the care network and it helps when clinicians know the patient.
Implementation implications: Efforts will be needed to minimize burden for clinicians, such as enrolling the patient in the continuity HF clinic so that they are familiar to staff (an action that was implemented), pulling in key information such as medications, lab values, etc. from the EHR into the monitoring note template to reduce the need for chart review, and possibly asking the patients to send in their vital signs and weight before the follow-up call if increased monitoring was selected in response to a notification.
Implementation tracking metrics
Designing RE-AIM constructs to be used during the randomized phase of the pragmatic trial was an important implementation strategy (Table 4). The identified metrics were recommended to be used by the implementation team to monitor specific strategies and to track progress.
Discussion
Summary of findings
This pilot implementation phase of the LINK-HF2 pragmatic trial identified important items for implementation planning of an ambulatory monitoring intervention that provides clinicians with an AI generated notification indicating high likelihood of incipient HF exacerbation. Behavioral tracking of clinician’s response to notifications showed adequate response times. Observation of the 2 early intervention sites showed varied workflow patterns, communication tools and barriers to integrating the process into workflow. The results of the implementation work were used to modify study procedures in the randomized phase of the pragmatic trial. These changes included mapping communication technologies and notification displays to different roles, enhancing information displays to improve explainability, reducing work burden by adding an “Increased Monitoring” response option and increasing provider and patient education. In addition, the results informed the development of implementation process metrics to be used during the randomized phase of the study.
Social structures and communication patterns related to AI decision support
Our findings highlighted the necessity to tailor processes related to AI facilitated decision support to the particular healthcare setting, since we identified notable differences even among just 2 clinics. Variations in clinic structure affected how patient care was managed, the team dynamics, collaboration, and the use of communication technologies. Crucial elements in the implementation of AI decision support include social influences and group dynamics, resonating with Shaw and Greenhalgh’s concept of technology adoption as a social process.12,37–39 Effective implementation in clinics relies on establishing trust and interpreting AI notifications within their proper context.40
The utility of AI notifications varies according to how they are integrated into the clinical expertise of different roles, such as pharmacists, nurses, and doctors, each adapting the information to their specific needs.38,39,41 Implementation strategies need to accommodate these diverse applications. While clinicians generally anticipate AI to enhance workflow and decrease mental burden, actual responses to AI notifications highlighted a need for clearer actionable guidance.
The primary challenge in implementation is to strengthen the link between AI notifications and clear action steps. A lack of explicit direction raises the potential for automation bias and increased cognitive strain.42 Employing visual aids and automated documentation tools may be crucial in alleviating cognitive demands.40,42,43
Clinician mental models of AI
Another major finding was the variety of cognitive interpretations and mental models clinicians employed in response to AI notifications. The notification itself came with limited information or decision rules. The clinicians tended to engage in habitual schematic responses when calling the patient. If the patient had no current symptoms, no change in vital signs, and no increase in weight, then the incentive to make a therapeutic change was lower, and the increased monitoring choice was a frequent response (this was especially true for nurses). However, since the AI analytics are likely to detect HF exacerbation before overt changes in signs and symptoms, a timely intervention in the absence of other clinical warning signs might be more effective but requires very explicit guidance.
Implementing AI decision support is unique, as the guidance provided through analytics is often viewed as a “black box.” The concept of “explainability” has emerged in previous studies as critical to physician trust and adoption, and a variety of recommendations on how to address it have been proposed.44–48 Combi et al. described a model of explainability that includes interpretability—the user can predict what the AI recommendation would do, understandability—the user has an accurate mental model of how it works, usefulness—the recommendations are useful and guide action, and usability—the system is easy to use.49
Further work is needed on design an interface to improve all aspects of AI decision support, including role-based targeted explainability, links to role-based specific actions, and careful attention to the cognitive load needed for interpretation.45 In our study, clinicians had an option to access a clinician portal with granular data, but few did. The notification did not have guidance on which specific action should be taken, nor was it integrated with other EHR data, resulting in increased cognitive load. Similar barriers have also been described by others.45 Further, the assessments of provider attitudes to AI-enabled decision support should go beyond determining whether the attitudes are positive or negative, and explore provider understanding of the AI data. Holzinger et al. suggest new human-AI interfaces are needed to enable clinicians to re-enact and retrace AI results and as such improve their level of understanding of the results.50
Our study has certain limitations. We only used 2 sites in the pilot work, and the patient and clinician sample were relatively small. However, we interviewed all team members that interacted with the system and the sites were geographically diverse and varied in their work processes. Although the clinician participants received training in the use of the remote monitoring system, no evaluation of skills was conducted.
Conclusion
We used implementation results from a pilot phase of a study of AI-based analytics in HF to develop strategies to address communication technology, patient and clinician education and EHR integration. The resulting modifications will increase efficiency and reduce shareholder demand in the randomized phase of the study. Our results also provide important insights into a broader application of AI analytics into clinical workflow.
Supplementary Material
Contributor Information
Konstantinos Sideris, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Charlene R Weir, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Department of Biomedical Informatics, School of Medicine, University of Utah, Salt Lake City, UT 84108, United States.
Carsten Schmalfuss, Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States.
Heather Hanson, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Matt Pipke, PhysIQ, Inc., Chicago, IL 60563, United States.
Po-He Tseng, PhysIQ, Inc., Chicago, IL 60563, United States.
Neil Lewis, Cardiology Section, Medical Service, Hunter Holmes McGuire Veterans Medical Center, Richmond, VA 23249, United States; Department of Internal Medicine, Division of Cardiovascular Disease, Virginia Commonwealth University, Richmond, VA 23249, United States.
Karim Sallam, Cardiology Section, Medical Service, VA Palo Alto Health Care System, Palo Alto, CA 94304, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, Stanford University School of Medicine, Stanford, CA 94305, United States.
Biykem Bozkurt, Cardiology Section, Medical Service, Michael E. DeBakey VA Medical Center, Houston, TX 77030, United States; Section of Cardiology, Department of Medicine, Baylor College of Medicine, Houston, TX 77030, United States.
Thomas Hanff, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Richard Schofield, Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States.
Karen Larimer, PhysIQ, Inc., Chicago, IL 60563, United States.
Christos P Kyriakopoulos, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Iosif Taleb, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Lina Brinker, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Tempa Curry, Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States.
Cheri Knecht, Cardiology Section, Medical Service, Malcom Randall VA Medical Center, Gainesville, FL 32608, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Florida College of Medicine, Gainesville, FL 32610, United States.
Jorie M Butler, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Department of Biomedical Informatics, School of Medicine, University of Utah, Salt Lake City, UT 84108, United States.
Josef Stehlik, Cardiology Section, Medical Service, George E. Wahlen Department of Veterans Affairs Medical Center, Salt Lake City, UT 84148, United States; Division of Cardiovascular Medicine, Department of Internal Medicine, University of Utah, Salt Lake City, UT 84112, United States.
Author contributions
Study concept and design: C.W., K.S., C.S., M.P., N.L., K.S., B.B., T.H., R.S., C.P.K., I.T., L.B., J.M.B., J.S.; data collection and analysis: K.L., P.H.T.; recruitment of participants: H.S., T.C., C.K.; drafting of the manuscript: C.W., K.S., J.M.B., J.S.
Supplementary material
Supplementary material is available at Journal of the American Medical Informatics Association online.
Funding
This work was supported by Veterans Health Administration HSR&D Merit Review (1 I01 HX002922-01A2, Principal Investigator: Stehlik J).
Conflict of interest
J.S. serves as a consultant to Natera, Medtronic, and TransMedics and received research support from Natera and Merck. M.P., K.L., and P.-H.T. are employees of PhysIQ, the provider of the platform. The remaining authors have nothing to disclose.
Data availability
The data underlying this article are available in the article and its online supplementary material.
References
- 1. Heidenreich PA, Albert NM, Allen LA, et al. ; Stroke Council. Forecasting the impact of heart failure in the United States: a policy statement from the American Heart Association. Circ Heart Fail. 2013;6(3):606-619. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2. Dharmarajan K, Hsieh AF, Lin Z, et al. Diagnoses and timing of 30-day readmissions after hospitalization for heart failure, acute myocardial infarction, or pneumonia. JAMA. 2013;309(4):355-363. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 3. Khan MS, Sreenivasan J, Lateef N, et al. Trends in 30- and 90-day readmission rates for heart failure. Circ Heart Fail. 2021;14(4):e008335. [DOI] [PubMed] [Google Scholar]
- 4. Virani SS, Alonso A, Aparicio HJ, et al. ; American Heart Association Council on Epidemiology and Prevention Statistics Committee and Stroke Statistics Subcommittee. Heart disease and stroke statistics-2021 update: a report from the American Heart Association. Circulation. 2021;143(8):e254-e743. [DOI] [PubMed] [Google Scholar]
- 5. Friedman MM, Quinn JR.. Heart failure patients’ time, symptoms, and actions before a hospital admission. J Cardiovasc Nurs. 2008;23(6):506-512. [DOI] [PubMed] [Google Scholar]
- 6. Koehler F, Winkler S, Schieber M, et al. ; Telemedical Interventional Monitoring in Heart Failure Investigators. Impact of remote telemedical management on mortality and hospitalizations in ambulatory patients with chronic heart failure: the telemedical interventional monitoring in heart failure study. Circulation. 2011;123(17):1873-1880. [DOI] [PubMed] [Google Scholar]
- 7. Mortara A, Pinna GD, Johnson P, et al. ; HHH Investigators. Home telemonitoring in heart failure patients: the HHH study (home or hospital in heart failure). Eur J Heart Fail. 2009;11(3):312-318. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8. Stehlik J, Schmalfuss C, Bozkurt B, et al. Continuous wearable monitoring analytics predict heart failure hospitalization: the LINK-HF multicenter study. Circ Heart Fail. 2020;13(3):e006513. [DOI] [PubMed] [Google Scholar]
- 9. Hogg HDJ, Al-Zubaidy M, Talks J, et al. ; Technology Enhanced Macular Services Study Reference Group. Stakeholder perspectives of clinical artificial intelligence implementation: systematic review of qualitative evidence. J Med Internet Res. 2023;25:e39742. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10. Yin J, Ngiam KY, Teo HH.. Role of artificial intelligence applications in real-life clinical practice: systematic review. J Med Internet Res. 2021;23(4):e25759. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11. Zhang J, Whebell S, Gallifant J, et al. An interactive dashboard to track themes, development maturity, and global equity in clinical artificial intelligence research. Lancet Digit Health. 2022;4(4):e212-e213. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12. Shaw J, Rudzicz F, Jamieson T, Goldfarb A.. Artificial intelligence and the implementation challenge. J Med Internet Res. 2019;21(7):e13659. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13. Novak LL, Russell RG, Garvey K, et al. Clinical use of artificial intelligence requires AI-capable organizations. JAMIA Open. 2023;6(2):ooad028. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14. Huo W, Yuan X, Li X, Luo W, Xie J, Shi B.. Increasing acceptance of medical AI: the role of medical staff participation in AI development. Int J Med Inform. 2023;175:105073. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15. Nilsen P, Svedberg P, Nygren J, Frideros M, Johansson J, Schueller S.. Accelerating the impact of artificial intelligence in mental healthcare through implementation science. Implement Res Pract. 2022;3:26334895221112033. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16. King H, Wright J, Treanor D, Williams B, Randell R.. What works where and how for uptake and impact of artificial intelligence in pathology: review of theories for a realist evaluation. J Med Internet Res. 2023;25:e38039. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17. Curran GM, Bauer M, Mittman B, Pyne JM, Stetler C.. Effectiveness-implementation hybrid designs: combining elements of clinical effectiveness and implementation research to enhance public health impact. Med Care. 2012;50(3):217-226. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18. Landes SJ, McBain SA, Curran GM.. An introduction to effectiveness-implementation hybrid designs. Psychiatry Res. 2019;280:112513. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 19. Powell BJ, Waltz TJ, Chinman MJ, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):21. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20. Waltz TJ, Powell BJ, Fernández ME, Abadie B, Damschroder LJ.. Choosing implementation strategies to address contextual barriers: diversity in recommendations and future directions. Implement Sci. 2019;14(1):42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21. Baker R, Camosso‐Stefinovic J, Gillies C, et al. ; Cochrane Effective Practice and Organisation of Care Group. Tailored interventions to address determinants of practice. Cochrane Database of Syst Rev. 2015;2015(4):CD005470. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22. Wensing M, Bosch M, Grol R. Selecting, tailoring, and implementing knowledge translation interventions. In: Straus SE, Tetroe J, Graham ID, eds. Knowledge Translation in Health Care: Moving from Evidence to Practice. Wiley-Blackwell; 2009:94-113. [Google Scholar]
- 23. D’Lima D, Soukup T, Hull L.. Evaluating the application of the RE-AIM planning and evaluation framework: an updated systematic review and exploration of pragmatic application. Front Public Health. 2022;9:755738. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24. Kelkar AA, Spertus J, Pang P, et al. Utility of patient-reported outcome instruments in heart failure. JACC Heart Fail. 2016;4(3):165-175. [DOI] [PubMed] [Google Scholar]
- 25. Ravera A, Santema BT, Sama IE, et al. Quality of life in men and women with heart failure: association with outcome, and comparison between the Kansas City Cardiomyopathy Questionnaire and the EuroQol 5 Dimensions Questionnaire. Eur J Heart Fail. 2021;23(4):567-577. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26. Mott J, Pipke M. Similarity-based modeling of aircraft flight paths. In: 2004 IEEE Aerospace Conference Proceedings(IEEE Cat. No.04TH8720), Big Sky, MT, USA; Vol. 3. 2004:1580-1593. doi: 10.1109/AERO.2004.1367933 [DOI]
- 27. Harvey G, Kitson A.. PARIHS revisited: from heuristic to integrated framework for the successful implementation of knowledge into practice. Implement Sci. 2016;11:33. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28. Harvey G, Kiston A.. Implementing Evidence-Based Practice in Healthcare: A Facilitation Guide. Routledge; 2015. [Google Scholar]
- 29. Hunter SC, Kim B, Kitson AL.. Mobilising implementation of i-PARIHS (Mi-PARIHS): development of a facilitation planning tool to accompany the integrated promoting action on research implementation in health services framework. Implement Sci Commun. 2023;4(1):2. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30. Gale NK, Heath G, Cameron E, Rashid S, Redwood S.. Using the framework method for the analysis of qualitative data in multi-disciplinary health research. BMC Med Res Methodol. 2013;13:117. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31. O’Brien BC, Harris IB, Beckman TJ, Reed DA, Cook DA.. Standards for Reporting Qualitative Research: a synthesis of recommendations. Acad Med. 2014;89(9):1245-1251. [DOI] [PubMed] [Google Scholar]
- 32. Morse JM, Barrett M, Mayan M, Olson K, Spiers J.. Verification strategies for establishing reliability and validity in qualitative research. Int J Qual Methods. 2002;1(2):13-22. [Google Scholar]
- 33. Glasgow RE, Estabrooks PE.. Pragmatic applications of RE-AIM for health care initiatives in community and clinical settings. Prev Chronic Dis. 2018;15:E02. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 34. Glasgow RE, Vogt TM, Boles SM.. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89(9):1322-1327. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35. McCreight MS, Rabin BA, Glasgow RE, et al. Using the practical, robust implementation and sustainability model (PRISM) to qualitatively assess multilevel contextual factors to help plan, implement, evaluate, and disseminate health services programs. Transl Behav Med. 2019;9(6):1002-1011. [DOI] [PubMed] [Google Scholar]
- 36. Holtrop JS, Estabrooks PA, Gaglio B, et al. Understanding and applying the RE-AIM framework: clarifications and resources. J Clin Transl Sci. 2021;5(1):e126. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37. Liu S, Reese TJ, Kawamoto K, Del Fiol G, Weir C.. A theory-based meta-regression of factors influencing clinical decision support adoption and implementation. J Am Med Inform Assoc. 2021;28(11):2514-2522. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38. Khanijahani A, Iezadi S, Dudley S, Goettler M, Kroetsch P, Wise J.. Organizational, professional, and patient characteristics associated with artificial intelligence adoption in healthcare: A systematic review. Health Policy Technol. 2022;11(1):100602. [Google Scholar]
- 39. Lambert SI, Madi M, Sopka S, et al. An integrative review on the acceptance of artificial intelligence among healthcare professionals in hospitals. NPJ Digit Med. 2023;6(1):111. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40. Shaw J, Shaw S, Wherton J, Hughes G, Greenhalgh T.. Studying scale-up and spread as social practice: theoretical introduction and empirical case study. J Med Internet Res. 2017;19(7):e244. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 41. Scott IA, Carter SM, Coiera E.. Exploring stakeholder attitudes towards AI in clinical practice. BMJ Health Care Inform. 2021;28(1):e100450. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42. Lyell D, Coiera E.. Automation bias and verification complexity: a systematic review. J Am Med Inform Assoc. 2017;24(2):423-431. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43. Greenhalgh T, Wieringa S.. Is it time to drop the ‘knowledge translation’ metaphor? A critical literature review. J R Soc Med. 2011;104(12):501-509. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44. Arbelaez Ossa L, Starke G, Lorenzini G, Vogt JE, Shaw DM, Elger BS.. Re-focusing explainability in medicine. Digit Health. 2022;8:20552076221074488. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 45. Markus AF, Kors JA, Rijnbeek PR.. The role of explainability in creating trustworthy artificial intelligence for health care: A comprehensive survey of the terminology, design choices, and evaluation strategies. J Biomed Inform. 2021;113:103655. [DOI] [PubMed] [Google Scholar]
- 46. Merry M, Riddle P, Warren J.. A mental models approach for defining explainable artificial intelligence. BMC Med Inform Decis Mak. 2021;21(1):344. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47. Nyrup R, Robinson D.. Explanatory pragmatism: a context-sensitive framework for explainable medical AI. Ethics Inf Technol. 2022;24(1):13. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48. Shortliffe EH, Sepúlveda MJ.. Clinical decision support in the era of artificial intelligence. JAMA. 2018;320(21):2199-2200. [DOI] [PubMed] [Google Scholar]
- 49. Combi C, Amico B, Bellazzi R, et al. A manifesto on explainability for artificial intelligence in medicine. Artif Intell Med. 2022;133:102423. [DOI] [PubMed] [Google Scholar]
- 50. Holzinger A, Carrington A, Müller H.. Measuring the quality of explanations: the system causability scale (SCS). Kunstliche Intell (Oldenbourg). 2020;34(2):193-198. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data underlying this article are available in the article and its online supplementary material.