Abstract
Objective:
Despite recent calls for home healthcare (HHC) to integrate informatics, the application of machine learning in HHC is relatively unknown. Thus, this study aimed to synthesize and appraise the literature describing the application of machine learning to predict adverse outcomes (e.g., hospitalizations or mortality) using electronic health record (EHR) data in the HHC setting. Our secondary aim was to evaluate the comprehensiveness of the predictors used in the machine learning algorithms guided by the Biopsychosocial Model.
Methods:
We conducted a literature search in four databases: PubMed, Embase, CINAHL, and Scopus on March 2022. Inclusion criteria were 1) describing services provided in the HHC setting, 2) applying machine learning algorithms to predict adverse outcomes, defined as outcomes related to patient deterioration, 3) using EHR data and 4) focusing on adult population. Predictors were mapped to the Biopsychosocial Model. A risk of bias analysis was conducted using the Prediction Model Risk Of Bias Assessment Tool.
Results:
The final sample included 20 studies. Eighteen studies used predictors from standardized assessments integrated into the EHR. The most common outcome of interest was hospitalization (55%), followed by mortality (25%). Psychological predictors were frequently excluded (35%). Tree based algorithms were most frequently applied (75%). Most studies demonstrated high or unclear risk of bias (75%).
Conclusion:
Future studies in HHC should consider incorporating machine learning algorithms into clinical decision support systems to identify patients at risk. Based on the Biopsychosocial model, psychological and interpersonal characteristics should be used along with biological characteristics to enhance risk prediction. To facilitate the widespread adoption of machine learning, stakeholders should encourage standardization in the HHC setting.
Keywords: Machine Learning, Home Health Care, Adverse Events, Prediction, Nursing Informatics
Introduction
In the United States, home healthcare (HHC) is provided to approximately 5.2 million Medicare beneficiaries through 7.4 million home visits per year.1 HHC patients include individuals recently discharged from an acute hospital stay or individuals who are managing chronic conditions in the community. The services provided during HHC include skilled nursing care, physical therapy, occupational therapy, speech therapy, and social work.2 As part of routine HHC visits, clinicians assess for early signs of deterioration to reduce the risk of adverse events.3 Despite comprehensive assessments and interventions, nearly one in five patients are hospitalized or visit the emergency department (ED) during the time they receive HHC services.4–6 Other initiatives aimed at hospitalization reduction include financial incentives;4 however, readmission rates have not improved.7 Estimates show that approximately 40% of hospitalizations and ED visits are due to preventable ambulatory care-sensitive conditions8 that could be avoided with timely and tailored interventions.9
In recent years, there has been growing evidence that machine learning algorithms can predict the risk of deterioration in patients by analyzing electronic health record (EHR) documentation.10 HHC agencies have their own EHR which provides a central place to store information describing patients’ diagnosis, problems, interventions, and outcomes. In HHC, nurses are the largest group of clinicians, hence nursing documentation accounts for a majority of EHR documentation. Other users of the EHR in HHC include physical, occupational, and speech therapists, and social workers.11 Machine learning algorithms can identify patients who may experience an adverse event by discovering patterns in previously documented data.12 The value of machine learning has been demonstrated in the acute care setting to detect various adverse events13,14 such as mortality15 and the activation of rapid response teams.16 This has led to the development of clinical decision support systems that proactively notify clinicians to patients at risk for experiencing an adverse outcome.17,18 However, the extent to which machine learning has been applied in the HHC setting for risk prediction has not been previously reported.
Another consideration when assessing the value of machine learning is understanding the predictors driving the algorithm. Previous studies have demonstrated the versatility of using machine learning to examine a wide range of predictors from symptoms to social determinants of health.19,20 However, while it is known that social determinants21 and psychological factors influence health outcomes,22 the extent to which these factors are included in machine learning algorithms in the HHC setting is unknown.19,20
Recent literature has called for an increase in informatics, such as machine learning, in the HHC setting to bring documentation-driven evidence to clinicians at the point of care.23–25 Despite HHC being one of the fastest growing sectors of healthcare26 and the benefits of machine learning to predict patients at risk,27 no previous studies have summarized the state of the science on machine learning applied to EHR data in the HHC setting. To address these knowledge gaps, this study aims to critically appraise and synthesize information on the application of machine learning to predict a priori specified adverse outcomes using EHR data in the HHC setting. Our secondary aim was to summarize the different dimensions of predictors (i.e., biological, psychological, and interpersonal) used to build the machine learning algorithms guided by the Biopsychosocial Model.
Materials and methods
This scoping review was conducted and reported according to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews (PRISMA-ScR) guidelines.28 The protocol describing this study was registered in Open Science Framework (osf.io/g5zrj). Studies were eligible to be included if they addressed skilled health services provided in HHC, applied machine learning algorithms to predict adverse outcomes a priori, used EHR data, focused on the adult population, and were published in English. Full text papers and conference proceedings covering quantitative machine learning algorithms were included. For this study, adverse events were defined as outcomes related to patient deterioration.29 Studies were excluded if they took place in a nursing home or skilled nursing facility or focused on smart homes, sensors, or robots because they were outside the scope of this manuscript. Qualitative studies, dissertation papers, and research proposals were excluded.
For a comprehensive synthesis of machine learning methodological approaches, we refined a previous literature query that examined the application of artificial intelligence in healthcare.30 Four databases were searched (PubMed, Embase, CINAHL, and Scopus) in March 2022. There was no restriction on the publication date. The full search strategy can be found in Appendix A. Our screening team included two registered nurses and two advanced practice nurses with informatics expertise (MH, JS, DS, MT). After duplicates were removed, all relevant studies were screened by two reviewers. Any conflicts were resolved through discussion between the three reviewers until consensus was reached. This process was repeated with full-text studies.
The data extraction template was drafted by two co-authors (MH, MT). Other co-authors were invited to modify the template prior to charting the items. The extraction template included information about the study and sample characteristics, algorithm overview, machine learning approaches, and algorithm performance. Upon consensus with the appropriate variables to extract (e.g., purpose, data source, best-performing algorithm), two reviewers (JS, DS) independently charted the data using the software, Covidence. A third reviewer went through and resolved any disagreements (MH).
Data were synthesized by mapping predictors to the Biopsychosocial Model.31–33 This model assumes that to understand health holistically, a range of characteristics should be considered, including an individual’s social, psychological, and behavioral dimensions.31,32 Lehman et al adapted this model to include the following dimensions: biological, psychological, and interpersonal.33 The biological dimension encompasses medical histories influencing the physical body; the psychological dimension encompasses coping and health behaviors; and the interpersonal dimension encompasses social support.33 Two reviewers were tasked with categorizing the predictors into one of the three dimensions of the Biopsychosocial Model. Discrepancies were resolved through consensus meetings with the screening team (MH, JS, DS).
Four reviewers (MH, JS, DS, MT) performed a risk of bias analysis using the Prediction Model Risk of Bias Assessment Tool (PROBAST).34 One reviewer (MH) resolved disagreements after discussing the conflicts with another member of the screening team who has expertise in machine learning (MT). PROBAST facilitates risk of bias assessment in four domains: participants, predictors, outcome, and analysis, and is specific to prediction model studies. Per the PROBAST guidelines, if a study has at least one criterion indicating an unclear or high risk of bias, the overall risk of bias for that study would be unclear or high.
Results
Our initial search yielded 1,644 studies from four databases (PubMed, Embase, CINAHL, and Scopus). After duplicates were removed (n = 190), there were 1,454 studies eligible for title and abstract screening. A total of 1,378 studies were excluded, and 76 studies were moved to full-text review. Of those, 20 studies were included for synthesis. Among the excluded studies (n = 56), studies were most commonly excluded because they did not predict an a priori specified adverse event (n = 32) or did not use EHR data (n = 17). The PRISMA flow diagram35 presented in Figure 1 summarizes the study selection process.
Figure 1.

PRISMA Diagram
Study characteristics
Results revealed the global advances in integrating machine learning into the HHC setting. Six countries were represented in this review (30%). Ten studies (50%) did not include a clear definition of HHC in their region. Most studies took place in the United States (n = 9; 45%)36–44 and Canada (n = 4; 20%)45–48 and used data retrospectively collected from 2015 and earlier (n = 13; 65%)36–39,42–44,46,47,49–52 although some studies did not specify dates (n = 3; 15%).40,48,53 The majority of studies were published between 2019 and 2022 (n = 13; 65%).37,40–44,46,47,49,51,52,54,55
Sample characteristics
Most studies did not focus on patients with a specific disease (n = 12; 60%). Among studies that did focus on a disease-specific population, patients with heart failure (n = 3; 15%)36,39,53 and advanced cancer (n = 2; 10%)50,55 were the most common. The number of participants included in the studies ranged from 284 to 317,621. A summary of study and sample characteristics is provided in Table 1.
Table 1.
Study & Sample Characteristics
| Author | Purpose | Disease Focus | Participants | Country | Retrospective Data Collection Period | HHC Definition |
|---|---|---|---|---|---|---|
| Calvo et al., 201951 | To evaluate sampling approaches using blood tests to detect successful versus unsuccessful (i.e., hospitalization) HHC stays. | General HHC population | 1,951 | Spain | January 2012 – December 2015 | Home-based medical and nursing services are provided when patients no longer need hospital facilities, but still require active and complex surveillance. |
| Calvo et al., 202052 | To assess the performance of models for predicting mortality and in-hospital admission | General HHC population | 1,936 | Spain | January 2009 – December 2015 | A service delivered by trained hospital personnel providing acute, home-based, complex interventions for a period that is not longer than the expected length of hospital stay. |
| Jones et al., 201845 | To compare the performance of machine learning models predicting ED and hospital utilization among HHC patients. | General HHC population | 88,345 | Canada | 2014 – 2016 | Not specified |
| Jones et al., 201942 | To develop a model to predict hospital readmissions from HHC in Medicare beneficiaries. | General HHC population | 43,407 | United States | January 2012 – January 2013 | Home-based skilled services, including nursing and therapies (e.g., physical, occupational, and speech) for patients after hospitalization. |
| Kang et al., 201636 | To examine risk factors for hospitalization among patients with heart failure who received telehomecare during their HHC episode. | Heart failure | 552 | United States | 2010 | Monitoring patients remotely using telehomecare. |
| Kuspinar et al., 201946 | To develop an algorithm to predict falls among HHC patients. | General HHC population | 126,703 | Canada | 2002 – 2014 | Not specified |
| Lo et al., 201940 | To identify factors that predict fall risks. | General HHC population | 59,006 | United States | Not specified | Not specified |
| Madigan and Curet, 200639 | To determine the drivers of HHC service outcomes (i.e., discharge destination, length of stay). | COPD, Heart failure, Hip fractures | 580 | United States | 2000 | Not specified |
| McArthur et al., 202047 | To develop a risk model to predict one-year incidence of hip fractures. | General HHC population | 317,621 | Canada | April 2011 – March 2015 | Not specified |
| Nabal et al., 201450 | To develop a model to predict survival based on symptoms identified by palliative HHC teams. | Advanced cancer | 698 | Spain | April 2009 – July 2009 | Not specified |
| Nuutinen et al., 202049 | To develop a model that predicts nursing home admission risk. | General HHC population | 2,523 | Finland | January 2012 – September 2015 | Home-based services that are provided to patients for longer than 12 months. |
| Olson et al., 201638 | To characterize patients at risk for medication-related hospital readmissions. | General HHC population | 911 | United States | 2004 | Not specified |
| Song et al., 202144 | To identify risk factors for wound infection-related hospitalization or ED visits among patients who received HHC. | Wounds | 54,316 | United States | January 2014 – December 2014 | HHC episodes typically last for 30 to 60 days or until the patient is discharged from HHC or admitted to the hospital or ED. |
| Song et al., 202241 | To compare the performance of risk models that predict hospitalization and ED visits in HHC. | General HHC population | 66,317 | United States | January 2015 – December 2017 | HHC episodes include services provided to a patient received between admission and discharge (30–60 days). |
| Sullivan et al., 202037 | The determine how well the OASIS-SQ predicts ED mortality risk over time compared to alternative OASIS items. | General HHC population | 69,097 | United States | January 2012 – 2013 | Nurses and therapists assess older adults residing in the community using the Outcomes and Assessment Information Set (OASIS). |
| Topaz et al., 202043 | To identify patients at high risk for hospitalization or ED visits. | General HHC population | 89,459 | United States | January 2014 – December 2014 | HHC encounters (e.g., nursing visits) typically happen a few days apart, and most of the data is generated by nurses and physical or occupational therapists. |
| Witt et al., 202254 | To evaluate the performance of the model used to predict hospitalization in HHC patients. | General HHC population | 1,282 | Denmark | January 2016 – December 2017 | Not specified |
| Yang et al., 202155 | To predict survival and medical expenses among patients with advanced cancer receiving HHC. | Advanced cancer | 310 | China | January 2016 – December 2018 | The Homecare Service Program for advanced cancer patients provides free medical services in patients’ homes. |
| Zhang et al., 201353 | To identify patients recently discharged from the hospital at risk for death or hospitalization related to heart failure. | Heart failure | 284 | Germany, UK, Netherlan ds | Not specified | Not specified |
| Zhu et al., 200748 | To explore if machine learning can predict rehabilitation potential. | General HHC population | 24,724 | Canada | 1 year (no dates cited) | Not specified |
HHC: Home healthcare; ED: Emergency department; OASIS: Outcome and Assessment Information Set
Algorithm overview
While all studies used data from the EHR, some studies included standardized assessments stored in the EHR (n = 18; 90%).36–50,52,54,55 Standardized assessments, housed within the EHR, include a set of standardized data elements allowing for easier comparison across different settings.56 For example, the United States’ federally mandated Outcome and Assessment Information Set (OASIS) assesses the patient’s functional status by collecting data about activities of daily living using a standardized set of questions such as describing the patient’s ability to “transfer in and out of bed.”57 In the United States, EHR data (e.g., demographics, clinical assessments) are typically collected by nurses within the HHC agency,11 siloed from healthcare data collected in other care settings.23 The most cited standardized assessment was the OASIS (n = 8; 40%)36–38,40–44 and Residential Assessment Instrument - Home Care (RAI-HC) (n = 5; 25%).45–49 OASIS is a patient-specific assessment tool required by Medicare in the United States to guide a patient’s plan of care, determine reimbursement, and collect quality measures in HHC.57 Similarly, the RAI-HC is used in Canada to assess “long stay home care clients’ health status, need for care, and basic background on housing and informal caregivers,” helping to guide care and collect quality indicators.58 In the included studies, only three studies included clinical notes as a data source (15%).41,43,44
Most studies considered variables from all three domains of the Biopsychosocial Model (n = 12; 60%): biological (e.g., age, hematocrit, fall history), psychological (e.g., depression, memory deficit, coping), and interpersonal (e.g., lives alone, social support, informal caregiver status).38,40–42,44–50,52 Predictors that did not map to one of the three domains were categorized as “Other” which included variables such as the number of days in the hospital, hospital utilization history, and the number of HHC visits. A list of variables from each study mapped to the Biopsychosocial Model is provided in Appendix B. Biological characteristics were included in all the studies. Psychological characteristics were excluded in seven studies.36,37,39,43,51,54,55 Interpersonal characteristics were excluded in three studies.37,51,53
Most studies were aimed at predicting hospitalization (n = 11; 55%).36,38,41–45,51–54 The next most common outcome of interest was mortality (n = 5; 25%).37,50,52,53,55 Though less frequent, other studies focused on predicting additional adverse events including falls,40,46 nursing home admission,39,49 hip fractures,39,47 and lack of rehabilitation potential.48 Table 2 includes details about each study’s data source, predictors and outcomes included in the algorithms.
Table 2.
Algorithm Overview
| Study | Data Source | Predictors’ Categories (Biopsychosocial Model) | Outcome of Interest |
|---|---|---|---|
| Calvo et al., 201951 | EHR | Biological | Hospitalization |
| Calvo et al., 202052 | EHR, Standardized assessments (GMA, SF-36, Barthel Index) | Biological, Psychological, Interpersonal, Other | Mortality and Hospitalization |
| Jones et al., 201845 | EHR, Standardized assessments (RAI-HC) | Biological, Psychological, Interpersonal | Unplanned ED visit, Hospitalization |
| Jones et al., 201942 | EHR, Standardized assessment (OASIS, Patient health questionnaire, Elixhauser comorbidity index), Medicare claims data, MEDPAR | Biological, Psychological, Interpersonal, Other | Hospitalization |
| Kang et al., 2016†36 | Standardized assessments (OASIS) | Biological, Psychological, Interpersonal | Hospitalization |
| Kuspinar et al., 201946 | Standardized assessments (RAI-HC) | Biological, Psychological, Interpersonal | Fall |
| Lo et al., 201940 | EHR, Standardized assessments (OASIS) | Biological, Psychological, Interpersonal | Fall |
| Madigan and Curet, 200639 | Standardized assessments (2000 National home and hospice care survey) | Biological, Interpersonal | Discharge destination, Length of stay |
| McArthur et al., 202047 | Standardized assessments (RAI-HC) | Biological, Psychological, Interpersonal | Hip fracture |
| Nabal et al., 201450 | EHR, Standardized assessments (Charlson scale, Cullum scale, Karnofsky index, Barthel index) | Biological, Psychological, Interpersonal | Mortality |
| Nuutinen et al., 202049 | EHR, Standardized assessments (RAI-HC) | Biological, Psychological, Interpersonal, Other | Nursing home admission |
| Olson et al., 201638 | EHR, Standardized assessments (OASIS) | Biological, Psychological, Interpersonal, Other | Hospitalization |
| Song et al., 202144 | EHR, Standardized assessments (OASIS), Clinical notes | Biological, Psychological, Interpersonal, Other | Wound related hospitalization |
| Song et al., 202241 | EHR, Standardized assessments (OASIS), Clinical notes | Biological, Psychological, Interpersonal, Other | Hospitalization or ED visit |
| Sullivan et al., 2020†37 | Standardized assessment (OASIS) | Biological | Mortality |
| Topaz et al., 2020†43 | Standardized assessment (OASIS), Clinical notes | Biological, Interpersonal | Hospitalization or ED visit |
| Witt et al., 202254 | EHR, Danish national patient register, Standardized assessment (Triaged changing table) | Biological, Interpersonal | Hospitalization |
| Yang et al., 202155 | EHR, Standardized assessment (Numeric pain rating scale, Karnofsky performance scale, Quality of life score) | Biological, Interpersonal | Mortality, Medical expenses |
| Zhang et al., 201353 | EHR | Biological | Mortality, Hospitalization |
| Zhu et al., 200748 | Standardized assessments (RAI-HC) | Biological, Psychological, Interpersonal | Lack of rehabilitation potential |
EHR: Electronic Health Record; NA: Not Applicable; GMA: Adjusted Morbidity Groups; SF-36: Short Form Health Survey; RAI-HC: Residential Assessment Instrument-Home Care; ED: Emergency department; OASIS: Outcome and Assessment Information Set; MEDPAR: Medicare Provider Analysis Review; COPD: Chronic obstructive pulmonary disease
Although standardized assessments (e.g., OASIS) have biological, psychological, and interpersonal dimensions, some studies only used a subset from one or more domains as inputs for the predictive models.
In this table, the term “EHR” covers variables not specific to standardized assessments. Standardized assessments (e.g., OASIS, RAI-HC) are listed separately.
Machine learning approaches
Of the machine learning approaches applied, most were categorized as supervised machine learning (n = 17; 85%).36,39–48,50–55 Supervised machine learning uses a deductive approach with a human-labeled dataset as a gold standard to train classification algorithms; unsupervised machine learning uses an inductive approach that does not include a labeled dataset and instead focuses on topic discovery.59 Among the machine learning algorithms, tree based algorithms (e.g., Decision tree, ADABoost, Classification and regression trees, Random forest) were most commonly applied (n = 15; 75%).36,39–44,46,47,50–53,55 Decision tree,36,39,43,46,47,52,53 Logistic regression,41,42,44,45,52–54 and Random forest40,41,43–45,52,55 were each applied in seven studies (35%). Neural networks were applied in three studies (15%).44,45,55 Alternatively, unsupervised approaches such as Clustering (n = 3; 15%) were less frequently applied in the included studies.37,38,49 Three studies discussed using natural language processing (NLP) which supports the analysis of narrative text (e.g., clinical notes) (n = 3; 15%).41,43,44
Algorithm performance
Area under the curve (AUC) was the most reported performance metric (n = 9; 45%).36,38,40,44–47,49,50,52–54 For the performance metrics, F-score, AUC, and c-statistics values ranged between 0–1, with higher values denoting better-performing algorithms.60,61 The best performing algorithm varied across the different studies, as shown in Table 3 with the corresponding performance metrics.
Table 3.
Machine Learning Approaches and Algorithm Performance
| Study | Machine Learning Algorithms | Best Algorithm Performance |
|---|---|---|
| Calvo et al., 201951 | ADABoost | Precision (0.10) Specificity (0.12) Sensitivity (1) F-score (0.11) |
| Calvo et al., 202052 | Decision trees, Logistic regression, Random forest |
Random forest AUC (0.80) Sensitivity (0.75) Specificity (0.71) |
| Jones et al., 201845 | Logistic regression, Neural networks, Random forest, Gradient boosted trees |
Gradient boosted trees Logarithmic score (−0.30) Brier score (0.17) AUC (0.68) |
| Jones et al., 201942 | Logistic regression, Gradient boosted trees |
Gradient boosted trees c-statistic (0.67) Brier score (0.12) |
| Kang et al., 201636 | Decision trees | c-statistic (0.59) Specificity (0.65) Sensitivity (0.49) |
| Kuspinar et al., 201946 | Decision trees | c-statistic (0.60) |
| Lo et al., 201940 | Random forest | AUC (0.67) Precision (0.10) Balanced Accuracy (0.62) |
| Madigan and Curet, 200639 | Decision trees, Classification and regression trees |
Classification and regression trees Classification rate (0.71) |
| McArthur et al., 202047 | Decision trees |
Decision trees c-statistic (0.66) |
| Nabal et al., 201450 | Classification and regression trees |
Decision trees AUC (0.88) |
| Nuutinen et al., 202049 | Principle component analysis, K means clustering | ACC (0.80) AUC (0.80) TPR (0.44) FPR (0.12) FNR (0.56) TNR (0.88) |
| Olson et al., 201638 | K means clustering, Hierarchical clustering | AUC (0.77) |
| Song et al., 202144 | Logistic regression, Neural networks, NLP, Random forest |
Logistic regression Sensitivity (0.88) Specificity (0.64) PPV (0.03) NPV (0.99) AUC (0.82) |
| Song et al., 202241 | Logistic regression, SVM, Random forest, Naïve bayes, NLP, Bayesian network |
Random forest Sensitivity (0.93) PPV (0.72) F-score (0.81) PRC (0.86) |
| Sullivan et al., 202037 | K means clustering, Survival analysis | PPV (0.49) NPV (0.72) K (0.20; 95%CI: [0.19, 0.21]) |
| Topaz et al., 202043 | Decision trees, Random forest, Naïve bayes |
Random forest Recall (0.81) Precision (0.83) F-score (0.82) PRC (0.76) |
| Witt et al., 202254 | Logistic regression, RUSBoost |
RUSBoost AUC (0.99) PRC (0.71) |
| Yang et al., 202155 | Neural networks, SVM, Random forest |
Random forest Accuracy (0.82) Normalized mean square error (0.42) |
| Zhang et al., 201353 | Chi-square automatic interaction detector decision tree, Logistic regression |
Chi-square automatic interaction detector decision tree AUC (0.80) |
| Zhu et al., 200748 | K nearest neighbor | FPR (0.34) FNR (0.36) DLR+ (1.88) DLR− (0.55) |
AUC: Aera Under the Curve; TPN: True-positive-rate; FPN: False positive-rate; FNR: False-negative-rate; TPN: True-negative-rate; PPV: Positive Predictive Value; NPV: Negative Predictive Value; NLP: Natural language processing; SVM: Support Vector Machines; RUSBoost: Random UnderSampling and Boosting; PRC: Precision-Recall Curve; FPR: False positive rate; FNR: False negative rate; DLR+: Positive diagnostic likelihood ratio: DLR−: Negative diagnostic likelihood ratio
Risk of bias analysis
Similar to recent literature, most studies demonstrated high or unclear risk of bias.62,63 Of the studies screened with PROBAST, five (25%) demonstrated an overall low risk of bias,36,42,43,45,52 seven demonstrated an overall unclear risk of bias (35%),38,40,46,47,49,50,53 and eight demonstrated an overall high risk of bias (40%).37,39,41,44,48,51,54,55 Studies most frequently (n = 8; 40%) had an overall unclear risk of bias38,40,46,47,49,50,53 because there was a lack of detail about overfitting,37,46,53 handling of missing data,40,46,47,49,50,53 use of univariable analysis,47,49,50 and handling of complexities in the data.38,40,46,47,49,50 Most commonly, studies were designated high risk of bias because they selected predictors using univariable analysis,41,44,51 did not account for model overfitting,39 assigned weights in the final model based on the multivariable analysis,39 did not handle missing data appropriately,54 and did not evaluate model performance measures appropriately.48,55 Full results of the PROBAST risk of bias analysis are provided in Table 4.
Table 4.
Risk of bias analysis
| Study | Participants | Predictors | Outcome | Analysis | Overall |
|---|---|---|---|---|---|
| Calvo et al., 201951 | Unclear | Low | Low | High | High |
| Calvo et al., 202052 | Low | Low | Low | Low | Low |
| Jones et al., 201845 | Low | Low | Low | Low | Low |
| Jones et al., 201942 | Low | Low | Low | Low | Low |
| Kang et al., 201636 | Low | Low | Low | Low | Low |
| Kuspinar et al., 201946 | Low | Low | Low | Unclear | Unclear |
| Lo et al., 201940 | Low | Low | Low | Unclear | Unclear |
| Madigan and Curet, 200639 | Low | Low | Unclear | High | High |
| McArthur et al., 202047 | Low | Unclear | Unclear | Unclear | Unclear |
| Nabal et al., 201450 | Low | Low | Low | Unclear | Unclear |
| Nuutinen et al., 202049 | Unclear | Unclear | Unclear | Unclear | Unclear |
| Olson et al., 201638 | Low | Unclear | Unclear | Unclear | Unclear |
| Song et al., 202144 | Low | Low | Unclear | High | High |
| Song et al., 202241 | Low | Low | Unclear | High | High |
| Sullivan et al., 202037 | Low | Low | High | Unclear | High |
| Topaz et al., 202043 | Low | Low | Low | Low | Low |
| Witt et al., 202254 | Low | Low | Low | High | High |
| Yang et al., 202155 | Low | Low | Low | High | High |
| Zhang et al., 201353 | Low | Low | Low | Unclear | Unclear |
| Zhu et al., 200748 | Unclear | Low | Low | High | High |
In this table, low indicates low bias indicating better quality
Discussion
This scoping review included 20 studies that described the application of machine learning to predict a priori specified adverse outcomes using EHR data in the HHC setting. Although similar reviews were conducted in the acute and ambulatory settings,64,65 this study was the first to focus on the HHC setting, which represents one of the fastest-growing healthcare sectors.26
Most studies (65%) included in this review were published between 2019 and 2022, suggesting a recent uptake of machine learning in the HHC setting.64,66 With that, 65% of the studies used data collected from 2015 or earlier, indicating a delay between retrospective data availability and analysis. This may be related to the lack of regulation and standardization of EHR systems and data collected in the HHC setting globally, which may increase the difficulty and time needed to extract, clean, and analyze the data.67,68 Generating global policies similar to the Improving Medicare Post-Acute Care Transformation (IMPACT) Act, which advances standardized data in post-acute care settings69 could lead to improved interoperability and analytical efficiency in the near future.
This review paper highlights a gap in current research around the implementation of machine learning algorithms into the HHC clinical setting.30 A previous systematic review describing prediction algorithms in post-acute care found that among the 37 models reviewed, very few were ever implemented in practice.70 While clinical validation is underway in the acute care setting through randomized controlled trials,71–73 there is still a significant gap in the HHC setting. Shifting focus from an end goal of performance optimization to practical clinical implementation is imperative to achieve the intention of almost all health-related machine learning algorithms – to advance health and improve outcomes. Seneviratne et al recommends that when building machine learning algorithms, developers should proactively consider clinical actionability, patient safety, and cost-utility to optimize implementation in the clinical setting.74
Although this study examined the global scope of machine learning in HHC across six countries, it is important to highlight some regional differences. For example, in the United States, HHC is defined as a period of 30 to 60 days where skilled services such as nursing, physical, occupational, and speech therapy are provided to help patients transition back to the community.42,44 Alternatively, in Spain, the purpose of HHC is to provide interventions aimed at substituting inpatient hospitalization in a period of time that does not exceed the expected length of an inpatient hospitalization.52 Concerningly, half of the studies included in this review did not provide detailed information on the structure of HHC in their country. The differences in HHC structure and purpose also influence EHR documentation75 which was not thoroughly discussed in all articles. Given the variation of HHC and EHRs globally,76–78 future studies should provide a detailed description of their regional HHC structure and EHR systems to help distinguish differences in patient populations, goals, timelines, and care structures. Providing further clarity around regional differences in the structure of HHC and EHR systems could lead to better tailoring and predictor selection when developing machine learning algorithms globally.
Standardized assessments, such as OASIS and the RAI-HC, are commonly stored in the EHR. Using standardized assessments can support interoperability, or the ability to exchange health information, across systems or within a health system.79 In addition, the utilization of standardized assessments has led to more evidence-based, consistent, and holistic assessments among clinicians.80,81 In fact, all three domains of the Biopsychosocial Model were aligned with studies that used a standardized assessment tool as their data source, suggesting that standardized assessments lead to more comprehensive documentation. Thus, incorporating standardized assessments into the EHR could improve comparability, access to data, interoperability, and generalizability in the HHC setting.
Our results reveal that most studies include predictors from all three domains of the Biopsychosocial Model. Unsurprisingly, and similar to a previous study,82 biological characteristics were included in all studies. Fewer studies included psychological (e.g., mental health) and/or interpersonal (e.g., social determinants of health) predictors in their machine learning algorithms. Of the studies that did include these dimensions, they were often from the structured standardized assessments. For psychological predictors, this may be related to using an assessment tool that does not comprehensively consider psychological characteristics. For interpersonal predictors, this might be related to limited data capture and documentation of social determinants of health in EHR systems.83 Previous studies have suggested that including social determinants of health can improve machine learning prediction among individuals from racial or ethnic minorities.21,84,85 The inclusion of psychological and interpersonal characteristics in machine learning algorithms could lead to improved equity in risk identification85 and proactive interventions for patients with mental health conditions or social risk factors who are often stigmatized in healthcare.86,87
The most frequently cited outcome of interest was hospitalization. In HHC, one of the primary goals is to prevent avoidable hospitalizations.88 Hospitalization rates reflect care quality,4 with reducing readmission being financially incentivized;4 thus, we anticipate the number of machine learning algorithms focused on hospitalization to continue to be common. Other outcomes specific to HHC, such as self-management, should be explored in future studies to investigate other ways to reduce the risk of specific adverse outcomes prior to hospitalization.89 In terms of disease conditions, a majority of the studies in this review did not focus on patients with a specific disease but rather on the general HHC population. The few studies that specified a condition focused on patients with heart failure or cancer. Future studies might examine specific disease-related predictors to identify characteristics that put specific populations at risk and inform the development of tailored interventions to improve patient outcomes.
The wide range of sample sizes in the included studies demonstrates the versatility of machine learning algorithms in studies with various sample sizes.90 Although small samples can be problematic for specific algorithms, such as Neural networks, other algorithms, such as Naïve bayes, can better accommodate smaller samples since it assumes class independence.91 However, data-demanding algorithms like Neural networks may be more prone to overfitting with smaller samples, leading to overestimated algorithm performance.92 Thus, we encourage researchers to consider sample size when determining the appropriate algorithm to apply to the data.
Tree based algorithms (e.g., Decision tree, ADABoost, Classification and regression trees, Random forest) were frequently cited in this review and are known to be a reliable method in biomedical literature.93 An advantage of a Decision tree is its interpretability,94 meaning that humans can understand the informative features driving the algorithm’s performance.95 Alternatively, “black box” algorithms, such as Neural networks, are less interpretable, which may indicate why they were used less frequently94 in studies included in this review. However, there has been a rise in the use of Neural networks96,97 applied to EHR data, given its increased performance and reduction in preprocessing and feature engineering compared to other explainable algorithms.97 Recent literature has published model-agnostic methods to increase the interpretability of “black box” models, such as Shapley Values, which help to explain features that are more informative (i.e., feature importance).95 Therefore, future studies may be more apt to apply Neural networks in the HHC setting. However, the studies in this review that applied Neural networks did not outperform alternative algorithms, suggesting that other models should not be abandoned altogether.
Additionally, NLP, an algorithm used to process unstructured text in clinical notes, was infrequently applied in studies included in this review. NLP is a method that takes text contained in clinical notes and makes it accessible to machine learning algorithms by transforming the text into a structured form.98 Eighty percent of healthcare data is unstructured data (e.g., clinical notes).99 Previous studies have demonstrated the value of analyzing clinical notes to reflect clinician concerns to improve risk prediction in HHC.44,100–102 More specifically, studies have found that additional information about patients’ social determinants103 and mental health104 can be found in clinical notes. Thus, future studies in HHC might explore using NLP to extract additional information about a patient’s biological, psychological, and interpersonal characteristics to improve risk predictive performance.
Overall, there were variations in algorithm performance among the included studies; however, comparing algorithms with different performance metrics is challenging. Each performance metric has its advantages and disadvantages. F-score, which handles imbalanced data, gives more weight to the majority class regardless of if it is the outcome of interest.105 AUC is better for ranking overall prediction because it equally weights both classes.105 The variety of measures used to report algorithm performance suggests no general consensus on a particular measure of best use.106 Therefore, it is important that the selection of performance metrics is thoughtfully considered and explained in the manuscript.107
Limitations
A few limitations of this scoping review need to be addressed. First, we recognize that sub-optimal machine learning performance may be driven by the dataset’s quality and influence the results. Additionally, the development of machine learning in recent years may have contributed to improved machine learning performance. The scope of this study did not include an in-depth analysis of how the data was collected, but future studies should consider the bias contained in these data. Similarly, no studies rigorously evaluated bias (e.g., racial, age) in algorithm performance recognizing an important gap to consider in future research. This review did not find HHC studies that compared prediction of machine learning models with clinical expert judgment and we recommend exploring this comparison in the future studies. This scoping review contains multiple studies published by the same research team, which suggests that minimal teams have published in this area within the last few years, potentially skewing the results. We mapped predictors to the Biopsychosocial Model to get a comprehensive overview of predictors considered in the included machine learning models; however, an alternative framework may have led to a more granular categorization of predictors.
Conclusion
With the increased demand for healthcare provided in the home, we anticipate the number of machine learning algorithms in HHC to continue to increase. Results from this review demonstrate the feasibility and potential of machine learning to support clinical care in the HHC setting. Future studies should consider how these machine learning algorithms can be packaged into decision support systems that can guide clinical decision-making at the point of care to reduce adverse events, such as hospitalizations. In addition, incorporating psychological and interpersonal characteristics into machine learning models will support comprehensive, holistic risk prediction. To enable widespread adoption of machine learning, stakeholders should consider how they can further promote the standardization of documentation elements in the HHC setting.
Supplementary Material
Summary table.
What is already known on the topic
Machine learning is a powerful technique used to help predict adverse events.
Machine learning applied to EHR data in HHC is bringing documentation-driven evidence to clinicians at the point of care.
What this study adds to our knowledge
Summarizes the literature of machine learning used in the HHC setting to predict adverse events.
Explores the basic components of the machine learning algorithms applied to EHR data in the HHC setting to predict adverse events.
Acknowledgements
This work was supported by the National Institute of Nursing Research (NINR) [grant T32NR007969 (MH, DS), R01 NR018831]; the Jonas Scholarship (MH), and the Agency for Healthcare Research and Quality [grant R01 HS027742].
Footnotes
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Conflict of interest
Declarations of interest for all authors: none
Statement of conflict of interest
Authors have no conflict of interests to disclose.
Contributor Information
Mollie Hobensack, Columbia University School of Nursing, Address: 560 W 168th Street, New York, NY, USA 10032.
Jiyoun Song, Columbia University School of Nursing, New York City, NY, USA.
Danielle Scharp, Columbia University School of Nursing, New York City, NY, USA.
Kathryn H. Bowles, University of Pennsylvania School of Nursing, Department of Biobehavioral Health Sciences Philadelphia, PA, USA; Vice President and Director of the Center for Home Care Policy & Research, VNS Health, New York City, NY, USA.
Maxim Topaz, Columbia University School of Nursing, New York City, NY, USA; Data Science Institute, Columbia University, New York City, NY, USA; Center for Home Care Policy & Research, VNS Health, New York City, NY, USA.
References
- 1.Home health quality reporting program. Centers for Medicare and Medicaid Services. Published 2022. Accessed March 23, 2022. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/HomeHealthQualityInits
- 2.Centers for Medicare and Medicaid Services. Medicare and Home Health Care. US Department of Health and Human Services; 2003. Accessed March 23, 2022. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/HomeHealthQualityInits/Downloads/HHQIHHBenefits.pdf
- 3.Gray E, Currey J, Considine J. Hospital in the Home nurses’ recognition and response to clinical deterioration. J Clin Nurs. 2018;27(9–10):2152–2160. doi: 10.1111/JOCN.14076 [DOI] [PubMed] [Google Scholar]
- 4.Siclovan DM, Bang JT, Yakusheva O, et al. Effectiveness of home health care in reducing return to hospital: Evidence from a multi-hospital study in the US. Int J Nurs Stud. 2021;119:103946. doi: 10.1016/j.ijnurstu.2021.103946 [DOI] [PubMed] [Google Scholar]
- 5.Jacobson G, Freed M, Damico A, Neuman T. A Dozen Facts About Medicare Advantage in 2019. Kaiser Family Foundation. Published 2019. Accessed August 14, 2022. https://www.kff.org/medicare/issue-brief/a-dozen-facts-about-medicare-advantage-in-2019/ [Google Scholar]
- 6.Busby J, Purdy S, Hollingworth W. A systematic review of the magnitude and cause of geographic variation in unplanned hospital admission rates and length of stay for ambulatory care sensitive conditions. BMC Health Services Research 2015 15:1. 2015;15(1):1–15. doi: 10.1186/S12913-015-0964-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Home Health Quality Measures. Centers for Medicare and Medicaid. Published 2022. Accessed October 1, 2021. https://www.cms.gov/Medicare/Quality-Initiatives-Patient-Assessment-Instruments/HomeHealthQualityInits/Home-Health-Quality-Measures
- 8.Fraser I, Davies SM, Gritz M, et al. AHRQ Quality Indicators Guide to Prevention Quality Indicators: Hospital Admission for Ambulatory Care Sensitive Conditions. Published online 2001.
- 9.Solberg LI, Ohnsorg KA, Parker ED, et al. Potentially preventable hospital and emergency department events: Lessons from a large innovation project. Perm J. 2018;22:17–102. doi: 10.7812/TPP/17-102 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Wood C, Chaboyer W, Carr P, C W, W C, P C. How do nurses use early warning scoring systems to detect and act on patient deterioration to ensure patient safety? A scoping review. 2019;94:166–178. doi: 10.1016/J.IJNURSTU.2019.03.012 [DOI] [PubMed] [Google Scholar]
- 11.Sockolow PS, Bowles KH, Adelsberger MC, Chittams JL, Liao C. Impact of Homecare Electronic Health Record on Timeliness of Clinical Documentation, Reimbursement, and Patient Outcomes. Appl Clin Inform. 2014;5(2):445. doi: 10.4338/ACI-2013-12-RA-0106 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Ossisanwo FY, Akinsola JET, Awodele O, Hinmikaiye JO, Olakanmi O, Akinjobi J. Supervised machine learning algorithms: Classification and comparison. International Journal of Computer Trends and Technology. 2017;48. doi: 10.14445/22312803/IJCTTV48P126 [DOI] [Google Scholar]
- 13.Ohu I, Benny PK, Rodrigues S, Carlson JN. Applications of machine learning in acute care research. J Am Coll Emerg Physicians Open. 2020;1(5):766. doi: 10.1002/EMP2.12156 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Shillan D, Sterne JAC, Champneys A, Gibbison B. Use of machine learning to analyse routinely collected intensive care unit data: A systematic review. Crit Care. 2019;23(1):1–11. doi: 10.1186/S13054-019-2564-9/FIGURES/5 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Mpanya D, Celik T, Klug E, Ntsinjana H. Predicting mortality and hospitalization in heart failure using machine learning: A systematic literature review. IJC Heart & Vasculature. 2021;34:100773. doi: 10.1016/J.IJCHA.2021.100773 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Korach ZT, Cato KD, Collins SA, et al. Unsupervised Machine Learning of Topics Documented by Nurses about Hospitalized Patients Prior to a Rapid-Response Event. Appl Clin Inform. 2019;10(5):952–963. doi: 10.1055/s-0039-3401814 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Carroll W Artificial intelligence: Optimizing patient care in acute and postacute settings. Nurs Manage. 2021;52(11):29–32. doi: 10.1097/01.NUMA.0000795584.59335.F7 [DOI] [PubMed] [Google Scholar]
- 18.Awaysheh A, Wilcke J, Elvinger F, Rees L, Fan W, Zimmerman KL. Review of medical decision support and machine-learning methods. Vet Pathol. 2019;56(4):512–525. doi: 10.1177/0300985819829524 [DOI] [PubMed] [Google Scholar]
- 19.Bompelli A, Wang Y, Wan R, et al. Social and behavioral determinants of health in the era of artificial intelligence with electronic health records: A scoping review. Health Data Science. Published online January 22, 2021. doi: 10.34133/2021/9759016 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 20.Koleck TA, Dreisbach C, Bourne PE, Bakken S. Natural language processing of symptoms documented in free-text narratives of electronic health records: a systematic review. Journal of the American Medical Informatics Association. 2019;26(4):364–379. doi: 10.1093/JAMIA/OCY173 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Hammond G, Johnston K, Huang K, Joynt Maddox KE. Social Determinants of Health Improve Predictive Accuracy of Clinical Risk Models for Cardiovascular Hospitalization, Annual Cost, and Death. Circ Cardiovasc Qual Outcomes. Published online 2020:290–299. doi: 10.1161/CIRCOUTCOMES.120.006752 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Kommescher M, Wagner M, Pützfeld V, et al. Coping as a predictor of treatment outcome in people at clinical high risk of psychosis. Early Interv Psychiatry. 2016;10(1):17–27. doi: 10.1111/EIP.12130 [DOI] [PubMed] [Google Scholar]
- 23.Sockolow PS, Bowles KH, Topaz M, et al. The time is now: Informatics research opportunities in home health care. Appl Clin Inform. 2021;12(1):100–106. doi: 10.1055/S-0040-1722222/ID/JR200182IE-68 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Schopp LH, Hales JW, Brown GD, Quetsch JL. A rationale and training agenda for rehabilitation informatics: Roadmap for an emerging discipline. NeuroRehabilitation. 2003;18(2):159–170. doi: 10.3233/NRE-2003-18208 [DOI] [PubMed] [Google Scholar]
- 25.Aggarwal N, Ahmed M, Basu S, et al. Advancing Artificial Intelligence in Health Settings Outside the Hospital and Clinic. NAM Perspectives. Published online November 30, 2020. doi: 10.31478/202011F [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Geng F, Mansouri S, Stevenson DG, Grabowski DC. Evolution of the home health care market: The expansion and quality performance of multi-agency chains. Health Serv Res. 2020;55(S3):1073–1084. doi: 10.1111/1475-6773.13597 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Nayyar A, Gadhavi L, Zaman N. Machine learning in healthcare: Review, opportunities and challenges. In: Machine Learning and the Internet of Medical Things in Healthcare. Elsevier; 2021:23–45. doi: 10.1016/B978-0-12-821229-5.00011-2 [DOI] [Google Scholar]
- 28.Tricco AC, Lillie E, Zarin W, et al. PRISMA extension for scoping reviews (PRISMA-ScR): Checklist and explanation. Ann Intern Med. 2018;169(7):467–473. doi: 10.7326/M18-0850 [DOI] [PubMed] [Google Scholar]
- 29.Scott CM, Lubritz RR, Graham GF. Adverse Events. Dermatological Cryosurgery and Cryotherapy Published online April 1, 2022:221–223. doi: 10.1007/978-1-4471-6765-5_47 [DOI] [Google Scholar]
- 30.von Gerich H, Moen H, Block LJ, et al. Artificial Intelligence-based technologies in nursing: A scoping literature review of the evidence. Int J Nurs Stud. 2022;127:104153. doi: 10.1016/J.IJNURSTU.2021.104153 [DOI] [PubMed] [Google Scholar]
- 31.Borell-Carrió F, Suchman AL, Epstein RM. The Biopsychosocial Model 25 Years Later: Principles, Practice, and Scientific Inquiry. Ann Fam Med. 2004;2(6):576. doi: 10.1370/AFM.245 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Engel GL. The Need for a New Medical Model: A Challenge for Biomedicine. Science (1979). 1977;196(4286):129–136. doi: 10.1126/SCIENCE.847460 [DOI] [PubMed] [Google Scholar]
- 33.Lehman BJ, David DM, Gruber JA. Rethinking the biopsychosocial model of health: Understanding health as a dynamic system. Soc Personal Psychol Compass. 2017;11(8):e12328. doi: 10.1111/SPC3.12328 [DOI] [Google Scholar]
- 34.Wolff RF, Moons KGM, Riley RD, et al. PROBAST: A tool to assess the risk of bias and applicability of prediction model studies. Ann Intern Med. 2019;170(1):51–58. doi: 10.7326/M18-1376/SUPPL_FILE/M18-1376_SUPPLEMENT.PDF [DOI] [PubMed] [Google Scholar]
- 35.Page MJ, McKenzie JE, Bossuyt PM, et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. The BMJ. 2021;372. doi: 10.1136/bmj.n71 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Kang Y, McHugh MD, Chittams J, Bowles KH. Utilizing home health care electronic health records for telehomecare patients with heart failure: a decision tree approach to detect associations with rehospitalizations. Comput Inform Nurs. 2016;34(4):175. doi: 10.1097/CIN.0000000000000223 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Sullivan SS, Casucci S, Li CS. Eliminating the Surprise Question Leaves Homecare Providers with Few Options for Identifying Mortality Risk. Am J Hosp Palliat Care. 2020;37(7):542. doi: 10.1177/1049909119892830 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Olson CH, Dey S, Kumar V, Monsen KA, Westra BL. Clustering of elderly patient subgroups to identify medication-related readmission risks. Int J Med Inform. 2016;85(1):43–52. doi: 10.1016/J.IJMEDINF.2015.10.004 [DOI] [PubMed] [Google Scholar]
- 39.Madigan EA, Curet OL. A data mining approach in home healthcare: Outcomes and service use. BMC Health Serv Res. 2006;6(1):1–10. doi: 10.1186/1472-6963-6-18/FIGURES/4 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 40.Lo Y, Lynch SF, Urbanowicz RJ, et al. Using Machine Learning on Home Health Care Assessments to Predict Fall Risk. Stud Health Technol Inform. 2019;264:684–688. doi: 10.3233/SHTI190310 [DOI] [PubMed] [Google Scholar]
- 41.Song J, Hobensack M, Bowles KH, et al. Clinical notes: An untapped opportunity for improving risk prediction for hospitalization and emergency department visit during home health care. J Biomed Inform. 2022;128:104039. doi: 10.1016/J.JBI.2022.104039 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 42.Jones CD, Falvey J, Hess E, et al. Predicting hospital readmissions from home healthcare in medicare beneficiaries. J Am Geriatr Soc. 2019;67(12):2505–2510. doi: 10.1111/JGS.16153 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 43.Topaz M, Woo K, Ryvicker M, Zolnoori M, Cato K. Home healthcare clinical notes predict patient hospitalization and emergency department visits. Nurs Res. 2020;69(6):448–454. doi: 10.1097/NNR.0000000000000470 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 44.Song J, Woo K, Shang J, Ojo M, Topaz M. Predictive risk models for wound infection-related hospitalization or ED visits in home health care using machine-learning algorithms. Adv Skin Wound Care. 2021;34(8):1–12. doi: 10.1097/01.ASW.0000755928.30524.22 [DOI] [PubMed] [Google Scholar]
- 45.Jones A, Costa AP, Pesevski A, McNicholas PD. Predicting hospital and emergency department utilization among community-dwelling older adults: Statistical and machine learning approaches. PLoS One. 2018;13(11). doi: 10.1371/JOURNAL.PONE.0206662 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 46.Kuspinar A, Hirdes JP, Berg K, McArthur C, Morris JN. Development and validation of an algorithm to assess risk of first-time falling among home care clients. BMC Geriatr. 2019;19(1):1–8. doi: 10.1186/S12877-019-1300-2/FIGURES/2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 47.McArthur C, Ioannidis G, Jantzi M, et al. Development and validation of the fracture risk scale home care (FRS-HC) that predicts one-year incident fracture: an electronic record-linked longitudinal cohort study. BMC Musculoskelet Disord. 2020;21(1). doi: 10.1186/S12891-020-03529-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 48.Zhu M, Chen W, Hirdes JP, Stolee P. The K-nearest neighbor algorithm predicted rehabilitation potential better than current Clinical Assessment Protocol. J Clin Epidemiol. 2007;60(10):1015–1021. doi: 10.1016/J.JCLINEPI.2007.06.001 [DOI] [PubMed] [Google Scholar]
- 49.Nuutinen M, Leskelä RL, Torkki P, Suojalehto E, Tirronen A, Komssi V. Developing and validating models for predicting nursing home admission using only RAI-HC instrument data. https://doi.org/101080/1753815720191656212. 2019;45(3):292–308. doi: 10.1080/17538157.2019.1656212 [DOI] [PubMed] [Google Scholar]
- 50.Nabal M, Bescos M, Barcons M, Torrubia P, Trujillano J, Requena A. New Symptom-Based Predictive Tool for Survival at Seven and Thirty Days Developed by Palliative Home Care Teams. J Palliat Med. 2014;17(10):1158. doi: 10.1089/JPM.2013.0630 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 51.Calvo M, Cano I, Hernandez C, et al. Class Imbalance Impact on the Prediction of Complications during Home Hospitalization: A Comparative Study. Proceedings of the Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBS. Published online July 1, 2019:3446–3449. doi: 10.1109/EMBC.2019.8857746 [DOI] [PubMed] [Google Scholar]
- 52.Calvo M, González R, Seijas N, et al. Health Outcomes from Home Hospitalization: Multisource Predictive Modeling. J Med Internet Res 2020;22(10):e21367 https://www.jmir.org/2020/10/e21367. 2020;22(10):e21367. doi: 10.2196/21367 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 53.Zhang J, Goode KM, Rigby A, Balk AHMM, Cleland JG. Identifying patients at risk of death or hospitalisation due to worsening heart failure using decision tree analysis: Evidence from the Trans-European Network-Home-Care Management System (TEN-HMS) Study. Int J Cardiol. 2013;163(2):149–156. doi: 10.1016/J.IJCARD.2011.06.009 [DOI] [PubMed] [Google Scholar]
- 54.Witt UF, Nibe SM, Ole H, Lebech CS. A novel approach for predicting acute hospitalizations among elderly recipients of home care? A model development study. Int J Med Inform. 2022;160:104715. doi: 10.1016/J.IJMEDINF.2022.104715 [DOI] [PubMed] [Google Scholar]
- 55.Yang C, Yu R, Ji H, Jiang H, Yang W, Jiang F. Application of data mining in the provision of in-home medical care for patients with advanced cancer. Transl Cancer Res. 2021;10(6):3013. doi: 10.21037/TCR-21-896 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 56.Maritz R, Ehrmann C, Prodinger B, Tennant A, Stucki G. The influence and added value of a Standardized Assessment and Reporting System for functioning outcomes upon national rehabilitation quality reports. International Journal for Quality in Health Care. 2020;32(6):379–387. doi: 10.1093/INTQHC/MZAA058 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 57.O’Connor M, Davitt JK. The Outcome and Assessment Information Set (OASIS): A Review of Validity and Reliability. Home Health Care Serv Q. 2012;31(4):267. doi: 10.1080/01621424.2012.703908 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 58.Wagner A, Schaffert R, Möckli N, Zúñiga F, Dratva J. Home care quality indicators based on the Resident Assessment Instrument-Home Care (RAI-HC): a systematic review. BMC Health Serv Res. 2020;20(1). doi: 10.1186/S12913-020-05238-X [DOI] [PMC free article] [PubMed] [Google Scholar]
- 59.Jiang T, Gradus JL, Rosellini AJ. Supervised machine learning: A brief primer. Behav Ther. 2020;51(5):675. doi: 10.1016/J.BETH.2020.05.002 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 60.Caetano SJ, Sonpavde G, Pond GR. C-statistic: A brief explanation of its construction, interpretation and limitations. Eur J Cancer. 2018;90:130–132. doi: 10.1016/j.ejca.2017.10.027 [DOI] [PubMed] [Google Scholar]
- 61.Seliya N, Khoshgoftaar TM, van Hulse J. A study on the relationships of classifier performance metrics. In: Proceedings - International Conference on Tools with Artificial Intelligence, ICTAI. ; 2009:59–66. doi: 10.1109/ICTAI.2009.25 [DOI] [Google Scholar]
- 62.Venema E, Wessler BS, Paulus JK, et al. Large-scale validation of the prediction model risk of bias assessment Tool (PROBAST) using a short form: high risk of bias models show poorer discrimination. J Clin Epidemiol. 2021;138:32–39. doi: 10.1016/j.jclinepi.2021.06.017 [DOI] [PubMed] [Google Scholar]
- 63.Jong Y, Ramspek CL, Zoccali C, Jager KJ, Dekker FW, Diepen M. Appraising prediction research: a guide and meta-review on bias and applicability assessment using the Prediction model Risk Of Bias ASsessment Tool (PROBAST). Nephrology. 2021;26(12):939–947. doi: 10.1111/nep.13913 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 64.Muralitharan S, Nelson W, Di S, et al. Machine Learning–Based Early Warning Systems for Clinical Deterioration: Systematic Scoping Review. J Med Internet Res 2021;23(2):e25187 https://www.jmir.org/2021/2/e25187. 2021;23(2):e25187. doi: 10.2196/25187 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 65.Mahmoudi E, Kamdar N, Kim N, Gonzales G, Singh K, Waljee AK. Use of electronic medical records in development and validation of risk prediction models of hospital readmission: systematic review. BMJ. 2020;369. doi: 10.1136/BMJ.M958 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 66.Huang Y, Talwar A, Chatterjee S, Aparasu RR. Application of machine learning in predicting hospital readmissions: a scoping review of the literature. BMC Med Res Methodol. 2021;21(1):1–14. doi: 10.1186/S12874-021-01284-Z/TABLES/3 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 67.Sockolow PS, Adelsberger MC, Bowles KH. Identifying Certification Criteria for Home Care EHR Meaningful Use. AMIA Annual Symposium Proceedings. 2011;2011:1280. Accessed July 8, 2022. [PMC free article] [PubMed] [Google Scholar]
- 68.Sockolow P, Zakeri I, hD P, Bowles KH, Chaney K. Barriers and Facilitators to Implementation and Adoption of EHR in Home Care - Final Report. Accessed July 8, 2022. www.ahrq.gov
- 69.Ptaszek A, Deutsch A, Li Q, et al. Policy and Quality: The Impact Act of 2014 and Development of the Skilled Nursing Facility Quality Reporting Program. Innov Aging. 2018;2(Suppl 1):26. doi: 10.1093/GERONI/IGY023.098 [DOI] [Google Scholar]
- 70.Kennedy EE, Bowles KH, Aryal S. Systematic review of prediction models for postacute care destination decision-making. Journal of the American Medical Informatics Association. 2021;29(1):176–186. doi: 10.1093/JAMIA/OCAB197 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 71.Keim-Malpass J, Ratcliffe SJ, Moorman LP, et al. Predictive Monitoring-Impact in Acute Care Cardiology Trial (PM-IMPACCT): Protocol for a Randomized Controlled Trial. JMIR Res Protoc. 2021;10(7). doi: 10.2196/29631 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 72.Rossetti SC, Dykes PC, Knaplund C, et al. The Communicating Narrative Concerns Entered by Registered Nurses (CONCERN) Clinical Decision Support Early Warning System: Protocol for a Cluster Randomized Pragmatic Clinical Trial. JMIR Res Protoc. 2021;10(12). doi: 10.2196/30238 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 73.Yin J, Ngiam KY, Teo HH. Role of Artificial Intelligence Applications in Real-Life Clinical Practice: Systematic Review. J Med Internet Res. 2021;23(4). doi: 10.2196/25759 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 74.Seneviratne MG, Shah NH, Chu L. Bridging the implementation gap of machine learning in healthcare. BMJ Innov. 2020;6(2):45–47. doi: 10.1136/BMJINNOV-2019-000359 [DOI] [Google Scholar]
- 75.Gjevjon ER, Hellesø R. The quality of home care nurses’ documentation in new electronic patient records. J Clin Nurs. 2010;19(1–2):100–108. doi: 10.1111/J.1365-2702.2009.02953.X [DOI] [PubMed] [Google Scholar]
- 76.Li C, Zhou R, Yao N, Cornwell T, Wang S. Health Care Utilization and Unmet Needs in Chinese Older Adults With Multimorbidity and Functional Impairment. J Am Med Dir Assoc. 2020;21(6):806–810. doi: 10.1016/J.JAMDA.2020.02.010 [DOI] [PubMed] [Google Scholar]
- 77.Akhtar S, Loganathan M, Nowaczynski M, et al. Aging at Home: A Portrait of Home-Based Primary Care Across Canada. Healthc Q. 2019;22(1):30–35. doi: 10.12927/HCQ.2019.25839 [DOI] [PubMed] [Google Scholar]
- 78.KPMG International. Delivering Healthcare Services Closer to Home; 2019. Accessed July 8, 2022. https://assets.kpmg/content/dam/kpmg/tw/pdf/2019/11/kpmg-delivering-healthcare-services-closer-to-home.pdf
- 79.Choi J, Jenkins ML, Cimino JJ, White TM, Bakken S. Toward semantic interoperability in home health care: Formally representing OASIS items for integration into a concept-oriented terminology. Journal of the American Medical Informatics Association. 2005;12(4):410–417. doi: 10.1197/JAMIA.M1786/2/JAMIAM1786.F02.JPEG [DOI] [PMC free article] [PubMed] [Google Scholar]
- 80.Lyon AR, Ludwig K, Wasse JK, Bergstrom A, Hendrix E, McCauley E. Determinants and Functions of Standardized Assessment Use among School Mental Health Clinicians: A Mixed Methods Evaluation. Adm Policy Ment Health. 2016;43(1):122. doi: 10.1007/S10488-015-0626-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 81.Mcilfatrick S, Hasson F. Evaluating an holistic assessment tool for palliative care practice. J Clin Nurs. 2014;23(7–8):1064–1075. doi: 10.1111/JOCN.12320 [DOI] [PubMed] [Google Scholar]
- 82.Al-Shwaheen TI, Moghbel M, Hau YW, Ooi CY. Use of learning approaches to predict clinical deterioration in patients based on various variables: a review of the literature. Artif Intell Rev. 2022;55(2):1055–1084. doi: 10.1007/S10462-021-09982-2 [DOI] [Google Scholar]
- 83.Phuong J, Zampino E, Dobbins N, et al. Extracting Patient-level Social Determinants of Health into the OMOP Common Data Model. AMIA Annu Symp Proc. 2021;2021:989–998. [PMC free article] [PubMed] [Google Scholar]
- 84.Navathe AS, Zhong F, Lei VJ, et al. Hospital Readmission and Social Risk Factors Identified from Physician Notes. Health Serv Res. 2018;53(2):1110–1136. doi: 10.1111/1475-6773.12670 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 85.Chen M, Tan X, Padman R. Social determinants of health in electronic health records and their impact on analysis and risk prediction: A systematic review. Journal of the American Medical Informatics Association. 2020;27(11):1764–1773. doi: 10.1093/jamia/ocaa143 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 86.Turner-Musa J, Ajayi O, Kemp L. Examining Social Determinants of Health, Stigma, and COVID-19 Disparities. Healthcare. 2020;8(2):168. doi: 10.3390/healthcare8020168 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 87.Knaak S, Mantler E, Szeto A. Mental illness-related stigma in healthcare: Barriers to access and care and evidence-based solutions. Healthc Manage Forum. 2017;30(2):111–116. doi: 10.1177/0840470416679413 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 88.Ellenbecker CH, Samia L, Cushman MJ, Alster K. Patient Safety and Quality in Home Health Care. Agency for Healthcare Research and Quality (US); 2008. Accessed September 28, 2022. http://www.ncbi.nlm.nih.gov/pubmed/21328733 [PubMed] [Google Scholar]
- 89.Chae S, Song J, Ojo M, et al. Factors associated with poor self-management documented in home health care narrative notes for patients with heart failure. Heart & Lung. 2022;55:148–154. doi: 10.1016/J.HRTLNG.2022.05.004 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 90.Althnian A, AlSaeed D, Al-Baity H, et al. Impact of Dataset Size on Classification Performance: An Empirical Evaluation in the Medical Domain. Applied Sciences 2021, Vol 11, Page 796. 2021;11(2):796. doi: 10.3390/APP11020796 [DOI] [Google Scholar]
- 91.Howedi F, Mohd M. Text classification for authorship attribution using naive bayes classifier with limited training data. Computer Engineering and Intelligent Systems. 2014;5(4). Accessed January 26, 2022. www.iiste.org [Google Scholar]
- 92.Adadi A A survey on data-efficient algorithms in big data era. Journal of Big Data 2021 8:1. 2021;8(1):1–54. doi: 10.1186/S40537-021-00419-9 [DOI] [Google Scholar]
- 93.Olson RS, la Cava W, Mustahsan Z, Varik A, Moore JH. Data-driven advice for applying machine learning to bioinformatics problems. Pac Symp Biocomput. 2018;23(212669):192. doi: 10.1142/9789813235533_0018 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 94.Petch J, Di S, Nelson W. Opening the Black Box: The Promise and Limitations of Explainable Machine Learning in Cardiology. Canadian Journal of Cardiology. 2022;38(2):204–213. doi: 10.1016/J.CJCA.2021.09.004 [DOI] [PubMed] [Google Scholar]
- 95.Molnar C Interpretable Machine Learning Interpretable Machine Learning; 2019.
- 96.Tomašev N, Harris N, Baur S, et al. Use of deep learning to develop continuous-risk models for adverse event prediction from electronic health records. Nature Protocols 2021 16:6. 2021;16(6):2765–2787. doi: 10.1038/s41596-021-00513-5 [DOI] [PubMed] [Google Scholar]
- 97.Shickel B, Tighe PJ, Bihorac A, Rashidi P. Deep EHR: A Survey of Recent Advances in Deep Learning Techniques for Electronic Health Record (EHR) Analysis. IEEE J Biomed Health Inform. 2018;22(5):1589–1604. doi: 10.1109/JBHI.2017.2767063 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 98.Davenport T, Kalakota R. The potential for artificial intelligence in healthcare. Future Healthc J. 2019;6(2):94. doi: 10.7861/FUTUREHOSP.6-2-94 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 99.Kong HJ. Managing unstructured big data in healthcare system. Healthc Inform Res. 2019;25(1):1–2. doi: 10.4258/hir.2019.25.1.1 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 100.Woo K, Song J, Adams V, et al. Exploring prevalence of wound infections and related patient characteristics in homecare using natural language processing. Int Wound J. Published online June 2021. doi: 10.1111/iwj.13623 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 101.Song J, Hobensack M, Bowles KH, et al. Clinical notes: An untapped opportunity for improving risk prediction for hospitalization and emergency department visit during home health care. J Biomed Inform. 2022;128:104039. doi: 10.1016/J.JBI.2022.104039 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 102.Hobensack M, Ojo M, Barron Y, et al. Documentation of hospitalization risk factors in electronic health records (EHRs): a qualitative study with home healthcare clinicians. 2022;ocac023. doi: 10.1093/jamia/ocac023 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 103.Patra BG, Sharma MM, Vekaria V, et al. Extracting social determinants of health from electronic health records using natural language processing: A systematic review. Journal of the American Medical Informatics Association. 2021;28(12):2716–2727. doi: 10.1093/jamia/ocab170 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 104.le Glaz A, Haralambous Y, Kim-Dufor DH, et al. Machine learning and natural language processing in mental health: Systematic review. J Med Internet Res. 2021;23(5):e15708. doi: 10.2196/15708 [DOI] [PMC free article] [PubMed] [Google Scholar]
- 105.Powers DM. Evaluation: From Precision, Recall And F-measure to ROC, Informedness, Markedness & Correlation. 2011;2(1):37–63. Accessed July 21, 2022. http://www.bioinfo.in/contents.php?id=51 [Google Scholar]
- 106.Seliya N, Khoshgoftaar TM, van Hulse J. A study on the relationships of classifier performance metrics. Published online 2009. doi: 10.1109/ICTAI.2009.25 [DOI] [Google Scholar]
- 107.Jiao Y, Du P. Performance measures in evaluating machine learning based bioinformatics predictors for classifications. Quantitative Biology 2016 4:4. 2016;4(4):320–330. doi: 10.1007/S40484-016-0081-2 [DOI] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
