Abstract
Background
While many falls are preventable, they remain a leading cause of injury and death in older adults. Primary care clinics largely rely on screening questionnaires to identify people at risk of falls. Limitations of standard fall risk screening questionnaires include suboptimal accuracy, missing data, and non-standard formats, which hinder early identification of risk and prevention of fall injury. We used machine learning methods to develop and evaluate electronic health record (EHR)-based tools to identify older adults at risk of fall-related injuries in a primary care population and compared this approach to standard fall screening questionnaires.
Methods
Using patient-level clinical data from an integrated health care system consisting of 16-member institutions, we conducted a case-control study to develop and evaluate prediction models for fall-related injuries in older adults. Questionnaire-derived prediction with three questions from a commonly used fall risk screening tool were evaluated. We then developed four temporal machine learning models using routinely available longitudinal EHR data to predict future risk of fall injury. We also developed a fall injury prevention clinical decision support (CDS) implementation prototype to link preventative interventions to patient-specific fall injury risk factors.
Results
Questionnaire-based risk screening achieved AUC up to 0.59 with 23% to 33% similarity for each pair of three fall injury screening questions. EHR-based machine learning risk screening showed significantly improved performance (best AUROC=0.76), with similar prediction performance between 6-month and one-year prediction models.
Conclusions
The current method of questionnaire-based fall risk screening of older adults is suboptimal with redundant items, inadequate precision and no linkage to prevention. A machine learning fall injury prediction method can accurately predict risk with superior sensitivity while freeing up clinical time for initiating personalized fall prevention interventions. The developed algorithm and data science pipeline can impact routine primary care fall prevention practice.
Keywords: Fall and fall-related injury, Community dwelling older adults, Primary care, Machine learning, Risk screening
Introduction
Preventable falls are the leading cause of injury and accidental death among adults aged 65 and older.1 Despite growing evidence that falls can be prevented,2,3 deaths from falls continue to rise.4 A single fall frequently results in a fear of falling that can begin a downward spiral of reduced mobility, leading to loss of function, activity restriction and further risk for falls.5,6 The challenge is to ensure that older adults at risk of falling receive appropriate referrals, education and support from their primary care provider. Many primary care providers (PCPs) are unaware of their patients’ falls, as half of older adults do not discuss falls with their PCP.7
Primary care clinics largely rely on screening questionnaires to target their fall prevention practices.8 However, the utility of existing screening questionnaires is limited by suboptimal accuracy of patient-reported falls, missing data, and non-standard formats (Figure 1).9,10 Fall risk screening and assessment is often included in the Medicare Annual Wellness Visit (AWV), but more than half of older patients do not participate in AWVs, and fall risk is not routinely addressed in other types of health care encounters. Fall risk screeners also have limited clinical impact because they are broad and do not inform individualized fall prevention plans.
Figure 1.
Comparison between current questionnaire-based approach and a data-driven approach for screening of fall and related injury
Given the large number of risk screeners recommended during the Medicare AWV and limited clinicians’ time, machine-learning based screening could potentially help streamline and enable provision of more efficient care (Figure 1). EHR data, with its high-intensity measures of patient status, could provide a useful data source to automate or at least inform primary care screening practices and tailored interventions.11–13 Using EHR data and machine learning algorithms could achieve more accurate prediction power for complex clinical conditions14–17 Recent work has shown that EHR-based models have robust performance with predicting fall risk in inpatient and outpatient settings18–20, suggesting that EHR derived algorithms could be used to augment the accuracy and streamline the number of items required to adequately screen for fall and fall-related injury risk. However, previous studies have several limitations making it challenging to translate these algorithms to clinical practice. First, most EHR-based models used patient reported fall events, as the prediction outcome, making it difficult to validate the algorithms across studies and health systems where this outcome in not consistently captured. Second, patients report fall events mainly when they sustain an injury and seek medical care. This may cause a high number of missing cases and result in model bias. Third, fall-related injuries, which is a more objective measurement, is rarely used as the model outcome. Furthermore, questions remain about how to effectively use longitudinal and highly-granular EHR data (including temporal features) to develop more accurate models for fall and related injury events while avoiding overfitting.21 Despite the promising studies to date, investigation is clearly needed to standardize fall and fall-related injury model outcome definitions using routinely captured EHR features and to develop risk models that can be linked to preventative interventions and inform routine care.
The goal of this project was to conduct a case-control study to develop and evaluate different screening tools for fall-related injuries in a primary care population. We also aimed to develop a fall injury prevention clinical decision support (CDS) implementation prototype to link patient-specific determinants of risk to evidence-based interventions to increase the utility of machine learning models in primary care practice.
Methods
Data source
We used clinical databases within the Mass General Brigham (MGB) Healthcare system, which includes a centralized clinical data warehouse consisting of data from 16-member institutions including academic medical centers, specialty and community hospitals, a rehabilitation network, and a primary and specialty care provider network. Patient records from primary care practices were extracted for the study.
Study cohort development and temporal design
We applied temporal criteria during cohort development (Figure 2). An “index time” was identified for the first fall risk screening time from 2018 (e.g., the time that all patients in the cohort were asked fall risk screening questions). We extracted model input features from a multi-year duration (2015–2017) before index time. By consulting with clinical experts, we used various time durations to extract these EHR variables, based on their time-sensitive nature, e.g. time sensitive features versus static features (see details in below sections and Figure 2). We used patient records in 2019 to capture the model outcome: fall-related injury based on validated diagnosis codes.22
Figure 2.
Summary of temporal EHR-based machine learning prediction model
We then applied multiple data mining steps (Step 1 to 4 were applied for both case and control groups): First, we identified patients age 65 and above and with at least one primary care visit within MGB network from 2015 to 2019. We only used clinical data from before 2020 to avoid potential impact of the COVID epidemic on clinical operations and data collection. Second, we selected patients with both a primary care visit and related fall risk screening records in 2018. Third, we captured records for a total of 17 types of diagnoses and 5 types of medication use information from 2015 to 2017 (see Supplementary Table 2). Fourth, we captured records of hospitalization (defined as >24 hours stay) and physical therapy in 2017 to identify patients who were more likely to experience severe health problems and to have mobility issues including gait and balance problems. These two variables were selected based on clinical expert recommendations as they could be associated with higher risk for fall injury (model outcome). Fifth, we identified patients with at least one fall injury diagnosis record in 2019 as the case group. Controls did not have any fall injury records during the same period. Through these steps, patients with any missing record or fall injury diagnoses were excluded from the study cohort.
Model development
1. Questionnaire-based screening tool evaluation
Questionnaire feature extraction
The Mass General Brigham fall risk screening questions recommended by the Centers for Disease Control and Prevention (CDC),23 are widely used in practice,24,25 and include: 1) have you fallen and hurt yourself in the past year?, 2) are you afraid that you might fall because of balance and walking problems?, and 3) have you fallen 2 or more times in the past year? The information collected from the three questions focuses on different aspects of fall and fall related injury, such as fall history, timing of previous falls with injury and fear of falling. Using the MGB study cohort from 2018, we obtained patient answers (YES or NO) to all three questions. We then developed three binary variables based on patient responses (YES as 1 and NO as 0).
Pair-wise similarity evaluation among three fall risk screening questions
Due to the binary format for all three questions, we chose the Jaccard similarity coefficient,26 rather than the correlation coefficient, to estimate their relationships. We calculated it for each pair among three questions to estimate their similarities. The Jaccard coefficient is defined as the size of the intersection divided by the size of the union.
Evaluation of fall risk screening questions
Multiple logistic regression models were developed using all three screening questions (full model) or part of the three question with different combinations (sub-models with two or one question) as independent variables. We used question records from 2018 as input and fall injury within one year after index time as the model outcome. To evaluate the model performance, we used the area under the receiver operating characteristic curve (AUC), sensitivity and specificity. All metrics were calculated using the test set from train-test split.
2. EHR-based machine learning prediction models (summarized in Figure 2)
EHR feature selection
Based on the literature and clinician expert opinion, we selected 31 EHR-derived variables as the model input, including patient demographics, diagnosis of comorbidities and fall-risk related medications. Data cleaning processes were conducted and patients with missing model input/outcome information were removed to develop the final study cohort.
Prediction outcome time windows
We defined two sequential time windows to capture model prediction outcomes: 1) fall injury occurred within 6 months after index visit time and 2) fall injury occurred within 1 year after index visit time.
Model development and evaluation
Using input features and outcomes defined above, four fall injury prediction models were developed, including regularized logistic regression (LR), support vector machines (SVM), random forest (RF), and neural network (NN). To overcome the overfitting issue, we tuned models through cross-validation to select the best set of parameters and evaluated their performance on an independent test set (see detail descriptions in Supplementary Material).
Development of fall injury prevention clinical decision support (CDS) implementation prototype (Summarized in Figure 3)
Figure 3.
Fall injury prevention CDS implementation prototype
We integrated machine learning model design and real-world clinical scenarios to develop a CDS implementation prototype. First, we categorized model input variables into clinically actionable features (e.g. comorbidity diagnosis) or non-actionable features (e.g. gender), based on whether or not the features are treatable (e.g., can be mapped to fall prevention guideline recommendations); Second, our clinical team generated a set of guideline-based fall prevention treatment strategies27,28 for each actionable model feature; Third, clinical experts linked actionable features to specific interventions, including treatment of comorbidities, medication adjustment, referral to specialist, and recommendations for exercise and/or physical therapy.
Results
Evaluation of current fall risk screener based on the primary care practice questionnaire
1). Similarity evaluation among fall risk screening questions
The pair-wise Jaccard similarity coefficient was calculated to evaluate the relationships among three questions (Supplementary Table 1). Question 1 and question 3 were most similar -- 33% of patients had the same answer (either both YES or both NO), while question 2 and question 3 were least similar -- 23% of patients had the same answer to these two questions. Jaccard similarity suggests that there was moderate redundancy in these questions.
2). Evaluation model for fall risk screening questions
Using 1,000 cases and 2,000 control patients from MGB system, seven logistic regression predictive models were implemented to predict fall injury with a different number of questions as input variables, in which the full model used all three questions and sub-models used either two or single question with different combinations. The full model achieved AUC of 0.59, with sensitivity and specificity of 0.39 and 0.77, respectively. Among three sub-models with two questions, similar performance was observed with AUC ranging from 0.57 to 0.59. For sub-models with a single question, only the model with question 2 showed a decreased performance with AUC of 0.53, while models with question 1 or question 3 achieved a comparable performance (AUC of 0.56 and 0.57) with the full model and two-question model (Table 1).
Table 1.
Performance of questionnaire-based models (full model vs sub-models)
Input Variables | AUC (SD) | Sensitivity (SD) | Specificity (SD) | |
---|---|---|---|---|
Full Model | All Three Questions | 0.59 (0.01) | 0.39 (0.01) | 0.77 (0.01) |
Sub-Model with two questions | Question 1+Question 2 | 0.57 (0.01) | 0.29 (0.02) | 0.84 (0.01) |
Question 1+Question 3 | 0.59 (0.01) | 0.39 (0.01) | 0.77 (0.01) | |
Question 2+Question 3 | 0.57 (0.02) | 0.28 (0.02) | 0.84 (0.01) | |
Sub-Model with single question | Question 1 | 0.57 (0.01) | 0.28 (0.01) | 0.84 (0.01) |
Question 2 | 0.53 (0.01) | 0.12 (0.01) | 0.94 (0.02) | |
Question 3 | 0.56 (0.01) | 0.25 (0.02) | 0.86 (0.01) |
Machine learning models with EHR variables to predict six-month and 12-month fall injury
Based on previous literature18,20,29 and recommendations from clinical experts, we developed models to predict fall injury within 6 months or 1 year after the index visit time. For each outcome prediction, we obtained 1,000 cases and 2,000 controls for model development.
Four predictive models were developed using a list of 31 potential predictive variables, including demographics, comorbidities, and medications (Supplementary Table 2). All four models showed decent performance across two prediction time windows. The 6-month prediction showed slightly better performance with the best AUROC =0.76 compared to AUROC =0.74 for the best one-year prediction model. Similar performance was observed across all models on all metrics (Table 2 and Supplementary Figure 1). Among the statistically significant predictors from the logistic regression model, patient age, was found to be the most predictive feature of fall injury followed by previous ICD code-based fall injury history (within 3 years before index visit) in both prediction outcome time windows. Other important predictors included hospitalization (within one year before index visit), abnormalities of gait and mobility (within 3 years before index visit) and total number of total comorbidities (within 3 years before index visit).
Table 2.
Performance of EHR-based models and important predictors
Fall injury within 6 months | Fall injury within 1 year | |||||
---|---|---|---|---|---|---|
AUC (SD) | Sensitivity (SD) | Specificity (SD) | AUC (SD) | Sensitivity (SD) | Specificity (SD) | |
Logistic Regression | 0.75(0.03) | 0.65(0.06) | 0.76(0.06) | 0.74(0.02) | 0.62(0.06) | 0.76(0.06) |
Random Forest | 0.76(0.02) | 0.69(0.03) | 0.75(0.05) | 0.74(0.02) | 0.64(0.05) | 0.74(0.09) |
Support Vector Machine | 0.76(0.02) | 0.70(0.08) | 0.71(0.07) | 0.74(0.02) | 0.65(0.05) | 0.72(0.04) |
Neural Network | 0.76(0.02) | 0.70(0.04) | 0.73(0.07) | 0.73(0.02) | 0.66(0.07) | 0.71(0.05) |
Top Ten Significant Predictors (based on Logistic Regression) | ||||||
Age (Standardized coefficient =0.39), Fall injury history (0.21), Hospitalization (0.20), Abnormalities of gait and mobility (0.18), Total comorbidities (0.15), Has comorbidity history (0.11), Stroke (0.08), Gender (0.08), Orthostatic.hypotension (0.06), monoplegia_of_lower_Extr (-0.06) | Age (Standardized coefficient =0.59), Fall injury history (0.25), Abnormalities of gait and mobility (0.23), Hospitalization (0.21), Total comorbidities (0.11), Has comorbidity history (0.11), Parkinsons (0.11), Gender (0.10), Vision_impairment (0.09), Osteoporotic_fractures (0.08) |
Machine learning models with both EHR and questionnaire variables to predict fall injury
We developed a sub-cohort (750 cases and 1500 controls) with both EHR information and all three screening questions. The same four predictive models were developed using EHR variables and questionnaire variables. Similar performance was observed compared with EHR variable only models. Three screening questions showed negligible contribution to the model performance (Supplementary Table 3 and Supplementary Figure 2).
Fall injury prevention CDS implementation prototype
We developed a fall injury prevention CDS implementation prototype to demonstrate how our EHR-based machine learning model could link preventative interventions to patient-specific determinants of risk and facilitate real-world implementation of our algorithm (illustrated in Figure 3)
Discussion
Injury caused by a fall is a major contributor to morbidity and death among community dwelling older adults.30 Falls are also associated with significant costs to health organizations.31 Studies have shown that appropriate interventions, such as deprescribing fall risk increasing drugs, fall prevention exercises that address strength, balance and gait impairments, and treatment for osteoporosis can reduce fall risk.32 A personalized and accurate tool to predict high risk patients and their specific risk factors will be important for informing personalized clinical decisions and for applying timely interventions. The current standard approach for fall risk screening is based on multiple CDC-endorsed questions that relate to different aspects of fall risk.33 These questions are widely used in primary care practices across the country to address regulatory requirements34 but the accuracy of these screening questions is poor and they are too broad to inform preventive interventions. We found that a machine learning fall injury prediction method using longitudinal EHR-derived data was superior to the usual screening approach as the machine learning approach can accurately predict risk with superior sensitivity while reducing time needed in a clinical encounter and facilitating personalized fall prevention interventions.
While we identified that machine leaning-based screening can be more accurate than manual questionnaires, we also conducted a comparative analysis to assess the performance of manual screening questionnaires by comparing the predictive accuracy of the full version (includes all three fall risk screening questions) and shortened versions (includes two questions or one question). We found that a shortened fall risk screening questionnaire identifies the risk of fall injury with similar accuracy as the commonly used23 three-question screener. Shortening screening questionnaires without compromising accuracy presents an opportunity to alleviate documentation burden and enhance the quality of patient care. For example, during a Medicare Annual Wellness Visit, multiple screenings are required (summarized in Supplementary Figure 3). Reducing time spent completing screening questionnaires allows providers more time during the visit to address patient needs and concerns, thus improving provider-patient communication and care.35 Specifically, due to the high similarity among the three questions, a two-question screener could be used without compromising accuracy. Furthermore, question 1 (i.e. falling twice or more in the past year) could be used as a one-question screener with comparable performance. However, we found that the sensitivity of all of the fall risk screening questions is poor, ranging from .12–.39 (Table 1) versus .62 –.70 for EHR-based models (Table 2), suggesting that replacing the manual fall risk screening questionnaire with the machine learning model that can be linked to preventative interventions may be a more accurate, efficient, and effective solution.
We explored a data-driven approach using machine learning to improve clinical accuracy and efficiency of the fall risk prediction process. We used EHR variables as predictors (listed in Supplementary Table 2) extracted from a multi-site study cohort to develop prediction models for fall injury in a primary care population. A novel feature is that temporal criteria are defined using specific time windows for inputs and outcomes and included in the model design to reflect their sequential order, e.g., predictors must precede outcomes. By consulting clinical experts, we defined three levels of time constraints to select input features: 1) non-time sensitive variables, e.g. race and gender; 2) chronic features within a 3-year duration before screening time (index time), e.g., diagnosis of comorbidities; 3) time sensitive features within a 1-year duration before screening time, e.g., records of hospitalization (Figure 2).
Our machine learning model achieved decent performance (AUC=0.76 and 0.74, sensitivity .62–.70) for both short-term and long-term prediction models, suggesting the promising potential of developing and applying predictive models for fall injury risk screening using EHR data. We also identified multiple important EHR variables which could also be used to facilitate fall injury prevention decisions, thus making the resulting clinical decision support actionable.
Our results suggest that the fall injury screening questions have limitations in real-world practice. Considering that these questions are widely used in primary care practices and consuming providers’ and patients’ valuable time, significant adjustments of these questions, including question design, number of questions needed and best time for these questions to be asked, as well as the most beneficial patient populations, should be further studied. ln light of the improved performance we obtained in EHR-based models, the use of existing EHR variables and machine learning algorithms, could be an efficient and effective strategy for future fall risk and other screening in large patient populations.
Clinical actionability and generalizability, (e.g., “making machine learning matter to clinicians”36), are increasingly important topics for applying machine learning in health care.37,38 Our team included several elements to facilitate translation into clinical practice: 1) Our model uses standard EHR features as input variables, making it easy to implement in diverse EHR systems across organizations; 2) All variables are routinely captured during typical encounters, requiring no new data acquisition or additional documentation to initialize model use; 3) We also collaborated with clinicians and informaticians to develop a fall injury prevention CDS implementation prototype to link CDS output with guideline-based intervention recommendations to help providers efficiently initiate personalized fall prevention recommendations during a visit; As illustrated in Figure 3, during the primary care visit, the machine learning model can automatically identify high-risk patients using existing EHR records and provide corresponding alerts to providers during the clinical encounter. The CDS output will also include patient-level intervention recommendations based on patient-specific actionable features (e.g., each patient’s fall risk factors). Providers can then make timely decisions based on this information. Informed by the CDS, providers can recommend evidence-based preventative interventions, including treatment of comorbidities, medication adjustment and/or exercise recommendations, to reduce a patient’s actionable risk factors identified from our algorithm. These interventions are widely recommended by the CDC’s STEADI toolkit2,8,39 and other national fall prevention guidelines but currently are not well integrated into clinical practice;40 4) Based on our design, the fall injury prevention CDS will run in the background of the EHR system and will not require providers to alter their EHR workflow. The CDS can be applied for both individual and sub-population levels for risk screening. By outputting fall risk score during primary care encounters, this algorithm can be used on patient-level risk screening for individual providers. In addition, using the EHR information stored in existing large-scale EHR databases, our algorithm can also be used to provide a summative risk index for selected patient cohorts on the hospital or system levels, which can be used for population health purposes. Lastly, we expect the data science design pipeline and experience gained in this study can help other researchers to better translate their informatics tools to real-world clinical practice.
Limitations
Our study has several limitations. First, we only evaluated three fall risk screening questions. These questions were recognized by the CDC33 and different versions and combinations are commonly used in primary care practices. Thus, we believe these three questions are a good representation of the fall injury screening process commonly used in primary care practices across the country. Second, we evaluated one type of screener from primary care practices. There are many other types of screening tools which need to be evaluated (e.g. depression risk, substance abuse risk, deep vein thrombosis risk). As discussed above, our analytic pipeline could be used to evaluate different screeners due to the structural similarity of these tools. Third, we applied temporal design for EHR model development for both input and outcomes. We used these time components to reflect a real-world clinical time flow. To optimize these time window definitions, results from prospective evaluation is needed. We have obtained approval from MGB Predictive Analytics Committee to test our model in the real-time clinical environment as part of an existing NIH-funded fall prevention research study (5R61AG068926–02).
Conclusion
This work provides evidence that the current method of questionnaire-based fall risk screening in primary care is suboptimal and that a machine learning risk screening tool using EHR data showed improved performance in predicting fall injury events. Specifically, machine learning screening using EHR data had superior sensitivity to detect patients at risk of future fall injury events compared to the standard questionnaire-based models. We also utilized standard EHR variables to improve model generalizability and demonstrated how CDS could be used to link patient specific risk factors to evidence-based fall injury prevention interventions to facilitate real-world implementation. Next steps include building on these findings to help primary care practices eliminate the unnecessary elements of manual screeners (fall risk and other commonly used screening tools) and to evaluate our model in other time-constrained settings including inpatient and emergency departments. Furthermore, we will optimize fall prevention management in clinical practice by testing the effect of personalized interventions guided by machine learning algorithm outputs.
Supplementary Material
Supplementary Table 1. Pair-wise similarities (Jaccard similarity coefficient) among 3 screening questions
Supplementary Table 2. Input variables for EHR models
Supplementary Table 3. Performance of EHR-questionnaire integrated models
Supplementary Figure 1. ROC curve of EHR models
Supplementary Figure 2. ROC curve of EHR-questionnaire integrated models
Supplementary Figure 3. Summary of multi-component tasks for a health professional during Medicare annual wellness visit
Key Points.
Multiple limitations of current fall risk screening questionnaires hinder early identification of risk and prevention of fall injury.
In this case-control study, we developed and evaluated prediction models for fall-related injuries in a primary care population.
Our results suggest that the current method of questionnaire-based fall risk screening in primary care settings is suboptimal and a data-driven fall injury prediction method using longitudinal EHR-derived features should be used.
Why does this paper matter?
Primary care clinics largely rely on screening questionnaires to identify people at risk of falls. This work provides evidence that the current method of questionnaire-based fall risk screening in primary care is suboptimal and that a machine learning risk screening tool using EHR data showed improved performance in predicting fall injury events.
Acknowledgements
Founding sources
This study was supported by National Institute on Aging (NIA) 5R61AG068926-02 (Patricia C. Dykes).
Footnotes
Conflicts of interest
The authors have no conflicts of interest to declare.
Ethics approval
This project was reviewed and approved by the Mass General Brigham Human Subjects Committee
Data sharing statement
The datasets generated during the current study are not publicly available due to hospital IRB regulation and patient privacy.
References
- 1.Moreland B, Kakara R & Henry A Trends in Nonfatal Falls and Fall-Related Injuries Among Adults Aged >/=65 Years - United States, 2012–2018. MMWR Morb Mortal Wkly Rep 69, 875–881, doi: 10.15585/mmwr.mm6927a5 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Panel on Prevention of Falls in Older Persons, A. G. S. & British Geriatrics, S. Summary of the Updated American Geriatrics Society/British Geriatrics Society clinical practice guideline for prevention of falls in older persons. J Am Geriatr Soc 59, 148–157, doi: 10.1111/j.1532-5415.2010.03234.x (2011). [DOI] [PubMed] [Google Scholar]
- 3.Kruschke C & Butcher HK Evidence-Based Practice Guideline: Fall Prevention for Older Adults. J Gerontol Nurs 43, 15–21, doi: 10.3928/00989134-20171016-01 (2017). [DOI] [PubMed] [Google Scholar]
- 4.Hartholt KA, Lee R, Burns ER & van Beeck EF Mortality From Falls Among US Adults Aged 75 Years or Older, 2000–2016. JAMA 321, 2131–2133, doi: 10.1001/jama.2019.4185 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Tinetti ME, Richman D & Powell L Falls efficacy as a measure of fear of falling. J Gerontol 45, P239–243, doi: 10.1093/geronj/45.6.p239 (1990). [DOI] [PubMed] [Google Scholar]
- 6.Scheffer AC, Schuurmans MJ, van Dijk N, van der Hooft T & de Rooij SE Fear of falling: measurement strategy, prevalence, risk factors and consequences among older persons. Age Ageing 37, 19–24, doi: 10.1093/ageing/afm169 (2008). [DOI] [PubMed] [Google Scholar]
- 7.Stevens JA, Sleet DA & Rubenstein LZ The Influence of Older Adults’ Beliefs and Attitudes on Adopting Fall Prevention Behaviors. Am J Lifestyle Med 12, 324–330, doi: 10.1177/1559827616687263 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Force, U. S. P. S. T. et al. Interventions to Prevent Falls in Community-Dwelling Older Adults: US Preventive Services Task Force Recommendation Statement. JAMA 319, 1696–1704, doi: 10.1001/jama.2018.3097 (2018). [DOI] [PubMed] [Google Scholar]
- 9.Pagano L, McKeough Z, Wootton S, Zwar N & Dennis S Accuracy of the COPD diagnostic questionnaire as a screening tool in primary care. BMC Prim Care 23, 78, doi: 10.1186/s12875-022-01685-z (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Bowling A Mode of questionnaire administration can have serious effects on data quality. J Public Health (Oxf) 27, 281–291, doi: 10.1093/pubmed/fdi031 (2005). [DOI] [PubMed] [Google Scholar]
- 11.Huffstetler AN, Epling J & Krist AH The Need for Electronic Health Records to Support Delivery of Behavioral Health Preventive Services. JAMA 328, 707–708, doi: 10.1001/jama.2022.13391 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 12.Rotenstein LS et al. Association Between Electronic Health Record Time and Quality of Care Metrics in Primary Care. JAMA Netw Open 5, e2237086, doi: 10.1001/jamanetworkopen.2022.37086 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Micek MA, Rule A, Baltus JJ & Arndt B Association of Primary Care Physicians’ Ambulatory Full-time Equivalencies With Time in the Electronic Health Record. JAMA Netw Open 6, e2320032, doi: 10.1001/jamanetworkopen.2023.20032 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Song W et al. Predicting pressure injury using nursing assessment phenotypes and machine learning methods. J Am Med Inform Assoc 28, 759–765, doi: 10.1093/jamia/ocaa336 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Song W et al. Predicting Hospitalization of COVID-19 Positive Patients Using Clinician-guided Machine Learning Methods. J Am Med Inform Assoc, doi: 10.1093/jamia/ocac083 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Coombs L et al. A machine learning framework supporting prospective clinical decisions applied to risk prediction in oncology. NPJ Digit Med 5, 117, doi: 10.1038/s41746-022-00660-3 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Desai RJ, Wang SV, Vaduganathan M, Evers T & Schneeweiss S Comparison of Machine Learning Methods With Traditional Models for Use of Administrative Claims With Electronic Medical Records to Predict Heart Failure Outcomes. JAMA Netw Open 3, e1918962, doi: 10.1001/jamanetworkopen.2019.18962 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Ye C et al. Identification of elders at higher risk for fall with statewide electronic health records and a machine learning algorithm. Int J Med Inform 137, 104105, doi: 10.1016/j.ijmedinf.2020.104105 (2020). [DOI] [PubMed] [Google Scholar]
- 19.Oshiro CES et al. Fall Ascertainment and Development of a Risk Prediction Model Using Electronic Medical Records. J Am Geriatr Soc 67, 1417–1422, doi: 10.1111/jgs.15872 (2019). [DOI] [PubMed] [Google Scholar]
- 20.Lindberg DS et al. Identification of important factors in an inpatient fall risk prediction model to improve the quality of care using EHR and electronic administrative data: A machine-learning approach. Int J Med Inform 143, 104272, doi: 10.1016/j.ijmedinf.2020.104272 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Park J et al. Deep learning on time series laboratory test results from electronic health records for early detection of pancreatic cancer. J Biomed Inform 131, 104095, doi: 10.1016/j.jbi.2022.104095 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 22.Min L et al. Measurement of Fall Injury With Health Care System Data and Assessment of Inclusiveness and Validity of Measurement Models. JAMA Netw Open 2, e199679, doi: 10.1001/jamanetworkopen.2019.9679 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 23.Phelan EA, Mahoney JE, Voit JC & Stevens JA Assessment and management of fall risk in primary care settings. Med Clin North Am 99, 281–293, doi: 10.1016/j.mcna.2014.11.004 (2015). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 24.Bhasin S et al. Strategies to Reduce Injuries and Develop Confidence in Elders (STRIDE): A Cluster-Randomized Pragmatic Trial of a Multifactorial Fall Injury Prevention Strategy: Design and Methods. J Gerontol A Biol Sci Med Sci 73, 1053–1061, doi: 10.1093/gerona/glx190 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Gill TM et al. Screening, Recruitment, and Baseline Characteristics for the Strategies to Reduce Injuries and Develop Confidence in Elders (STRIDE) Study. J Gerontol A Biol Sci Med Sci 73, 1495–1501, doi: 10.1093/gerona/gly076 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Salvatore S et al. Beware the Jaccard: the choice of similarity measure is important and non-trivial in genomic colocalisation analysis. Brief Bioinform 21, 1523–1530, doi: 10.1093/bib/bbz083 (2020). [DOI] [PubMed] [Google Scholar]
- 27.Dautzenberg L et al. Interventions for preventing falls and fall-related fractures in community-dwelling older adults: A systematic review and network meta-analysis. J Am Geriatr Soc 69, 2973–2984, doi: 10.1111/jgs.17375 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Montero-Odasso MM et al. Evaluation of Clinical Practice Guidelines on Fall Prevention and Management for Older Adults: A Systematic Review. JAMA Netw Open 4, e2138911, doi: 10.1001/jamanetworkopen.2021.38911 (2021). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Kenis C et al. Incidence of falls and fall-related injuries and their predictive factors in frail older persons with cancer: a multicenter study. BMC Geriatr 22, 877, doi: 10.1186/s12877-022-03574-7 (2022). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 30.Burns E & Kakara R Deaths from Falls Among Persons Aged >/=65 Years - United States, 2007–2016. MMWR Morb Mortal Wkly Rep 67, 509–514, doi: 10.15585/mmwr.mm6718a1 (2018). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 31.Hoffman GJ, Hays RD, Shapiro MF, Wallace SP & Ettner SL The Costs of Fall-Related Injuries among Older Adults: Annual Per-Faller, Service Component, and Patient Out-of-Pocket Costs. Health Serv Res 52, 1794–1816, doi: 10.1111/1475-6773.12554 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 32.Reuben DB et al. The Strategies to Reduce Injuries and Develop Confidence in Elders Intervention: Falls Risk Factor Assessment and Management, Patient Engagement, and Nurse Co-management. J Am Geriatr Soc 65, 2733–2739, doi: 10.1111/jgs.15121 (2017). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 33.Nithman RW & Vincenzo JL How steady is the STEADI? Inferential analysis of the CDC fall risk toolkit. Arch Gerontol Geriatr 83, 185–194, doi: 10.1016/j.archger.2019.02.018 (2019). [DOI] [PubMed] [Google Scholar]
- 34.Commission, t. J. Fall Reduction Program - Inclusion of All Populations, <https://www.jointcommission.org/standards/standard-faqs/nursing-care-center/national-patient-safety-goals-npsg/000001602/> (2021).
- 35.Abbo ED, Zhang Q, Zelder M & Huang ES The increasing number of clinical items addressed during the time of adult primary care visits. J Gen Intern Med 23, 2058–2065, doi: 10.1007/s11606-008-0805-8 (2008). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 36.Ehrmann DE, Joshi S, Goodfellow SD, Mazwi ML & Eytan D Making machine learning matter to clinicians: model actionability in medical decision-making. NPJ Digit Med 6, 7, doi: 10.1038/s41746-023-00753-7 (2023). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 37.Shah P et al. Artificial intelligence and machine learning in clinical development: a translational perspective. NPJ Digit Med 2, 69, doi: 10.1038/s41746-019-0148-3 (2019). [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Sauer CM et al. Leveraging electronic health records for data science: common pitfalls and how to avoid them. Lancet Digit Health 4, e893–e898, doi: 10.1016/S2589-7500(22)00154-6 (2022). [DOI] [PubMed] [Google Scholar]
- 39.Ganz DA & Latham NK Prevention of Falls in Community-Dwelling Older Adults. N Engl J Med 382, 734–743, doi: 10.1056/NEJMcp1903252 (2020). [DOI] [PubMed] [Google Scholar]
- 40.Dykes PC et al. Evaluation of a Patient-Centered Fall-Prevention Tool Kit to Reduce Falls and Injuries: A Nonrandomized Controlled Trial. JAMA Netw Open 3, e2025889, doi: 10.1001/jamanetworkopen.2020.25889 (2020). [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Supplementary Table 1. Pair-wise similarities (Jaccard similarity coefficient) among 3 screening questions
Supplementary Table 2. Input variables for EHR models
Supplementary Table 3. Performance of EHR-questionnaire integrated models
Supplementary Figure 1. ROC curve of EHR models
Supplementary Figure 2. ROC curve of EHR-questionnaire integrated models
Supplementary Figure 3. Summary of multi-component tasks for a health professional during Medicare annual wellness visit
Data Availability Statement
The datasets generated during the current study are not publicly available due to hospital IRB regulation and patient privacy.