Skip to main content
Annals of Family Medicine logoLink to Annals of Family Medicine
. 2025 Jul-Aug;23(4):294–301. doi: 10.1370/afm.240316

Predicting Missed Appointments in Primary Care: A Personalized Machine Learning Approach

Wen-Jan Tuan 1,, Yifang Yan 2, Bilal Abou Al Ardat 1, Todd Felix 1, Qiushi Chen 2
PMCID: PMC12306982  PMID: 40721333

Abstract

PURPOSE

Factors influencing missed appointments are complex and difficult to anticipate and intervene against. To optimize appointment adherence, we aimed to use personalized machine learning and big data analytics to predict the risk of and contributing factors for no-shows and late cancellations in primary care practices.

METHODS

We conducted a retrospective longitudinal study leveraging geolinked clinical, care utilization, socioeconomic, and climate data from 15 family medicine clinics at a regional academic health center in Pennsylvania from January 2019 to June 2023. We developed multiclass machine learning models using gradient boost, random forest, neural network, and logistic regression to predict appointment outcomes, followed by feature importance analysis to identify contributing factors for no-shows or late cancellations at the population and patient levels. We performed stratified analysis to evaluate the prediction performance by sex and race/ethnicity to ensure the fairness of the final model among sensitive features.

RESULTS

The analysis consisted of 109,328 patients and 1,118,236 appointments, including 77,322 (6.9%) no-shows and 75,545 (6.8%) late cancellations. The gradient boost model achieved the best performance with an area under the receiver operating characteristic curve of 0.852 for predicting no-shows and 0.921 for late cancellations. No bias against patient characteristics was detected. Schedule lead time was identified as the most important predictor of missed appointments.

CONCLUSIONS

Missed appointments remain a challenge for primary care. This study provided a practical and robust framework to predict missed appointments, laying the foundation for developing personalized strategies to improve patients’ adherence to primary care appointments.

Key words: no-show, late cancellation, continuity of care, machine learning, health disparity

INTRODUCTION

Continuity of care plays a vital role in primary care practice, to achieve better health outcomes and minimize overall medical costs.1-3 The longitudinal nature of continuity of care enables clinicians to optimize treatment strategies via a comprehensive view of the patient’s medical history and socioeconomic contexts that affect the patient’s ongoing illness and future health.4 Failure to attend primary care visits could disrupt essential care for patients and also interrupts clinical workflows and strains health care resources.2,5,6 Delayed or missed appointments for essential care can lead to serious health consequences and increased disparities among populations already experiencing health inequities.6-8 Despite major efforts to maintain the care continuum by building patient-clinician relationships and offering more access options, many health systems continue to struggle with patients missing their appointments.

The health services literature often identifies missed appointments when patients fail to show up or cancel within 24 hours before their appointments,5,9 though the nature of and contributing factors for no-shows and late cancellations could be different and require separate measurements.10,11 Whereas recent studies have provided insights into the demographic and clinical-related underpinnings of missed appointments,9,12 the effects of the care continuum and geosocial characteristics on missed appointments are not well understood.11 In addition, research on missed appointments has mainly relied on conventional statistical analyses of information available in electronic health records (EHRs) or patient surveys. There is limited knowledge regarding how to leverage machine learning (ML) to optimize previsit planning and appointment adherence in the primary care setting.

The objective of this study was to apply ML modeling using personal, health care utilization, geosocial, and climate data to assess the risk of no-shows and late cancellations among patients managed by primary care practices. This study could provide valuable insights into the barriers underlying missed appointments, enabling care teams to design personalized interventions that improve patient appointment adherence.

METHODS

Study Design and Setting

We applied a retrospective longitudinal design using integrated clinical, geosocial, and climate data. We retrieved patients’ clinical characteristics and appointment information from the EHR database of a regional academic medical center in southcentral Pennsylvania, which were then were geocoded and linked to corresponding US Census Bureau statistics and national weather reporting databases. Our analysis included all primary care visits and consultations scheduled from January 2019 to June 2023 at 15 family medicine clinics of the medical center. The outcome of each appointment was extracted from the scheduling table of the EHR database and categorized as no-show, late cancellation (cancelled within 24 hours before appointment), or completed visit. Persons deceased or living outside of Pennsylvania were excluded from the study. This research was approved by the academic medical center’s institutional review board.

Risk Factors for Missed Appointments

Demographic and Clinical Characteristics

Patient characteristics included sex, age at visit, race/ethnicity, whether English was the preferred speaking language, insurance type, and comorbid conditions (more details are listed in Supplemental Table 1). Race/ethnicity was classified into the following 5 subgroups: Asian, Hispanic, non-Hispanic Black, non-Hispanic White, and other. Insurance types consisted of commercial, Medicare, Medicaid, and uninsured. The complexity of each patient’s health condition was measured by the number of health conditions present according to the Elixhauser comorbidity measure (29 medical, psychiatric, and lifestyle-related indicators),13,14 which is considered a reliable predictor of health care needs.15

Health Care Utilization

Health care utilization characteristics included scheduled lead time (number of days from a patient’s appointment request to the appointment date), prior appointment history, mode of visit (in-person, telemedicine), clinician type (physician, physician assistant/nurse practitioner), clinician’s years of practice, primary care physician (PCP) relationship, and continuity of care indices. Prior appointment history was measured by prior completed visit, no-show, and late cancellation rates, representing the ratio of the number of specific outcome events to the number of all prior appointments within 3 years. In addition, 3 continuity indices, including the usual provider of care index, the continuity of care index, and the sequential continuity of care index, were used to assess the density, dispersion, and sequential nature of care continuity within a 3-year period (more details are listed in Supplemental Table 1).16

Geosocial and Environmental Contexts

Each patient’s home address was geocoded and matched to a corresponding census block group. The distance from each patient’s home to their primary care clinic was estimated in miles using the Euclidean method. Each patient’s rural status was determined by linking their address to the Census Bureau’s urban and rural classification mapping file.17 Area deprivation index was extracted at the census block–group level to represent the composite socioeconomic status (eg, poverty, education, and housing quality) of the patient in the neighborhood context.18 Two additional Census statistics, the percentage of people without a vehicle and people graduated from high school at the block-group level, were obtained as proxies for transportation access and education. We extracted local climate information for each clinic, including daily average temperature, rain precipitation (inches), and snowfall (inches), from the national weather database.19

Machine Learning Modeling and Evaluation

Model Selection

We used multiclass ML modeling approaches, including gradient boost, random forest, neural network, and least absolute shrinkage and selection operator logistic regression, to predict appointment outcomes (no-show, late cancellation, and completion). We randomly categorized the study sample at the patient level into a training data set (80% of the total samples) and a testing data set (20%). Prediction performance on the testing set was evaluated using the area under the receiver operating characteristic curve (AUROC) and F1 score for each class (vs other classes). The macro-average AUROC and macro-average F1 score were also calculated as the average of the AUROC and F1 score for all classes. We further calculated the misclassification rate of no-shows and late cancellations, respectively, to denote the percentage of incorrect predictions. The 95% CI of each metric was calculated to provide the accuracy of the performance measure. Details on model development are provided in Supplemental Appendix 1.

Feature Importance

The Shapley Additive Explanations (SHAP)20 value was calculated on the basis of game theory to assess each factor’s contribution to the predicted output at the patient level.20,21 Factors positively influencing the predicted outcome generally have positive SHAP values for that outcome and vice versa. The global importance of each factor was also calculated by averaging absolute SHAP values across all individuals to indicate the overall importance of each factor for predicted appointment outcomes. In addition, SHAP values captured the heterogeneous effects of each factor, enabling a personalized interpretation of predictive factors and outcomes.

Fairness Check

To assess for potential bias of the ML models associated with sensitive features, we conducted stratified analysis by evaluating the prediction performance for each subpopulation (sex, race, ethnicity) to ensure the fairness of the final model. Feature importance analysis was also reassessed by each subgroup.

RESULTS

Study Cohort Characteristics

The study included 109,328 patients and 1,118,236 appointments, including 77,322 (6.9%) no-shows and 75,545 (6.8%) late cancellations (Table 1). Overall, patients who missed appointments tended to be female, younger, under/uninsured, less fluent in English, and in ethnic minority groups. They also experienced longer lead times, greater prior missed appointment rates, and more socioeconomic challenges. Clinicians of missed appointments tended to have fewer years of practice. The percentages of no-shows, late cancellations, and completed visits also varied by visit mode, clinician type, and whether the visit was with the patient’s PCP.

Table 1.

Characteristics of Study Cohort by Appointment Outcome

Characteristic Total Appointment outcome P value
No-show Late cancellation Completion
Total, No. (%) 1,118,236 (100.0)     77,322 (100.0)     75,545 (100.0) 965,369 (100.0)
Female, No. (%)   668,448 (59.8)     44,661 (57.8)     47,705 (63.1) 576,082 (59.7) <.001
Age at visit, y, mean (SD)               46.2 (23.7)               36.1 (21.5)               42.2 (22.6)               47.3 (23.7) <.001
Race/ethnicity, No. (%) <.001
    Asian     90,352 (8.1)       3,391 (4.4)       2,450 (3.2)     33,986 (3.5)
    Hispanic     39,827 (3.6)       7,370 (9.5)       5,126 (6.8)     45,106 (4.7)
    Non-Hispanic Black     57,602 (5.2)     13,035 (16.9)       8,034 (10.6)     69,283 (7.2)
    Non-Hispanic White   816,718 (73.0)     41,720 (54.0)     51,213 (67.8)   723,785 (75.0)
    Other   113,737 (10.2)     11,806 (15.3)       8,722 (11.5)     93,209 (9.7)
English as preferred speaking language, No. (%) 1,084,866 (97.0)     73,243 (94.7)     73,376 (97.1)   938,247 (97.2) <.001
Payer type, No. (%) <.001
    Commercial   546,189 (48.8)     30,692 (39.7)     34,831 (46.1)   480,666 (49.8)
    Medicaid   316,160 (28.3)     29,263 (37.8)     19,942 (26.4)   166,787 (17.3)
    Medicare   215,992 (19.3)     11,096 (14.4)     16,971 (22.5)   288,093 (29.8)
    Uninsured     39,895 (3.6)       6,271 (8.1)       3,801 (5.0)     29,823 (3.1)
No. of comorbidities, mean (SD)                 2.0 (1.8)                 2.0 (1.8)                 2.2 (1.9)                 2.0 (1.8) <.001
Visit physician vs other, No. (%)   615,103 (55.0)     40,343 (52.2)     38,107 (50.4)   536,653 (55.6) <.001
Visit patient’s PCP, No. (%)   484,699 (43.3)     34,101 (44.1)     33,450 (44.3)   417,148 (43.2) <.001
Years of practice, y, mean (SD)               17.9 (12.6)               14.7 (10.2)               15.4 (10.8)               18.3 (12.9) <.001
Telemedicine visit, No. (%)     58,253 (5.2)       1,823 (2.4)       3,872 (5.1)     52,558 (5.4) <.001
Scheduled lead time, d, mean (SD)               46.9 (83.0)               64 (88.8)               57.7 (91.1)               44.7 (81.6) <.001
Prior appointment history, mean (SD)
    Completion                 0.8 (0.3)                 0.7 (0.3)                 0.7 (0.3)                 0.8 (0.3) <.001
    No-show              0 (0.1)                 0.1 (0.2)                 0.1 (0.1)              0 (0.1) <.001
    Late cancellation                 0.1 (0.1)                 0.1 (0.2)                 0.1 (0.2)                 0.1 (0.1) <.001
Continuity of care index, mean (SD)
    Usual provider of care index                 0.4 (0.4)                 0.3 (0.3)                 0.3 (0.3)                 0.4 (0.4) <.001
    Continuity of care index                 0.5 (0.4)                 0.6 (0.4)                 0.6 (0.4)                 0.6 (0.4) <.001
    Sequential continuity of care index                 0.3 (0.3)                 0.3 (0.3)                 0.3 (0.3)                 0.3 (0.3) <.001
Initial visit of patient, No. (%)   106,369 (9.5)       8,249 (10.7)       9,739 (12.9)   124,357 (12.9) <.001
Distance from home to clinic, miles, mean (SD)                 7.8 (11.6)                 8.4 (12.1)                 8.7 (12.3)                 7.7 (11.5) <.001
Living in rural area, No. (%)   216,980 (19.4)     11,042 (14.3)     13,109 (17.4)   192,829 (20.0) <.001
Area deprivation index, mean (SD)               51.8 (17.5)               56.9 (19.1)               53.8 (17.8)               51.3 (17.2) <.001
Percentage of population with no vehicle in neighborhood, mean (SD)                 6.2 (6.9)                 8.0 (8.4)                 6.7 (7.3)                 6.0 (6.7) <.001
Percentage of population graduated with high school in neighborhood, mean (SD)               33.0 (11.1)               33.9 (10.5)               33.7 (10.7)               32.8 (11.1) <.001
Weather on day of appointment, mean (SD)
    Temperature, °F               51.2 (16.5)               51.4 (16.5)               50.3 (16.6)               51.2 (16.5) <.001
    Rainfall, in                 0.1 (0.3)                 0.1 (0.3)                 0.1 (0.4)                 0.1 (0.3) <.001
    Snowfall, in                 0.1 (0.4)               0 (0.4)                 0.1 (0.7)               0 (0.4) <.001

PCP = primary care physician.

Performance of Prediction Model

The gradient boost model showed the best performance compared with the other ML models (Table 2). The model achieved a macro-averaged AUROC of 0.881 (95% CI, 0.881-0.882) for overall model performance, along with an AUROC of 0.852 (95% CI, 0.851-0.852) for predicting no-shows, 0.921 (95% CI, 0.920-0.921) for late cancellations, and 0.872 (95% CI, 0.871-0.872) for completed appointments (Figure 1). The gradient boost model also showed better balance between precision and recall, with the highest F1 scores, as well as low misclassification rates for no-shows and late cancellations (Table 2).

Table 2.

Comparison of Performance in Predicting Missed Appointments by Different Machine Learning Models

Machine learning model
Gradient boost Random forest Neural network Logistic regression with LASSO
AUROC (95% CI)a
    No-show vs other 0.852 (0.851-0.852) 0.826 (0.825-0.826) 0.833 (0.833-0.834) 0.757 (0.756-0.757)
    Late cancellation vs other 0.921 (0.920-0.921) 0.825 (0.824-0.825) 0.732 (0.731-0.733) 0.652 (0.651-0.652)
    Completed visit vs other 0.872 (0.871-0.872) 0.833 (0.833-0.834) 0.796 (0.796-0.797) 0.713 (0.713-0.714)
    Macro-averaged 0.881 (0.881-0.882) 0.828 (0.828-0.828) 0.787 (0.787-0.788) 0.707 (0.707-0.708)
F1 score (95% CI)a,b
    No-show vs other 0.318 (0.317-0.318) 0.264 (0.263-0.264) 0.300 (0.299-0.302) 0.244 (0.243-0.245)
    Late cancellation vs other 0.387 (0.386-0.388) 0.231 (0.230-0.232) 0.178 (0.177-0.18) 0.154 (0.154-0.155)
    Completed visit vs other 0.802 (0.801-0.802) 0.686 (0.685-0.686) 0.786 (0.782-0.79) 0.727 (0.726-0.728)
    Macro-averaged 0.502 (0.502-0.503) 0.393 (0.393-0.394) 0.422 (0.420-0.423) 0.375 (0.375-0.376)
Misclassification rate (95% CI)a,b
    No-show misclassified as completed visit 0.173 (0.172-0.174) 0.106 (0.105-0.106) 0.175 (0.171-0.178) 0.228 (0.227-0.229)
    Late cancellation misclassified as completed visit 0.091 (0.091-0.092) 0.096 (0.095-0.097) 0.310 (0.304-0.315) 0.361 (0.36-0.362)
    No-show misclassified as late cancellation 0.177 (0.176-0.179) 0.230 (0.229-0.232) 0.206 (0.197-0.215) 0.229 (0.227-0.230)
    Late cancellation misclassified as no-show 0.204 (0.202-0.205) 0.361 (0.360-0.363) 0.356 (0.348-0.365) 0.327 (0.326-0.328)

AUROC = area under the receiver operating characteristic curve; LASSO = least absolute shrinkage and selection operator.

a

The 95% CIs for all performance metrics were calculated based on results from 50 independent replications of model training and testing.

b

To calculate F1 scores and misclassification rates, we determined the predicted category (among the 3 appointment outcomes) in 2 steps: (1) calculate the ratio of predicted probability and the population average for each appointment outcome and (2) select the outcome showing the greatest ratio. For example, if an individual had a predicted probability (0.5, 0.4, 0.1), and the population average was (0.1, 0.8, 0.1) for no-show, late cancellation, and completed visit, respectively, we calculated the ratio as (5.0, 0.5, 1.0) for each appointment outcome, respectively. We then classified this patient as a no-show because it exhibited the greatest ratio of 5.0 compared with the others.

Figure 1.

Figure 1.

ROC Curve for Each Outcome vs Others and Macro-Averaged ROC Curve of 3 ROC Curves as the Overall Model Performance

AUROC = area under the ROC curve; ROC = receiver-operating characteristic curve.

Feature Importance

The global importance of the key factor, expressed as average SHAP values, in predicting appointment outcomes is shown in Figure 2. The most influential factors included scheduling lead time, providers’ years of practice, patient age at visit, and whether the visit was with the patient’s PCP, followed by the history of prior appointment outcomes (completed, no-show rates), geosocial contexts (area deprivation index, distance from home to clinic), and environmental factors (temperature). The degree of importance of each factor also varied by target outcome. For example, patient age at visit and prior no-show rate were found to play a strong role in predicting no-shows but were less important for predicting late cancellations.

Figure 2.

Figure 2.

Feature Importance of Top 15 Most Impactful Predictors in Gradient Boost Prediction Model

PCP = primary care physician; SECOC = sequential continuity of care index; UPC = usual provider of care index.

Figure 3 shows the distribution of the SHAP values for the top contributing factors on the no-show risk at the individual patient level. The wide divergence of SHAP values for several factors indicates large heterogeneity in the study population. For example, although lead time played an important role in predicting no-shows, lead time effect on the predicted outcome could vary by person and other characteristics of the person. Feature importance analysis allowed clinicians to evaluate the individualized effect and how the predicted outcomes were driven by each person’s unique features, for which we provide more examples in Supplemental Figure 1 and Supplemental Figure 2.

Figure 3.

Figure 3.

Heterogeneity of SHAP Values on Predicted No-Show Risk for 15 Most Influential Predictors Across Individuals

ADI = area deprivation index; COCI = continuity of care index; SHAP = Shapley Additive Explanations.

Note: Each point corresponds to an individual visit in the data set, with the color denoting the numeric value of the corresponding predictor feature. The horizontal position of each point represents the individual’s SHAP value of the corresponding predictor feature, which illustrates the size and direction of how it affects the no-show risk.

Fairness Check

The fairness of the gradient boost model among sex and racial/ethnic subgroups was validated by stratified analyses. Figure 4 shows that the macro-averaged AUROC was consistently high across sex (0.881-0.882) and race/ethnicity subgroups (0.846-0.882). The AUROC for each target appointment outcome remained >0.8, with modest variation within ±0.25 across all subgroups. All subgroup analyses also showed similar results in other performance measures (Supplemental Table 2) and shared important features among the 10 most influential factors for each subpopulation (Supplemental Figure 3).

Figure 4.

Figure 4.

Stratified Analysis of Prediction Performance Measured by Area Under the ROC Curve Across Subpopulations

ROC = receiver-operating characteristic.

DISCUSSION

Missed appointments have long posed challenges to optimizing the care continuum, clinical workflow, and resource allocation for primary care practices. Our analysis showed a 13.7% missed appointment rate, within the 8% to 24% range reported in other primary care settings.22,23 The differences could be attributed to variations in demographics, practice settings, or health care access. This study developed multiclass ML models using integrated clinical, demographic, geosocial, and climate data to predict appointment outcomes at primary care clinics. Whereas all of the proposed ML models achieved good performance, with an overall AUROC >0.8 in predicting appointment outcomes,24 the gradient boost model outperformed random forest, neural network, and logistic regression models. In addition, a fairness check on the model showed similar model performance after stratifying by sex and ethnic/racial subgroups, indicating that predicted results were not biased against specific patient characteristics.

Feature importance analysis revealed that lead time was the most important predictor of appointment outcome, consistent with prior research.9,25 Our study found large variability in how lead time influences predicted risk. Essentially, longer lead times (>60 days) are associated with a greater risk of missed appointments, whereas the effect of shorter lead times (<30 days) can be complex, owing to potential interactions with factors such as age, race/ethnicity, socioeconomic status, and prior no-show rates. Given the strong effect of lead time, clinics could prioritize shorter wait times for high-risk patients. Given that short lead times had variable effects, ML-based analytics could help clinicians anticipate patient-specific needs, personalize outreach efforts, and proactively facilitate appointment scheduling. For example, patients with no-show history or from disadvantaged backgrounds might benefit from additional outreach such as text reminders or transportation assistance.

Several demographic characteristics, including female sex, younger age, non-English speaking, number of chronic conditions, and distance to the clinic, were also identified as key contributors to appointment outcomes. Essentially, patients’ sex, age, and language fluency played an important role in predicting no-shows, whereas home-to-clinic distance had a strong influence on late cancellations. The findings suggest that health systems should prioritize strategies for decreasing lead time, increasing interpreter availability, and offering transportation assistance for better appointment adherence. Whereas telemedicine was not identified among the most influential predictors of appointment adherence in our analysis, health systems should continue offering it as a convenient and accessible option26,27 to decrease missed appointments and improve access to care.

Among geosocial factors, the socioeconomic deprivation level of the neighborhood where patients lived was found to be the most important predictor of missed appointments. Studies have linked missed appointments to low socioeconomic status, which is often associated with a lack of health care literacy, financial stability, and social support.5,28 Among environmental factors, temperature was shown to have the greatest effect, especially on late cancellations. The influence of these factors on missed appointments has been reported9,29; however, it might not be viable for health systems to tackle these risks directly within the scope of health care operations. Nonetheless, incorporating geosocial and weather data into predictive modeling is beneficial for accounting for seasonality effect, offering valuable insights into patient behavior that can inform clinic operations and patient navigation efforts.

Several care utilization factors were also shown to be important predictors of appointment outcome. The clinician’s years of practice, prior completed visit rates, continuity of care indices, and whether an appointment was the first visit at a clinic were strong indicators of whether future visits would be completed. Previous no-show and late cancellation rates were considered strong predictors of no-shows and late cancellations, respectively. Primary care practices generally view these factors as core measures of continuity of care, closely associated with the quality, cost, and outcome of health care. Our analysis revealed that the history of care continuum could further serve as a powerful predictor of missed appointments, which is intuitive and practical given that this metric is already monitored in day-to-day clinical operations. Health systems could enhance continuity of care via better clinician-patient communication, customized team-based care,30 patient engagement, and shared decision making.31

Unlike earlier studies that relied on a single EHR or administrative database to predict patient no-shows,32-36 the present study augmented clinical information with geosocial/environmental contexts and leveraged ML modeling, particularly its ability to effectively capture complex nonlinear relations within high-dimensional data, accommodating the substantial heterogeneity of the study population. In contrast, traditional regression models emphasize global effects across the entire population, potentially overlooking key insights for individual patients. Integrating personalized predictive models with system-wide initiatives, such as automated reminders, patient navigations, and late cancellation fees, offers opportunities to help care teams create more targeted, effective, and personalized interventions for better appointment adherence.

Study Limitations

Our study has several limitations. First, despite a large sample size, the study population was based on a single academic health care center, potentially limiting the findings’ generalizability. Second, our study period encompassed the COVID-19 pandemic, which might have affected patients’ appointment adherence. Our supplementary analysis (Supplemental Appendix 1) showed that despite changes in appointment adherence before and after the pandemic, the prediction model performed robustly across periods, supporting the framework’s feasibility for future predictions. Third, ML models provide a limited explanation of how the models make predictions. Thus, we applied SHAP analysis at both the population and individual levels to describe the factors’ contributions to the prediction results. Fourth, this analysis was based on a single gradient boost model, which had the best performance internally validated with our data. Future research should explore ensemble learning to optimize prediction accuracy and robustness. Lastly, our analysis predicted the outcomes of each patient’s appointments separately instead of treating them as time-series data. To capture the temporal nature of appointment data, researchers can consider more sophisticated deep learning approaches, such as recurrent neural networks, to assess individual’s dynamic risk of missed appointments, helping health systems to make more informative decisions.

CONCLUSION

Missed appointments remain a challenge for primary care practices, despite great efforts to improve access and continuity of care. This study developed novel multiclass ML models using integrated clinical, geosocial, and environmental data for personalized prediction of no-show, late cancellation, and completion of primary care visit appointments. The gradient boost model showed the best performance, with little bias toward sensitive characteristics (sex and race/ethnicity), resulting in fair and equitable outcomes. Feature importance analysis further identified a core set of contributing predictors of missed appointments, enabling health systems to prioritize efforts in minimizing no-shows and late cancellations. This analytical framework lays a foundation for health systems to assess individual risk of missed appointments and design personalized strategies to help patients adhere to primary care appointments.

Supplementary Material

Tuan_Supp_Apps_Tables_Figures.pdf
Tuan_VA_final.pdf
Tuan_VA_final.pdf (172.8KB, pdf)

Footnotes

References

  • 1.Bodenheimer T, Ghorob A, Willard-Grace R, Grumbach K.. The 10 building blocks of high-performing primary care. Ann Fam Med. 2014; 12(2): 166-171. doi: 10.1370/afm.1616 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Saultz JW, Lochner J.. Interpersonal continuity of care and care outcomes: a critical review. Ann Fam Med. 2005; 3(2): 159-166. doi: 10.1370/afm.285 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Casalino LP, Pesko MF, Ryan AM, et al. Small primary care physician practices have low rates of preventable hospital admissions. Health Aff (Millwood). 2014; 33(9): 1680-1688. doi: 10.1377/hlthaff.2014.0434 [DOI] [PubMed] [Google Scholar]
  • 4.Rosser W, Schultz K.. Promoting continuity of care should be integral to any health care system. CMAJ. 2007; 177(11): 1385-1386. doi: 10.1503/cmaj.071303 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.Kheirkhah P, Feng Q, Travis LM, Tavakoli-Tabasi S, Sharafkhaneh A.. Prevalence, predictors and economic consequences of no-shows. BMC Health Serv Res. 2016; 16: 13. doi: 10.1186/s12913-015-1243-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Hwang AS, Atlas SJ, Cronin P, et al. Appointment “no-shows” are an independent predictor of subsequent quality of care and resource utilization outcomes. J Gen Intern Med. 2015; 30(10): 1426-1433. doi: 10.1007/s11606-015-3252-3 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Nuti LA, Lawley M, Turkcan A, et al. No-shows to primary care appointments: subsequent acute care utilization among diabetic patients. BMC Health Serv Res. 2012; 12: 304. doi: 10.1186/1472-6963-12-304 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Greig EC, Gonzalez-Colaso R, Nwanyanwu K.. Racial, ethnic, and socioeconomic disparities drive appointment no-show in patients with chronic eye disease. J Racial Ethn Health Disparities. 2023; 10(4): 1790-1797. doi: 10.1007/s40615-022-01363-x [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Marbouh D, Khaleel I, Al Shanqiti K, et al. Evaluating the impact of patient no-shows on service quality. Risk Manag Healthc Policy. 2020; 13: 509-517. doi: 10.2147/RMHP.S232114 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Norris JB, Kumar C, Chand S, Moskowitz H, Shade SA, Willis DR.. An empirical investigation into factors affecting patient cancellations and no-shows at outpatient clinics. Decis Support Syst. 2014; 57: 428-443. doi: 10.1016/j.dss.2012.10.048 [DOI] [Google Scholar]
  • 11.Tuan WJ, Weems A, Leong SL.. Personal, health system, and geosocial disparities in appointment nonadherence at family medicine clinics in southcentral Pennsylvania, United States. J Gen Fam Med. 2024; 25(4): 214-223. doi: 10.1002/jgf2.698 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Parsons J, Bryce C, Atherton H.. Which patients miss appointments with general practice and the reasons why: a systematic review. Br J Gen Pract. 2021; 71(707): e406-e412. doi: 10.3399/BJGP.2020.1017 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.Elixhauser A, Steiner C, Harris DR, Coffey RM.. Comorbidity measures for use with administrative data. Med Care. 1998; 36(1): 8-27. doi: 10.1097/00005650-199801000-00004 [DOI] [PubMed] [Google Scholar]
  • 14.Moore BJ, White S, Washington R, Coenen N, Elixhauser A.. Identifying increased risk of readmission and in-hospital mortality using hospital administrative data: the AHRQ Elixhauser Comorbidity Index. Med Care. 2017; 55(7): 698-705. doi: 10.1097/MLR.0000000000000735 [DOI] [PubMed] [Google Scholar]
  • 15.Farley JF, Harley CR, Devine JW.. A comparison of comorbidity measurements to predict healthcare expenditures. Am J Manag Care. 2006; 12(2): 110-119. [PubMed] [Google Scholar]
  • 16.Chan CL, You HJ, Huang HT, Ting HW.. Using an integrated COC index and multilevel measurements to verify the care outcome of patients with multiple chronic conditions. BMC Health Serv Res. 2012; 12: 405. doi: 10.1186/1472-6963-12-405 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Ratcliffe M, Burd C, Holder K, Fields A.. Defining rural at the U.S. Census Bureau. American Community Survey and geography brief. US Census Bureau. Published Dec 2016. Accessed April 25, 2025. https://www2.census.gov/geo/pdfs/reference/ua/Defining_Rural.pdf [Google Scholar]
  • 18.Kind AJH, Jencks S, Brock J, et al. Neighborhood socioeconomic disadvantage and 30-day rehospitalization: a retrospective cohort study. Ann Intern Med. 2014; 161(11): 765-774. doi: 10.7326/M13-2946 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.US National Oceanic and Atmospheric Administration . NOWData - NOAA online weather data. Accessed Aug 23, 2023. https://www.weather.gov/wrh/Climate?wfo=ctp
  • 20.Lundberg SM, Lee SI.. A unified approach to interpreting model predictions. In: Guyon I, Von Luxburg U, Bengio S, et al., eds. Advances in Neural Information Processing Systems 30 (NlPS 2017). Curran Associates, Inc; 2018: 4765-4774. https://papers.nips.cc/paper/7062-a-unified-approach-to-interpreting-model-predictions.pdf [Google Scholar]
  • 21.Lundberg SM, Erion G, Chen H, et al. From local explanations to global understanding with explainable AI for trees. Nature Machine Intelligence. 2020; 2: 56-67. doi: 10.1038/s42256-019-0138-9 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.Goffman RM, Harris SL, May JH, et al. Modeling patient no-show history and predicting future outpatient appointment behavior in the Veterans Health Administration. Mil Med. 2017; 182(5): e1708-e1714. doi: 10.7205/MILMED-D-16-00345 [DOI] [PubMed] [Google Scholar]
  • 23.Shour A, Onitilo AA.. Distance matters: investigating no-shows in a large rural provider network. Clin Med Res. 2023; 21(4): 177-191. doi: 10.3121/cmr.2023.1853 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Muller MP, Tomlinson G, Marrie TJ, et al. Can routine laboratory tests discriminate between severe acute respiratory syndrome and other causes of community-acquired pneumonia? Clin Infect Dis. 2005; 40(8): 1079-1086. doi: 10.1086/428577 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Drewek R, Mirea L, Adelson PD.. Lead time to appointment and no-show rates for new and follow-up patients in an ambulatory clinic. Health Care Manag (Frederick). 2017; 36(1): 4-9. doi: 10.1097/HCM.0000000000000148 [DOI] [PubMed] [Google Scholar]
  • 26.Sumarsono A, Case M, Kassa S, Moran B.. Telehealth as a tool to improve access and reduce no-show rates in a large safety-net population in the USA. J Urban Health. 2023; 100(2): 398-407. doi: 10.1007/s11524-023-00721-2 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Drerup B, Espenschied J, Wiedemer J, Hamilton L.. Reduced no-show rates and sustained patient satisfaction of telehealth during the COVID-19 pandemic. Telemed J E Health. 2021; 27(12): 1409-1415. doi: 10.1089/tmj.2021.0002 [DOI] [PubMed] [Google Scholar]
  • 28.Chapman KA, Machado SS, van der Merwe K, Bryson A, Smith D.. Exploring primary care non-attendance: a study of low-income patients. J Prim Care Community Health. 2022; 13: 21501319221082352. doi: 10.1177/21501319221082352 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Liu D, Shin WY, Sprecher E, et al. Machine learning approaches to predicting no-shows in pediatric medical appointment. NPJ Digit Med. 2022; 5(1): 50. doi: 10.1038/s41746-022-00594-w [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.Ljungholm L, Edin-Liljegren A, Ekstedt M, Klinga C.. What is needed for continuity of care and how can we achieve it? - Perceptions among multiprofessionals on the chronic care trajectory. BMC Health Serv Res. 2022; 22(1): 686. doi: 10.1186/s12913-022-08023-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Montori VM, Ruissen MM, Hargraves IG, Brito JP, Kunneman M.. Shared decision-making as a method of care. BMJ Evid Based Med. 2023; 28(4): 213-217. doi: 10.1136/bmjebm-2022-112068 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Mohammadi I, Wu H, Turkcan A, Toscos T, Doebbeling BN.. Data analytics and modeling for appointment no-show in community health centers. J Prim Care Community Health. 2018; 9: 2150132718811692. doi: 10.1177/2150132718811692 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Milicevic AS, Mitsantisuk K, Tjader A, Vargas DL, Hubert TL, Scott B.. Modeling patient no-show history and predicting future appointment behavior at the Veterans Administration’s outpatient mental health clinics: NIRMO-2. Mil Med. 2020; 185(7-8): e988-e994. doi: 10.1093/milmed/usaa095 [DOI] [PubMed] [Google Scholar]
  • 34.Tarabichi Y, Higginbotham J, Riley N, Kaelber DC, Watts B.. Reducing disparities in no show rates using predictive model-driven live appointment reminders for at-risk patients: a randomized controlled quality improvement initiative. J Gen Intern Med. 2023; 38(13): 2921-2927. doi: 10.1007/s11606-023-08209-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Ahmad Hamdan AF, Abu Bakar A.. Machine learning predictions on outpatient no-show appointments in a Malaysia major tertiary hospital. Malays J Med Sci. 2023; 30(5): 169-180. doi: 10.21315/mjms2023.30.5.14 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Dunstan J, Villena F, Hoyos JP, et al. Predicting no-show appointments in a pediatric hospital in Chile using machine learning. Health Care Manag Sci. 2023; 26(2): 313-329. doi: 10.1007/s10729-022-09626-z [DOI] [PMC free article] [PubMed] [Google Scholar]

Associated Data

This section collects any data citations, data availability statements, or supplementary materials included in this article.

Supplementary Materials

Tuan_Supp_Apps_Tables_Figures.pdf
Tuan_VA_final.pdf
Tuan_VA_final.pdf (172.8KB, pdf)

Articles from Annals of Family Medicine are provided here courtesy of Annals of Family Medicine, Inc.

RESOURCES