Skip to main content
PLOS One logoLink to PLOS One
. 2024 Dec 30;19(12):e0314989. doi: 10.1371/journal.pone.0314989

Ascertaining provider-level implicit bias in electronic health records with rules-based natural language processing: A pilot study in the case of prostate cancer

Ashwin Ramaswamy 1, Michael Hung 1, Joe Pelt 1, Parsa Iranmahboub 1, Lina P Calderon 1, Ian S Scherr 1, Gerald Wang 1, David Green 1, Neal Patel 1, Timothy D McClure 1, Christopher Barbieri 1, Jim C Hu 1, Charlotta Lindvall 2,3, Douglas S Scherr 1,*
Editor: Alvaro Galli4
PMCID: PMC11684669  PMID: 39775249

Abstract

Purpose

Implicit, unconscious biases in medicine are personal attitudes about race, ethnicity, gender, and other characteristics that may lead to discriminatory patterns of care. However, there is no consensus on whether implicit bias represents a true predictor of differential care given an absence of real-world studies. We conducted the first real-world pilot study of provider implicit bias by evaluating treatment parity in prostate cancer using unstructured data–the most common way providers document granular details of the patient encounter.

Methods and findings

Patients ≥18 years with a diagnosis of very-low to favorable intermediate-risk prostate cancer followed by 3 urologic oncologists from 2010 through 2021. The race Implicit Association Test was administered to all providers. Natural language processing screened human annotation using validated regex ontologies evaluated each provider’s care on four prostate cancer quality indicators: (1) active surveillance utilization; (2) molecular biomarker discussion; (3) urinary function evaluation; and (4) sexual function evaluation. The chi-squared test and phi coefficient were utilized to respectively measure the statistical significance and the strength of association between race and four quality indicators. 1,094 patients were included. While Providers A and B demonstrated no preference on the race Implicit Association Test, Provider C showed preference for White patients. Provider C recommended active surveillance (p<0.01, φ = 0.175) and considered biomarkers (p = 0.047, φ = 0.127) more often in White men than expected, suggestive of treatment imparity. Provider A considered biomarkers (p<0.01, φ = 0.179) more often in White men than expected. Provider B demonstrated treatment parity in all evaluated quality indicators (p>0.05).

Conclusions

In this pilot study, providers’ practice patterns were associated with both patient race and implicit racial preferences in prostate cancer. Alerting providers of existing implicit bias may restore parity, however future assessments are needed to validate this concept.

Introduction

Implicit biases in medicine are personal attitudes or beliefs about race, ethnicity, sex, disability, and other characteristics that may unconsciously influence clinical decision making and lead to discriminatory patterns of care. Though rarely harboring overtly explicit prejudices, the majority of providers hold implicit bias against marginalized groups [13]. These stereotyped beliefs, which discriminate against those already vulnerable to their health conditions, are learned over time through cultural exposure to the modern healthcare environment and are notoriously difficult to change [4, 5]. This is especially troubling as there is widespread evidence reaffirming a clear association between minority group status and negative clinical outcomes across medicine [68]. Though rarely harboring overtly explicit prejudices, most providers hold implicit bias against marginalized groups that are notoriously difficult to change [1]. These stereotyped beliefs are especially troubling given widespread evidence associating minority group status with negative clinical outcomes across medicine.

However, subsequent studies have not reached consensus on whether implicit bias represents a true predictor of differential care. While some series demonstrate that provider implicit bias independently influences care delivery [9], others argue that “physicians’ implicit bias does not impact clinical decision making” [10]. Importantly, this unclear relationship is belied by the near-total absence of studies using real-world data [11].

Unstructured clinical narratives, which make up nearly 80% of clinical information in the electronic health record (EHR) [12], represents a unique opportunity to capture implicit bias as it represents the most convenient way providers document granular details of clinical management not well-populated in discrete datasets [13]. Applying recent advances in natural language processing (NLP) to real-world EHRs [14], we explored the relationship between providers’ implicit bias and their respective treatment parity in prostate cancer, a condition associated with guideline-based care [15], validated functional outcomes, and well-documented disparities in care [1618].

Methods

Study design

This was a retrospective, observational pilot study of new patients ≥18 years presenting with an elevated PSA or prostate cancer to all three urologic oncologists (average 17 years in practice, average h-index of 51) affiliated with Weill Cornell Medical Center from 2010 to 2021. All note-level data was extracted from the EHR (Epic®, Verona, WI) including new and follow-up visit notes, as well as telephone encounters, MyChart messages, pathology reports, and radiology reports. Notes were associated with patient age, race, ethnicity, body mass index (BMI), laboratory values, and procedure history. Patient race was categorized as White or Not White as determined by patient-reported designations on clinical intake forms. Patients were excluded if they did not have an associated prostate cancer diagnosis or a missing PSA associated with the first encounter.

Population

We decided to study men with very low-risk or low-risk prostate cancer and men with favorable intermediate-risk prostate cancer, as specified by the National Comprehensive Cancer Network (NCCN) Prostate Cancer Guidelines [19]. These guidelines stratify risk based on histologic grade group (GG), which ranges from 1 (most favorable) to 5 (least favorable), PSA, tumor volume, and other associated measures. These men are considered to be highly appropriate for active surveillance (AS) by most clinicians [20] and is the preferred management for very low-risk or low-risk prostate cancer as recommended by the American Urologic Association, American Society for Radiation Oncology, Society of Urologic Oncology, and NCCN, and is a viable option for men with favorable intermediate-risk prostate cancer with a PSA < 10 or low volume disease (defined as < 50% positive biopsy cores) [19, 21, 22]. For the purposes of this study, we categorized any man with GG1 prostate cancer & a visit-matched PSA < 10 to be low-risk (including very low-risk), and any man with GG1 prostate cancer & a visit-matched PSA > 10 or GG2 prostate cancer with < 50% positive cores to be favorable intermediate-risk.

Outcomes

Four quality indicators reflecting guideline-concordant [19, 21] prostate cancer care–which are uniquely [13] and accurately [23] captured in EHR unstructured clinical narratives–were chosen to evaluate each provider’s clinical decision making in the aforementioned patient cohort: (1) utilization of AS, (2) discussion of prognostic (e.g., Decipher [24]) prostate cancer biomarkers, use of which is associated with improved risk stratification [25], (3) discussion of the patients’ urinary function, and (4) discussion of the patients’ sexual (i.e., erectile) function [21]. This information has historically been inaccessible in the literature, given the laborious process of manual chart review required to extract discrete and usable data elements.

Natural language processing (NLP), here defined as work that computationally represents, transforms, or utilizes text, was employed to iteratively identify provider performance in the four quality indicators of prostate cancer care across each patient’s unstructured clinical narrative documentation. Specifically, we employed NLP-screened human annotation to ascertain provider performance in the four quality indicators for each patient using validated regular expressions (regex) ontologies to display only the relevant outcomes-in-context enabling considerably expedited data abstraction [14]. Ontology development erred towards improved recall to maximize retrieval of relevant positive “hits”, as “misses” would not be reviewed. Final ontologies included various potential abbreviations, word phrasing, and punctuation; furthermore, utilizing regex enabled us to create context-dependent rules, such as excluding those notes in which the text “proceed with prostatectomy” either preceded or succeeded the text “active surveillance The four quality outcomes were measured in all unstructured data from the first visit to 1 year thereafter. Provider-specific dot phrases were excluded to prevent automated text from being considered as implemented care. An in-depth description of regex ontologies, validation, and the software used is in S1 Appendix.

Lastly, the race Implicit Association Test (IAT) was administered to all providers to assess implicit bias for White vs. Black people’s faces. IAT results were administered after data collection was performed. The IAT is the most widely used test to evaluate unconscious preferences across a spectrum of social categories [26]. The IAT is a series of computerized categorization tasks in which participants sort stimuli into opposing categories as quickly and accurately as possible by rapidly pressing “I” and “E” keys; in the race IAT, the participant is tasked with sorting pictures of White or Black people’s faces with words with positive (e.g., smile, diamond) or negative (e.g., grief, crash) connotation presented in a random order. The test is predicated on the assumption that the ease with which an individual sorts concepts together is indicative of implicit bias. The IAT is scored on a 7-point relative preference measure of strong, moderate, slight, preference for either race with no racial preference set in the middle. IAT results were not known to the providers prior to data collection.

Statistical analysis

Baseline patient characteristics were assessed using descriptive patient-level statistics; the Mann-Whitney U test and chi-squared test of independence were used to compare continuous and categorical variables, respectively, at an alpha of 0.05. We calculated a phi coefficient (φ) to measure the strength of association between race and the quality indicators and to enable standardized comparison of preferential treatment of White men between the three providers.

All analyses were anonymized to patient name, medical record number, location data, and date of birth prior to access on March 13, 2023; study authors did not have access to de-anonymized data during nor after analysis was performed. Statistical analysis was performed in R 3.4.1 (R Foundation for Statistical Computing, Vienna, Austria). All analyses adhered to STROBE guidelines and were approved by the Weill Cornell Medicine Institutional Review Board (#22–03024547); the requirement for informed consent was waived as our study met the criteria of a minimal risk study under 45 CFR 46.116(f). All authors declare no financial or non-financial competing interests.

Results

We identified 5,981 (associated with 211,345 data entry points; 119,438 unstructured, 91,907 structured) patients who were diagnosed with elevated PSA or prostate cancer from January 2010 –December 2021. Of the 1,758 patients who had complete data for PSA, race/ethnicity, and prostate biopsy pathology, 1,094 were identified as having very low, low, or favorable intermediate risk prostate cancer per NCCN risk stratification and were included for analysis: of this sample, 67% and 33% of patients were White and non-White, respectively. Additionally, while Providers A and B demonstrated no racial preference on the race IAT, Provider C had a slight preference for White patients.

In aggregate across all providers, (1) 31.0% patients were recommended to initiate or continue AS, (2) molecular biomarkers were considered in 27.2% patients, (3) urinary function was evaluated in 58.1% patients, and (4) erectile function was evaluated in 56.4% patients (Table 1). Older age was significantly associated with providers’ decision to initiate or continue AS (median [IQR], 67 [61–72] vs. 64 [58–69] years, p<0.001), evaluate urinary function (66 [59–70.5] vs. 64 [58–70] years, p = 0.023), and not evaluate erectile function (67 [60–72] vs. 64 [58–69], p<0.001). Molecular biomarker discussion (p<0.001) and urinary function evaluation (p = 0.033) occurred more often in White men than expected, suggestive of treatment imparity. Neither BMI, baseline PSA, nor income were associated with the quality indicators.

Table 1. Quality Indicators for men with very low to favorable intermediate risk prostate cancer.

Quality Indicators Active Surveillance (AS) Molecular Biomarkers Discussion of Urinary Function (UF) Discussion of Erectile Function (EF)
AS–No AS–Yes P Biomarker–No Biomarker–Yes p UF–No UF–Yes p EF–No EF–Yes p
No. 898 401 946 353 544 755 567 732
Age, years (median (IQR)) 64.0 [58.0, 69.0] 67.0 [61.0, 72.0] <0.001* 65.0 [59.0, 70.0] 65.0 [59.0, 70.0] 0.737 64.0 [58.0, 70.0] 66.0 [59.0, 70.5] 0.023* 67.0 [60.0, 72.0] 64.0 [58.0, 69.0] <0.001*
BMI (median (IQR)) 26.6 [24.4, 29.3] 26.1 [24.3, 28.7] 0.086 26.5 [24.4, 29.3] 26.3 [24.4, 28.7] 0.723 26.5 [24.4, 29.2] 26.5 [24.4, 29.1] 0.895 26.4 [24.4, 29.2] 26.5 [24.4, 29.0] 0.962
PSA ng/mL (median (IQR)) 5.1 [3.7, 7.0] 5.2 [3.9, 7.2] 0.411 5.1 [3.7, 7.0] 5.3 [4.0, 7.4] 0.051 5.2 [3.9, 7.5] 5.1 [3.7, 6.9] 0.090 5.1 [3.8, 7.2] 5.2 [3.8, 7.0] 0.907
Race Group 0.067 <0.001* 0.033* 0.150
    White 584 (65.0) 287 (71.6) 605 (64.0) 266 (75.4) 343 (63.1) 528 (69.9) 364 (64.2) 507 (69.3)
    Not White 198 (22.0) 73 (18.2) 221 (23.4) 50 (14.2) 128 (23.5) 143 (18.9) 130 (22.9) 141 (19.3)
    Unknown 116 (12.9) 41 (10.2) 120 (12.7) 37 (10.5) 73 (13.4) 84 (11.1) 73 (12.9) 84 (11.5)
Income Group 0.092 0.213 0.508 0.495
    < 60k 121 (13.8) 41 (10.4) 127 (13.7) 35 (10.2) 66 (12.5) 96 (12.9) 75 (13.6) 87 (12.1)
    60k - 200k 712 (81.4) 341 (86.3) 761 (82.2) 292 (84.9) 435 (82.4) 618 (83.3) 451 (81.6) 602 (84.0)
    >200K 42 (4.8) 13 (3.3) 38 (4.1) 17 (4.9) 27 (5.1) 28 (3.8) 27 (4.9) 28 (3.9)

Numbers are No. (%) unless otherwise noted. SD = standard deviation, IQR = interquartile range

Provider A considered or utilized molecular biomarkers (p = 0.001, φ = 0.179) more often in White men than expected (Table 2 and Fig 1). Provider C recommended initiating or continuing AS (p = 0.006, φ = 0.175) and considered molecular biomarkers (p = 0.046, φ = 0.127) more often in White men than expected. Provider B demonstrated treatment parity in all evaluated quality indicators (p>0.05). All providers demonstrated treatment parity in urinary and erectile function evaluation (p>0.05).

Table 2. Contingency table of prostate cancer quality indicators by race, stratified by provider.

(Race IAT) All Providers Provider A Provider B Provider C
(No preference) (No preference) (Slight White preference)
White Not White p White Not White p White Not White p White Not White p
No. 871 271 - 290 122 - 366 70 - 193 53 -
(1) Active Surveillance 0.063 0.690 0.857 0.006*
    Observed 287 (33.0) 73 (26.9) 101 (34.8) 40 (32.8) 82 (22.4) 15 (21.4) 103 (53.4) 17 (32.1)
    Expected 274.6 (31.5) 85.4 (31.5) 99.3 (34.2) 41.8 (34.2) 81.4 (22.3) 15.6 (22.3) 94.2 (48.8) 25.9 (48.8)
(2) Molecular Biomarkers 0.001* 0.001* 0.333 0.046*
    Observed 266 (30.5) 50 (18.5) 113 (39.0) 25 (20.5) 76 (20.8) 11 (15.7) 76 (39.4) 13 (24.5)
    Expected 241.0 (27.7) 75.0 (27.7) 97.1 (33.5) 40.9 (33.5) 73.0 (20.0) 14.0 (20.0) 69.8 (36.2) 19.2 (36.2)
(3) Urinary Function 0.022* 0.772 0.551 0.127
    Observed 528 (60.6) 143 (52.8) 97 (33.4) 39 (32.0) 259 (70.8) 52 (74.3) 163 (84.5) 40 (75.5)
    Expected 511.8 (58.8) 159.2 (58.8) 95.7 (33.0) 40.3 (33.0) 261.0 (71.3) 49.9 (71.3) 159.3 (82.5) 43.7 (82.5)
(4) Erectile Function 0.073 0.928 0.824 0.610
    Observed 507 (58.2) 141 (52.0) 125 (43.1) 52 (42.6) 246 (67.2) 48 (68.6) 131 (67.8) 34 (64.2)
    Expected 494.2 (56.7) 153.8 (56.7) 124.6 (43.0) 52.4 (43.0) 246.8 (67.4) 47.2 (67.4) 129.5 (67.1) 35.7 (67.1)

Numbers are No. (%) unless otherwise noted. SD = standard deviation, IQR = interquartile range

Fig 1. Degree of association between provider quality indicators (active surveillance utilization [green], molecular biomarker discussion [orange], urinary function evaluation [blue], and sexual function evaluation [pink]) and race among each provider evaluated via the phi coefficient with the asterisks indicating statistical significance at an alpha of 0.05.

Fig 1

Discussion

In this novel study utilizing NLP in the evaluation of unstructured EHRs, we found significant racial variation in the management of men with prostate cancer. In the cumulative delivery of guideline-concordant urologic care, providers considered molecular biomarkers and evaluated urinary function more often in White men than expected; in contrast, providers demonstrated racial parity in both utilizing AS and evaluating erectile function. However, our analysis of granular unstructured data revealed provider-level racial imparity in clinical decision making which both corroborates and belies clinic-level observations. For example, while Provider A & C’s molecular biomarker imparity corroborates observed clinic-level imparity, Provider C’s AS imparity belies our finding that White and non-White men receive care equity in aggregate. Notably, unlike previous studies using clinical vignettes, IAT results imperfectly correlate with provider treatment parity.

Despite the best intentions by providers and health systems to ensure the delivery of equitable care, disparities in healthcare outcomes exist in nearly every disease even after controlling for factors such as access to care, socioeconomic status, or insurance coverage [27]. This determination, made in the Institute of Medicine (IOM)’s Unequal Treatment 2002 Report to Congress, is suggested to be caused in part by “bias, stereotyping, and prejudice on the part of healthcare providers”, although direct evidence supporting this association was considered lacking at the time of the report’s publication [27]. In ensuing decades, researchers continued to call attention to the pervasiveness of unequal care in a broad range of medical disciplines, while concurrently demonstrating that the majority of healthcare providers do harbor some degree of implicit bias [13]. However, the abundant use of hypothetical clinical vignettes and the near-total absence of real-world data in implicit bias studies has yielded only circumstantial–and conflicting–evidence regarding the possible relationship between provider implicit bias and observed disparities in patient outcomes [28, 29].

The challenges of using real-world patient data to evaluate implicit bias is limited by an imperfect methodology by which to do so. For example, the only study utilizing actual health outcomes inferred provider racial preferences in hypertensive treatment intensification solely on the basis of pharmacy refills, a poor proxy of clinician intent [10, 30]. Furthermore, not all disparities between patient populations are necessarily discriminatory. To attribute a difference to a disparity requires a granular understanding of clinical context which can be uniquely characterized by unstructured EHR data [31]. Use of these records enables the reliable interpretation of a provider’s clinical decision-making as tailored to a specific disease profile across an individual patient’s trajectory through a healthcare system [23, 32]. Extracting usable data elements from clinical narratives has traditionally been infeasible given the laborious process of “gold standard” manual chart review. However, as demonstrated in this study, rules-based NLP-assisted human annotation enables the efficient abstraction of EHR outcomes that is broadly validated [3335], widely generalizable to a range of clinical contexts [14, 36], demands minimal investment in computing resources and upfront cost [37], and not limited by non-interrogable “black box” models common in machine learning (ML) algorithms [38]. Compared to ML, rules-based algorithms better suited to evaluating treatment parity using EHR data: they can flexibly be adapted to the interpretation of unformatted “messy” data in small populations in which significant domain knowledge, high precision, and transparency are required. Furthermore, rules-based methodologies as described are foundational in a future of automated, accurate interpretation of healthcare data.

Prostate cancer is a well-suited candidate condition to evaluate the association of racial implicit bias and provider treatment parity using unstructured EHR data. First, there are increasingly apparent racial disparities in prostate cancer care–lower rates of AS [3942], higher rates of treatment regret [43], higher cancer-specific mortality [44]–which are broadly attributable to social, rather than biologic, determinants of health. Second, prostate cancer only affects patients with male anatomy, thereby broadly removing sex as a confounding study variable (the notable exception being intersex or transgender patients). Third, prostate cancer screening is generally considered only within a narrow age range (approximately 50–69 years) with guidelines discouraging routine testing in older men, thereby largely removing age as a confounding study variable [45]. Lastly, the quality of prostate cancer care is measurable by outcomes found sparingly in structured data sources, such as whether a provider assessed a patients’ baseline-disease specific function (e.g., urinary, sexual function) [21]. However, these outcomes are richly populated in unstructured clinical documentation, which effectively make existing structured data sources–either claims or registry–incompletely represent prostate cancer quality [46] and altogether inadequate at explaining observed provider-level deviations from guideline-concordant care [47].

Our study has important implications for clinical practice. First, by utilizing a previous validated methodology using NLP on unstructured EHR data, we conducted the first study of implicit bias to observe providers in their real work environments, without which one cannot understand how implicit bias translates into differential patient care [28]. Our novel approach enables a means to granularly measure healthcare disparities at the provider-level that could be integrated into routine clinical documentation inquiry and hospital reporting frameworks. Furthermore, unlike automated deep-learning models, our outlined process of NLP-screened human annotation is uncomplicated, customizable, interrogable, and–perhaps most importantly–utilizes supervised “human-in-the-loop” determinations. Second, our results naturally suggest clinician-facing, data-driven interventions to improve awareness of implicit bias. Provider awareness of implicit bias has been shown to encourage more egalitarian patterns of care [9]. Similar clinician-facing prompting interventions have been demonstrated to effectively promote behavior change, particularly among minoritized patients [48]. Notably, no current bias-reducing efforts, ranging from mental-debiasing interventions to group-administered trainings, have been demonstrated to reduce implicit bias [31, 49]. The approach outlined by our study may help healthcare systems periodically identify potentially clinically significant provider-level disparities in care, tailor personalized prompting interventions, further evaluate the effect of race-concordant care, as well as measure the success of any corrective efforts to reduce discriminatory disparities.

Our study has several limitations. First, the NLP model was developed at a single urban, tertiary-care, academic institution among three providers and should be viewed as a pilot study; generalizability to more providers and other health systems may require a separate validation process. Second, treatment algorithms are not absolute; shared decision making plays an integral role in prostate cancer management where the decision to pursue AS may be highly personalized. Additionally, there are patient factors that may influence treatment that are not captured by our analysis including a family history of malignancy, anxiety about a cancer diagnosis, performance status, and medical comorbidities. To limit these confounders, our analysis included only men with very low to favorable intermediate risk prostate cancer for whom AS is a recommended option by guidelines. Third, we acknowledge that our NLP method relies on accurate documentation during a patient encounter. Failure to explicitly document the contents of shared decision making or findings such as normal urinary and erectile function despite assessments made during an encounter cannot be interpreted by our NLP method. However, this is a limitation of EHR documentation that exists even with manual chart review. Fourth, the IAT results were administered at one point in time and may not be a predictor of care across the entire time interval of our data. Similarly, we were not able to adjust for year of counseling as provider practices may have changed over time. Fifth, use of racial categories are artificial, social constructs that may poorly reflect the experiences of the categorized people; our choice to compare “White” and “non-White” was predicated on a recommendation to understand race not as a measure of biologic difference, but rather as a proxy for exposure to racism [50], which in the United States has historically been driven by White supremacy [51]. Additionally, there is added challenges in racial categorization based on EHR documentation, which further encouraged us to avoid extensive disaggregation of racial designations [52]. Lastly, while we were able to rule out age, weight, and income as potential confounders, we were unable to account for a range of additional socioeconomic factors that may contribute to treatment imparity, such as disability, education, or religion.

Conclusions

In this pilot study, overall clinical practice patterns were associated with both patient race and implicit provider racial preferences among urologic oncologists in the management of prostate cancer. Further studies involving more quality indicators and providers may be warranted. Alerting providers of existing implicit bias can restore parity in clinical care but future assessments are needed to validate this concept.

Supporting information

S1 Appendix

(DOCX)

pone.0314989.s001.docx (2.3MB, docx)

Data Availability

The EHR datasets cannot be shared for the privacy of the individuals whose data were used in this study given it is entirely populated with protected health information which can be used to identify specific individuals, even after our efforts to anonymize patient name, medical record number, and date of birth. Human research participant data includes: names, location data, date of birth, examination dates, sex, ethnicity, and race. Data sets may be available (with limitations) on request through Weill Cornell Medicine Human Research Protections at 575 Lexington Avenue, New York, NY 10022, Phone: (646)-962-8200.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1.Haider AH, Sexton J, Sriram N, Cooper LA, Efron DT, Swoboda S, et al. Association of unconscious race and social class bias with vignette-based clinical assessments by medical students. JAMA. 2011. Sep 7;306(9):942–51. doi: 10.1001/jama.2011.1248 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Haider AH, Schneider EB, Sriram N, Scott VK, Swoboda SM, Zogg CK, et al. Unconscious Race and Class Biases among Registered Nurses: Vignette-Based Study Using Implicit Association Testing. J Am Coll Surg. 2015. Jun;220(6):1077–1086.e3. doi: 10.1016/j.jamcollsurg.2015.01.065 [DOI] [PubMed] [Google Scholar]
  • 3.Haider AH, Schneider EB, Sriram N, Dossick DS, Scott VK, Swoboda SM, et al. Unconscious race and social class bias among acute care surgical clinicians and clinical treatment decisions. JAMA Surg. 2015. May;150(5):457–64. doi: 10.1001/jamasurg.2014.4038 [DOI] [PubMed] [Google Scholar]
  • 4.Lai CK, Skinner AL, Cooley E, Murrar S, Brauer M, Devos T, et al. Reducing implicit racial preferences: II. Intervention effectiveness across time. J Exp Psychol Gen. 2016. Aug;145(8):1001–16. doi: 10.1037/xge0000179 [DOI] [PubMed] [Google Scholar]
  • 5.van Ryn M, Hardeman R, Phelan SM, Burgess DJ, Dovidio JF, Herrin J, et al. Medical School Experiences Associated with Change in Implicit Racial Bias Among 3547 Students: A Medical Student CHANGES Study Report. J Gen Intern Med. 2015. Dec;30(12):1748–56. doi: 10.1007/s11606-015-3447-7 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6.Barnett ML, Meara E, Lewinson T, Hardy B, Chyn D, Onsando M, et al. Racial Inequality in Receipt of Medications for Opioid Use Disorder. New England Journal of Medicine. 2023. May 11;388(19):1779–89. doi: 10.1056/NEJMsa2212412 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7.Chan AK, McGovern RA, Brown LT, Sheehy JP, Zacharia BE, Mikell CB, et al. Disparities in Access to Deep Brain Stimulation Surgery for Parkinson Disease: Interaction Between African American Race and Medicaid Use. JAMA Neurology. 2014. Mar 1;71(3):291–9. doi: 10.1001/jamaneurol.2013.5798 [DOI] [PubMed] [Google Scholar]
  • 8.McGrath CL, Bettinger B, Stimpson M, Bell SL, Coker TR, Kronman MP, et al. Identifying and Mitigating Disparities in Central Line–Associated Bloodstream Infections in Minoritized Racial, Ethnic, and Language Groups. JAMA Pediatrics. 2023. Jul 1;177(7):700–9. doi: 10.1001/jamapediatrics.2023.1379 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Green AR, Carney DR, Pallin DJ, Ngo LH, Raymond KL, Iezzoni LI, et al. Implicit Bias among Physicians and its Prediction of Thrombolysis Decisions for Black and White Patients. J Gen Intern Med. 2007. Sep;22(9):1231–8. doi: 10.1007/s11606-007-0258-5 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Dehon E, Weiss N, Jones J, Faulconer W, Hinton E, Sterling S. A Systematic Review of the Impact of Physician Implicit Racial Bias on Clinical Decision Making. Acad Emerg Med. 2017. Aug;24(8):895–904. doi: 10.1111/acem.13214 [DOI] [PubMed] [Google Scholar]
  • 11.Zestcott CA, Blair IV, Stone J. Examining the Presence, Consequences, and Reduction of Implicit Bias in Health Care: A Narrative Review. Group Process Intergroup Relat. 2016. Jul;19(4):528–42. doi: 10.1177/1368430216642029 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.Murdoch TB, Detsky AS. The Inevitable Application of Big Data to Health Care. JAMA. 2013. Apr 3;309(13):1351–2. doi: 10.1001/jama.2013.393 [DOI] [PubMed] [Google Scholar]
  • 13.Yang X, Chen A, PourNejatian N, Shin HC, Smith KE, Parisien C, et al. A large language model for electronic health records. npj Digit Med. 2022. Dec 26;5(1):1–9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14.Lindvall C, Deng CY, Moseley E, Agaronnik N, El-Jawahri A, Paasche-Orlow MK, et al. Natural Language Processing to Identify Advance Care Planning Documentation in a Multisite Pragmatic Clinical Trial. J Pain Symptom Manage. 2022. Jan;63(1):e29–36. doi: 10.1016/j.jpainsymman.2021.06.025 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15.Carroll PH, Mohler JL. NCCN Guidelines Updates: Prostate Cancer and Prostate Cancer Early Detection. Journal of the National Comprehensive Cancer Network. 2018. May 1;16(5S):620–3. doi: 10.6004/jnccn.2018.0036 [DOI] [PubMed] [Google Scholar]
  • 16.Nyame YA, Cooperberg MR, Cumberbatch MG, Eggener SE, Etzioni R, Gomez SL, et al. Deconstructing, Addressing, and Eliminating Racial and Ethnic Inequities in Prostate Cancer Care. European Urology. 2022. Oct 1;82(4):341–51. [DOI] [PubMed] [Google Scholar]
  • 17.Vince RA Jr, Jiang R, Bank M, Quarles J, Patel M, Sun Y, et al. Evaluation of Social Determinants of Health and Prostate Cancer Outcomes Among Black and White Patients: A Systematic Review and Meta-analysis. JAMA Network Open. 2023. Jan 11;6(1):e2250416. doi: 10.1001/jamanetworkopen.2022.50416 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Mahal BA, Gerke T, Awasthi S, Soule HR, Simons JW, Miyahira A, et al. Prostate Cancer Racial Disparities: A Systematic Review by the Prostate Cancer Foundation Panel. Eur Urol Oncol. 2022. Feb;5(1):18–29. doi: 10.1016/j.euo.2021.07.006 [DOI] [PubMed] [Google Scholar]
  • 19.Schaeffer EM, Srinivas S, Adra N, An Y, Barocas D, Bitting R, et al. NCCN Guidelines® Insights: Prostate Cancer, Version 1.2023. J Natl Compr Canc Netw. 2022. Dec;20(12):1288–98. [DOI] [PubMed] [Google Scholar]
  • 20.Cher ML, Dhir A, Auffenberg GB, Linsell S, Gao Y, Rosenberg B, et al. Appropriateness Criteria for Active Surveillance of Prostate Cancer. Journal of Urology. 2017. Jan;197(1):67–74. doi: 10.1016/j.juro.2016.07.005 [DOI] [PubMed] [Google Scholar]
  • 21.Eastham JA, Auffenberg GB, Barocas DA, Chou R, Crispino T, Davis JW, et al. Clinically Localized Prostate Cancer: AUA/ASTRO Guideline, Part I: Introduction, Risk Assessment, Staging, and Risk-Based Management. J Urol. 2022. Jul;208(1):10–8. doi: 10.1097/JU.0000000000002757 [DOI] [PubMed] [Google Scholar]
  • 22.Al Hussein Al Awamlh B, Barocas DA, Zhu A, Tosoian JJ, Ponsky L, Negoita S, et al. Use of Active Surveillance vs Definitive Treatment Among Men With Low- and Favorable Intermediate-Risk Prostate Cancer in the US Between 2010 and 2018. JAMA Intern Med. 2023. Jun 1;183(6):608–11. doi: 10.1001/jamainternmed.2022.7100 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.Ye F, Moon DH, Carpenter WR, Reeve BB, Usinger DS, Green RL, et al. Comparison of Patient Report and Medical Records of Comorbidities: Results From a Population-Based Cohort of Patients With Prostate Cancer. JAMA Oncol. 2017. Aug 1;3(8):1035–42. doi: 10.1001/jamaoncol.2016.6744 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Spratt DE, Dai DLY, Den RB, Troncoso P, Yousefi K, Ross AE, et al. Performance of a Prostate Cancer Genomic Classifier in Predicting Metastasis in Men with Prostate-specific Antigen Persistence Postprostatectomy. Eur Urol. 2018. Jul;74(1):107–14. doi: 10.1016/j.eururo.2017.11.024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Spratt DE, Zhang J, Santiago-Jiménez M, Dess RT, Davis JW, Den RB, et al. Development and Validation of a Novel Integrated Clinical-Genomic Risk Group Classification for Localized Prostate Cancer. J Clin Oncol. 2018. Feb 20;36(6):581–90. doi: 10.1200/JCO.2017.74.2940 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26.Greenwald AG, Nosek BA, Banaji MR. Understanding and using the Implicit Association Test: I. An improved scoring algorithm. Journal of Personality and Social Psychology. 2003;85(2):197–216. doi: 10.1037/0022-3514.85.2.197 [DOI] [PubMed] [Google Scholar]
  • 27.Institute of Medicine (US) Committee on Understanding and Eliminating Racial and Ethnic Disparities in Health Care. Unequal Treatment: Confronting Racial and Ethnic Disparities in Health Care [Internet]. Smedley BD, Stith AY, Nelson AR, editors. Washington (DC): National Academies Press (US); 2003. [cited 2023 Jul 10]. Available from: http://www.ncbi.nlm.nih.gov/books/NBK220358/ [PubMed] [Google Scholar]
  • 28.Samuels EA, Boatright D, Sanchez LD, Heron SL, Liferidge AT, Wilson T, et al. Clinical Vignettes Inadequate to Assess Impact of Implicit Bias: Concerning Limitations of a Systematic Review. Academic Emergency Medicine. 2017;24(12):1531–2. doi: 10.1111/acem.13317 [DOI] [PubMed] [Google Scholar]
  • 29.Dehon E, Sterling SA, Weiss N. Clinical Vignettes Inadequate to Assess Impact of Implicit Bias: Concerning Limitations of a Systematic Review. Academic Emergency Medicine. 2017;24(12):1533–4. doi: 10.1111/acem.13316 [DOI] [PubMed] [Google Scholar]
  • 30.Blair IV, Steiner JF, Hanratty R, Price DW, Fairclough DL, Daugherty SL, et al. An Investigation of Associations Between Clinicians’ Ethnic or Racial Bias and Hypertension Treatment, Medication Adherence and Blood Pressure Control. J Gen Intern Med. 2014. Jul;29(7):987–95. doi: 10.1007/s11606-014-2795-z [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Greenwald AG, Dasgupta N, Dovidio JF, Kang J, Moss-Racusin CA, Teachman BA. Implicit-Bias Remedies: Treating Discriminatory Bias as a Public-Health Problem. Psychol Sci Public Interest. 2022. May;23(1):7–40. doi: 10.1177/15291006211070781 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 32.Jensen K, Soguero-Ruiz C, Oyvind Mikalsen K, Lindsetmo RO, Kouskoumvekaki I, Girolami M, et al. Analysis of free text in electronic health records for identification of cancer patient trajectories. Sci Rep. 2017. Apr 7;7(1):46226. doi: 10.1038/srep46226 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.Agaronnik N, Lindvall C, El-Jawahri A, He W, Iezzoni L. Use of Natural Language Processing to Assess Frequency of Functional Status Documentation for Patients Newly Diagnosed With Colorectal Cancer. JAMA Oncol. 2020. Oct;6(10):1628–30. doi: 10.1001/jamaoncol.2020.2708 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Udelsman BV, Corey KE, Lindvall C, Gee DW, Meireles OR, Hutter MM, et al. Risk factors and prevalence of liver disease in review of 2557 routine liver biopsies performed during bariatric surgery. Surgery for Obesity and Related Diseases. 2019. Jun 1;15(6):843–9. doi: 10.1016/j.soard.2019.01.035 [DOI] [PubMed] [Google Scholar]
  • 35.Udelsman B, Lee K, Qadan M, Lillemoe KD, Chang D, Lindvall C, et al. Management of Pneumoperitoneum: Role and Limits of Nonoperative Treatment. Annals of Surgery. 2021. Jul;274(1):146. doi: 10.1097/SLA.0000000000003492 [DOI] [PubMed] [Google Scholar]
  • 36.Lilley EJ, Lindvall C, Lillemoe KD, Tulsky JA, Wiener DC, Cooper Z. Measuring Processes of Care in Palliative Surgery: A Novel Approach Using Natural Language Processing. Annals of Surgery. 2018. May;267(5):823. doi: 10.1097/SLA.0000000000002579 [DOI] [PubMed] [Google Scholar]
  • 37.Lee RY, Kross EK, Torrence J, Li KS, Sibley J, Cohen T, et al. Assessment of Natural Language Processing of Electronic Health Records to Measure Goals-of-Care Discussions as a Clinical Trial Outcome. JAMA Network Open. 2023. Mar 2;6(3):e231204. doi: 10.1001/jamanetworkopen.2023.1204 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Savage N. Breaking into the black box of artificial intelligence. Nature [Internet]. 2022. Mar 29 [cited 2023 Aug 13]; Available from: https://www.nature.com/articles/d41586-022-00858-1 doi: 10.1038/d41586-022-00858-1 [DOI] [PubMed] [Google Scholar]
  • 39.Cooperberg MR, Meeks W, Fang R, Gaylis FD, Catalona WJ, Makarov DV. Time Trends and Variation in the Use of Active Surveillance for Management of Low-risk Prostate Cancer in the US. JAMA Netw Open. 2023. Mar 2;6(3):e231439. doi: 10.1001/jamanetworkopen.2023.1439 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.Butler S, Muralidhar V, Chavez J, Fullerton Z, Mahal A, Nezolosky M, et al. Active Surveillance for Low-Risk Prostate Cancer in Black Patients. New England Journal of Medicine. 2019. May 23;380(21):2070–2. doi: 10.1056/NEJMc1900333 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 41.Krishna S, Fan Y, Jarosek S, Adejoro O, Chamie K, Konety B. Racial Disparities in Active Surveillance for Prostate Cancer. Journal of Urology. 2017. Feb;197(2):342–9. doi: 10.1016/j.juro.2016.08.104 [DOI] [PubMed] [Google Scholar]
  • 42.Vigneswaran HT, Mittelstaedt L, Crippa A, Eklund M, Vidal A, Freedland SJ, et al. Progression on active surveillance for prostate cancer in Black men: a systematic review and meta-analysis. Prostate Cancer Prostatic Dis. 2022. Feb;25(2):165–73. doi: 10.1038/s41391-021-00425-1 [DOI] [PubMed] [Google Scholar]
  • 43.Collingwood SA, McBride RB, Leapman M, Hobbs AR, Kwon YS, Stensland KD, et al. Decisional regret after robotic-assisted laparoscopic prostatectomy is higher in African American men. Urologic Oncology: Seminars and Original Investigations. 2014. May 1;32(4):419–25. doi: 10.1016/j.urolonc.2013.10.011 [DOI] [PubMed] [Google Scholar]
  • 44.Benjamins MR, Hunt BR, Raleigh SM, Hirschtick JL, Hughes MM. Racial Disparities in Prostate Cancer Mortality in the 50 Largest US Cities. Cancer Epidemiology. 2016. Oct 1;44:125–31. doi: 10.1016/j.canep.2016.07.019 [DOI] [PubMed] [Google Scholar]
  • 45.Wei JT, Barocas D, Carlsson S, Coakley F, Eggener S, Etzioni R, et al. Early Detection of Prostate Cancer: AUA/SUO Guideline Part I: Prostate Cancer Screening. Journal of Urology. 2023. Jul;210(1):46–53. doi: 10.1097/JU.0000000000003491 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Cooperberg MR, Fang R, Schlossberg S, Wolf JS, Clemens JQ. The AUA Quality Registry: Engaging Stakeholders to Improve the Quality of Care for Patients with Prostate Cancer. Urology Practice. 2017. Jan;4(1):30–5. [DOI] [PubMed] [Google Scholar]
  • 47.Luckenbaugh AN, Auffenberg GB, Hawken SR, Dhir A, Linsell S, Kaul S, et al. Variation in guideline concordant active surveillance follow-up in diverse urology practices. J Urol. 2017. Mar;197(3 Pt 1):621–6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Curtis JR, Lee RY, Brumback LC, Kross EK, Downey L, Torrence J, et al. Intervention to Promote Communication About Goals of Care for Hospitalized Patients With Serious Illness: A Randomized Clinical Trial. JAMA. 2023. Jun 20;329(23):2028–37. doi: 10.1001/jama.2023.8812 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 49.Sabin JA. Tackling Implicit Bias in Health Care. New England Journal of Medicine. 2022. Jul 14;387(2):105–7. doi: 10.1056/NEJMp2201180 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Lett E, Asabor E, Beltrán S, Cannon AM, Arah OA. Conceptualizing, Contextualizing, and Operationalizing Race in Quantitative Health Sciences Research. Ann Fam Med. 2022;20(2):157–63. doi: 10.1370/afm.2792 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 51.Bailey ZD, Krieger N, Agénor M, Graves J, Linos N, Bassett MT. Structural racism and health inequities in the USA: evidence and interventions. Lancet. 2017. Apr 8;389(10077):1453–63. doi: 10.1016/S0140-6736(17)30569-X [DOI] [PubMed] [Google Scholar]
  • 52.Freed GL, Bogan B, Nicholson A, Niedbala D, Woolford S. Error Rates in Race and Ethnicity Designation Across Large Pediatric Health Systems. JAMA Network Open. 2024. Sep 3;7(9):e2431073. doi: 10.1001/jamanetworkopen.2024.31073 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Alvaro Galli

Transfer Alert

This paper was transferred from another journal. As a result, its full editorial history (including decision letters, peer reviews and author responses) may not be present.

17 Sep 2024

PONE-D-24-33113Ascertaining provider-level implicit bias in electronic health records with rules-based natural language processing: a pilot study in the case of prostate cancerPLOS ONE

Dear Dr. Ramaswamy,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by Oct 31 2024 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Alvaro Galli

Academic Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, we expect all author-generated code to be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

3. We note that you have indicated that there are restrictions to data sharing for this study. For studies involving human research participant data or other sensitive data, we encourage authors to share de-identified or anonymized data. However, when data cannot be publicly shared for ethical reasons, we allow authors to make their data sets available upon request. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions

Before we proceed with your manuscript, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information, data are owned by a third-party organization, etc.) and who has imposed them (e.g., a Research Ethics Committee or Institutional Review Board, etc.). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of recommended repositories, please see https://journals.plos.org/plosone/s/recommended-repositories. You also have the option of uploading the data as Supporting Information files, but we would recommend depositing data directly to a data repository if possible.

Please update your Data Availability statement in the submission form accordingly.

4. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: The authors have submitted a thoughtful and compelling study. It is novel, thoughtful, and merits publication.

The authors examined whether implicit biases among providers affect treatment decisions in prostate cancer care using natural language processing analysis of medical records. Using case records of three urologic oncologists to over 1,000 patients, they analyzed care quality and intensity in four domains: active surveillance, molecular biomarker discussion, and evaluations of urinary and sexual function. The providers took an Implicit Association Test to measure racial biases. The authors offer that the implicit association test results showed that two of the providers did not have implicit bias favoring White patients while one did. The 4 care criteria provided differed for two of the providers in statistically significant ways, one of which was the provider with implicit bias. These findings again emphasize that CaP care may reflect aspects of the provider rather than the patients and that implicit biases might be included as such a factor. While IAT testing has gained in its use, most have recognized that the results reflect beliefs rather than actions. This paper adds an important perspective by looking at such actions (care provided).

The authors appropriately argue that these results are appropriately a pilot study and highlights the need for further research- as well as potential interventions- to address biases in care.

I offer suggestions, both major and minor, that might be helpful to the manuscript.

Major:

1. Three providers: The N for the study is really 3 providers rather than 1000+ patients.

a. Recommend that this be expressed as the first limitation in discussion rather than the fourth.

2. IAT raw scores: If we are speaking about the same IAT, the test generates scores on a continuous scale that are subsequently categorized (-3 to 3)

a. The raw scores may be informative for Providers A and B, even if they are in the 0-0.15 range

b. The authors should state specifically that these numeric scores were either negative (a Black implicit bias) or positive numbers (a White implicit bias)

3. Non-White versus Black: The relationship between the independent variable IAT testing (White/Black bias) as a predictor of the dependent variable- patient-by-patient care provided as analyzed by NLP analysis are not matched (White/Non-White).

a. In addition to those self-identifying as Black, exactly who is non-White? Can this be further broken down?

i. Can the analysis also be run for just patients who have identified as Black and White?

b. The breakdown of populations into “white” and “non-white” has a troublesome history (one-drop theory, social hierarchy, justification for slavery, etc.).

i. I do not believe that this is what the authors are implying, and a brief description of what is NOT meant could be in the limitations.

Minor:

1. Self-identified race of the providers: We learn some facts about the 3 providers, but not about their race

a. We learn years of experience and a metric of publication impact (h-factor), which perhaps may have been included to suggest that experienced scientists cannot be biased. This potentially can be included in the limitations section.

b. Interestingly for a publication about race and bias, we do not learn the self-identified race of the individuals. This is not meant to assign an explanation for their bias, but it allows a very brief mention in the discussion as to who mainly provides care to Black patients (and to point out that having more providers- and a diverse sample-in future studies allows for a deeper inquiry of racial concordant/racial discordant care).

2. SES variables: We have income, but not education. Educational attainment of the patients could be important (bias against low education, educational attainment being a confounder)

a. Education might be an important variable. If not available for analysis, that should be indicated as a limitation.

3. More about GG: GG1 and GG2 could be more explicitly defined as this article may be read widely and outside of urological circles

4. IAT results: More information needed.

a. Knowing that the IAT results were not known to the providers would be important

b. Likely the IAT test was done once with each provider. At what point was this done? That bias at one testing session is a predictor for care over 12+ years is a difficult issue and simply can be discussed as a limitation.

i. It may have been done late in the timeframe of the cohort, meaning that one’s attitude in 2020 was used to predict care delivered in 2010

ii. Bias may not be constant over time (we hope it gets better!)

5. NLP will likely not be able to verify to what extent shared decision making actually took place. We're relying on what the provider included in the note to represent this. This is a limitation.

6. Getting a PSA test is also subject to bias: Everyone in this study already has a “positive” PSA value. That every patient getting into this analysis passed that barrier might suggest that the effect of bias in CaP care is likely larger than just these 4 criteria used (i.e., the need to consider the entire cascade of care framework).

Reviewer #2: Ramaswamy et al. present data on implicit bias evaluation in prostate cancer patients using natural language processing. I´ve really enjoyed reading this manuscript and the authors should be congratulated for this really good use NLP. However, sample size (3 providers, 1000 patients) and granularity of data is limited.

Major issue #1:

One of the significant findings was that " Provider C recommended active surveillance (p<0.01, �=0.175) and considered biomarkers (p=0.047, �=0.127) more often in White men than expected, suggestive of treatment imparity". I would consider caution on that conclusion. Afro-Americans are often diagnosed at younger age and have more aggressive disease. I would recommend to re-run the analysis applying a more granular approach on race/ethnicity (Caucasian, Hispanics, Afro-American, Others).

Major issue #2:

Because MRI fusion biopsy cause artificial stage migration to some degree, I was wondering, if the proportion of fusion biopsy was equal in the different population or if this was considered during analysis. There are some works, finding lower rates of MRI fusion biopsy in disadvantaged population. Because active surveillance is more safe in an MRI fusion biopsy pathway, AS is more frequently recommended if a fusion biopsy was performed. This could have introduced a systematic bias.

Major issue #3:

Did you adjust the analysis for the year of counselling? (biomarker and AS were more implemented in recent years) Did you adjust the analysis on socioeconomic status and insurance coverage of biomarkers etc?

Minor issue #1:

I think it should be mentioned in the limitations that all providers were highly proliferated trained academics in a large cancer center which might influenced results. In rural areas, these differences might be more pronounced.

Minor issue #2:

I would recommend to introduce sub-section in the methods parts to make it more reader-friendly.

Minor issue #3:

I would recommend to add all tables / figures at the end to make it more easy to access them.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes: Theodore M. Johnson II

Reviewer #2: Yes: Fabian Falkenbach

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2024 Dec 30;19(12):e0314989. doi: 10.1371/journal.pone.0314989.r002

Author response to Decision Letter 0


3 Nov 2024

We have provided this as an attachment that will be easier to understand given the depth of response and revisions prompted by the Decision Letter.

Attachment

Submitted filename: PLOS Reviewer Comments_Final.docx

pone.0314989.s002.docx (126.9KB, docx)

Decision Letter 1

Alvaro Galli

20 Nov 2024

Ascertaining provider-level implicit bias in electronic health records with rules-based natural language processing: a pilot study in the case of prostate cancer

PONE-D-24-33113R1

Dear Dr. Ramaswamy,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice will be generated when your article is formally accepted. Please note, if your institution has a publishing partnership with PLOS and your article meets the relevant criteria, all or part of your publication costs will be covered. Please make sure your user information is up-to-date by logging into Editorial Manager at Editorial Manager® and clicking the ‘Update My Information' link at the top of the page. If you have any questions relating to publication charges, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Alvaro Galli

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #2: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #2: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #2: The authors have addressed all my comments in detail. The manuscript fulfills the quality criteria for publication.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #2: Yes: Fabian Falkenbach

**********

Acceptance letter

Alvaro Galli

13 Dec 2024

PONE-D-24-33113R1

PLOS ONE

Dear Dr. Ramaswamy,

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now being handed over to our production team.

At this stage, our production department will prepare your paper for publication. This includes ensuring the following:

* All references, tables, and figures are properly cited

* All relevant supporting information is included in the manuscript submission,

* There are no issues that prevent the paper from being properly typeset

If revisions are needed, the production department will contact you directly to resolve them. If no revisions are needed, you will receive an email when the publication date has been set. At this time, we do not offer pre-publication proofs to authors during production of the accepted work. Please keep in mind that we are working through a large volume of accepted articles, so please give us a few weeks to review your paper and let you know the next and final steps.

Lastly, if your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

If we can help with anything else, please email us at customercare@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Alvaro Galli

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Appendix

    (DOCX)

    pone.0314989.s001.docx (2.3MB, docx)
    Attachment

    Submitted filename: PLOS Reviewer Comments_Final.docx

    pone.0314989.s002.docx (126.9KB, docx)

    Data Availability Statement

    The EHR datasets cannot be shared for the privacy of the individuals whose data were used in this study given it is entirely populated with protected health information which can be used to identify specific individuals, even after our efforts to anonymize patient name, medical record number, and date of birth. Human research participant data includes: names, location data, date of birth, examination dates, sex, ethnicity, and race. Data sets may be available (with limitations) on request through Weill Cornell Medicine Human Research Protections at 575 Lexington Avenue, New York, NY 10022, Phone: (646)-962-8200.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES