Skip to main content
PLOS One logoLink to PLOS One
. 2023 Jan 25;18(1):e0280931. doi: 10.1371/journal.pone.0280931

Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs

Danne C Elbers 1,2,3,*, Jennifer La 2, Joshua R Minot 1, Robert Gramling 4, Mary T Brophy 2,5, Nhan V Do 2,5, Nathanael R Fillmore 2,3,6, Peter S Dodds 1, Christopher M Danforth 1
Editor: Nabil Elhadi Elsayed Ali Omar7
PMCID: PMC9876289  PMID: 36696437

Abstract

Natural language processing of medical records offers tremendous potential to improve the patient experience. Sentiment analysis of clinical notes has been performed with mixed results, often highlighting the issue that dictionary ratings are not domain specific. Here, for the first time, we re-calibrate the labMT sentiment dictionary on 3.5M clinical notes describing 10,000 patients diagnosed with lung cancer at the Department of Veterans Affairs. The sentiment score of notes was calculated for two years after date of diagnosis and evaluated against a lab test (platelet count) and a combination of data points (treatments). We found that the oncology specific labMT dictionary, after re-calibration for the clinical oncology domain, produces a promising signal in notes that can be detected based on a comparative analysis to the aforementioned parameters.

Introduction

Estimating the sentiment expressed by text corpora has become popular in the social sciences, especially with the increasing prevalence of social media platforms such as Twitter and Facebook. The Hedonometer, developed by Dodds and Danforth [1, 2], estimates daily global happiness based on Twitter messages [3], and has been shown to reflect collective attention to global events. The instrument has been successfully utilized in a wide range of domains from quantifying happiness in green spaces [4] to identifying story arcs in palliative care conversations [5] and from understanding social amplification [6] to presidential engagement [7].

Here, for the first time, we explore the possibility of re-calibrating the Hedonometer sentiment scoring instrument to the clinical oncology domain and utilizing it to identify aspects of the patient trajectory through clinical notes. Understanding how care plans and communication are perceived by patients is crucial for continuous improvement of healthcare systems and are currently often quantified only through narrative or survey based studies [810]. Sentiment scoring has been explored in the healthcare domain with mixed results. For example, Weissman et al. explored several sentiment scoring methodologies on notes recorded in the ICU [11] and made available through the MIMIC-III data set [12]. They found high variability between methods, but point out a strong association between sentiment and death, that had been seen previously [13]. Weissman et al. suggest that sentiment scoring methodology needs to be more strongly tailored to the healthcare domain and addressed with coverage of specific medical terminology. This argument was highlighted as well by McCoy et al. [14], who performed sentiment scoring on clinical notes utilizing a generic vocabulary scored for polarity (negative vs positive).

We agree that the sentiment associated with medical vocabulary differs greatly from its more common layperson usage. For example, the words ‘positive’ and ‘negative’ have a very different meaning in a clinical context then they do in generic use. A ‘positive’ clinical test often signifies an diagnostic indication of a health issue and is almost never a good event in medicine. In this manuscript we address this gap through execution of a data-driven approach to re-calibration of the original labMT sentiment dictionary. Secondly, we evaluate the domain-specific Hedonometer against several oncologic data points. Our goal is to gauge if the Hedonometer, after re-calibration, can detect a signal indicating clinical trajectory of cancer care using medical record notes.

Materials and methods

Selected data

The cohort used in this analysis consists of 10,000 randomly selected patients whose healthcare is managed by the Department of Veterans’ Affairs (VA), and who were diagnosed with lung cancer between 2017 and 2019. A lung cancer diagnosis was determined based on the VA’s Tumor Registry [15]. All notes (3,500,000+) were extracted from the VA’s Corporate Data Warehouse (CDW) from date of diagnosis until 2 years after. To properly validate the signal strength of the re-calibrated Hedonometer instrument, no sub-selection was made based on note types, i.e. all notes were included.

This study was approved by the VA Boston Healthcare System Institutional Review Board (3209), under a waiver of informed consent. Access to raw data from the United States Department of Veterans Affairs’ corporate data warehouse was granted by the VA Boston Healthcare System Institutional Review Board and the Veterans Health Administration National Data Systems.

Re-calibration of the Hedonometer

First, we aimed to revise the LabMT word list for this clinical context, specifically so the sentiment scores were oriented towards the cancer domain. The original study produced the Language Assessment by Mechanical Turk sentiment ratings using an online survey on Amazon’s crowdsourced work platform [2]. For the present study, we asked 5 health care providers (3 MDs, 2 nurses) with 10+ years experience in oncology to (re)assess 200 high-importance words (details on their selection will be provided later). In accordance with the original LabMT word list, words are scored on a scale from 1–9, with 9 being the happiest or most positive value and 1 being the least positive or most unhappy value. The instructions given to the SME’s were the following:

The results of this survey will be used to measure the happiness of words in the context of lung oncology care notes. The overall aim is to asses how providers feel about individual words in their clinical context. Please rate the following individual words on a 9 point ‘unhappy - happy’ scale with 5 being neutral. 1. Read the word. 2. Observe your emotional response. 3. Select score in accordance with your emotional response’.

SME’s were able to select a radio button response for each word, with a reminder of score meaning at 1 (’unhappy’), 5 (’neutral’), 9 (’happy’).

After exclusion of stop words [16], the list of high-importance words for domain specific re-scoring was selected as follows. We aimed to use both (a) word frequency and (b) the likely difference between layperson and clinical context to identify labMT words most mismatched in sentiment. We chose 100,000 random notes from the larger dataset, and used these notes to design two distinct mechanisms for re-evaluation.

First, we counted all non-labMT words in the notes, and sorted them by frequency. The top 5 words in this lists were ‘tab’, ‘medication’, ‘prn’, ‘reviewed’ and ‘provider’. Second, we parsed the notes for appearances of each anchor labMT word wi, with labMT sentiment hi, and identified the adjacent 5 words before and after in the notes. The average sentiment of these neighboring words across all notes was calculated to be hiamb, the so-called ambient sentiment, and compared with the original labMT rating hi. Words wi were then sorted by the magnitude of hiδ=|hi-hiamb| such that outliers for which the medical context sentiment hiamb deviated substantially from the context independent sentiment hi could be identified. Words not scored in the original labMT study such as ‘medication’, found through frequency ranking, were artificially assigned an hiδ of 1. To combine these two word lists, we multiplied each word’s frequency in the notes by hiδ, and truncated after the top 200. The resulting 200 words were deemed most important to re-evaluate, given their prominence in the medical notes and their absent or poor context matching in labMT. Among the 200 words to be scored, a total of 66 new medical domain words were identified, and 134 labMT words were re-scored.

Calculating sentiment

Sentiment for each note is calculated using the re-calibrated Hedonometer score list, which includes the original words, the original words with new scores, and the new words. In order to focus on the more informative sentiment contributions, words with values between 4 and 6 in the original word list [2] have been excluded from calculating a note’s score. Due to the domain specific focus, words scored by the SME’s have been included, regardless of score. A note’s average score is subsequently calculated by, after excluding stop words [16], obtaining the frequency of the unique words occurring in the note, looking up their score in the re-calibrated word list, and multiplying the frequency of the word by its score. After which the sum is taken and the mean for the individual note is calculated [2], thus resulting in an average score for each note that can be used for further analysis. Words that are not scored are ignored in all calculations and in line with the Hedonometer strategy no further word cleaning, such as stemming, is performed [2].

Comparative analysis

In order to validate the calibration of the Hedonometer instrument, several external parameters were selected to detect a signal in notes. Parameters that were selected for comparison included objective data in the form of lab test outcomes; platelet count, and a combination parameter based on medications, see Table 1. This specific set of parameters was chosen to assess the performance of the Hedonometer, since they are either very objective and often performed, such as the laboratory results known to be indicative of a patient’s well being and cancer prognosis (platelet counts) [1719], or a combination of medication and date associations that allow for comparison (treatments) [2022]. In addition to validating the re-calibrated Hedonometer, the original Hedonometer was also evaluated for comparison.

Table 1. Data collection.

Data Type Start End Iteration
Notes Date of Diagnosis Date of Diagnosis + 24 months all
Treatment Start of Treatment Start of Treatment + 6 weeks daily
Platelet Count Day of Test - 1 week Day of Test + 1 week all

Platelet count

Platelet count was selected as an objective or hard measure, since it has been shown to be an indicator of how well a patient is feeling and has been linked extensively to cancer [1719]. The hypothesis is that if a patient is not feeling well, which could relate to an abnormal platelet count, this ‘unwell feeling’ is recorded in clinical notes resulting in a drop in the sentiment score. Additionally, this lab test is ordered quite often, especially in cancer care, varying from every two weeks to daily depending on a patient’s state. Since platelet count lab test outcomes come in a multitude of units, data was cleaned up and converted to one unit type (109/L) before inclusion. For this analysis, platelet count results were divided into three groups: low, normal and high. A low platelet count means below the threshold of 160 109/L, a normal platelet count is between 160 109/L and 410 109/L, and a high platelet count above 410 109/L. All clinical notes one week before and one week after a platelet count test was performed were selected and the average score per note was calculated as described above. These notes were placed in the group in accordance with the lab test outcome. The three groups were subsequently tested for normality with Shapiro’s as well as Anderson’s tests. The outcome of these tests determines whether a one-way ANOVA or Kruskal-Wallis test is most appropriate to assess significance. If a significant difference between groups is found, a post-hoc test in the form of Tukey (parametric) or Conover (non-parametric) will be performed to find out which specific between-strata comparisons of platelet count appear to account for the overall difference in sentiment.

Treatments

As a combined data point to validate the re-calibrated Hedonometer against, cancer treatments were chosen. Specifically, chemotherapy, platinum (a form of chemotherapy), targeted and checkpoint therapies were evaluated. First, 39 medications were mapped to the treatments (see S3 Table), subsequently all these medications and the start of treatment were pulled from the Inpatient, Outpatient, Pharmacy and IV domains in VA’s Corporate Data Warehouse (CDW). As treatments might be given in combination or sequence, patients can be counted multiple times. Notes were analyzed until six weeks after the start of the specific medication and associated with the treatment. The mean note score for each day was calculated, and a visual inspection of the mean note score fluctuation over time was provided and analyzed with SME’s in clinical oncology. Additionally, word shift graphs [23] were created to gather understanding of which words influenced clear sentiment fluctuations. Lastly, the analysis was ran a second time, utilizing the LabMT dictionary only, to evaluate the difference with the re-calibrated dictionary.

Software and systems

Cohort selection has been performed and data has been extracted from the VA’s CDW using MSSQL. All subsequent calculations and analysis has been performed in Python 3.8, for analysis specifically the scipy, statsmodels and scikit packages have been used. Shared code can be found under the Code Availability section.

Results

Selected data

The selected cohort of 10,000 lung cancer patients resulted in a mean of 886 notes per patient per year and a standard deviation of 2180.

Re-calibration of the Hedonometer

Please see supplemental information for a complete list of the 200 words selected based on their surrounding sentiment and frequency to be re-scored by SME’s. Fig 1 shows that our selection of 200 words will cover 30% of note text.

Fig 1. Words are ranked based on the product of their clinical note coverage and the difference in word score to a word’s ambient sentiment score.

Fig 1

Mean standard deviation of re-scoring the high-importance words between the SME’s was 0.51 on the 9-point scale. The mean of re-scored words, 4.93, is lower then the mean of the original labMT dataset, namely 5.37. Full scores and standard deviation are available in S2 Table. The addition and updates of the oncology domain specific words result in a word list of 10,253, an increase of 66 words from the original 10,187. Examination of the re-scored words, see the left of Fig 2, shows that word ‘positive’ is scored less high by the SME’s then in LabMT (5.6 vs. 7.8), while the word ‘negative’ is scored much higher (5.8 vs. 2.4). Other words standing out include ‘discharge’, ‘pain’, ‘cancer’, ‘veteran’ and ‘family’. To compare, the right of Fig 2 shows the top 50 words with the largest shifts in word score for both the SME’s evaluation and based on the surrounding sentiment calculation. A subset of random clinical notes was evaluated for confirmatory no’s; standard questions generally answered with the word ‘no’. Questions similar to ‘have you recently travelled outside of the country?’, with answers of ‘no’ were identified, however did not occur in high frequency. It appears that repetitive questions are often answered with the word ‘negative’, e.g. ‘negative for fever, chills, blurring of vision, redness of eye, nausea’ a word that was re-scaled by the SME’s. The word ‘no’ itself is ranked 410, and was thus not highly influential on score calculation.

Fig 2. Word score shifts due to calibration by SME’s.

Fig 2

The figure on the left compares the LabMT scores to the scores assigned by the SME’s. The right figure compares the ambient sentiment for each anchor term and the eventual assessment by the SME’s.

Comparative analysis

Platelet count

Every group with platelet count results, low, normal and high, was comprised evenly to contain 39882 unique notes. This number was equivalent to the group with the lowest count of notes (high group), for the larger groups a random subset was selected to create equal groups. All three groups failed both the Shapiro and Anderson normality tests. Although visual inspection of QQ-plots in addition to histograms did appear to come close to normality, the high number of samples might have played a factor in failing the tests. To be on the safe side, it was decided to test non-parametric, thus a Kruskal-Wallis was subsequently performed. The result of the Kruskal-Wallis test showed a significant difference (p = 2.79 × 10−06) and post-hoc Conover results are displayed in Table 2.

Table 2. Platelet count—Post hoc conover test.
High Low Normal
High 1 0.56 3.49 × 10−06
Low 0.56 1 4.84 × 10−05
Normal 3.49 × 10−06 4.84 × 10−05 1

A significant difference in note score is present between the high platelet count group and the normal platelet count group, as well as between the low platelet count group and the normal platelet count group. However no difference was found between the high and low platelet count group. The high platelet count group had a median note score of 5.308; mean of 5.355, the low platelet count group a median note score of 5.316; mean of 5.356, while the normal platelet count group had a slightly higher median note score of 5.325; mean of 5.363. The original Hedonometer produced no significant result (p = 0.62) after the non-parametric test. This data similarly failed the Shapiro and Anderson normality tests.

Treatments

Evaluating the scoring of daily notes generated during the six weeks after the start of either one of the four different treatments; chemotherapy (633160 notes), platinum (317392 notes), checkpoint (230500 notes) and targeted (139908 notes), shows a cyclical weekly pattern, see Fig 3. Day 40 for targeted therapy has the lowest count of data points, namely 1980 notes.

Fig 3. Note scores on a day to day basis per treatment, starting at the day of treatment till six weeks from date of start of treatment.

Fig 3

To better identify what creates the dip in note scores on day 21, word shift plots are created, see Fig 4. Words related to lung cancer treatment such as ‘lung’, ‘cancer’, ‘treatment’, ‘dose’, ‘mouth’ and ‘chemotherapy’ appear to drive the score down. While more generic words, such as ‘care’, ‘support’, ‘activity’, ‘patient’ and ‘independent’, related to patient care seem to positively influence the note scores on peak days.

Fig 4. Using notes authored on Day 21 of treatment as a reference, word-shift graphs detail the words influencing the drop in sentiment when compared with day 19 (left, Checkpoint Treatment) and and 17 (right, Chemotherapy Treatment).

Fig 4

Looking at the comparison between days 19 and 21 on the left, words appearing on the left side of the graph contribute positively to day 21, while words on the right side contribute positively to day 19 (there are many more of this type). For example, the relatively positive words ‘support’, ‘discharge’, and ‘independent’ are more common on day 19. The relatively negative words ‘disease’ and ‘metastatic’ are less common on day 19. Going against the overall trend, the relatively positive words ‘today’, ‘well’, and ‘stable’ are more common on day 21. The relatively negative words ‘risk’, ‘pressure’, and ‘fall’ are less common on day 21.

Rerunning the analysis with the original labMT dictionary showed a similar cyclical pattern, however the overall note scores were 0.3 points higher. This difference was attributed to the prevalence of missing clinical words in labMT, which were generally scored low by the SME’s.

Discussion

We have designed and implemented a data-driven method for re-calibration of an existing sentiment scoring instrument, the Hedonometer, in a specific healthcare domain. This re-calibration has been proposed numerous times in previous research on sentiment in clinical notes [11, 13, 14] as a solution to the mixed results found when utilizing existing vocabularies in this context. However, to the authors’ awareness, re-calibration has not been done before. The re-calibration of the labMT sentiment dictionary for the oncology healthcare domain resulted in 200 re-scored words, which together cover 30% of clinical note text for the oncology domain, see Fig 1. Overall, this improves the coverage of words included in calculating the average score for a clinical note. A limitation of our re-calibration methodology is that only five SME’s were included in the re-calibration, while a higher number would have improved its value. Identifying highly skilled SME’s in the lung oncology domain with sufficient expertise is difficult and we are thankful for their given time.

The successful application of the domain-specific Hedonometer in our comparative analysis suggests that a signal is present and measurable in clinical notes. Validating the instrument on laboratory test outcomes, we found a significant difference in normal versus low or high platelet counts. This difference was not present when utilizing the original Hedonometer instrument. Additionally, a consistent cyclic effect is visible when scoring notes in relation to cancer treatment timelines, showing a comparable cycle for all treatments. This cycle can be explained by weekly provider visits discussing side effects and treatment itself, with a clear dip on day 21 for the chemotherapy (incl. platinum) and checkpoint treatments indicating the side-effects being at their worst. Targeted therapy tends to produce fewer side-effects and is often given to patients in better health, which could explain the better overall score and reduced fluctuation. Discussion with clinical SME’s underscores this belief. When examining the score shift with day 21, treatment related words indeed appear to drive the note score down. Notably the text size is larger on day 21, associating more volume of note generation during visits. This negative shift is also consistent with Portier et al. [22], who perform sentiment analysis on online cancer support forums and find that side-effects of treatments are a topic with a clear negative sentiment shift. Though an explanation for the low dip in checkpoint treatment specifically was not presented by the SME’s, it can be hypothesized that this treatment often includes further advanced cancer cases or cases in palliative care. It would be interesting to see if splitting out durvalumab, often given to less advanced cases, would generate higher sentiment. The outcome of the validation of the Hedonometer on these two measures does not preclude the sentiment changes from being related to other events in patients’ care. Merely it shows presence of a measurable and explainable signal.

Our evaluation of the domain-specific Hedonometer is far from exhaustive, and future research should be done to understand better how and where the instrument can be utilized and what its limitations are in the oncology and healthcare domain. The finding that a signal appears present, detectable, and consistent is encouraging and the authors hypothesize next research steps as stratification of the cohort and researching difference between note types, both of which have not been addressed here, and evaluation against other parameters. Other parameters could include, for example, the the VA - Frailty Index [24, 25], a well-validated performance measurement based on a large set of variables across different healthcare domains, or the VA-CAN score [26], surgery or neutrophil counts. If the signal remains robust, clinical and research implementation can vary from quality control of a new medication or treatment plan, to better understanding certain subgroups to quantifying differences between hospital, provider and nurse care. We hope that the domain-specific Hedonometer could provide informative value as a non-invasive, real-time lens into systems of care, making use of all notes entered into the Electronic Health Care Records (EHR). To foster future research in this area, the re-calibrated score list and associated code is made openly available along with this study.

Supporting information

S1 Table. Top 40 words based on rank of surrounding sentiment * text coverage.

(CSV)

S2 Table. Re-scoring outcomes from SME’s.

(CSV)

S3 Table. Medication treatment matrix.

(CSV)

Acknowledgments

The authors like to thank Dr. Albert Lin, Dr. David Tuck, Karen Murray and Karen Visnaw for their help in re-calibration, and Dr. Albert Lin and Dr. Mikaela Fudolig for their input and suggestions. The views expressed are those of the authors and do not necessarily reflect the position or policy of the Department of Veterans Affairs or the U.S. government.

Data Availability

Patient related and note data cannot be shared publicly because it involves sensitive human subject data. Data may be available for researchers who meet the criteria for access to confidential data after evaluation from VA Research and Development Committees. As a VA national legal policy (VHA Directive 1605.01), VA will only share patient data if there is a fully executed contract in place for the specific project. A common contractual mechanism utilized for this type of sharing is a ‘Cooperative Research and Development’ (CRADA) agreement. These contracts are typically negotiated in collaboration with VA national Office of General Council (OGC) and lawyers from the collaborating institution. These national sharing policies and standards also apply to deidentified data. In addition, if a contract is in place allowing sharing of deidentified data outside of VA, then VA national policy (VHA Directive 1605.01), states that deidentification certification needs to be met by Expert Determination. The expert determination requires independent assessment from an experienced master or PhD in biostatistics, from a third party not involved in the project, and may require outside funding to support. In addition, for an outside entity to preform research on VA patient data, IRB as well as VA Research and Development Committee approval is required for the specific project. Data requests may be sent to: VA Information Resource Center (VIReC) Building 18 Hines VA Hospital (151V) 5000 S. 5th Avenue Hines, IL 60141-3030 708-202-2413 708-202-2415 (fax) virec@va.gov.

Funding Statement

The author(s) received no specific funding for this work.

References

  • 1. Dodds PS, Danforth CM. Measuring the Happiness of Large-Scale Written Expression: Songs, Blogs, and Presidents. Journal of Happiness Studies. 2010;11(4):441–456. doi: 10.1007/s10902-009-9150-9 [DOI] [Google Scholar]
  • 2. Dodds PS, Harris KD, Kloumann IM, Bliss CA, Danforth CM. Temporal patterns of happiness and information in a global social network: Hedonometrics and Twitter. PLoS ONE. 2011;6(12):e26752. doi: 10.1371/journal.pone.0026752 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Dodds PS, Danforth CM. Hedonometer;. Available from: https://hedonometer.org.
  • 4. Schwartz AJ, Dodds PS, O’Neil‐Dunne JPM, Danforth CM, Ricketts TH. Visitors to urban greenspace have higher sentiment and lower negativity on Twitter. People and Nature. 2019;1(4):476–485. doi: 10.1002/pan3.10045 [DOI] [Google Scholar]
  • 5. Ross L, Danforth CM, Eppstein MJ, Clarfeld LA, Durieux BN, Gramling CJ, et al. Story Arcs in Serious Illness: Natural Language Processing features of Palliative Care Conversations. Patient Education and Counseling. 2020;103(4):826–832. doi: 10.1016/j.pec.2019.11.021 [DOI] [PubMed] [Google Scholar]
  • 6. Alshaabi T, Dewhurst DR, Minot JR, Arnold MV, Adams JL, Danforth CM, et al. The growing amplification of social media: Measuring temporal and social contagion dynamics for over 150 languages on Twitter for 2009–2020. EPJ data science. 2021;10(1):15. doi: 10.1140/epjds/s13688-021-00271-0 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Minot JR, Arnold MV, Alshaabi T, Danforth CM, Dodds PS. Ratioing the President: An exploration of public engagement with Obama and Trump on Twitter. PLOS ONE. 2021;16(4):e0248880. doi: 10.1371/journal.pone.0248880 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Jha AK, Orav EJ, Zheng J, Epstein AM. Patients’ Perception of Hospital Care in the United States. New England Journal of Medicine. 2008;359(18):1921–1931. doi: 10.1056/NEJMsa0804116 [DOI] [PubMed] [Google Scholar]
  • 9. Broadbent E, Wilkes C, Koschwanez H, Weinman J, Norton S, Petrie KJ. A systematic review and meta-analysis of the Brief Illness Perception Questionnaire. Psychology & Health. 2015;30(11):1361–1385. doi: 10.1080/08870446.2015.1070851 [DOI] [PubMed] [Google Scholar]
  • 10. Ruiz-Ceamanos A, Spence C, Navarra J. Individual Differences in Chemosensory Perception Amongst Cancer Patients Undergoing Chemotherapy: A Narrative Review. Nutrition and Cancer. 2022;0(0):1–15. doi: 10.1080/01635581.2021.2000625 [DOI] [PubMed] [Google Scholar]
  • 11. Weissman GE, Ungar LH, Harhay MO, Courtright KR, Halpern SD. Construct validity of six sentiment analysis methods in the text of encounter notes of patients with critical illness. Journal of Biomedical Informatics. 2019;89:114–121. doi: 10.1016/j.jbi.2018.12.001 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12. Johnson AEW, Pollard TJ, Shen L, Lehman LwH, Feng M, Ghassemi M, et al. MIMIC-III, a freely accessible critical care database. Scientific Data. 2016;3(1):160035. doi: 10.1038/sdata.2016.35 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13. Waudby-Smith IER, Tran N, Dubin JA, Lee J. Sentiment in nursing notes as an indicator of out-of-hospital mortality in intensive care patients. PLOS ONE. 2018;13(6):e0198687. doi: 10.1371/journal.pone.0198687 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 14. McCoy TH, Castro VM, Cagan A, Roberson AM, Kohane IS, Perlis RH. Sentiment Measured in Hospital Discharge Notes Is Associated with Readmission and Mortality Risk: An Electronic Health Record Study. PLOS ONE. 2015;10(8):e0136341. doi: 10.1371/journal.pone.0136341 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 15. Zullig LL, Jazowski SA, Chawla N, Williams CD, Winski D, Slatore CG, et al. Summary of Veterans Health Administration Cancer Data Sources Journal of Registry Management. 2019;46(3);76–83. [PubMed] [Google Scholar]
  • 16. Wilbur WJ, Sirotkin K. The automatic identification of stop words. Journal of Information Science. 1992;18(1):45–55. doi: 10.1177/016555159201800106 [DOI] [Google Scholar]
  • 17. Sylman JL, Boyce HB, Mitrugno A, Tormoen GW, Thomas IC, Wagner TH, et al. A Temporal Examination of Platelet Counts as a Predictor of Prognosis in Lung, Prostate, and Colon Cancer Patients. Scientific Reports. 2018;8(1):6564. doi: 10.1038/s41598-018-25019-1 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18. Zhang X, Lv Z, Yu H, Zhu J. The clinicopathological and prognostic role of thrombocytosis in patients with cancer: A meta-analysis. Oncology Letters. 2017;13(6):5002–5008. doi: 10.3892/ol.2017.6054 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19. Maráz A, Furák J, Varga Z, Kahán Z, Tiszlavicz L, Hideghéty K. Thrombocytosis Has a Negative Prognostic Value in Lung Cancer. Anticancer Research. 2013;33(4):1725–1729. [PubMed] [Google Scholar]
  • 20. Howlader N, Forjaz G, Mooradian MJ, Meza R, Kong CY, Cronin KA, et al. The Effect of Advances in Lung-Cancer Treatment on Population Mortality. New England Journal of Medicine. 2020;383(7):640–649. doi: 10.1056/NEJMoa1916623 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Junaid Nabi, Quoc-Dien Trinh. New Cancer Therapies Are Great—But Are They Helping Everyone? | Health Affairs Forefront;. Available from: https://www.healthaffairs.org/do/10.1377/forefront.20190410.590278/full/.
  • 22. Portier K, Greer GE, Rokach L, Ofek N, Wang Y, Biyani P, et al. Understanding Topics and Sentiment in an Online Cancer Survivor Community. JNCI Monographs. 2013;2013(47):195–198. doi: 10.1093/jncimonographs/lgt025 [DOI] [PubMed] [Google Scholar]
  • 23. Gallagher RJ, Frank MR, Mitchell L, Schwartz AJ, Reagan AJ, Danforth CM, et al. Generalized word shift graphs: a method for visualizing and explaining pairwise comparisons between texts. EPJ Data Science. 2021;10(1):4. doi: 10.1140/epjds/s13688-021-00260-3 [DOI] [Google Scholar]
  • 24. Cheng D, DuMontier C, Yildirim C, Charest B, Hawley CE, Zhuo M, et al. Updating and Validating the U.S. Veterans Affairs Frailty Index: Transitioning From ICD-9 to ICD-10. The Journals of Gerontology: Series A. 2021;76(7):1318–1325. doi: 10.1093/gerona/glab071 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. DuMontier C, Fillmore NR, Yildirim C, Cheng D, La J, Orkaby AR, et al. Contemporary Analysis of Electronic Frailty Measurement in Older Adults with Multiple Myeloma Treated in the National US Veterans Affairs Healthcare System. Cancers. 2021;13(12):3053. doi: 10.3390/cancers13123053 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Osborne TF, Veigulis ZP, Arreola DM, Röösli E, Curtin CM. Automated EHR score to predict COVID-19 outcomes at US Department of Veterans Affairs. PLOS ONE. 2020;15(7):e0236554. doi: 10.1371/journal.pone.0236554 [DOI] [PMC free article] [PubMed] [Google Scholar]

Decision Letter 0

Maria Elisabeth Johanna Zalm

17 Aug 2022

PONE-D-22-11193Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs.PLOS ONE

Dear Dr. Elbers,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Your manuscript has been evaluated by one peer-reviewer and their comments are appended below.  The reviewer has raised a number of concerns that need your attention. They request additional information on methodological aspects of the study and the assessment of the results. In addition, the reviewer comments that the discussion section needs a deeper discussion of the limitations of the study. 

Could you please revise the manuscript to carefully address the concerns raised? Please note that we have only been able to secure a single reviewer to assess your manuscript. We are issuing a decision on your manuscript at this point to prevent further delays in the evaluation of your manuscript. Please be aware that the editor who handles your revised manuscript might find it necessary to invite additional reviewers to assess this work once the revised manuscript is submitted. However, we will aim to proceed on the basis of this single review if possible. 

Please submit your revised manuscript by Sep 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Maria Elisabeth Johanna Zalm, Ph.D

Editorial Office

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at 

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Please include an ethics statement in your Methods section, clarifying whether or not the study was approved by an IRB, and provide the name of the IRB and any approval number.

3. Please note that PLOS ONE has specific guidelines on code sharing for submissions in which author-generated code underpins the findings in the manuscript. In these cases, all author-generated code must be made available without restrictions upon publication of the work. Please review our guidelines at https://journals.plos.org/plosone/s/materials-and-software-sharing#loc-sharing-code and ensure that your code is shared in a way that follows best practice and facilitates reproducibility and reuse.

4. Please provide additional details regarding participant consent. In the ethics statement in the Methods and online submission information, please ensure that you have specified (1) whether consent was informed and (2) what type you obtained (for instance, written or verbal, and if verbal, how it was documented and witnessed). If your study included minors, state whether you obtained consent from parents or guardians. If the need for consent was waived by the ethics committee, please include this information.

If you are reporting a retrospective study of medical records or archived samples, please ensure that you have discussed whether all data were fully anonymized before you accessed them and/or whether the IRB or ethics committee waived the requirement for informed consent. If patients provided informed written consent to have data from their medical records used in research, please include this information.

5. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. 

In your revised cover letter, please address the following prompts:

a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent.

b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories.

We will update your Data Availability statement on your behalf to reflect the information you provide.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Partly

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: No

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This study re-calibrated the labMT sentiment dictionary on 3.5M clinical notes describing 10,000 patients diagnosed with lung cancer at the Department of Veterans Affairs. This is an interesting research topic in oncology domain, could be helpful for patient care in practice in the future. Some of my concerns focus on the following points:

1. Your goal is to gauge if the Hedonometer, after re-calibration, can detect a signal indicating clinical trajectory of cancer care using medical record notes. Have you compared with the original Hedonometer before re-calibration? Is the original Hedonometer capable of detecting such a signal? Then what the advantage of the Hedonometer after re-calibration?

2. You calculated the sentiment score of notes and evaluated against a lab test and treatment. Is this evaluation based on the assumption that the sentiment from notes is associated with lab tests or treatments? Would it be possible that the sentiment comes from other events? How did you deal with this possibility?

3. What is the rationale to determine the word frequency in your method: “obtaining the frequency of the unique words occurring in the note, looking up their score in the re-calibrated word list, and multiplying the frequency of the word by its score.”

4. In the sentence: “All clinical notes one week before and one week after a platelet count test was performed were selected and individually scored.” Who scored these clinical notes? What is the inter annotator agreement?

5. “According to SME’s in clinical lung cancer, this cyclical pattern can be explained due to patients’ weekly visits with their providers at which point they are asked about side-effects, bringing down the sentiment score of generated notes.” My question regarding this is if the dictionary before re-calibration can also produce the same result?

6. What do you mean here ? “The re-calibration of the labMT sentiment dictionary for the oncology healthcare domain resulted in 200 re-scored words, which together cover 30% of clinical note text for the oncology domain”

7. “Additionally, a consistent cyclic effect is visible when scoring notes in relation to cancer treatment timelines, showing a comparable cycle for all treatments.” Again, what is the result from the dictionary before re-calibration?

8. What is your limitation in this study?

English errors:

As treatments might be given in combination or sequence, patients can be counted more then once.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

PLoS One. 2023 Jan 25;18(1):e0280931. doi: 10.1371/journal.pone.0280931.r002

Author response to Decision Letter 0


14 Oct 2022

This study re-calibrated the labMT sentiment dictionary on 3.5M clinical notes describing 10,000 patients diagnosed with lung cancer at the Department of Veterans Affairs. This is an interesting research topic in oncology domain, could be helpful for patient care in practice in the future. Some of my concerns focus on the following points:

1. Your goal is to gauge if the Hedonometer, after re-calibration, can detect a signal indicating clinical trajectory of cancer care using medical record notes. Have you compared with the original Hedonometer before re-calibration? Is the original Hedonometer capable of detecting such a signal? Then what the advantage of the Hedonometer after re-calibration?

We compared with the original Hedonometer before re-calibration and have included these results in the updated manuscript.

Briefly, the original Hedonometer was not capable of detecting a signal in the platelet comparison (p = 0.62), in contrast to the re-calibrated version. The re-calibrated Hedonometer also performed better than the original Hedonometer in regard to treatment analysis, as detailed in our response to item 5 below. We have updated the methods, results, and discussion to add comparison to the original Hedonometer.

2. You calculated the sentiment score of notes and evaluated against a lab test and treatment. Is this evaluation based on the assumption that the sentiment from notes is associated with lab tests or treatments? Would it be possible that the sentiment comes from other events? How did you deal with this possibility?

Our rationale for selecting platelet count as an objective measure is that it has been shown to be an indicator of how well a patient is feeling and has been linked extensively to cancer. The hypothesis is that if a patient is not feeling well, which could relate to an abnormal platelet count, this ‘unwell feeling’ is recorded in clinical notes, resulting in a drop in the sentiment score. However, we do acknowledge that the outcome of the validation of the Hedonometer on these two measures does not preclude the sentiment changes from being related to other events in patients' care. It merely shows presence of a measurable signal. We have added these points to the methods (under Methods – Platelet Count) and discussion.

3. What is the rationale to determine the word frequency in your method: “obtaining the frequency of the unique words occurring in the note, looking up their score in the re-calibrated word list, and multiplying the frequency of the word by its score.”

The rationale behind word score frequency is utilizing it to calculate the average score of a single note, this score is then further utilized in the analysis. We’ve updated the methods section accordingly with this.

In addition, we’ve made the code available for calculating note scores through Github, see: https://github.com/delbers/SentimentAnalysisOfMedicalRecordNotes/blob/main/Code/score_note.py.

4. In the sentence: “All clinical notes one week before and one week after a platelet count test was performed were selected and individually scored.” Who scored these clinical notes? What is the inter annotator agreement?

Average note scores are calculated utilizing the recalibrated Hedonometer Word Score methodology as described under 3. As this is an automated method, inter annotator agreement is not applicable. Though slightly different from IAA, but comparable, the standard deviation of individual re-scored words by the Subject Matter Experts is made available in Table S2. The calculation method has been more strongly referenced in the methods section.

5. “According to SME’s in clinical lung cancer, this cyclical pattern can be explained due to patients’ weekly visits with their providers at which point they are asked about side-effects, bringing down the sentiment score of generated notes.” My question regarding this is if the dictionary before re-calibration can also produce the same result?

The cyclical pattern was visible as well before recalibration; however it is not possible to find out what the oncology and healthcare specific words were that attributed to the cyclical pattern as highlighted in Fig 4, as these are not part of the original dictionary. Moreover, the average score was 0.3 points higher. This difference was attributed to the prevalence of missing clinical words in labMT, which were generally scored low by the SME’s. This information is present in the results section.

6. What do you mean here ? “The re-calibration of the labMT sentiment dictionary for the oncology healthcare domain resulted in 200 re-scored words, which together cover 30% of clinical note text for the oncology domain”

This references back to the first section of the Results, where we describe that the 200 words selected for inclusion in the oncology specific word dictionary, and scored by Subject Matter Experts, cover 30% of all words in the clinical notes in the lung oncology domain within the VA. Thus, if we had only used the original Hedonometer dictionary the coverage of words being used to calculate the average sentiment for a clinical note would have been (on average) 30% less. The discussion has been updated with this reference and with a reference to Fig 1.

7. “Additionally, a consistent cyclic effect is visible when scoring notes in relation to cancer treatment timelines, showing a comparable cycle for all treatments.” Again, what is the result from the dictionary before re-calibration?

Addressed in 5.

8. What is your limitation in this study?

A limitation of our re-calibration methodology is that only five SME's were included in the re-calibration, while a higher number would have improved its value. Identifying highly skilled SME's in the lung oncology domain with sufficient expertise is difficult and we are thankful for their given time.

Additionally, the outcome of the validation of the Hedonometer on these two measures does not preclude the sentiment changes from being related to other events in patients' care. Merely it shows presence of a measurable signal.

We like to emphasize that this is an explorative manuscript and is by no means exhaustive, future research should be done to understand better how and where the instrument can be utilized and what its limitations are in the oncology and healthcare domain. The finding that a signal appears present, detectable, and consistent is encouraging. These points have been added to the discussion.

English errors:

As treatments might be given in combination or sequence, patients can be counted more then once.

We thank the reviewer for noticing this mistake in English grammar and have updated the sentence as follows:

‘As treatments might be given in combination or sequence, patients can be counted multiple times’

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 1

Nabil Elhadi Elsayed Ali Omar

28 Dec 2022

PONE-D-22-11193R1Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs.PLOS ONE

Dear Dr. Elbers,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please submit your revised manuscript by 11/01/2023. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.

  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.

  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Nabil Elhadi Elsayed Ali Omar, PharmD.,BCOP.,PhD(C)

Academic Editor

PLOS ONE

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #2: (No Response)

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #2: Thank you for your efforts, please find below my comments for improvements and some suggestions

Title: looks good and reflects the methodology and aim of the research

Introduction:

Good explanation of sentiment analysis in a simple way

Part of sentiment analysis negotiation and that you agree with them better present in the discussions and limit the introduction for the currently available evidence and use of this type of analysis in healthcare systems. Also, you may add information about how the previous sentimental analysis helped and affected the healthcare systems. You have to use evidence with references.

You mentioned what you have done in the introduction, again this should be in the methodology and the discussion part. Try assessing where is the gap you aim to fill with your research and mention it in the introduction.

I would suggest that you paraphrase and clearly state your aim at the end of the introduction section.

Methods:

In the selected data section, I suggest you clearly mention your inclusion and exclusion criteria.

Have you obtained Ethics Review Board? Please mention this in the methodology

I suggest moving this “This resulted in a mean of 886 notes per patient per year and a standard 52

deviation of 2180.” To the results section.

Where is the reference for lines 58,59?

What systems, programs, or program languages have you used? Please mention that

Please justify and explain lines 96-99 in more detail.

What did you use to perform the statistical analysis?

Results:

Line 152 is missing table number

I suggest that explaining the meaning of the results like in line 192 to be moved to the discussion section.

Discussion

I appreciate that you compared your data to other studies

Please elaborate more on the explanation of the result’s meaning and how it can affect the practice. Also, you may mention some of the limitations of your study. Furthermore, clear mention of what further studies need to be done in the future would be important and would enrich your manuscript.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: Yes

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

<quillbot-extension-portal></quillbot-extension-portal>

PLoS One. 2023 Jan 25;18(1):e0280931. doi: 10.1371/journal.pone.0280931.r004

Author response to Decision Letter 1


10 Jan 2023

Response to Reviewers (PONE-D-22-11193)

Based on comments from the reviewers the manuscript ‘Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs.’ has been adjusted. Our response can be found below.

Reviewer 1

All comments have been addressed, no additional comments.

Reviewer 2

Thank you for your efforts, please find below my comments for improvements and some suggestions

Title: looks good and reflects the methodology and aim of the research

Introduction:

Good explanation of sentiment analysis in a simple way

Part of sentiment analysis negotiation and that you agree with them better present in the discussions and limit the introduction for the currently available evidence and use of this type of analysis in healthcare systems. Also, you may add information about how the previous sentimental analysis helped and affected the healthcare systems. You have to use evidence with references.

You mentioned what you have done in the introduction, again this should be in the methodology and the discussion part. Try assessing where is the gap you aim to fill with your research and mention it in the introduction.

I would suggest that you paraphrase and clearly state your aim at the end of the introduction section.

We have restructured the paper, in particular the introduction and discussion, as suggested.

Methods:

In the selected data section, I suggest you clearly mention your inclusion and exclusion criteria.

This has been updated in the Methods – Selected Data section with the addition of the VA’s Tumor Registry codes used to determine lung cancer. A reference has been added as well.

Have you obtained Ethics Review Board? Please mention this in the methodology

The Ethics Review Board (Institutional Review Board) is added including the number of the study, in accordance with PLOS ONE’s guidelines, under Methods – Selected Data.

I suggest moving this “This resulted in a mean of 886 notes per patient per year and a standard 52 deviation of 2180.” To the results section.

This has been moved to the Results

Where is the reference for lines 58,59?

The appropriate reference has been added.

What systems, programs, or program languages have you used? Please mention that

A new subheader ‘Software and Systems’ had been added to Methods.

Please justify and explain lines 96-99 in more detail.

This is further expanded upon in the Methods – Calculating Sentiment paragraph and includes a reference to the general practice of the Hedonometer, referencing the descriptive methodology paper by Dodds et al. (2011).

What did you use to perform the statistical analysis?

Python 3.8, specifically the scipy, statsmodels and scikit packages. This has been added to the Software and Systems paragraph.

Results:

Line 152 is missing table number

The table number has been added.

I suggest that explaining the meaning of the results like in line 192 to be moved to the discussion section.

This was moved to the Discussion.

Discussion

I appreciate that you compared your data to other studies

Please elaborate more on the explanation of the result’s meaning and how it can affect the practice. Also, you may mention some of the limitations of your study. Furthermore, clear mention of what further studies need to be done in the future would be important and would enrich your manuscript.

This has been elaborated upon and limitations have been added into the Discussion.

Attachment

Submitted filename: Response to Reviewers.docx

Decision Letter 2

Nabil Elhadi Elsayed Ali Omar

12 Jan 2023

Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs.

PONE-D-22-11193R2

Dear Dr. Elbers,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Nabil Elhadi Elsayed Ali Omar, PharmD.,BCOP.,PhD(C)

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

<quillbot-extension-portal></quillbot-extension-portal>

Acceptance letter

Nabil Elhadi Elsayed Ali Omar

13 Jan 2023

PONE-D-22-11193R2

Sentiment analysis of medical record notes for lung cancer patients at the Department of Veterans Affairs.

Dear Dr. Elbers:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Nabil Elhadi Elsayed Ali Omar

Academic Editor

PLOS ONE

Associated Data

    This section collects any data citations, data availability statements, or supplementary materials included in this article.

    Supplementary Materials

    S1 Table. Top 40 words based on rank of surrounding sentiment * text coverage.

    (CSV)

    S2 Table. Re-scoring outcomes from SME’s.

    (CSV)

    S3 Table. Medication treatment matrix.

    (CSV)

    Attachment

    Submitted filename: Response to Reviewers.docx

    Attachment

    Submitted filename: Response to Reviewers.docx

    Data Availability Statement

    Patient related and note data cannot be shared publicly because it involves sensitive human subject data. Data may be available for researchers who meet the criteria for access to confidential data after evaluation from VA Research and Development Committees. As a VA national legal policy (VHA Directive 1605.01), VA will only share patient data if there is a fully executed contract in place for the specific project. A common contractual mechanism utilized for this type of sharing is a ‘Cooperative Research and Development’ (CRADA) agreement. These contracts are typically negotiated in collaboration with VA national Office of General Council (OGC) and lawyers from the collaborating institution. These national sharing policies and standards also apply to deidentified data. In addition, if a contract is in place allowing sharing of deidentified data outside of VA, then VA national policy (VHA Directive 1605.01), states that deidentification certification needs to be met by Expert Determination. The expert determination requires independent assessment from an experienced master or PhD in biostatistics, from a third party not involved in the project, and may require outside funding to support. In addition, for an outside entity to preform research on VA patient data, IRB as well as VA Research and Development Committee approval is required for the specific project. Data requests may be sent to: VA Information Resource Center (VIReC) Building 18 Hines VA Hospital (151V) 5000 S. 5th Avenue Hines, IL 60141-3030 708-202-2413 708-202-2415 (fax) virec@va.gov.


    Articles from PLOS ONE are provided here courtesy of PLOS

    RESOURCES