Skip to main content
Lippincott Open Access logoLink to Lippincott Open Access
. 2024 Mar 7;42(14):1607–1611. doi: 10.1200/JCO.23.02439

Promise and Perils of Large Language Models for Cancer Survivorship and Supportive Care

Danielle S Bitterman 1,2,, Andrea Downing 3, Julia Maués 4, Maryam Lustberg 5
PMCID: PMC11095890  PMID: 38452323

Abstract

A call to action to bring stakeholders together to plan for the future of LLM-enhanced cancer survivorship.

Introduction

Cancer symptom toxicity affects patients and caregivers of all cancer types and stages, with effects on physiologic, psychologic, financial, and social well-being.1 Barriers to symptom assessment and management of patients with cancer include lack of survivorship care access and resources, incomplete data for risk-stratification, inadequate means for timely communication about symptoms, fragmented health care across multiple specialists, misconceptions and insufficient patient and clinician education, and changing needs over the cancer trajectory.1-7 Large language models (LLMs) are uniquely poised to address these barriers because of their ability to process, transfer, and transform high-dimensional knowledge, providing more individualized management and enabling multilevel communication that can adapt to users' needs. Here, we discuss the current status and future applications of LLMs for cancer symptom management and call for cross-disciplinary collaboration that centers the needs of patients and caregivers. Table 1 provides an overview of key opportunities and existing challenges.

TABLE 1.

Key Opportunities and Existing Challenges for LLMs in Cancer Survivorship and Supportive Care

Key Opportunity Example Challenges
EHR-based symptom/adverse event collection An LLM fine-tuned to identify adverse event mentions in clinical notes is used to build a large supportive care and survivorship registry Underreporting of symptoms/adverse events in EHR
Fragmented health care and data siloes
Ambiguities in definition and description of symptoms and adverse events
Outcome prediction An LLM fine-tuned to predict the occurrence of severe platinum-induced peripheral neuropathy provides risk estimates for individualized shared decision making about treatment options Symptom/adverse event outcomes are needed to fine-tune LLMs but are underdocumented
Institutional buy-in for operationalization and monitoring needed for successful implementation
Lack of interpretability into how LLMs arrive at their prediction
Algorithmic bias arising from historic differences in outcomes and access to care across population
Predictions could be used by other parties not in the service of patients (eg, denying insurance coverage)
Access to survivorship care LLM-enabled EHR automatically identifies billing codes and drafts notes, reducing clinicians' administrative burden and creating more time for them to see patients LLMs hallucinate and have risk of harmful language/bias learned from on pretraining data
Lack of automated metrics to evaluate quality of LLM performance over time
Risk of over-reliance on LLM output, automation bias, and clinician deskilling
Remote symptom monitoring and data collection An LLM chatbot proactively asks patients about symptoms on the basis of the treatments they have received. The chatbot adapts to patient responses and asks clarifying questions for high-quality reporting. Results are fed back to the EHR for clinician review Existing out-of-the-box LLMs often provide incorrect recommendations and are not ready for clinical use
Inequitable access to technologies (eg, computers, phones, broadband Internet, cellular service)
Careful protections needed to protect patient privacy for remote data collection and transfer
EHR vendor and institutional buy-in needed for integrating data from external apps
Not all patients and caregivers may want to interact with LLMs for their cancer care
Digital remote care could place more burden on patients and caregivers
LLMs hallucinate and have risk of harmful language/bias learned from on pretraining data
LLMs are sensitive to small changes in how they are prompted
Lack of task-specific, patient-centered evaluation metrics for longitudinal monitoring
Remote supportive care and survivorship management An LLM chatbot answers patient's questions about managing symptoms from home and providers appointment reminders. High-risk responses are triaged for urgent review by a clinician
Patient and clinician education An LLM chatbot provides a user-friendly interface to identify relevant supportive care and survivorship guidelines Could shift care burden from healthcare institutions and clinicians onto patients and caregivers
LLMs hallucinate and have risk of harmful language/bias learned from on pretraining data
LLMs are sensitive to small changes in how they are prompted
Lack of task-specific, patient-centered evaluation metrics for longitudinal monitoring
Patient-clinician communication An LLM translates clinical language into simpler language or other forms that are most preferred by each patient
Patient navigation An LLM summarizes recommendations from different specialists and future appointments

Abbreviations: EHR, electronic health record; LLM, large language model.

What are LLMs?

Natural language processing (NLP) is a field of artificial intelligence (AI) that enables computers to understand human language. NLP methods range from simple rules-based systems to statistical models to the current generation of machine learning models on the basis of deep neural networks.8,9 Language models refer to deep neural networks trained on vast amounts of text to process human language. Since 2018, a type of neural architecture known as transformers has achieved state-of-the-art NLP performance.10,11 Using previously unavailable computational power, transformer-based language models have been trained with increasing text data curated from the Internet, leading us to the current LLMs that are orders of magnitude larger than their predecessors, exemplified by ChatGPT.12,13

These newest LLMs are a type of generative AI, which refers to AI methods that create new content. LLMs have received most attention for their ability to carry out generative tasks, for example question-answering (eg, chatbot-type applications), summarization, and translation. LLMs can pass the US Medical Licensing Examination,14-17 provide empathetic responses to medical questions posted on Reddit,18 and translate biomedical and clinical jargon into more accessible summaries.19 LLMs can also be used for nongenerative tasks such as predicting outcomes20 and extracting information from text for automated real-world data collection.21

Potential of LLMs

There is a significant body of NLP literature on developing methods to automatically extract symptoms and adverse events from the electronic health records. Previous work relevant to cancer survivorship has used smaller language models to automate detection of provider-reported adverse events,22-24 symptoms,25 and patient-reported outcome characterization.26 Developing these models historically required a large quantity of manually labeled data to fine-tune the model for specific adverse events or symptoms—an expensive, expertise-intensive task. Better performing LLMs may require much less labeled fine-tuning data for this type of information extraction. We therefore anticipate LLM-driven advancements in our ability to automate electronic health record–based symptom surveillance. This could benefit patients at the individual level because their symptoms could be tracked more thoroughly, allowing more timely and precise management.8,9 Automated collection of symptoms and adverse events will enable larger, more comprehensive postmarket surveillance studies, reduce the resources required to track symptom toxicity on trials, and provide real-world evidence to better understand risks, associations between symptoms, and management strategies. However, symptoms are underreported in the electronic health record, and better collection of patient-reported symptoms is needed to realize this promise. As we discuss below, LLMs could also improve collection of these data.

An emerging application of LLMs for cancer symptom management is in developing models to predict outcomes. Recent work has shown that LLMs trained on clinical documents can predict in-hospital mortality, comorbidity index, 30-day readmission, and insurance denial prediction.20 Symptom risk prediction could support treatment decision making and personalized monitoring schedules by providing more individualized estimates of the risks and benefits of treatment. Multimodal models that combine clinical text information with other types of health data, including imaging, structured data, patient-reported outcomes, and patient-generated health data is a promising area of active research.27-32

LLMs may be most transformative in enabling patient-centered remote supportive care and symptom management and data collection, which are essential to improving cancer outcomes.33-36 Some studies have demonstrated a quality of life and survival benefit of digital health solutions for symptom monitoring and management, although results vary across studies and specific outcomes.35-37 Studies incorporating digital health solutions with AI have focused broadly on education, behavior change, and symptom management, particularly pain, anxiety, and depression. Patients report improved empowerment and communication with health care providers and real-time symptom monitoring when using these technologies.

LLMs could facilitate personalized interaction on the basis of up-to-date guidelines and proactive collection of patient-reported outcomes. Compared with chatbots for symptom collection, chatbots that help patients self-manage and triage are higher risk of harm because they directly affect management38 but could potentially reduce health care utilization and help patients avoid urgent care. Key opportunities for improvement include uniform standards for measuring patient-related outcomes, consistent research methods across studies, and integrated patient and clinician involvement.36 Well-designed studies are still needed to show effectiveness of LLMs in symptom management and usability.

The inherent strengths of LLMs in transforming language for summarization and translation create a promising foundation for automating, scaling, and personalizing patient-centered digital health. Methods that summarize and simplify our lengthy, jargon-full guidelines that are difficult even for experts to manage could be a major step forward for patient education and could be tailored to patient preference and background. Automated LLM summarization of clinical visits could potentially improve the longstanding challenge of communication between different clinical specialties, patients, and families—although data silos will need to be overcome for significant advances in this area. LLMs are excellent at machine translation, which has clear implications for improving access. LLMs and similar models are advancing to handle multiple data types,27-31 providing an avenue to further personalize content to how each patient prefers to consume it (eg, in writing, video, audio, infographic). Similarly, patient- and clinician-facing LLMs could improve education by navigating emerging evidence, guidelines, and management strategies in the context of each patient's history—all of which is increasingly important in our ever-evolving complex cancer care landscape.

Perils of LLMs

However, the promises of LLMs will only be realized if their harms are effectively studied and mitigated. Several technical, clinical, structural, and ethical challenges need to be addressed for effective deployment for cancer symptom management. Most importantly, patients and caregivers should contribute to LLM governance to prioritize ethical, sustainable development and implementation.39 Notably, patients' voices have been largely excluded from the AI conversation. Patients may experience real safety risks, including physical, emotional, and financial harm.40 This is especially concerning as it is the patients' data that are needed to develop and refine models, and their health care and privacy are most at risk with these applications. Patient-led advocacy groups such as The Light Collective41 are exploring ways that patient-centered governance of technology can be operationalized. In collaboration with stakeholders, more regulatory oversight from national and governmental organizations is needed to guide and safeguard implementation.42

Patients should provide foundational input on how and whether they want LLMs integrated into their care and on how they want their health data and public data used to develop models. Transparency is paramount—patients and the public in general should be made aware of how their digital data are used. Especially given the unique ability of LLMs to mimic human interaction, it should also be made clear to users when they are interacting with AI, and informed consent should be obtained as appropriate.39

Patients and their caregivers will have varying levels of interest and aptitude in learning and using this technology. New advances should expand and integrate—not replace—avenues for patients to engage in their care, including the option to opt-out of LLM-enabled digital health. Patients already need to manage and access multiple different applications and patient portal systems during their care. New digital health technologies should simplify systems and not add yet another complication to an already burdensome digital landscape. A few studies are beginning to ask important questions about overall acceptability of such advances including use of AI platforms to monitor distress.43

Patient engagement also requires aligning goals so that their health data are not exploited for the financial gain of developers and institutions while patients continue to face escalating cancer costs. Nefarious use and those not aligned with patient needs and wellbeing need to be anticipated. For example, predictive models could be used by payers and other parties to limit access to care for certain populations if adequate protections are not in place.

Technical issues also need to be addressed for effective clinical translation. One challenge arises from the fact that many LLMs are trained to give a response that a human wants to hear, not necessarily one that is factually correct. While this may be okay for casual interaction, it could be dangerous for health care—especially because LLMs can sound very convincing.44,45 Emerging evidence shows that while AI chatbots provide cancer treatment information that superficially appears reasonable, results are sub-par when compared held with clinical standards.46 How to ensure LLMs are updated as guidelines evolve is an open question and will require close collaboration between professional societies, researchers, and AI developers. LLMs can be sensitive to how they are prompted to give a response, and small changes in the phrasing of a question can yield vastly different responses.46 How to obtain consistent and factually correct information from generative LLMs is a research gap. Maintaining a human-in-the-loop will be an important safety check until there is clinical evidence that technologies are safe and effective without additional oversight.

As clinical translation proceeds, all stakeholders will need to work together to design evaluations. Importantly, there is a need for new standards for evaluating generative output. Computed metrics for generative text do not meet the bar for clinical acceptability,47,48 and patient-centered outcomes and user preferences should remain the standard by which we evaluate. There are no benchmark clinical data sets for cancer survivorship and these are urgently to ensure that best methods are prioritized for translation. In addition, federal and institutional regulatory guidance for longitudinal performance and usability monitoring will need to be carried out on these models given the inevitably of performance drift over time.49

While LLMs could democratize cancer care by making self-management and patient education resources widely available, they may also amplify cancer inequities and biases. Like all AI, LLMs learn the biases, racism, and prejudices present in the data that they are trained on.50 For example, LLMs could perpetuate ongoing inequities by providing differential recommendations on the basis of previous learned distributions of treatments, monitoring, and management practices across patient populations. In addition, technologies are not equally accessible and often require access and ability to use computers, broadband Internet, or cellular services.6,51,52 Investments in LLM technologies must be balanced with investments in access for those who need it most.

Our Call for Collaboration and Multidisciplinary Efforts

Cancer survivorship and supportive care is founded in patient-centered communication and education, uniquely poising LLMs to help advance the field. Patients, clinicians, and developers should each play a role throughout the development to deployment pathway. Systems should be designed around the needs of patients with cancer and not around the data that are most easily available. Clinicians and developers will need to work and continuously touch base with patients, including a wide and diverse range of voices, to define and prioritize efforts that align with these needs. Developers can help guide data and model design decisions that affect downstream performance and risks while clinicians and patients should provide insight into data availability, acceptable performance thresholds and tradeoffs, and appropriate evaluations. All stakeholders have critical experience and expertise to contribute to cocreating structures that govern LLMs.

Holding ourselves to high standards for evaluation, safety, and ethics will help avoid early missteps leading to medical errors and loss of patient and clinician trust in LLMs. Now is the time to bring all the stakeholders together as we envision the future of personalized supportive care and symptom management enhanced by evidence-based integration of AI approaches to deliver holistic and equitable care.

Danielle S. Bitterman

Employment: Brigham and Women's Hospital

Honoraria: Harvard Medical School

Other Relationship: AACR Project GENIE (Inst)

Uncompensated Relationships: HemOnc.org

Uncompensated Relationships: HemOnc.org

Open Payments Link: https://openpaymentsdata.cms.gov/physician/2607557

Julia Maués

Consulting or Advisory Role: MBQ Pharma

Maryam Lustberg

Honoraria: Novartis

Consulting or Advisory Role: Pfizer, AstraZeneca, Lilly, Gilead Sciences, Osmol Therapeutics

Other Relationship: Cynosure/Hologic

No other potential conflicts of interest were reported.

SUPPORT

Supported by NIH-NCI: U54CA274516-01A1 (D.S.B.).

AUTHOR CONTRIBUTIONS

Conception and design: All authors

Financial support: Danielle S. Bitterman

Provision of study materials or patients: Andrea Downing

Manuscript writing: All authors

Final approval of manuscript: All authors

AUTHORS' DISCLOSURES OF POTENTIAL CONFLICTS OF INTEREST

Promise and Perils of Large Language Models for Cancer Survivorship and Supportive Care

The following represents disclosure information provided by authors of this manuscript. All relationships are considered compensated unless otherwise noted. Relationships are self-held unless noted. I = Immediate Family Member, Inst = My Institution. Relationships may not relate to the subject matter of this manuscript. For more information about ASCO's conflict of interest policy, please refer to www.asco.org/rwc or ascopubs.org/jco/authors/author-center.

Open Payments is a public database containing information reported by companies about payments made to US-licensed physicians (Open Payments).

Danielle S. Bitterman

Employment: Brigham and Women's Hospital

Honoraria: Harvard Medical School

Other Relationship: AACR Project GENIE (Inst)

Uncompensated Relationships: HemOnc.org

Uncompensated Relationships: HemOnc.org

Open Payments Link: https://openpaymentsdata.cms.gov/physician/2607557

Julia Maués

Consulting or Advisory Role: MBQ Pharma

Maryam Lustberg

Honoraria: Novartis

Consulting or Advisory Role: Pfizer, AstraZeneca, Lilly, Gilead Sciences, Osmol Therapeutics

Other Relationship: Cynosure/Hologic

No other potential conflicts of interest were reported.

REFERENCES

  • 1. Lustberg MB, Kuderer NM, Desai A, et al. Mitigating long-term and delayed adverse events associated with cancer treatment: Implications for survivorship. Nat Rev Clin Oncol. 2023;20:527–542. doi: 10.1038/s41571-023-00776-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2. Nayak M, George A, Vidyasagar MS, et al. Symptoms experienced by cancer patients and barriers to symptom management. Indian J Palliat Care. 2015;21:349–354. doi: 10.4103/0973-1075.164893. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3. Yeom H-E, Heidrich SM. Effect of perceived barriers to symptom management on quality of life in older breast cancer survivors. Cancer Nurs. 2009;32:309–316. doi: 10.1097/NCC.0b013e31819e239e. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4. Nayak M, George A, Vidyasagar MS. Perceived barriers to symptoms management among family caregivers of cancer patients. Indian J Palliat Care. 2018;24:202–206. doi: 10.4103/IJPC.IJPC_27_18. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5. Dupuis LL, Grimes A, Vettese E, et al. Barriers to symptom management care pathway implementation in pediatric Cancer. BMC Health Serv Res. 2021;21:1068. doi: 10.1186/s12913-021-07047-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 6. Nguyen H, Butow P, Dhillon H, et al. A review of the barriers to using Patient-Reported Outcomes (PROs) and Patient‐Reported Outcome Measures (PROMs) in routine cancer care. J Med Radiat Sci. 2021;68:186–195. doi: 10.1002/jmrs.421. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 7. Schepers SA, Haverman L, Zadeh S, et al. Healthcare professionals’ preferences and perceived barriers for routine assessment of patient-reported outcomes in pediatric oncology practice: Moving toward international processes of change. Pediatr Blood Cancer. 2016;63:2181–2188. doi: 10.1002/pbc.26135. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8. Bitterman DS, Miller TA, Mak RH, et al. Clinical natural language processing for radiation oncology: A review and practical primer. Int J Radiat Oncol Biol Phys. 2021;110:641–655. doi: 10.1016/j.ijrobp.2021.01.044. [DOI] [PubMed] [Google Scholar]
  • 9. Savova GK, Danciu I, Alamudun F, et al. Use of natural language processing to extract clinical cancer phenotypes from electronic medical records. Cancer Res. 2019;79:5463–5470. doi: 10.1158/0008-5472.CAN-19-0579. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Vaswani A, Shazeer N, Parmar N, et al. NeurIPS Proceedings. 2017. Attention is all you need.https://proceedings.neurips.cc/paper_files/paper/2017/file/3f5ee243547dee91fbd053c1c4a845aa-Paper.pdf [Google Scholar]
  • 11.Devlin J, Chang M-W, Lee K, et al. Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers) Minneapolis, Minnesota: Association for Computational Linguistics; 2019. BERT: Pre-Training of deep bidirectional transformers for language understanding; pp. 4171–4186. [Google Scholar]
  • 12.OpenAI API. http://platform.openai.com [Google Scholar]
  • 13.Ouyang L, Wu J, Jiang X, et al. NeurIPS Proceedings. 2022. Training language models to follow instructions with human feedback.https://proceedings.neurips.cc/paper_files/paper/2022/hash/b1efde53be364a73914f58805a001731-Abstract-Conference.html [Google Scholar]
  • 14.Singhal K, Tu T, Gottweis J, et al. Towards Expert-Level Medical Question Answering with Large Language Models. 2023. http://arxiv.org/abs/2305.09617 arXiv [csCL] [Google Scholar]
  • 15.Singhal K, Azizi S, Tu T, et al. Large Language Models Encode Clinical Knowledge. 2022. http://arxiv.org/abs/2212.13138 arXiv [csCL] [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 16.Nori H, King N, McKinney SM, et al. Capabilities of GPT-4 on Medical Challenge Problems. 2023. http://arxiv.org/abs/2303.13375 arXiv [csCL] [Google Scholar]
  • 17.Nori H, Lee YT, Zhang S, et al. Can Generalist Foundation Models Outcompete Special-Purpose Tuning? Case Study in Medicine. 2023. http://arxiv.org/abs/2311.16452 arXiv [csCL] [Google Scholar]
  • 18. Ayers JW, Poliak A, Dredze M, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183:589–596. doi: 10.1001/jamainternmed.2023.1838. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 19.Goldsack T, Luo Z, Xie Q, et al. The 22nd Workshop on Biomedical Natural Language Processing and BioNLP Shared Tasks. Toronto, Canada: Association for Computational Linguistics; 2023. BioLaySumm 2023 shared task: Lay summarisation of biomedical research articles; pp. 468–477. [Google Scholar]
  • 20. Jiang LY, Liu XC, Nejatian NP, et al. Health system-scale language models are all-purpose prediction engines. Nature. 2023;619:357–362. doi: 10.1038/s41586-023-06160-y. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Guevara M, Chen S, Thomas S, et al. Large Language Models to Identify Social Determinants of Health in Electronic Health Records. NPJ Digit Med 7:62024 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22. Hong JC, Fairchild AT, Tanksley JP, et al. Natural language processing for abstraction of cancer treatment toxicities: Accuracy versus human experts. JAMIA Open. 2020;3:513–517. doi: 10.1093/jamiaopen/ooaa064. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23. Chen S, Guevara M, Ramirez N, et al. Natural language processing to automatically extract the presence and severity of esophagitis in notes of patients undergoing radiotherapy. JCO Clin Cancer Inform. doi: 10.1200/CCI.23.00048. 10.1200/CCI.23.00048 [DOI] [PubMed] [Google Scholar]
  • 24. Gupta S, Belouali A, Shah NJ, et al. Automated identification of patients with immune-related adverse events from clinical notes using word embedding and machine learning. JCO Clin Cancer Inform. doi: 10.1200/CCI.20.00109. 10.1200/CCI.20.00109 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25. Lindvall C, Deng C-Y, Agaronnik ND, et al. Deep learning for cancer symptoms monitoring on the basis of electronic health record unstructured clinical notes. JCO Clin Cancer Inform. doi: 10.1200/CCI.21.00136. 10.1200/CCI.21.00136 [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 26. Lu Z, Sim J-A, Wang JX, et al. Natural Language processing and machine learning methods to characterize unstructured patient-reported outcomes: Validation study. J Med Internet Res. 2021;23:e26777. doi: 10.2196/26777. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27. Meskó B. The impact of multimodal large language models on health care’s future. J Med Internet Res. 2023;25:e52865. doi: 10.2196/52865. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 28. Moor M, Banerjee O, Abad ZSH, et al. Foundation models for generalist medical artificial intelligence. Nature. 2023;616:259–265. doi: 10.1038/s41586-023-05881-4. [DOI] [PubMed] [Google Scholar]
  • 29.Moor M, Huang Q, Wu S, et al. Med-flamingo: A Multimodal Medical Few-Shot Learner. 2023. http://arxiv.org/abs/2307.15189 arXiv [csCV] [Google Scholar]
  • 30.Yang Z, Li L, Lin K, et al. The Dawn of LMMs: Preliminary Explorations with GPT-4V(ision) 2023. http://arxiv.org/abs/2309.17421 arXiv [csCV] [Google Scholar]
  • 31.Radford A, Kim JW, Hallacy C, et al. Learning Transferable Visual Models from Natural Language Supervision. 2021. http://arxiv.org/abs/2103.00020 arXiv [csCV] [Google Scholar]
  • 32. Lin H, Ginart JB, Chen W, et al. OncoBERT: Building an interpretable transfer learning bidirectional encoder representations from transformers framework for longitudinal survival prediction of cancer patients. Research Square. 2023 https://www.researchsquare.com/article/rs-3158152/v1 [Google Scholar]
  • 33. Jordan K, Aapro M, Kaasa S, et al. European Society for Medical Oncology (ESMO) position paper on supportive and palliative care. Ann Oncol. 2018;29:36–43. doi: 10.1093/annonc/mdx757. [DOI] [PubMed] [Google Scholar]
  • 34. Kaasa S, Loge JH, Aapro M, et al. Integration of oncology and palliative care: A lancet oncology commission. Lancet Oncol. 2018;19:e588–e653. doi: 10.1016/S1470-2045(18)30415-7. [DOI] [PubMed] [Google Scholar]
  • 35. Aapro M, Bossi P, Dasari A, et al. Digital health for optimal supportive care in oncology: Benefits, limits, and future perspectives. Support Care Cancer. 2020;28:4589–4612. doi: 10.1007/s00520-020-05539-1. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36. Pan L-C, Wu X-R, Lu Y, et al. Artificial intelligence empowered digital health technologies in cancer survivorship care: A scoping review. Asia Pac J Oncol Nurs. 2022;9:100127. doi: 10.1016/j.apjon.2022.100127. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37. Shaffer KM, Turner KL, Siwik C, et al. Digital health and telehealth in cancer care: A scoping review of reviews. Lancet Digit Health. 2023;5:e316–e327. doi: 10.1016/S2589-7500(23)00049-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Chen S, Guevara M, Moningi S, et al. The impact of responding to patient messages with large language model assistance. 2023. http://arxiv.org/abs/2310.17703 arXiv [csCL] [Google Scholar]
  • 39. Perni S, Lehmann LS, Bitterman DS. Patients should be informed when AI systems are used in clinical trials. Nat Med. 2023;29:1890–1891. doi: 10.1038/s41591-023-02367-8. [DOI] [PubMed] [Google Scholar]
  • 40. Perakslis ED, Ranney ML, Goldsack JC. Characterizing cyber harms from digital health. Nat Med. 2023;29:528–531. doi: 10.1038/s41591-022-02167-6. [DOI] [PubMed] [Google Scholar]
  • 41.Light Collective—No Aggregation without Representation. https://lightcollective.org/ [Google Scholar]
  • 42. Meskó B, Topol EJ. The imperative for regulatory oversight of large language models (or generative AI) in healthcare. NPJ Digit Med. 2023;6:120. doi: 10.1038/s41746-023-00873-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43. Zhang A, Hu H, Acquati C, et al. Evaluating the feasibility and acceptability of an artificial-intelligence-enabled and speech-based distress screening mobile app for adolescents and young adults diagnosed with cancer: A study protocol. Cancers. 2022;14:1450. doi: 10.3390/cancers14040914. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Si C, Goyal N, Wu ST, et al. Large Language Models Help Humans Verify Truthfulness—Except when They Are Convincingly Wrong. 2023. http://arxiv.org/abs/2310.12558 arXiv [csCL] [Google Scholar]
  • 45.Sharma M, Tong M, Korbak T, et al. Towards Understanding Sycophancy in Language Models. 2023. http://arxiv.org/abs/2310.13548 arXiv [csCL] [Google Scholar]
  • 46. Chen S, Kann BH, Foote MB, et al. Use of artificial intelligence chatbots for cancer treatment information. JAMA Oncol. 2023;9:1459–1462. doi: 10.1001/jamaoncol.2023.2954. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Lin C-Y. Text Summarization Branches Out. Barcelona, Spain: Association for Computational Linguistics; 2004. Rouge: A package for automatic evaluation of summaries; pp. 74–81. [Google Scholar]
  • 48.Zhang T, Kishore V, Wu F, et al. BERTScore: Evaluating Text Generation with BERT. 2019. http://arxiv.org/abs/1904.09675 arXiv [csCL] [Google Scholar]
  • 49. Finlayson SG, Subbaswamy A, Singh K, et al. The clinician and dataset shift in artificial intelligence. N Engl J Med. 2021;385:283–286. doi: 10.1056/NEJMc2104626. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Zhao J, Wang T, Yatskar M, et al. Proceedings of the 2017 Conference on Empirical Methods in Natural Language Processing. Copenhagen, Denmark: Association for Computational Linguistics; 2017. Men also like shopping: Reducing gender bias amplification using corpus-level constraints; pp. 2979–2989. [Google Scholar]
  • 51. Potdar R, Thomas A, DiMeglio M, et al. Access to internet, smartphone usage, and acceptability of mobile health technology among cancer patients. Support Care Cancer. 2020;28:5455–5461. doi: 10.1007/s00520-020-05393-1. [DOI] [PubMed] [Google Scholar]
  • 52. Berkowitz CM, Zullig LL, Koontz BF, et al. Prescribing an app? Oncology providers’ views on mobile health apps for cancer care. JCO Clin Cancer Inform. doi: 10.1200/CCI.17.00107. 10.1200/CCI.17.00107 [DOI] [PubMed] [Google Scholar]

Articles from Journal of Clinical Oncology are provided here courtesy of Wolters Kluwer Health

RESOURCES