Abstract
Background: As ChatGPT becomes a primary information source for college students, its performance in providing dietary advice is under scrutiny. This study assessed ChatGPT’s performance in providing nutritional guidance to college students. Methods: ChatGPT’s performance on dietary advice was evaluated by 30 experienced dietitians and assessed using an objective nutrition literacy (NL) test. The dietitians were recruited to assess the quality of ChatGPT’s dietary advice, including its NL achievement and response quality. Results: The results indicate that ChatGPT’s performance varies across scenarios and is suboptimal for achieving NL with full achievement rates from 7.50% to 37.56%. While the responses excelled in readability, they lacked understandability, practicality, and completeness. In the NL test, ChatGPT showed an 84.38% accuracy rate, surpassing the NL level of Taiwanese college students. The top concern among the dietitians, cited 52 times in 242 feedback entries, was that the “response information lacks thoroughness or rigor, leading to misunderstandings or misuse”. Despite the potential of ChatGPT as a supplementary educational tool, significant gaps must be addressed, especially in detailed dietary inquiries. Conclusion: This study highlights the need for improved AI educational approaches and suggests the potential for developing ChatGPT teaching guides or usage instructions to train college students and support dietitians.
Keywords: ChatGPT, dietary dvice, health education, nutrition literacy
1. Introduction
In an era marked by the pervasive influence of digital technologies, the health and wellness landscape has undergone a notable transformation because of the emergence of artificial intelligence (AI)-driven platforms [1]. Among these advancements are conversational large language models (LLMs) such as ChatGPT, which have revolutionized public health education [2,3]. Notably, ChatGPT serves as a significant digital assistant, offering users a diverse range of information and personalized recommendations, including health advice customized to their specific needs and preferences [4].
Platforms such as ChatGPT have gained worldwide attention for their impressive performance in producing well-structured, logical, and informative responses [5]. The integration of AI into health counseling offers promising benefits for encouraging individuals to adopt healthier practices. In a medical study, experienced thoracic surgical clinicians assessed the feasibility of using ChatGPT for perioperative patient education regarding thoracic surgery. The results indicated that 92% of the responses met the qualification criteria, indicating ChatGPT’s potential feasibility for patient education [6]. Additionally, studies related to cancer education [4], dermatology education [7], diabetes education [8], or other diseases [9] found ChatGPT to be applicable for clinical health education in various evaluation indicators.
However, in another study assessing ChatGPT’s patient-education materials for implant-based breast reconstruction, it was found that although ChatGPT could generate materials more rapidly, its readability was poorer compared to official materials. Additionally, the ChatGPT-generated content had 50% accuracy with most errors being information errors [10]. Another study focusing on men’s health found ChatGPT-generated content to have worse understandability than traditional patient-education materials [11]. ChatGPT was also found to be better than Google Search for providing general medical knowledge but worse for medical recommendations [12].
In the realm of dietary education, ChatGPT-3.5 demonstrates potential as an efficient tool for renal dietary planning in patients with chronic kidney disease [13]. This underscores its potential despite its lower accuracy rate compared to ChatGPT-4.0. However, even ChatGPT-4.0 has produced errors, for example in creating menus for vegetarians [14]. One study delved into its credibility in providing dietary advice for individuals with food allergies and found that, while generally accurate, ChatGPT had a propensity to generate harmful dietary recommendations. Common errors often involve inaccuracies in food portions, calorie estimations for meals, or overall diet composition [15]. While AI integration, such as using ChatGPT, in health counseling shows promise for healthier practices, limitations such as quality and accuracy issues in specific contexts must be addressed for better outcomes.
During the transition to college, young adults frequently develop unhealthy eating habits, as shown by research in diverse regions, including Taiwan [16,17,18]. These habits increase the risk of rapid weight gain and nutritional deficiencies with potential long-term health consequences [19]. This makes it crucial to establish healthy eating habits [20]. Nutrition education becomes imperative in responding to these obstacles by employing diverse approaches to promote healthy eating habits and cultivating positive behaviors regarding nutrition [21].
With the rise of AI technology, college students are also beginning to utilize ChatGPT for learning. College students have expressed concerns about the quality and reliability of information sources; however, overall, they hold a positive attitude toward using ChatGPT [22]. Therefore, the use of this tool for nutrition education should be considered as a future intervention trend. Given the specialized knowledge and expertise possessed by dietitians in the nutrition field, their perspectives are invaluable for evaluating the efficacy and reliability of AI-generated dietary recommendations, especially in vulnerable populations such as college students.
This study’s primary objective was to evaluate ChatGPT’s performance in providing dietary advice to college students utilizing a comprehensive approach that included dietitians’ perspectives on nutrition literacy (NL) achievement, quality indicators, and an objective NL test. Multidimensional evaluation enables the assessment of dietary recommendations from both objective and subjective perspectives. It also identifies potential limitations and areas for improvement. These findings are expected to guide the development of AI-driven dietary counseling tools and emphasize the importance of expert perspectives in evaluating digital health interventions.
2. Methods
This study implemented a multidimensional evaluation methodology, as illustrated in Figure 1. The figure presents the objective and subjective assessments adopted in this study as well as the sources and acquisition methods of the data used for evaluating ChatGPT. The details are described in the following Section 2.1, Section 2.2 and Section 2.3.
2.1. ChatGPT Input and Data Sources
Given that NL significantly influences healthy dietary behavior, the NL indicators for college students in Taiwan were established by dietitians through the Delphi consensus process [23]. Subsequently, a scenario-based online program was developed and evaluated based on these corresponding indicators [24]. To explore the common dietary challenges faced by college students, this study employed five realistic scenarios from the program. The central hypothesis posited that students would turn to ChatGPT to assist in resolving these scenarios. To ensure that the prompts used in ChatGPT were suitable for nonprofessional individuals, this study recruited 20 students from nonmedical departments using convenience sampling. These participants had prior experience using any version of ChatGPT to seek answers. In the online survey, the participants were asked to envision themselves as protagonists within the scenarios and articulate how they would seek solutions from ChatGPT by recording their inquiry methods (prompts). The research team then analyzed the collected data, identifying the most prevalent prompts used by the participants when consulting ChatGPT (see Supplementary Materials).
2.2. Response Generation with ChatGPT
In this study, ChatGPT-3.5 was chosen for use because of its status as a free version, making it more accessible as a tool for public health education and consultation compared to the paid versions. The “New Chat” function was used to input questions sequentially and independently to facilitate each question’s processing. Subsequently, the responses generated by ChatGPT were recorded and documented. The Supplementary Materials include prompts and their corresponding ChatGPT-generated responses for the five scenarios.
2.3. Assessing ChatGPT’s Response and NL Test Performance
Thirty dietitians meeting the following specific eligibility criteria were invited to participate: (1) possessing a valid Taiwanese dietitian license and (2) having at least three years of experience in nutrition counseling. These participants were recruited through the personal networks of the research team members using a snowball sampling method. An online survey was conducted to evaluate the responses generated by ChatGPT. The questionnaire consisted of three parts. First, the participants assessed the achievement level of the NL indicators by rating the extent to which ChatGPT responses in various scenarios aligned with the corresponding NL indicators for college students. The ratings ranged from “not achieved at all” to “partially achieved” and “fully achieved”. Second, seven criteria used in previous studies evaluating online health information [25] were selected according to ChatGPT characteristics to assess the ChatGPT response quality. These criteria include accuracy (whether a source or information is consistent with agreed-upon scientific findings), currency (whether a source or information is up to date), completeness (whether necessary or expected aspects of a subject or topic are provided), understandability (whether a source or information has appropriate depth, quantity, specificity, and error-free), readability (whether the information is presented in a form that is easy to read), relevance (whether the information is relevant to the topic of interest or information seekers’ situation and background), and practicality (whether the information can be readily applied by an individual). The respondents rated each criterion on a Likert scale ranging from 1 to 10 with higher scores indicating better performance. The ratings greater than 7 were labeled as “acceptable”. Finally, the participants provided open-ended feedback expressing their appreciation and concerns regarding the ChatGPT responses. They also shared opinions on using ChatGPT for nutritional education among college students.
To further assess ChatGPT’s ability to provide nutrition education, this study employed a published test designed to evaluate the NL of Taiwanese college students [26]. The test questions were presented to ChatGPT, and the proportion of correct answers given by ChatGPT was compared with the results of previous studies on Taiwanese college students’ NL.
2.4. Statistical Analysis
Due to the small sample size of this study and the focus on understanding the evaluation performance, the collected data were subjected to descriptive statistical analysis. Both parametric and nonparametric methods were employed to provide a comprehensive view of the data. Parametric measures such as means and standard deviations were used to illustrate central tendencies and variability, while nonparametric measures such as medians and ranges were included to highlight the diversity of evaluations among the participants. Additionally, the distribution of ChatGPT’s performance ratings was illustrated using a stacked bar chart and a line chart to provide a clear visual representation of the data. The correct rate (%) of ChatGPT’s ability on the NL test is also demonstrated. The perspectives of the dietitians regarding ChatGPT were organized, and their evaluation of the information was summarized. Additionally, the NL test accuracy rates were compared between ChatGPT and college students.
3. Results
3.1. The Achievement Level of NL Indicators
As presented in Table 1, the participants evaluated the achievement of NL indicators in dietary advice provided by ChatGPT across five scenarios. Among the corresponding indicators in each scenario, the fully achieved rates ranged from 7.50% to 37.56%. Conversely, in situations where achievement was not attained at all, the percentages ranged from 20.00% to 63.33%. This suggests that the participants generally believe there is room for improvement in ChatGPT’s response content to meet the expected corresponding NL indicators. Additionally, in terms of indicator achievement, Scenario 4: Fewer processed foods showed the best performance (fully achieved: 35.56%; not achieved at all: 20.00%).
Table 1.
Not Achieved at All | Partially Achieved | Fully Achieved | |
---|---|---|---|
Scenario (No. of Indicators) | n (%) | n (%) | n (%) |
S1: Balanced Eating (8) | 137 (57.08) | 77 (32.08) | 26 (10.38) |
S2: Fit Weight (8) | 151 (62.92) | 71 (29.58) | 18 (7.50) |
S3: Dining Out Well (5) | 65 (43.33) | 46 (30.67) | 39 (26.00) |
S4: Fewer Processed Foods (3) | 18 (20.00) | 40 (44.44) | 32 (35.56) |
S5: Limit Sugary Drinks (4) | 76 (63.33) | 34 (28.33) | 10 (8.33) |
Note: The evaluation of ChatGPT’s performance was completed by 30 dietitians. The total number of evaluation result items for each scenario is equal to the number of corresponding indicators multiplied by 30.
3.2. Evaluation of Response Quality
Table 2 and Figure 2 and Figure 3 show that, among the seven evaluation indicators, readability consistently performed well across various scenarios (mean: 7.97–8.27, median: 8). The acceptability ratings (>7) fell within the 67%–80% range. In contrast, the three indicators with relatively poor performance were understandability (mean: 6.23–7.40, median: 7–8), practicality (mean: 6.30–7.67, median: 7–7.5), and completeness (mean: 6.70–7.90, median: 6.5–8). These three indicators received acceptable ratings with understandability and practicality having ratings not exceeding 50% in the four scenarios and completeness having ratings below 50% in three scenarios.
Table 2.
S1: Balanced Eating | S2: Fit Weight | S3: Dining Out Well | S4: Fewer Processed Foods | S5: Limit Sugary Drinks | |||||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
Criteria | Range | M (SD) | Median | Range | M (SD) | Median | Range | M (SD) | Median | Range | M (SD) | Median | Range | M (SD) | Median |
Accuracy | 4–10 | 7.90 (1.73) | 8 | 3–10 | 7.53 (1.94) | 8 | 3–10 | 7.47 (2.00) | 8 | 2–10 | 7.90 (1.85) | 8 | 2–10 | 6.60 (2.14) | 7 |
Currency | 3–10 | 7.70 (1.73) | 8 | 3–10 | 7.27 (1.91) | 7 | 2–10 | 7.27 (2.29) | 8 | 2–10 | 7.73 (1.93) | 8 | 3–10 | 6.97 (1.69) | 7 |
Completeness | 3–10 | 7.27 (1.89) | 8 | 3–10 | 6.70 (1.93) | 6.5 | 3–10 | 6.73 (2.29) | 7 | 3–10 | 7.90 (1.73) | 8 | 3–10 | 6.80 (1.96) | 7 |
Understandability | 1–10 | 7.00 (2.20) | 7 | 1–10 | 6.23 (1.98) | 6.5 | 1–10 | 6.37 (2.36) | 7 | 3–10 | 7.40 (1.96) | 8 | 2–10 | 6.70 (2.25) | 7.5 |
Readability | 4–10 | 8.03 (1.85) | 8 | 4–10 | 8.03 (1.71) | 8 | 3–10 | 7.97 (2.04) | 8 | 3–10 | 8.27 (1.72) | 8 | 3–10 | 8.00 (1.84) | 8 |
Relevance | 1–10 | 7.43 (2.27) | 8 | 1–10 | 6.57 (2.49) | 7 | 3–10 | 6.83 (2.17) | 7 | 3–10 | 7.83 (1.90) | 8 | 3–10 | 6.83 (1.95) | 7 |
Practicality | 1–10 | 6.77 (2.37) | 7.5 | 1–10 | 6.30 (2.14) | 7 | 3–10 | 6.63 (2.19) | 7 | 3–10 | 7.67 (1.88) | 8 | 3–10 | 7.23 (1.91) | 7.5 |
Note: M: mean; SD: standard deviation.
Table 3 shows that the participants provided 242 feedback entries regarding the use of ChatGPT for nutrition education among college students. Among these, positive opinions occurred 87 times (35.95%), whereas concerns occurred 155 times (64.05%). Positive opinions that appeared more than 20 times include “A02 Provides popularized, preliminary, and easy-to-understand dietary advice” (29 times) and “A04 Provides comprehensive and detailed health education information” (21 times). For the former, the dietitians mostly believe that ChatGPT’s responses can “provide basic dietary concepts and suggestions using simple replies” (#4, representing dietitian No.4, and the following as well) or “quickly convey needed nutritional knowledge for the general public” (#9). Regarding the latter, the dietitians stated that the information is “sufficiently clear” (#19) and “can rapidly organize usable educational information with high completeness” (#17).
Table 3.
Positive Opinions (Occurrences = 87, 35.95%) | Concerns (Occurrences = 155, 64.05%) |
---|---|
|
|
Note: A total of 242 feedback entries.
In the concern category, “B02 Provides health education information lacking thoroughness or rigor, leading to misunderstandings or misuse” appeared 52 times, making it the most frequent concern. Related suggestions include “S5: Suggesting homemade fruit juice as a thirst-quenching method, but it may lead to a higher risk of hyperlipidemia or fatty liver in the long run” (#21), “S1: When mentioning lean meat, there is no specific clarification on which part of beef or pork is considered lean meat, which may cause misunderstanding” (#14), and “S3: Due to the lack of clear guidance on nutritional labeling, sodium recommendations, or nutritional claim regulations, readers still do not understand which products to choose after reading” (#1). The second most frequent concern is “B07 Health education information is incorrect, especially regarding food categorization” (23 times). The dietitians pointed out various errors in ChatGPT’s responses, such as “S1: Food categorization does not include the category of ‘legumes’” (#31), “S4: Pumpkin is not classified as a vegetable” (#13), “S3: Grilled meat involves the use of barbecue sauce, and it should not be listed as a healthy option” (#14), and “S4: The option of ham in high-quality protein choices is contradictory to the content of reducing processed food intake, as it is a high-calorie and high-sodium food” (#14). Another concern mentioned 20 times is “B06 Unable to provide personalized (dietary or exercise) analysis and recommendations”. The dietitians believe that ChatGPT’s responses have limitations in providing personalized recommendations, as expressed in comments like “Lacks personalized dietary analysis and recommendations” (#5), “Does not provide recommendations based on individual blood biochemical values, which may cause harm to the body” (#6), and “Can only provide very basic nutritional directions and does not adequately address individual needs” (#31).
3.3. ChatGPT’s Performance on the NL Test
As indicated in Table 4, the NL test used here comprises 32 items. A survey conducted among college students in Taiwan [26] revealed an average correct score of 77.40%. In comparison, ChatGPT exhibited an accuracy rate of 84.38% (27 out of 32 items correct), surpassing the NL level of Taiwanese college students. That said, among the incorrectly answered questions, the percentage of correct answers among the college students in Taiwan ranged from 54.2% to 73.6%. This suggests that ChatGPT did not perform well, even regarding moderately difficult questions.
Table 4.
Items and Correct Answer | ChatGPT Responses | The Correct Rate of Taiwanese College Students [26] |
---|---|---|
10. According to Rui’s diet list today †, what do you think about his dietary choices: A. Vegetable servings are just right. B. Too many sugary drinks. C. Calories from fats are too high (T). |
B | 73.6% |
11. “Research shows that being overweight or obese (i.e., BMI ≥ ___) is a major risk factor for chronic diseases such as diabetes, cardiovascular disease, and malignant tumors”. The number in the blank should be †: A. 24 (T) B. 27 C. 30 |
C | 54.2% |
14. Rui spends most of his time either attending classes or returning to the dormitory to study. He hasn’t participated in any extracurricular activities and does not have a regular exercise routine. Based on Table †, Rui’s daily calorie requirement should be: A. 2450 kcal B. 2100 kcal C. 1400~1750 kcal (T) |
B | 54.2% |
18. Which of the following products † has the highest total calorie content? A. Product: Lemon-filled cookies, Weight: 80 g (T) B. Product: Braised beef noodles, Weight: 100 g C. Product: Chocolate milk, Volume: 300 milliliters |
B | 62.3% |
29. Since Rui has never had a habit of exercising in the past, he is just starting to develop an exercise routine. Which of the following exercise plans is more suitable for Rui? A. Take a 40-min walk around the campus after dinner every day with the principle of slight breathlessness and slightly increased heart rate (T). B. Take a 30-min walk around the campus after dinner every day with the principle of no breathlessness and no increased heart rate. C. Engage in more strenuous and breathless exercise at the gym for at least 60 min every day. |
B | 70.1% |
Correct rates of 32 items | 84.38% | 77.40% |
Note: † 10-provided with the diet list; 11-provided with the BMI classification; 14-provided with the daily activity level, weight, and calorie requirement comparison data; 18-provided with the product calorie labeling for each product.
4. Discussion
This is the first study to evaluate the quality of ChatGPT-generated diet-related responses using a multidimensional evaluation approach. The primary findings revealed that, while ChatGPT is proficient in providing prompt dietary guidance, it exhibits variable performance when evaluated by registered dietitians. These findings underscore AI’s potential as a supplementary tool in nutrition education but also emphasize the need to enhance its performance to manage nuanced and personalized dietary inquiries.
Regarding the dietitians’ perspective on ChatGPT’s NL achievement, the rates for fully meeting the NL indicators were all below 40 percent with significant variability across different scenarios. This indicates ChatGPT’s inconsistency and relatively low effectiveness in providing fully adequate educational information depending on the dietary scenario being addressed. NL education is an innovative element in nutrition education today that enhances individuals’ ability to make healthy dietary decisions [27]. Effective NL intervention can significantly alter college students’ dietary behaviors [24]. Consequently, an increasing number of nutrition education programs have been using enhancements in NL as their main intervention strategy [27,28,29].
While using ChatGPT-3.5 as a source of health education information, this study found that it could only provide corresponding nutrition knowledge responses. It has yet to enhance the level of its educational approach. These results suggest that future AI design and training should actively integrate multiple health behavior theories to offer additional specific operational suggestions for healthy dietary behaviors [30]. Currently, because ChatGPT can only serve to provide educational information, the NL-related education strategies should be provided by dietitians or health-related educators. Therefore, it is important to develop courses and guidelines to deliver NL intervention strategies.
Furthermore, among the seven indicators established here, the dietitians found that ChatGPT’s responses performed best for readability, whereas its understandability, practicality, and completeness were the least effective. These outcomes were similarly reflected in the dieticians’ feedback, which noted that, despite the generalized nature of ChatGPT responses, they often lacked the specificity and depth required for individual dietary counseling. Despite its strengths, a higher proportion of concerns were raised regarding the use of ChatGPT in nutrition education. Numerous studies have indicated that the materials produced by ChatGPT are highly readable [6,7,31], making this a likely ChatGPT characteristic. It can present information in an easily readable and organized manner, making it more accessible and absorbable.
In professional health education, however, nutrition counseling is specialized. Dietitians work as licensed healthcare professionals tasked with safeguarding public dietary health. In this context, the most crucial aspect is to tailor dietary plans to individual health needs [32]. This study found that, as most inquirers did not adequately disclose their personal health conditions to ChatGPT, the responses produced did not suit individual physiological characteristics. This poses a significant risk in nutritional counseling. Over one-third of the dietitians’ concerns about ChatGPT involved its “lacking thoroughness or rigor, leading to misunderstandings or misuse”. Consequently, dietitians often believe that such information cannot be adequately understood or utilized without professional assistance, a finding consistent with previous research [33,34,35].
This study employed prompts that closely mirror those typically used by college students when querying ChatGPT. However, these prompts may not adhere to the established principles of crafting effective ChatGPT prompts [36], which significantly influences response quality. Given that ChatGPT has become a prevalent source of information for college students [22], a critical need exists to develop and implement educational programs that teach college students how to effectively use it as a source of health information. Simultaneously, the awareness of ChatGPT’s limitations must be enhanced [4]. This approach will not only enable ChatGPT to serve as a supplementary tool in health education but also mitigate the risks associated with using it. Additionally, future research could examine the correlation between the quality and depth of prompts and the quality of health education information generated by ChatGPT, which would further underscore the importance of such educational programs.
The results from the objective NL test also suggested that, although ChatGPT generally surpasses average college students regarding NL knowledge, accuracy issues exist. The dietitians in this study highlighted that ChatGPT’s responses, whether in food categorization, portion calculation, or even the use of professional terminology, have accuracy problems. These results indicate that, without professional oversight, ChatGPT has substantial limitations in practical application. Compared to previous studies, while the NL test accuracy rate was not inferior to that of other professional medical assessments [37,38,39], the test was primarily set at a basic difficulty level [26]. Better performance is expected, as in the perfect scores observed in other studies such as the diabetes knowledge questionnaire (24-DKQ) [8]. At this stage, ChatGPT’s role should be supportive. Considering the heavy workload of dietitians [40], establishing a practical model in which ChatGPT assists in completing dietitians’ tasks is a crucial objective. Furthermore, considering that young students might have a more open attitude towards AI, incorporating AI education into dietitian training curricula at universities should be a future trend. Therefore, designing and evaluating effective AI-assisted dietitian training programs is also a priority for future research.
Limitations
This study has several limitations that warrant consideration. First, the use of predefined scenarios may not encompass the broad range of dietary challenges faced by college students. This may affect our findings’ generalizability. Second, the small sample size and potential lack of diversity among the participants limit the scope of the conclusions. Additionally, the potential misunderstanding of this technology by the dietitians may have introduced bias into their responses. The results revealed that the dietitians’ evaluations of the same quality indicators were not concentrated, suggesting variability in their familiarity and proficiency with this tool. Future research should further assess the dietitians’ use of ChatGPT to build on these findings and address the limitations identified in this study.
5. Conclusions
This study offers a pioneering evaluation of ChatGPT’s performance in providing dietary advice to college students. The findings reveal that, while ChatGPT demonstrates proficiency in providing quick and accessible information, it falls short in delivering personalized, in-depth dietary counseling, which is essential for addressing unique nutritional needs. The inconsistency in the achievement of NL indicators and variability in response quality across different scenarios underscore AI technology’s current limitations in adapting to complex dietary inquiries.
Moreover, feedback from the dietitians highlighted significant concerns regarding the accuracy of ChatGPT responses, especially in areas requiring precise knowledge, such as food categorization, portion sizes, and the use of professional terminology. These shortcomings emphasize the crucial role of professional oversight in integrating AI tools into nutritional education to ensure that the advice provided is accurate and safe.
Despite these challenges, AI’s potential to support dietary practices cannot be overlooked. ChatGPT has shown capabilities that, if further developed and refined, can significantly enhance the efficiency and reach of nutritional counseling, particularly in settings burdened by high client volumes and limited resources. Future research should focus on enhancing the personalization capabilities of AI systems, improving the understanding of complex nutritional concepts, and seamlessly integrating these tools into professional healthcare practices.
In conclusion, AI, such as ChatGPT, has the potential to become a valuable tool in nutrition education. Its current application, however, should be approached with caution, ensuring that it complements, rather than replaces, the nuanced judgment of skilled dietitians. AI’s evolution in dietetics promises a future in which technology and human expertise collaborate to more effectively foster healthier dietary behaviors.
Supplementary Materials
The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/nu16121939/s1.
Author Contributions
Conceptualization and methodology, I.-J.L. and L.-L.L.; software, validation, and formal analysis, L.-C.C.; investigation, resources, and data curation, L.-C.C.; writing—original draft preparation, I.-J.L.; writing—review and editing, L.-L.L.; project administration, I.-J.L. All authors have read and agreed to the published version of the manuscript.
Institutional Review Board Statement
This study was approved by the National Cheng Kung University Human Research Ethics Committee (NCKU HREC-E-112-573-2, 25 January 2024).
Informed Consent Statement
Dietitians participating in the study were informed about its purpose, agreed to participate anonymously, and had the option to withdraw at any time.
Data Availability Statement
The data presented in this study are available upon request from the corresponding author. Due to privacy and ethical concerns, neither the data nor the source of the data can be made publicly available.
Conflicts of Interest
The authors declare no conflicts of interest.
Funding Statement
This research received no external funding.
Footnotes
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
References
- 1.Wang F., Preininger A. AI in Health: State of the Art, Challenges, and Future Directions. Yearb. Med. Inform. 2019;28:16–26. doi: 10.1055/s-0039-1677908. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Ghadban Y.A., Lu H., Adavi U., Sharma A., Gara S., Das N., Kumar B., John R., Devarsetty P., Hirst J.E. Transforming Healthcare Education: Harnessing Large Language Models for Frontline Health Worker Capacity Building using Retrieval-Augmented Generation. medRxiv. 2023 doi: 10.1101/2023.12.15.23300009. [DOI] [Google Scholar]
- 3.Gangavarapu A. LLMs: A Promising New Tool for Improving Healthcare in Low-Resource Nations; Proceedings of the 2023 IEEE Global Humanitarian Technology Conference (GHTC); Villanova, PA, USA. 12–15 October 2023; pp. 252–255. [Google Scholar]
- 4.Hopkins A., Logan J., Kichenadasse G., Sorich M. Artificial intelligence chatbots will revolutionize how cancer patients access information: ChatGPT represents a paradigm-shift. JNCI Cancer Spectr. 2023;7:pkad010. doi: 10.1093/jncics/pkad010. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 5.Zhai X. ChatGPT User Experience: Implications for Education. 2022. [(accessed on 1 April 2024)]. Available online: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4312418.
- 6.Shao C.-Y., Li H., Liu X.-L., Li C., Yang L.-Q., Zhang Y.-J., Luo J., Zhao J. Appropriateness and Comprehensiveness of Using ChatGPT for Perioperative Patient Education in Thoracic Surgery in Different Language Contexts: Survey Study. Interact. J. Med. Res. 2023;12:e46900. doi: 10.2196/46900. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 7.Mondal H., Mondal S., Podder I. Using ChatGPT for Writing Articles for Patients’ Education for Dermatological Diseases: A Pilot Study. Indian Dermatol. Online J. 2023;14:482–486. doi: 10.4103/idoj.idoj_72_23. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 8.Nakhleh A., Spitzer S., Shehadeh N. ChatGPT’s Response to the Diabetes Knowledge Questionnaire: Implications for Diabetes Education. Diabetes Technol. Ther. 2023;25:571–573. doi: 10.1089/dia.2023.0134. [DOI] [PubMed] [Google Scholar]
- 9.Van Bulck L., Moons P. What if your patient switches from Dr. Google to Dr. ChatGPT? A vignette-based survey of the trustworthiness, value and danger of ChatGPT-generated responses to health questions. Eur. J. Cardiovasc. Nurs. 2023;23:95–98. doi: 10.1093/eurjcn/zvad038. [DOI] [PubMed] [Google Scholar]
- 10.Hung Y.-C., Chaker S.C., Sigel M., Saad M., Slater E.D. Comparison of Patient Education Materials Generated by Chat Generative Pre-Trained Transformer Versus Experts: An Innovative Way to Increase Readability of Patient Education Materials. Ann. Plast. Surg. 2023;91:409–412. doi: 10.1097/SAP.0000000000003634. [DOI] [PubMed] [Google Scholar]
- 11.Shah Y.B., Ghosh A., Hochberg A.R., Rapoport E., Lallas C.D., Shah M.S., Cohen S.D. Comparison of ChatGPT and Traditional Patient Education Materials for Men’s Health. Urol. Pract. 2024;11:87–94. doi: 10.1097/UPJ.0000000000000490. [DOI] [PubMed] [Google Scholar]
- 12.Ayoub N.F., Lee Y.-J., Grimm D., Divi V. Head-to-Head Comparison of ChatGPT Versus Google Search for Medical Knowledge Acquisition. Otolaryngol.–Head Neck Surg. 2023;170:1484–1491. doi: 10.1002/ohn.465. [DOI] [PubMed] [Google Scholar]
- 13.Qarajeh A., Tangpanithandee S., Thongprayoon C., Suppadungsuk S., Krisanapan P., Aiumtrakul N., Garcia Valencia O.A., Miao J., Qureshi F., Cheungpasitporn W. AI-Powered Renal Diet Support: Performance of ChatGPT, Bard AI, and Bing Chat. Clin. Pract. 2023;13:1160–1172. doi: 10.3390/clinpract13050104. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Göktaş L.S. The role of ChatGPT in vegetarian menus. Tour. Recreat. 2023;5:79–86. doi: 10.53601/tourismandrecreation.1343598. [DOI] [Google Scholar]
- 15.Niszczota P., Rybicka I. The credibility of dietary advice formulated by ChatGPT: Robo-diets for people with food allergies. Nutrition. 2023;112:112076. doi: 10.1016/j.nut.2023.112076. [DOI] [PubMed] [Google Scholar]
- 16.Abraham S., Noriega B.R., Shin J.Y. College students eating habits and knowledge of nutritional requirements. J. Nutr. Hum. Health. 2018;2:13–17. doi: 10.35841/nutrition-human-health.2.1.13-17. [DOI] [Google Scholar]
- 17.Yang S.C., Luo Y.F., Chiang C.-H. Electronic health literacy and dietary behaviors in Taiwanese college students: Cross-sectional study. J. Med. Internet Res. 2019;21:e13140. doi: 10.2196/13140. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.Liao L.-L., Lai I.-J., Chang L.-C. Nutrition literacy is associated with healthy-eating behaviour among college students in Taiwan. Health Educ. J. 2019;78:756–769. doi: 10.1177/0017896919836132. [DOI] [Google Scholar]
- 19.Racette S.B., Deusinger S.S., Strube M.J., Highstein G.R., Deusinger R.H. Changes in weight and health behaviors from freshman through senior year of college. J. Nutr. Educ. Behav. 2008;40:39–42. doi: 10.1016/j.jneb.2007.01.001. [DOI] [PubMed] [Google Scholar]
- 20.Stok F.M., Renner B., Clarys P., Lien N., Lakerveld J., Deliens T. Understanding eating behavior during the transition from adolescence to young adulthood: A literature review and perspective on future research directions. Nutrients. 2018;10:667. doi: 10.3390/nu10060667. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Contento I.R. Nutrition education: Linking research, theory, and practice. Asia Pac. J. Clin. Nutr. 2008;17:176–179. [PubMed] [Google Scholar]
- 22.Ngo T.T.A. The perception by university students of the use of ChatGPT in education. Int. J. Emerg. Technol. Learn. 2023;18:4. doi: 10.3991/ijet.v18i17.39019. [DOI] [Google Scholar]
- 23.Liao L.L., Lai I.J. Construction of Nutrition Literacy Indicators for College Students in Taiwan: A Delphi Consensus Study. J. Nutr. Educ. Behav. 2017;49:734–742. doi: 10.1016/j.jneb.2017.05.351. [DOI] [PubMed] [Google Scholar]
- 24.Lai I.J., Chang L.C., Lee C.K., Liao L.L. Preliminary evaluation of a scenario-based nutrition literacy online programme for college students: A pilot study. Public Health Nutr. 2023;26:3190–3201. doi: 10.1017/s1368980023002471. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Sun Y., Zhang Y., Gwizdka J., Trace C.B. Consumer Evaluation of the Quality of Online Health Information: Systematic Literature Review of Relevant Criteria and Indicators. J. Med. Internet Res. 2019;21:e12522. doi: 10.2196/12522. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 26.Liao L.-L., Lai I.-J., Shih S.-F., Chang L.-C. Development and validation of the nutrition literacy measure for Taiwanese college students. Taiwan J. Public Health. 2018;37:582. doi: 10.6288/TJPH.201810_37(5).107054. [DOI] [Google Scholar]
- 27.Velpini B., Vaccaro G., Vettori V., Lorini C., Bonaccorsi G. What Is the Impact of Nutrition Literacy Interventions on Children’s Food Habits and Nutrition Security? A Scoping Review of the Literature. Int. J. Environ. Res. Public Health. 2022;19:3839. doi: 10.3390/ijerph19073839. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 28.Doustmohammadian A., Omidvar N., Shakibazadeh E. School-based interventions for promoting food and nutrition literacy (FNLIT) in elementary school children: A systematic review protocol. Syst. Rev. 2020;9:87. doi: 10.1186/s13643-020-01339-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Azevedo J., Padrão P., Gregório M.J., Almeida C., Moutinho N., Lien N., Barros R. A web-based gamification program to improve nutrition literacy in families of 3-to 5-year-old children: The Nutriscience Project. J. Nutr. Educ. Behav. 2019;51:326–334. doi: 10.1016/j.jneb.2018.10.008. [DOI] [PubMed] [Google Scholar]
- 30.Achterberg C., Miller C. Is One Theory Better than Another in Nutrition Education? A Viewpoint: More Is Better. J. Nutr. Educ. Behav. 2004;36:40–42. doi: 10.1016/S1499-4046(06)60127-9. [DOI] [PubMed] [Google Scholar]
- 31.Moons P., Van Bulck L. Using ChatGPT and Google Bard to improve the readability of written patient information: A proof of concept. Eur. J. Cardiovasc. Nurs. 2024;23:122–126. doi: 10.1093/eurjcn/zvad087. [DOI] [PubMed] [Google Scholar]
- 32.Marcason W. Dietitian, Dietician, or Nutritionist? J. Acad. Nutr. Diet. 2015;115:484. doi: 10.1016/j.jand.2014.12.024. [DOI] [PubMed] [Google Scholar]
- 33.Chatelan A., Clerc A., Fonta P.-A. ChatGPT and Future Artificial Intelligence Chatbots: What may be the Influence on Credentialed Nutrition and Dietetics Practitioners? J. Acad. Nutr. Diet. 2023;123:1525–1531. doi: 10.1016/j.jand.2023.08.001. [DOI] [PubMed] [Google Scholar]
- 34.Ponzo V., Goitre I., Favaro E., Merlo F.D., Mancino M.V., Riso S., Bo S. Is ChatGPT an Effective Tool for Providing Dietary Advice? Nutrients. 2024;16:469. doi: 10.3390/nu16040469. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 35.Garcia M.B. ChatGPT as a virtual dietitian: Exploring its potential as a tool for improving nutrition knowledge. Appl. Syst. Innov. 2023;6:96. doi: 10.3390/asi6050096. [DOI] [Google Scholar]
- 36.Liu C., Bao X., Zhang H., Zhang N., Hu H., Zhang X., Yan M. Improving chatgpt prompt for code generation. arXiv. 2023 doi: 10.48550/arXiv.2305.08360.2305.08360 [DOI] [Google Scholar]
- 37.Gilson A., Safranek C.W., Huang T., Socrates V., Chi L., Taylor R.A., Chartash D. How Does ChatGPT Perform on the United States Medical Licensing Examination? The Implications of Large Language Models for Medical Education and Knowledge Assessment. JMIR Med. Educ. 2023;9:e45312. doi: 10.2196/45312. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 38.Nikdel M., Ghadimi H., Tavakoli M., Suh D.W. Assessment of the Responses of the Artificial Intelligence–based Chatbot ChatGPT-4 to Frequently Asked Questions About Amblyopia and Childhood Myopia. J. Pediatr. Ophthalmol. Strabismus. 2024;61:86–89. doi: 10.3928/01913913-20231005-02. [DOI] [PubMed] [Google Scholar]
- 39.Wang Y.-M., Shen H.-W., Chen T.-J. Performance of ChatGPT on the pharmacist licensing examination in Taiwan. J. Chin. Med. Assoc. 2023;86:653–658. doi: 10.1097/JCMA.0000000000000942. [DOI] [PubMed] [Google Scholar]
- 40.Hewko S., Oyesegun A., Clow S., VanLeeuwen C. High turnover in clinical dietetics: A qualitative analysis. BMC Health Serv. Res. 2021;21:25. doi: 10.1186/s12913-020-06008-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The data presented in this study are available upon request from the corresponding author. Due to privacy and ethical concerns, neither the data nor the source of the data can be made publicly available.