Abstract
Background
Precision medicine has resulted in increasing complexity in the treatment of cancer. Web-based educational materials can help address the needs of oncology health care professionals seeking to understand up-to-date treatment strategies.
Objective
This study aimed to assess learning styles of oncology health care professionals and to determine whether learning style-tailored educational materials lead to enhanced learning.
Methods
In all, 21,465 oncology health care professionals were invited by email to participate in the fully automated, parallel group study. Enrollment and follow-up occurred between July 13 and September 7, 2015. Self-enrolled participants took a learning style survey and were assigned to the intervention or control arm using concealed alternating allocation. Participants in the intervention group viewed educational materials consistent with their preferences for learning (reading, listening, and/or watching); participants in the control group viewed educational materials typical of the My Cancer Genome website. Educational materials covered the topic of treatment of metastatic estrogen receptor-positive (ER+) breast cancer using cyclin-dependent kinases 4/6 (CDK4/6) inhibitors. Participant knowledge was assessed immediately before (pretest), immediately after (posttest), and 2 weeks after (follow-up test) review of the educational materials. Study statisticians were blinded to group assignment.
Results
A total of 751 participants enrolled in the study. Of these, 367 (48.9%) were allocated to the intervention arm and 384 (51.1%) were allocated to the control arm. Of those allocated to the intervention arm, 256 (69.8%) completed all assessments. Of those allocated to the control arm, 296 (77.1%) completed all assessments. An additional 12 participants were deemed ineligible and one withdrew. Of the 552 participants, 438 (79.3%) self-identified as multimodal learners. The intervention arm showed greater improvement in posttest score compared to the control group (0.4 points or 4.0% more improvement on average; P=.004) and a higher follow-up test score than the control group (0.3 points or 3.3% more improvement on average; P=.02).
Conclusions
Although the study demonstrated more learning with learning style-tailored educational materials, the magnitude of increased learning and the largely multimodal learning styles preferred by the study participants lead us to conclude that future content-creation efforts should focus on multimodal educational materials rather than learning style-tailored content.
Keywords: e-learning; Web-based Instruction; learning; teaching materials; information dissemination; online systems; education, distance; continuing education; medical oncology/education
Introduction
Precision medicine is the use of a patient’s molecular characteristics to determine disease risk, make a precise diagnosis, determine disease prognosis, and to select the best treatment plan for the patient. In the field of cancer, researchers have been working to develop new drugs and therapeutic strategies tailored to cancers harboring particular biomarkers. Breast cancer has a long history of biomarker-driven prediction of sensitivity to targeted therapies. This study used educational materials on inhibition of cyclin-dependent kinases 4/6 (CDK4/6) to block cell growth as a therapeutic strategy being investigated in patients with hormone receptor-positive breast cancer, including those who have developed resistance to endocrine therapy. These materials were used to investigate oncology health care professional learning styles and optimization of materials for Web-based learning.
Rapidly evolving information about precision cancer medicine creates a knowledge gap in the education of oncology health care professionals regarding complex and important precision cancer medicine concepts, along with approaches for identifying therapeutic strategies for individual patients [1-4]. For example, a large survey in 2011 showed that oncology nurses did not discuss mutation testing with patients because they felt they lacked the knowledge to do so [1]. More recently, an international survey demonstrated that a majority of lung cancer oncologists understand that improved survival is associated with therapies selected after epidermal growth factor receptor (EGFR) mutation testing, but a quarter of lung cancer oncologists do not consider the specific EGFR mutation detected in making therapeutic decisions [4]. This knowledge gap needs to be addressed quickly and effectively to bring the promise of precision cancer medicine to all cancer patients. The topic of CDK4/6 inhibitors in breast cancer was chosen because it was a timely, representative topic in the field of precision cancer medicine with an active knowledge gap.
Web-based tools provide an important platform for oncology health care professionals to address this knowledge gap. Websites such as UpToDate [5] are heavily used by physicians of all types. A wealth of clinical trial information can be found at ClinicalTrials.gov [6] or on the National Cancer Institute’s website [7]. We have developed the My Cancer Genome website [8] as a publicly accessible knowledge resource targeted at oncology health care providers. My Cancer Genome provides up-to-date information to oncology health care providers on the clinical relevance of mutations in cancers and gene-specific clinical trials. Launched in 2011, My Cancer Genome receives more than 10,000 site visits per week, from 211 countries and territories across the world, from an audience of health care providers, researchers, and patients/caregivers (usage statistics current as of January 2017). My Cancer Genome provides content pages on 23 cancer types, 823 genes, and 456 disease gene-variant relationships (content statistics current as of May 2017). Breast cancer educational content includes information on the therapeutic implications of alterations in several genes, including links to relevant clinical trials. A companion mobile app has also been available for Apple iOS devices since 2013.
An individual’s learning style refers to how that individual prefers to gather, interpret, organize, and think about information [9]. For example, individuals may vary in their preferences for how they receive educational information. Examples include preferences for learning through visual (eg, pictures, graphs, diagrams, charts), auditory, and text-based (eg, lists, bullets, or hierarchically organized text) formats [10]. Several learning style assessments have been developed that evaluate preferences according to different learning style models [11]. A person’s learning style preferences may change throughout their lifetime, and cultural factors and previous experiences may contribute to differences in preferences [11-14].
The literature on learning styles in the medical setting has predominantly focused on medical students and residents [13-18]. Learning style preferences may vary based on level of training [19,20]. For example, one small study reported differences in the prevalence of multimodal learners among Australian rural general practice registrars compared to rural medical students [19].
Tailoring education based on learning style may facilitate comprehension of the information. In a study evaluating the effect of providing medical students with instruction matched according to learning styles or in a standard format, tailored instruction was found to result in enhanced understanding of the material based on improvement in test scores, with statistically significant differences seen for kinesthetic learners [13]. Similarly, prior research has demonstrated that providing patients with health educational information customized to their health literacy and learning style preferences increased understanding and retention of the material [21-24]. However, in an updated systematic review and meta-analysis, Cook [25] found that evidence was lacking to support the use of adaptation to cognitive and learning styles in computer-assisted instruction. More rigorous studies are needed to better understand the effectiveness of tailored instruction on Web-based learning. To the best of our knowledge, no studies have been published reporting the learning style preferences of oncology health care professionals. Considering health care providers’ learning style preferences when developing educational content for Web-based tools may help accelerate understanding and retention of the information.
Objectives
In this study, we developed educational information tailored to different learning styles on the topic of treatment of metastatic estrogen receptor-positive (ER+) breast cancer using CDK4/6 inhibitors. The objectives of this study were to (1) assess learning style preferences of oncology health care professionals and (2) evaluate the effectiveness of providing educational materials customized to learning style preferences using a fully automated, controlled study design with concealed allocation.
Methods
The CONSORT-EHEALTH checklist for this study can be found in Multimedia Appendix 1. Technical details and changes after study commencement can be found in Multimedia Appendix 2.
Ethics
This study was approved by the Vanderbilt University Medical Center institutional review board (IRB). This study received a waiver of consent from the IRB. In place of the traditional consent process, when participants clicked the link in their invitation email, they arrived at the welcome page. This page provided participants with information about the study’s purpose, mechanics, and risks, with contact information for those with further questions. The welcome page also provided an estimate of the time commitment needed to participate in the study. Participants then entered their email address, personal identification number (PIN) as provided in their invitation email, and answered yes or no to the statement, “I agree to participate in this study.” We did not receive any phone calls or emails from participants prior to their agreement to participate.
Study Design
A parallel study design was used to examine the effectiveness of providing learning style-tailored materials compared to control materials (Figure 1). The study opened on July 13, 2015, and closed on September 7, 2015. Participants were allocated 1:1 to the intervention or control arms when they landed on the enrollment screen. Participants were asked to complete a learning style assessment and a knowledge pretest. Participants allocated to the intervention arm viewed materials consistent with their learning style preferences, and participants allocated to the control arm viewed materials in the format used on My Cancer Genome. Following review of the educational materials, all participants were asked to complete a knowledge posttest and a feedback survey. After 2 weeks, participants who had completed the posttest were asked to take the knowledge follow-up test. The 2-week interval was chosen for consistency with the authors’ related research [21,24]. After taking the knowledge follow-up test, participants could provide demographic information and fill out a form to provide the information needed to send the US $100 Amazon.com Gift Card incentive.
Data Security
Safety and security of participant data were ensured through use of REDCap (Research Electronic Data Capture) for administration of the study [26]. All study data were collected and managed using REDCap electronic data capture tools hosted at Vanderbilt University Medical Center. REDCap is a secure, Web-based app designed to support data capture for research studies, providing (1) an intuitive interface for validated data entry, (2) audit trails for tracking data manipulation and export procedures, (3) automated export procedures for seamless data downloads to common statistical packages, and (4) procedures for importing data from external sources.
Participants
The email marketing service Medical Marketing Service, Inc was used to recruit physicians, nurse practitioners, and physician assistants. The lists were comprised of professional society members who had indicated oncology as an area of focus. Participants were assumed to be computer and Internet literate. Prior to enrollment, participants received the invitation email and could view the welcome page in REDCap. Copies of both are provided in Multimedia Appendix 3. To ensure that participants had been invited to take part in the study, each invitation email contained a unique 10-character alphanumeric PIN. Details related to the PINs are described in Multimedia Appendix 2. Following agreement to participate in the study, participants answered an eligibility question; participants were required to be in active practice in an oncology setting.
Recruitment
Potential volunteers were sent email invitations to participate in the study on July 13 and 20, 2015. The second email was sent regardless of whether they had enrolled in the study or not. In order to protect participant privacy, Medical Marketing Service was not given the list of participants who had enrolled in the study. Enrollment and follow-up occurred through study close on September 7, 2015. The trial closed on September 7 because enrollments had dropped to almost zero and we believed there were no remaining active participants. After study close, we found one remaining active participant. This participant was given the gift card incentive, but the data from this participant were not used in the analysis. It was possible for participants to enroll more than once; for more information, see Multimedia Appendix 2.
Interventions and Outcomes
The codebook for all surveys is provided in Multimedia Appendix 4. Multimedia Appendix 5 contains information about participant tracking and related analyses.
Learning Style Assessment
Participants were asked to self-report their learning style preferences by selecting one of the following responses to complete the statement, “I am likely to remember something a year from now if: (1) I learn it by reading, (2) I learn it by listening, (3) I learn it by watching, (4) I learn it by reading and listening, (5) I learn it by reading and watching, (6) I learn it by listening and watching, (7) I learn it by reading, listening, and watching” (see also Multimedia Appendix 4). This question was modeled after a single-item self-report measure, which was used by the authors in previous research [23]. The model self-report measure determined learning style preferences based on participants’ responses to whether they would recall how to do something a year from now if they learned it by reading, listening, watching, or trying things on their own; participants could select all that applied [23]. The revised question wording used in our study reflects that our educational content does not involve skill-based or kinesthetic learning.
Knowledge Tests
A 10-question test was developed to evaluate participant knowledge of the educational materials. The questions were multiple choice or true/false, and each question included a response option of “don’t know.” Participants were asked to complete the test immediately before (pretest) and after (posttest) viewing the information and 2 weeks later (follow-up test). The same questions were used for all three surveys, although the order of the questions and the order of multiple-choice answers were changed for each survey. The order of true/false answers was not changed. The surveys were loaded into REDCap and tested before study commencement. Each set of educational materials contained the answers to all 10 questions.
Educational Materials
New content was developed for the project for both the control and intervention arms. Copies of the educational materials are shown in Multimedia Appendix 6. Educational materials were developed by the authors, who together brought experience in the adaptation of health information according to learning preferences and in-depth knowledge of oncology [21-24,27,28]. The materials were also reviewed by two experts in breast cancer research to further ensure accuracy of the information. Intervention materials with a watching component included slides with figures, captions, and limited bulleted text. Intervention materials with a listening component included an audio recording. Intervention materials with a reading component included text. The control materials included text, figures, and tables.
Feedback, Demographics, and Gift Card Surveys
The feedback survey was presented to participants following the knowledge posttest. The feedback survey evaluated whether participants thought the information in the learning materials was easy to understand and if they learned something new. The survey consisted of five questions (see Multimedia Appendixes 4 and 7). Options for four of the questions were on a rating scale from strongly agree to strongly disagree (strongly agree, disagree, neither agree or disagree, agree, or strongly agree). For the fifth question, participants could share any other thoughts about the materials in an open-text field. Answering questions in this survey was optional.
The demographics survey was presented to participants following the knowledge follow-up test. For all participants, the survey consisted of six questions about their practice type (academic, community, both, or other), percentage of patients seen with breast cancer (<25%, 25%-50%, >50%, or unknown), age, gender, race, and ethnicity. Answering questions in this survey was optional.
The gift card survey followed the demographics survey. This survey collected information required for institutional financial reporting requirements and US federal statutory requirements. Although completing this survey was optional, participants could not receive the gift card incentive without completing it.
Study Sample
The primary objective of this study was to examine the improvement of the knowledge test score (pretest vs posttest) between experimental group (matched learning materials according to learning style) and the control group (standard learning materials). According to the prospective sample size calculation, a sample size of 250 per group would provide at least 90% power to detect a conservative effect size of 0.3 with two-sided type I error of 5%. The effect size is defined as the ratio of mean difference of test score between study groups to the standard deviation.
From Medical Marketing Service, we learned that a good open rate—the likelihood that a recipient will open and view the email—for a marketing email to oncologists is 13% (Jane Stormzand, personal communication, February 24, 2015). We used this number to estimate our open rate for the invitation email. We did not have data on the level of attrition to expect during the course of the study itself, so we estimated using high and low attrition rates at 80% and 40% for each of four study events: (1) agreeing to participate in the study, (2) viewing the educational materials, (3) taking the knowledge posttest, and (4) taking the knowledge follow-up test. With a high attrition rate, we would have expected only five participants to complete the study; with a low attrition rate, we would have expected 438 to complete the study. As a result, we did not limit enrollment, and we had an IRB-approved upper limit of enrollment set at 1200. Participants who did not complete required portions of the study (learning style survey, pretest, posttest, and follow-up test) were excluded.
Allocation Concealment and Blinding
The study was fully automated; therefore, the study personnel did not have access to the list of individuals invited to participate in the study and the study personnel had no control over when any individual clicked on the survey link in the recruitment email (the action that creates a numbered record in REDCap). For these reasons, we decided to use alternating ABAB allocation rather than randomization. Although not randomized, the allocation was concealed. Participants with odd-numbered records were allocated to the intervention arm, whereas even-numbered records were allocated to the control arm.
Although no expectations were set about the content of the educational materials before allocation or review of educational materials, we do not consider the participants to have been blinded because after clicking the submit button on the REDCap page with the link to the educational materials, participants were shown the survey queue, which listed whether the surveys were on the control or intervention arm. Study personnel were not blinded; this was deemed unnecessary because the outcome measures were objective measures—test scores—and because the study was fully automated. The study statisticians were blinded to group assignment.
Statistical Methods
Multiple imputation was performed using the rms R package to account for missing data (practice type: 2.0%, 11/552; percent of breast cancer patients seen: 1.4%, 8/552; age 4.5%, 25/552; gender: 4.7%, 26/552; race: 13.9%, 77/552; physician specialty: 7.6%, 42/552). Multivariable linear regression was used to estimate the intervention effect on (1) knowledge posttest score and (2) knowledge follow-up test score, adjusted for knowledge pretest score (baseline assessment), as well as other covariates (practice type, percentage of breast cancer patients seen, age, gender, race, and physician specialty). Hierarchical cluster analysis and redundancy analysis were performed for data reduction. Years since completing residency/fellowship was dropped from the model because it could be predicted from other variables in the model. Residual analysis was used to check the linear regression assumptions of homogeneity for variance, normality, and linearity. For each individual learning style, the Wilcoxon signed rank test was performed to test for differences between (1) knowledge pretest score and knowledge posttest score, and (2) knowledge pretest score and knowledge follow-up test score. Multiple comparisons were corrected using the Bonferroni method. All tests were significant at the overall two-sided 5% level. All statistical analyses were performed in R version 3.1.2.
Results
Participant Flow
A total of 751 participants enrolled and completed the learning style survey (Figure 2). Of those, 384 were allocated to the control arm and 367 were allocated to the intervention arm. A total of 296 on the control arm and 256 on the intervention arm completed all required portions of the study, for a total of 552. Of those who completed the study on the control arm, 19 viewed educational materials on the control arm and the intervention arm due to multiple enrollments. Of those who completed the study on the intervention arm, 22 viewed educational materials on the control arm and the intervention arm, and 10 viewed multiple intervention arm educational materials due to multiple enrollments. Because participants could view any nonstudy materials they desired, in electronic or hard copy, and because participants were not prevented from re-enrolling to view additional on-study educational materials, participants who viewed multiple sets of educational materials were not excluded from analysis.
Based on the first recruitment email, the view rate of the recruitment email itself was 13.62% (2923/21,465), and the subsequent view rate of the welcome page of the study was 22.85% (668/2923). Participation rate, determined by the ratio of unique participants completing the learning style survey to those who clicked on a link in a recruitment email, was 66.70% (751/1126). Note that this underestimates the participation rate because some participants clicked links in both recruitment emails. Completion rate, determined by the ratio of participants completing the knowledge follow-up test to those who completed the learning style survey, was 73.50% (552/751). The attrition rate decreased at each step (see Figure 3), particularly following enrollment in the study. Note that the second two events underestimate attrition; these numbers represent the sum of individuals opening and clicking links in the two recruitment emails. Because some individuals opened or clicked links in both emails, the sums overestimate participation.
We found that several PINs were used for multiple study enrollments: 552 participants enrolled once, 172 participants enrolled twice, 23 participants enrolled three times, and 4 participants enrolled four times.
Baseline Data
Participant characteristics are shown in Table 1. Overall, the arms were well balanced for participant characteristics, with 296 participants in the control arm and 256 in the intervention arm. The majority of participants were physicians, with slightly more in the control arm versus the intervention arm (274 vs 237). There were four fewer physician assistants and one more nurse practitioner and other participant types in the intervention arm than the control arm. Half (50.6%, 274/541) of the participants in both arms worked in an academic setting, with 39.89% (215/541) working in a community setting, and 9.6% (52/541) reporting that their practice type was both academic and community based. For the majority of participants in both arms, less than 25% of patients in their practices were breast cancer patients, and for just under a third of participants, 25% to 50% of patients in their practices were breast cancer patients.
Table 1.
Participant characteristics | Na | Control (n=296) | Intervention (n=256) | P | |
Participant type, n (%) | 552 | .81b | |||
Physician | 274 (92.6) | 237 (92.6) | |||
Physician assistant | 16 (5.4) | 12 (4.7) | |||
Nurse and other | 6 (2.0) | 7 (2.7) | |||
Practice type, n (%) | 541 | .42b | |||
Academic | 541 | 151 (51.9) | 123 (49.2) | ||
Community | 109 (37.5) | 106 (42.4) | |||
Both | 31 (10.6) | 21 (8.4) | |||
Percentage of patients with breast cancer seen in participant’s practice, n (%) | 544 | .88b | |||
<25% | 170 (58.4) | 141 (55.7) | |||
25%-50% | 78 (26.8) | 74 (29.2) | |||
>50% | 37 (12.7) | 34 (13.4) | |||
Unknown | 6 (2.1) | 4 (1.6) | |||
Age (years) | |||||
Mean (SD) | 527 | 42.8 (9.8) | 43.3 (9.9) | .57c | |
Median (IQR) | 40 (35-50) | 40 (35-50) | |||
Gender, n (%) | 526 | .42b | |||
Male | 202 (70.9) | 163 (67.6) | |||
Female | 83 (29.1) | 78 (32.4) | |||
Race, n (%) | 475 | .95b | |||
White | 173 (67.8) | 150 (68.2) | |||
Asian | 65 (25.5) | 54 (24.6) | |||
Other/multiracial | 17 (6.7) | 16 (7.3) | |||
Hispanic, Latino, or Spanish origin, n (%) | 476 | .32b | |||
Hispanic or Latino | 8 (3.2) | 11 (4.9) | |||
Non-Hispanic | 245 (96.8) | 212 (95.1) | |||
Physician specialty, n (%) | 510 | .11b | |||
Resident | 6 (2.2) | 11 (4.7) | |||
Fellow | 39 (14.2) | 20 (8.5) | |||
Medical oncologist and/or hematologist | 164 (59.9) | 133 (56.4) | |||
Surgical oncologist | 10 (3.6) | 9 (3.8) | |||
Radiation oncologist | 35 (12.8) | 46 (19.5) | |||
Pathologist | 13 (4.7) | 13 (5.5) | |||
None of the above | 7 (2.5) | 4 (1.7) | |||
Years since completed residency/fellowship, n (%) | 510 | .10b | |||
<5 | 85 (31.1) | 81 (34.2) | |||
5-9 | 51 (18.7) | 56 (23.6) | |||
10-15 | 50 (18.3) | 29 (12.2) | |||
>15 | 64 (23.4) | 60 (25.3) | |||
Other | 23 (8.4) | 11 (4.6) | |||
Follow-up interval (days) | |||||
Mean (SD) | 552 | 15.3 (3.0) | 15.2 (2.5) | .58c | |
Median (IQR) | 14.2 (14.0-15.8) | 14.2 (14.0-15.5) |
a N is the number of nonmissing observations.
b The test used was Pearson chi-square test.
c The test used was Wilcoxon rank-sum test.
The demographics were fairly well balanced between arms; all P values were insignificant at the two-sided 5% significance level, as shown in Table 1. There were more men in the control arm (202 vs 163). The majority of participants in both arms were white (68%, 323/475), with an additional 25% (119/475) of participants reporting their race as Asian. The mean age was 42.8 (SD 9.8) years for the control arm and 43.3 (SD 9.9) years for the intervention arm.
More than half the participants were medical oncologists, with more in the control arm (164 vs 133). Radiation oncologists made up 15.9% (81/510) of the participants, with more in the intervention arm (35 vs 46). The number of years since physicians had completed residency/fellowship was similar in both arms, with one-third reporting less than 5 years and one-third more than 15 years.
Learning styles were also balanced between the intervention arm and the control arm. The largest group was the watching plus listening plus reading group, making up 38.8% (214/552) of participants. The next largest group was the watching plus reading group, with 18.7% (103/552) of participants selecting this group. More participants chose listening or listening plus reading in the control arm than the intervention arm (9 vs 3 for listening, 38 vs 29 for listening plus reading).
The median follow-up interval between posttest and follow-up test was equivalent between arms at 14.2 (IQR 14.0-15.8 for control arm and IQR 14.0-15.5 for intervention arm) days.
Outcomes and Estimation
Analysis was conducted by original assigned groups. Almost 80% (438/552) of oncology health care professionals were multimodal, with the most common learning style being watching plus listening plus reading (Table 2).
Table 2.
Learning style | All, n (%) (N=552) |
Control, n (%) (n=296) |
Intervention, n (%) (n=256) |
Watching | 19 (3.4) | 10 (3.4) | 9 (3.5) |
Listening | 12 (2.2) | 9 (3.0) | 3 (1.2) |
Reading | 83 (15.0) | 43 (14.5) | 40 (15.6) |
Watching & listening | 54 (9.8) | 23 (7.8) | 31 (12.1) |
Watching & reading | 103 (18.7) | 63 (21.3) | 40 (15.6) |
Listening & reading | 67 (12.1) | 38 (12.8) | 29 (11.3) |
Watching & listening & reading | 214 (38.8) | 110 (37.2) | 104 (40.6) |
There was a significant difference in both knowledge posttest score and knowledge follow-up test scores between the control arm and the intervention arm. The intervention arm showed a greater improvement of 0.4 points in the knowledge posttest score compared to the control group (the adjusted mean posttest scores were 7.861, SE 0.408 and 7.461, SE 0.414 for intervention arm and control arm, respectively; P=.004), and on average 0.3 points higher follow-up test score than the control group (the adjusted mean follow-up scores were 7.177, SE 0.400 and 6.805, SE 0.406 for the intervention arm and control arm, respectively; P=.02). Both analyses were adjusted for knowledge pretest score and other covariates. Variance analysis and parameter estimates for the knowledge posttest and knowledge follow-up test score regression models are shown in Multimedia Appendix 8. Among the seven individual learning styles, we detected a significant improvement in knowledge posttest score and a significant improvement in knowledge follow-up test score in all learning styles (adjusted P values <.001), except for watching and listening styles.
Participant and Investigator Feedback
Table 3 shows the results of the feedback survey by study arm. Multimedia Appendix 7 shows tables and diverging stacked bar charts of the results for each question of the feedback survey by study arm and by learning style. Of the participants who completed the feedback survey, 89.7% (489/545) agreed or strongly agreed that the information was new to them, 79.9% (436/546) found it satisfying, and 78.0% (425/545) found it easy to understand. Of the 543 participants who completed the question, 174 (32.0%) agreed or strongly agreed that the information was confusing.
Table 3.
Survey questions and responses | N | Control, n (%) (n=296) |
Intervention, n (%) (n=256) |
|
Was the information easy to understand? | 545 | |||
Strongly disagree | 10 (3.4) | 18 (7.1) | ||
Disagree | 19 (6.5) | 13 (5.1) | ||
Neither disagree nor agree | 41 (14.0) | 19 (7.5) | ||
Agree | 170 (58.2) | 127 (50.2) | ||
Strongly agree | 52 (17.8) | 76 (30.0) | ||
Was the information confusing? | 543 | |||
Strongly disagree | 20 (6.9) | 39 (15.5) | ||
Disagree | 97 (33.3) | 105 (41.7) | ||
Neither disagree nor agree | 61 (21.0) | 47 (18.6) | ||
Agree | 105 (36.1) | 55 (21.8) | ||
Strongly agree | 8 (2.8) | 6 (2.4) | ||
Was the information satisfying? | 546 | |||
Strongly disagree | 4 (1.4) | 13 (5.2) | ||
Disagree | 15 (5.1) | 6 (2.4) | ||
Neither disagree nor agree | 54 (18.4) | 18 (7.1) | ||
Agree | 175 (59.5) | 150 (59.5) | ||
Strongly agree | 46 (15.6) | 65 (25.8) | ||
Was the information new? | 545 | |||
Strongly disagree | 5 (1.7) | 10 (4.0) | ||
Disagree | 1 (0.3) | 4 (1.6) | ||
Neither disagree nor agree | 25 (8.5) | 11 (4.4) | ||
Agree | 173 (59.0) | 127 (50.4) | ||
Strongly agree | 89 (30.4) | 100 (39.7) |
Most positive comments on the feedback survey dealt with the educational materials: participants found them to be good, useful, organized, evidence-based, important, well designed, succinct, easy to understand, interesting, enjoyable, and exciting. Others appreciated the provision of a knowledge pretest or said that the information would help them explain treatment regimens to patients. We received an email from a colleague of a participant expressing interest in related studies. We received requests for the educational materials to be made available to participants following the study.
Most negative comments on the feedback survey also dealt with the educational materials: some found them to be poorly organized, busy, or confusing. Eighteen respondents suggested adding figures, tables, or videos. Of these, 14 participants had been allocated to the control arm. These participants were generally multimodal learners who perceived My Cancer Genome content as being text-based and thus lacking in graphical elements. The other four participants were allocated to the intervention arm. These participants were listening, reading, and listening plus reading learners; they each requested visual educational materials. Others felt the information was not up to date.
Discussion
The first objective of this study was to measure learning styles of oncology health care professionals. Compared to assessments of learning styles of medical students and allied health students, this cohort has a higher percentage of multimodal learners at 79.3% (436/552); those studies found rates of 61% (61/100) and 66% (90/137), respectively. The comparison is not direct because this study did not include kinesthetic as a learning style option and because this study used self-reporting to assess learning style [15,29].
The second objective of the study was to determine whether learning style-tailored learning materials fostered greater learning and retention than typical My Cancer Genome content. Although participants did learn more when viewing educational materials tailored to learning style, the mean benefit was only 0.40 points on the knowledge test. An opportunity for future research that would permit a conclusion as to whether the use of learning style-tailored materials facilitate learning in general would use materials inconsistent with learning style preferences rather than the standard format used in our study as the control [30]. However, care would need to be taken to account for demand characteristics, especially due to the higher level of scientific literacy of participants and the likelihood that the learning style survey at the beginning would provide a cue regarding the hypothesis being tested [31]. Instead, by providing control arm participants with access to My Cancer Genome standard content, some participants on the control arm received matched-style content, whereas others received mismatched style content. This may have reduced demand characteristics that would skew the control arm to poorer performance. Finally, a limitation of the study was that we did not assess how the quality of the materials may have impacted learning.
Several participants requested more visual elements to be incorporated into the control educational materials of typical My Cancer Genome content and into nonvisual educational materials in the intervention arm, even when those participants did not identify themselves as watching learners. Together, these participant comments, the high percentage (79.3%, 436/552) of multimodal learners, and the small but significant improvement in knowledge scores when presented with materials tailored to learning style highlight the need for Web-based educational materials to address watching, listening, and reading learners. Therefore, it will be important for My Cancer Genome to incorporate more watching and listening elements, including graphics and videos, in particular.
Given the relatively small improvement in test scores and the heavy resource demand required to create separate content for each learning style, it is probably a poor use of resources to generate new content for each learning style. Instead, the creation and embedding of graphical and video content to address multimodal learning styles may enhance current My Cancer Genome pages, address the learning gap for watching, listening, and multimodal learners, and provide the oncology community with resources for presentations and additional dissemination of learning materials. To this end, since the completion of this study, My Cancer Genome has begun to enhance its website with multimodal content. My Cancer Genome has improved its visual representations of molecular pathways involved in cancer, embedded graphics in new content related to diseases such as chronic lymphocytic leukemia and in new content related to new genetic testing modalities, such as digital droplet polymerase chain reaction, and embedded graphical-audio “knowledge pearls” explaining key concepts such as “mutation” [27]. Additionally, as a result of this study, My Cancer Genome is in the process of creating a 2.0 version of the My Cancer Genome website, which will heavily rely on multimodal data presentation to convey curated knowledge content.
In conclusion, the results of this study imply that Web-based educational materials should be multimodal: (1) most oncology health care professionals are multimodal learners, (2) the increase in learning when learning style-tailored educational materials were used is small, and (3) multiple requests for more multimodal materials were received from participants in both the control and intervention arms. These conclusions are corroborated by a significant body of evidence and confirmed by this work for oncology health care professionals [30].
Acknowledgments
We thank Yu Shyr for conducting the power analysis and sample size estimation as well as developing the initial data analysis and subgroup analysis plans. We thank Ingrid Mayer and Justin Balko for critical review of the educational materials. We thank Daniel Wenner, Ross Oreto, Daniel Carbone, Joey Schneider, and Scott Sobecki of the Vanderbilt University Medical Center Research Informatics Core for their development work to host the educational materials on My Cancer Genome. We thank the Vanderbilt REDCap team for their assistance with project development, project testing, data analysis, and bug fixes. We thank Kathleen Mittendorf for detailed review of the manuscript. This study was funded by Pfizer under an Independent Grant for Learning & Change. Pfizer had no role in the design, conduct, or publication of this study. REDCap is funded through grant support (UL1 TR000445 from National Center for Advancing Translational Sciences /National Institutes of Health).
Abbreviations
- CDK4
cyclin-dependent kinase 4 protein
- CDK6
cyclin-dependent kinase 6 protein
- CONSORT
Consolidated Standards of Reporting Trials
- EGFR
epidermal growth factor receptor
- ER+
estrogen receptor positive
- IRB
institutional review board
- PIN
personal identification number
- REDCap
Research Electronic Data Capture
CONSORT-EHEALTH checklist.
Technical Details and Changes After Study Commencement.
Emails Sent to Participants.
Codebook for All Surveys.
Participant Tracking and Related Analyses.
Educational Materials.
Feedback Survey Results.
Variance Analysis and Parameter Estimates.
Footnotes
Authors' Contributions: MAL, NBG, SVK, CMM, IAA, PL, and KJ conceived the project. MAL, SVK, CMM, IAA, PL, and KJ designed the project and developed the educational materials. CMM managed data acquisition. FY and S-CC provided statistical analysis of the data. CMM, MAL, SVK, IAA, PL, KJ, FY, and S-CC interpreted the data. CMM and IAA drafted the manuscript. All authors provided critical review of the manuscript and approved the publication version. All authors are accountable for all aspects of the manuscript, including its accuracy and integrity.
Conflicts of Interest: Authors developed and tested the educational materials, for which there are no financial interests. MAL declares two conflicts of interest: she is a member of the external scientific advisory board for Personalis, Inc and receives royalties from GenomOncology, Inc. No other conflicts are declared.
References
- 1.Boehringer Ingelheim Pharmaceuticals. 2011. [2016-04-22]. Surveys reveal wide gaps in knowledge of genetic mutation testing exist between oncologists, nurses and cancer patients http://us.boehringer-ingelheim.com/news_events/press_releases/press_release_archive/2011/november_16_2011.html .
- 2.Boehringer Ingelheim Pharmaceuticals. 2013. May 07, [2016-04-22]. Surveys reveal opportunity for medical community to fully embrace biomarker testing in lung cancer http://us.boehringer-ingelheim.com/news_events/press_releases/press_release_archive/2013/05-07-13-surveys-medical-community-biomarker-testing-lung-cancer.html .
- 3.Boehringer Ingelheim Pharmaceuticals. 2015. [2016-04-22]. New international survey of lung cancer oncologists highlights underutilization of personalized treatments http://newscentre.boehringer-ingelheim.com/diru/news_releases/press_releases/2015/17_april_2015_oncology.html .
- 4.Spicer J, Tischer B, Peters M. EGFR mutation testing and oncologist treatment choice in advanced NSCLC: global trends and differences. Ann Oncol; European Lung Cancer Conference (ELCC); Apr 11-15, 2015; Geneva, Switzerland. Annals of Oncology; 2015. p. i60. [Google Scholar]
- 5.UpToDate. 2016. [2016-03-31]. http://www.uptodate.com/contents/search .
- 6.ClinicalTrials.gov. 2016. [2016-03-31]. https://clinicaltrials.gov/
- 7.National Cancer Institute. 2016. [2016-03-31]. Find NCI-supported clinical trials http://www.cancer.gov/about-cancer/treatment/clinical-trials/search .
- 8.My Cancer Genome. 2016. [2016-03-31]. https://www.mycancergenome.org/
- 9.Davis BG. Tools for Teaching. San Francisco, CA: Jossey-Bass Publishers; 2009. [Google Scholar]
- 10.Fleming ND, Mills C. Not another inventory, rather a catalyst for reflection. Improve Acad. 1992;11:137–149. [Google Scholar]
- 11.Anderson I. Identifying different learning styles to enhance the learning experience. Nurs Stand. 2016 Oct 12;31(7):53–63. doi: 10.7748/ns.2016.e10407. [DOI] [PubMed] [Google Scholar]
- 12.Dunn R, Griggs SA. Multiculturalism and Learning Style: Teaching and Counseling Adolescents. Westport, CT: Praeger; 1998. [Google Scholar]
- 13.Anbarasi M, Rajkumar G, Krishnakumar S, Rajendran P, Venkatesan R, Dinesh T, Mohan J, Venkidusamy S. Learning style-based teaching harvests a superior comprehension of respiratory physiology. Adv Physiol Educ. 2015 Sep;39(3):214–217. doi: 10.1152/advan.00157.2014. http://advan.physiology.org/cgi/pmidlookup?view=long&pmid=26330041. [DOI] [PubMed] [Google Scholar]
- 14.Liew S, Sidhu J, Barua A. The relationship between learning preferences (styles and approaches) and learning outcomes among pre-clinical undergraduate medical students. BMC Med Educ. 2015 Mar 11;15:44. doi: 10.1186/s12909-015-0327-0. https://bmcmededuc.biomedcentral.com/articles/10.1186/s12909-015-0327-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Kharb P, Samanta PP, Jindal M, Singh V. The learning styles and the preferred teaching-learning strategies of first year medical students. J Clin Diagn Res. 2013 Jun;7(6):1089–1092. doi: 10.7860/JCDR/2013/5809.3090. http://europepmc.org/abstract/MED/23905110. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 16.Prithishkumar IJ, Michael SA. Understanding your student: using the VARK model. J Postgrad Med. 2014;60(2):183–186. doi: 10.4103/0022-3859.132337. http://www.jpgmonline.com/article.asp?issn=0022-3859;year=2014;volume=60;issue=2;spage=183;epage=186;aulast=Prithishkumar. [DOI] [PubMed] [Google Scholar]
- 17.Kim RH, Gilbert T. Learning style preferences of surgical residency applicants. J Surg Res. 2015 Sep;198(1):61–65. doi: 10.1016/j.jss.2015.05.021. [DOI] [PubMed] [Google Scholar]
- 18.Kim RH, Gilbert T, Ristig K. The effect of surgical resident learning style preferences on American Board of Surgery In-training Examination scores. J Surg Educ. 2015;72(4):726–731. doi: 10.1016/j.jsurg.2014.12.009. [DOI] [PubMed] [Google Scholar]
- 19.Brumpton K, Kitchener S, Sweet L. Learning styles in vertically integrated teaching. Clin Teach. 2013 Oct;10(5):282–286. doi: 10.1111/tct.12024. [DOI] [PubMed] [Google Scholar]
- 20.Samarakoon L, Fernando T, Rodrigo C. Learning styles and approaches to learning among medical undergraduates and postgraduates. BMC Med Educ. 2013 Mar 25;13:42. doi: 10.1186/1472-6920-13-42. https://bmcmededuc.biomedcentral.com/articles/10.1186/1472-6920-13-42. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Giuse NB, Koonce TY, Storrow AB, Kusnoor SV, Ye F. Using health literacy and learning style preferences to optimize the delivery of health information. J Health Commun. 2012 Oct;17 Suppl 3:122–140. doi: 10.1080/10810730.2012.712610. [DOI] [PubMed] [Google Scholar]
- 22.Epelbaum H, Epelbaum M, Giuse NB, Hurley S, Koonce TY, Wallstrom KA. Personalizing health communications: Using health literacy and learning style preferences to improve patients' diabetes knowledge. Medical Library Association Annual Meeting; May 16-21, 2014; Chicago, IL. 2014. http://www.mlanet.org/page/past-and-future-meetings. [Google Scholar]
- 23.Koonce TY, Giuse NB, Choemprayong S, Martin SL, Epelbaum M, Kusnoor SV. Using personalized education delivery to improve community clinic patients' knowledge of hypertension. Special Libraries Association Annual Meeting, Biomedical & Life Sciences Division; June 11, 2013; San Diego, CA. 2013. http://dbiosla.org/events/past_sla_conference/San%20Diego/Personalized%20Education%20Delivery_Final%20Paper_2013.pdf. [Google Scholar]
- 24.Koonce TY, Giuse NB, Kusnoor SV, Hurley S, Ye F. A personalized approach to deliver health care information to diabetic patients in community care clinics. J Med Libr Assoc. 2015 Jul;103(3):123–130. doi: 10.3163/1536-5050.103.3.004. http://europepmc.org/abstract/MED/26213503. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.Cook DA. Revisiting cognitive and learning styles in computer-assisted instruction: not so useful after all. Acad Med. 2012 Jun;87(6):778–784. doi: 10.1097/ACM.0b013e3182541286. [DOI] [PubMed] [Google Scholar]
- 26.Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. 2009 Apr;42(2):377–381. doi: 10.1016/j.jbi.2008.08.010. http://linkinghub.elsevier.com/retrieve/pii/S1532-0464(08)00122-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 27.Kusnoor SV, Koonce TY, Levy MA, Lovly CM, Naylor HM, Anderson IA, Micheel CN, Chen SC, Ye F, Giuse NB. My Cancer Genome: evaluating an educational model to introduce patients and caregivers to precision medicine information. AMIA Jt Summits Transl Sci Proc. 2016;2016:112–121. [PMC free article] [PubMed] [Google Scholar]
- 28.Giuse NB, Kusnoor SV, Koonce TY, Naylor HM, Chen SC, Blasingame MN, Anderson IA, Micheel CM, Levy MA, Ye F, Lovly CM. Guiding oncology patients through the maze of precision medicine. J Health Commun. 2016 Apr;21 Suppl 1:5–17. doi: 10.1080/10810730.2015.1131772. http://europepmc.org/abstract/MED/27043753. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 29.Good JP, Ramos D, D'Amore DC. Learning style preferences and academic success of preclinical allied health students. J Allied Health. 2013;42(4):e81–e90. [PubMed] [Google Scholar]
- 30.Pashler H, McDaniel M, Rohrer D, Bjork R. Learning styles: concepts and evidence. Psychol Sci Public Interest. 2008 Dec;9(3):105–119. doi: 10.1111/j.1539-6053.2009.01038.x. [DOI] [PubMed] [Google Scholar]
- 31.McCambridge J, de Bruin BM, Witton J. The effects of demand characteristics on research participant behaviours in non-laboratory settings: a systematic review. PLoS One. 2012;7(6):e39116. doi: 10.1371/journal.pone.0039116. http://dx.plos.org/10.1371/journal.pone.0039116. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
CONSORT-EHEALTH checklist.
Technical Details and Changes After Study Commencement.
Emails Sent to Participants.
Codebook for All Surveys.
Participant Tracking and Related Analyses.
Educational Materials.
Feedback Survey Results.
Variance Analysis and Parameter Estimates.