Abstract
Background
Contraception knowledge and attitudes are largely formed from conversations within one’s social network. More recently, this network has expanded to include social media. As the fastest growing social media platform, we aimed to assess popular contraception videos on TikTok for content understandability, actionability and accuracy.
Methods
This is a secondary analysis of the most viewed, contraception-specific TikTok videos that were previously coded as containing educational messaging in a content-analysis study. We assessed videos for understandability (ability to explain a key message) and actionability (identify what they can do to act on that message) using the Patient Education Materials Assessment Tool for Audiovisual Materials (PEMAT-A/V) instrument and for Currency, Relevance, Authority, Accuracy, and Purpose (CRAAP) using a modified instrument within the health information education literature.
Results
The 174 videos with educational content scored as understandable but not actionable in PEMAT-A/V assessment, with videos created by healthcare providers (HCPs) performing better than those by non-HCPs in six of eleven domains of understandability. Videos overall scored well in relevance and purpose within the CRAAP assessment. Videos created by HCP’s (n = 99) scored higher than those by non-HCPs (n = 75) in relevance (score = 3.9 versus 3.3), authority (score = 4.8 versus 1.7), accuracy (score = 6.0 versus 3.5), and purpose (score = 7.1 versus 5.7) [all p < 0.001]. Regarding video engagement, the median number of views among the videos in the cohort was 604,450. Of all video views, 79.2% were of HCP-created videos and 20.8% were of non-HCP videos.
Discussion
Much of the contraceptive educational messaging on TikTok is understandable, relevant, and accurate. HCP-created videos overall scored better as compared to other creators, though even HCP-created videos score fail to provide actionable recommendations. The videos created by HCP also had greater metrics of engagement. Given this, we encourage HCPs to use social media to better inform its users.
Supplementary Information
The online version contains supplementary material available at 10.1186/s40834-024-00324-5.
Keywords: Contraception, Social media, Health education, TikTok, Birth control, Misinformation, Accuracy, Relevance, PEMAT
Background
Patients rely on social networks for contraception decision making [1], networks that are now largely comprised of people who post on social media [2]. Family members, peers, and social media have all affect reproductive-aged individuals’ contraception attitudes and behaviors [3]. TikTok is the world’s fastest growing social media platform and users ages 18–24 years are its largest demographic [4, 5]. A prior content analysis study [6] of the 700 most viewed contraception TikTok videos categorized 36% of the videos as educational, defined as informing or teaching about a birth control method. Close to half of the educational videos (49.6%) were created by healthcare providers (HCPs), and these videos had much higher levels of user engagement (views, shares, likes, comments). The other half of educational content videos were created by non-HCPs informed, most commonly, by the personal experience as current or former birth control users rather than by professional expertise. Recent reports suggest a rise of contraceptive misinformation on social media, often urging users to switch to less effective contraception, particularly non-hormonal methods, such as “natural cycles” [7, 8]. Furthermore, those who primarily derive birth control information from the internet and social media tend to have greater concerns about hormonal methods than those whose main source of information is a healthcare provider or a school setting [9]. Younger people aged 18–24 years, the same demographic who most use TikTok, report lower levels of contraceptive access and use in recent years within the United States [10]. The potential for influential misinformation on social media is particularly alarming when reproductive options, particularly abortion, are limited through restrictive U.S. state laws. To gauge the degree of misinformation on TikTok, we performed a secondary analysis of the TikTok content analysis study by Stoddard et al. (2024) to evaluate the understandability, actionability, and scientific accuracy of those most-viewed videos containing contraception educational content [6]. These results represent the next step in understanding the potential cultural and personal influence of social media on people’s contraceptive decision making.
Methods
In Stoddard et al. (2024), we performed a content analysis of the 700 most viewed contraception TikTok videos at the time of download in December 2021 [6]. We searched based on a previously published list of hashtags associated with contraception on social media for each of the six birth control methods: intrauterine devices (#iud), implant (#nexplanon), oral contraceptive pills (#birthcontrolpills), vaginal ring (#nuvaring), birth control patch (#birthcontrolpatch) and birth control injection (#depo) [11]. In addition to this list, we added the search terms #birthcontrol and #planb [6]. From this original cohort, we coded 252 videos (36%) as educational, defined as “containing at least one take-home point or fact the creator is educating the viewer on” [6]. We limited this new analysis to just these videos coded in Stoddard et al. [6] as educational as we recognized we could assess accuracy only for those videos with factually derived, as opposed to emotionally derived, content (examples of the latter include jokes about birth control or political messaging).
To measure video understandability and actionability, we used fourteen validated measures within the Patient Education Materials Assessment Tool for Audiovisual Materials (PEMAT-A/V) [12], an instrument that has previously been used to evaluate the understandability of social media content related to the IUD [13] and vasectomy [14]. PEMAT-A/V describes understandable materials as those that allow individuals with varying educational backgrounds and levels of health literacy to “process and explain key messages” of the materials. Actionable materials allow for these individuals to “identify what they can do” with the information [12]. We rated each item on a binary scale of yes (1) or no (0). PEMAT scores were summed across individual videos and calculated as a percentage of the sum of the total points scored by the total points possible; higher percentages suggest greater understandability and actionability [12]. Two reviewers (JKC, MEH-K) independently scored videos, and a third (DB) adjudicated discrepancies reaching substantial agreement for all items (Cohen’s Kappa > 0.7).
To assess video reliability and accuracy, we used relevant questions from a previously modified version of the Currency, Relevance, Authority, Accuracy, and Purpose (CRAAP) test [15], which has been employed in other studies evaluating the accuracy of online health information [16–18]. Initially developed to qualitatively assess the reliability of online educational resources [19], the CRAAP test has been adapted for quantitative assessment of health information scored with points within the domains of currency (0–5), relevance (0–4), authority (0–5), accuracy (0–7), and purpose (0–8) (Supplement) [15]. Compared to other instruments to evaluate medical information, the CRAAP test’s criteria focus on broad patient education rather than discrete, specific medical content, and thus is more suitable for social media content [19]. Two board-certified gynecologists (DB, NJ) coded the videos, with discrepancies adjudicated if Cohen’s Kappa was < 0.7.
We reported frequencies for each question within each domain for both PEMAT-A/V and CRAAP assessments. We performed a Chi-square test to compare groups for each question. We calculated means and standard deviations and medians and interquartile ranges (IQR) for summative domain scores. We used an independent samples t-test to evaluate differences in means and a Wilcoxon rank sum test to test medians between groups. We used Stata version 16.0 for analyses. This study was declined for review by Mass General Brigham IRB as non-human research.
Results
Of the 252 videos coded as educational within the original dataset [6], 174 videos were included in this secondary analysis. Seventy-eight of the original videos were ineligible due to deletion (n = 46) or restriction (n = 16) of creators’ accounts since the original analysis, content being too brief or in a form that precluded assessment by the PEMAT or CRAAP instruments (n = 11), or nearly duplicate videos from the same creator (n = 5). Not all videos could be evaluated within each question of each domain, leading some videos to be scored as “not applicable.” Physicians, physician assistants, midwives, and nurses as HCPs created 57% (n = 99) of the eligible videos: these videos on averaged had more views than those created by non-HCPs. The total number of views for all videos was 424,306,029 and the median was 604,450 per video with 9.8% of videos having over 5 million views. Of the 17 videos having over 5 million views, 15 (88.2%) were HCP-created and 2 (11.8%) were non-HCP created videos (p = 0.008). Of all views, 79.2% were of HCP-created videos and 20.8% were of non-HCP videos. The total number of likes was 32,647,492 and the median number was 58,100 per video with 60.5% coming from HCP-created videos and 39.5% from non-HCP videos. The total number of shares was 1,192,437 and the median number was 915 per video with 55.4% from HCP-created videos and 44.6% from non-HCP videos.
The mean score of all educational videos on PEMAT-A/V criteria for understandability was 74.7% (SD ± 17.4%, range 28.6–100) and for actionability was 37.7% (SD ± 40.7%, range 0-100%) of the total possible points, with established thresholds set at “understandable” and “actionable” if scored at 70% or higher [12]. Specific domains of understandability and actionability are presented for all videos overall and when comparing HCP with non-HCPs. Educational videos created by HCPs were significantly better in three domains (purpose, logical sequence, and clear images) and significantly worse in three domains (using common language, defined medical terms, and clear wording) of understandability. Healthcare providers were significantly better in the actionability domain of breaking down information. There were no differences in the PEMAT-A/V summative scores by creator type (Table 1).
Table 1.
Patient education materials assessment tool for audiovisual materials differences in mean scores by TikTok video creator type, N = 174
Overall | HCP | Non-HCP | P-value | |
---|---|---|---|---|
Domains of Understandability, n(%)* | ||||
Purpose is completely evident | ||||
No | 91 (52.3) | 39 (39.4) | 52 (69.3) | |
Yes | 83 (47.7) | 60 (60.6) | 23 (30.7) | < 0.001 |
Uses common, everyday language | ||||
No | 48 (27.8) | 42 (42.9) | 6 (8.0) | |
Yes | 125 (72.3) | 56 (57.1) | 69 (92.0) | < 0.001 |
Medical terms used only to familiarize audience with the term and are defined | ||||
No | 34 (19.7) | 26 (26.5) | 8 (10.7) | |
Yes | 139 (80.4) | 72 (73.5) | 67 (89.3) | 0.009 |
Uses active voice | ||||
No | 8 (4.6) | 5 (5.1) | 3 (4.0) | |
Yes | 165 (95.4) | 93 (94.9) | 72 (96.0) | 1.000 |
Breaks information into short sections | ||||
No | 2 (50.0) | 0 (0.0) | 2 (100.0) | |
Yes | 2 (50.0) | 2 (100.0) | 0 (0.0) | 0.333 |
Sections have informative headers | ||||
No | 3 (75.0) | 1 (50.0) | 2 (100.0) | |
Yes | 1 (25.0) | 1 (50.0) | 0 (0.0) | 1.000 |
Information presented in a logical sequence | ||||
No | 7 (4.1) | 1 (1.0) | 6 (8.0) | |
Yes | 166 (96.0) | 97 (98.9) | 69 (92.0) | 0.044 |
Summary provided | ||||
No | 3 (75.0) | 1 (50.0) | 2 (100.0) | |
Yes | 1 (25.0) | 1 (50.0) | 0 (0.0) | 1.000 |
Text is easy to read | ||||
No | 51 (32.9) | 29 (33.0) | 22 (32.8) | |
Yes | 104 (67.1) | 59 (67.1) | 45 (67.2) | 0.988 |
Can hear the words clearly | ||||
No | 6 (5.6) | 6 (11.1) | 0 (0.0) | |
Yes | 101 (94.4) | 48 (88.9) | 53 (100.0) | 0.027 |
Illustrations and photographs are clear and uncluttered | ||||
No | 73 (42.9) | 27 (28.1) | 46 (62.2) | |
Yes | 97 (57.1) | 69 (71.9) | 28 (37.8) | < 0.001 |
Understandability Mean (SD) Score* | 74.7 (17.4) | 75.1 (18.7) | 74.2 (15.7) | 0.762 |
Domains of Actionability, n(%)† | ||||
Clearly defines at least one action the user can take | ||||
No | 97 (55.8) | 55 (55.6) | 42 (56.0) | |
Yes | 77 (44.3) | 44 (44.4) | 33 (44.0) | 0.953 |
Addresses the user directly when describing actions | ||||
No | 98 (56.3) | 58 (58.6) | 40 (53.3) | |
Yes | 76 (43.7) | 41 (41.1) | 35 (46.7) | 0.489 |
Breaks down any action into manageable, explicit steps | ||||
No | 130 (74.7) | 68 (68.7) | 62 (82.7) | |
Yes | 44 (25.3) | 31 (31.1) | 13 (17.3) | 0.036 |
Actionability Median (IQR) Score* | 33.3 (0-100) | 33.3 (0-100) | 33.3 (0-100) | 0.667 |
*Overall mean scores were calculated as number of points divided by number of possible points for a percentage out of 100%
†Not all videos could be evaluated within each question of each domain, leading some videos to be scored as “not applicable.” Hence, some questions have small numbers reported
HCP = Healthcare professional
SD = Standard deviation
On the CRAAP assessment, HCP-created content demonstrated significantly greater summative scores for relevance (3.9 versus 3.3), authority (4.8 versus 1.7), accuracy (6.0 versus 3.5), and purpose (7.1 versus 5.7) than non-HCP-created content (all p<0.001) [Table 2]. All videos scored well on domains of currency, no matter the creator type. Overall, videos scored well in domains of relevance, though 33% of non-HCP created videos had overgeneralized their content beyond the specific relevance of the video topic. As would be expected, videos created by HCP scored very differently on all three domains of authority since the video content represents their professional expertise. Overall, the majority of videos scored well on domains of accuracy, though 26.7% of non-HCP content contradicted or could not be corroborated by published evidence. Almost 99% of videos were clear in stating their intended purpose, though the purpose differed between groups, with 99% of HCP creating content stating facts compared to non-HCP who presented a mix of facts (48.7%), opinion (43.2%), and propaganda (8.1%).
Table 2.
Currency, relevance, authority, accuracy and purpose domains and differences in mean scores for contraception TikTok videos by creator type, N = 174
Overall | HCP | Non-HCP | P-value | |
---|---|---|---|---|
Domains of Currency, n(%) | ||||
Date created* | ||||
< 1 year | 112 (64.7) | 59 (59.6) | 53 (71.6) | |
1–5 years | 61 (35.3) | 40 (40.4) | 21 (28.4) | |
> 5 years | 0 (0) | 0 (0) | 0 (0) | 0.101 |
Information out of date | ||||
No | 160 (95.2) | 98 (99.0) | 62 (89.9) | |
Yes | 8 (4.8) | 1 (1.0) | 7 (10.1) | 0.009 |
Embedded links still accessible? | ||||
Yes | 27 (15.7) | 13 (13.3) | 14 (18.9) | |
No | 8 (4.7) | 3 (3.1) | 5 (6.8) | |
None listed | 137 (79.7) | 82 (83.7) | 55 (74.3) | 0.267 |
Overall Currency Mean (SD) Score, possible range 0–5 | 2.91 (0.95) | 2.88 (0.84) | 2.96 (1.10) | 0.580 |
Domains of Relevance, n (%) | ||||
Information answers central question | ||||
Yes | 159 (92.4) | 99 (100.0) | 60 (82.2) | |
No | 13 (7.6) | 0 (0.0) | 13 (17.8) | < 0.001 |
Information identifies intended audience | ||||
Yes | 171 (100.0) | 98 (100.0) | 73 (100.0) | |
No | 0 (0) | 0 (0) | 0 (0) | n/a |
Information appropriate for needs of intended audience | ||||
Yes | 159 (94.1) | 97 (100.0) | 62 (86.1) | |
No | 10 (5.9) | 0 (0.0) | 10 (13.9) | < 0.001 |
Information avoids overgeneralization | ||||
Yes | 144 (82.8) | 94 (95.0) | 50 (66.7) | |
No | 30 (17.2) | 5 (5.1) | 25 (33.3) | < 0.001 |
Overall Relevance Mean (SD) Score, possible range 0–4 | 3.64 (0.85) | 3.92 (0.37) | 3.27 (1.12) | < 0.001 |
Domains of Authority, n(%) | ||||
Identity of author/source | ||||
Expert in field | 97 (58.4) | 95 (96.0) | 2 (3.0) | |
Medication user/patient | 40 (24.1) | 1 (1.0) | 39 (58.2) | |
Unclear | 29 (17.5) | 3 (3.0) | 26 (38.8) | < 0.001 |
Author’s credentials | ||||
Licensed medical professional | 97 (57.1) | 95 (96.0) | 2 (2.8) | |
Lived experience | 41 (24.1) | 1 (1.0) | 40 (56.3) | |
Unclear | 32 (18.8) | 3 (3.0) | 29 (40.9) | < 0.001 |
Author qualified to discuss topic | ||||
Yes | 133 (94.3) | 95 (100.0) | 38 (82.6) | |
No | 8 (5.7) | 0 (0.0) | 8 (17.4) | < 0.001 |
Overall Authority Mean (SD) Score, possible range 0–5 |
3.46 (1.96) | 4.82 (0.88) | 1.67 (1.48) | < 0.001 |
Domains of Accuracy, n{%) | ||||
Derivation of information | ||||
Evidence-based review | 23 (13.6) | 16 (16.3) | 7 (9.9) | |
Professional experience | 80 (47.3) | 79 (80.6) | 1 (1.4) | |
Individual lived experience | 41 (24.3) | 2 (2.0) | 39 (54.9) | |
Unclear | 25 (14.8) | 1 (1.0) | 24 (33.8) | < 0.001 |
Information supported by evidence? | ||||
Yes | 153 (87.9) | 98 (99.0) | 55 (73.3) | |
No | 21 (12.1) | 1 (1.0) | 20 (26.7) | < 0.001 |
What kind of evidence supports the claim? | ||||
Published evidence-based guidelines | 137 (80.6) | 94 (95.9) | 43 (59.7) | |
Expert/consensus opinion | 8 (4.7) | 4 (4.1) | 4 (5.6) | |
Individual lived experience | 8 (4.7) | 0 (0.0) | 8 (11.1) | |
None | 17 (10.0) | 0 (0.0) | 17 (23.6) | < 0.001 |
Overall Accuracy Mean (SD) Score, possible range 0–7 |
4.93 (1.93) | 6.02 (0.76) | 3.49 (2.06) | < 0.001 |
Domains of Purpose, n(%) | ||||
Purpose of information | ||||
Teaching | 30 (17.4) | 22 (22.2) | 8 (11.0) | |
Informing | 125 (72.7) | 72 (72.7) | 53 (72.6) | |
Persuading/entertainment | 12 (7.0) | 2 (2.0) | 10 (13.7) | |
Advertising | 5 (2.9) | 3 (3.0) | 2 (2.7) | 0.008 |
Intentions/purpose clear? | ||||
Yes | 169 (98.8) | 98 (99.0) | 71 (98.6) | |
No | 2 (1.2) | 1 (1.0) | 1 (1.4) | 1.000 |
Nature of information | ||||
Facts | 134 (77.5) | 98 (99.0) | 36 (48.7) | |
Opinion | 33 (19.1) | 1 (1.0) | 32 (43.2) | |
Propaganda | 6 (3.5) | 0 (0.0) | 6 (8.1) | < 0.001 |
Point of view objective and impartial | ||||
Yes | 143 (84.1) | 97 (99.0) | 46 (63.9) | |
No | 27 (15.9) | 1 (1.0) | 26 (36.1) | < 0.001 |
Political, ideological, cultural, religious biases | ||||
No | 165 (96.5) | 98 (99.0) | 67 (93.1) | |
Yes | 6 (3.5) | 1 (1.0) | 5 (6.9) | 0.084 |
Overall Purpose Mean (SD) Score, possible range 0–8 |
6.49 (1.33) | 7.09 (0.70) | 5.71 (1.55) | < 0.001 |
*Scored from the time of the initial download in December, 2021
HCP = Healthcare professional
SD = Standard deviation
Discussion
Contraception content posted on TikTok for the purpose of teaching or informing was understandable, but not actionable, according to PEMAT-A/V thresholds when created by both HCPs and non-HCPs. Thus, according to this validated metric to evaluate health information for people with varying levels of literacy, most videos conveyed a clear message but did not provide viewers with further resources to act on or implement this message. Overall, videos demonstrated good currency, relevance, authority, accuracy, and purpose, though, as would be expected, HCP-created content had greater scores of relevance, authority, accuracy, and purpose. Given the U.S. cultural mores about reproduction and birth control, sometimes resulting in mis- or even deliberate dis-information, we specifically used the CRAAP instrument for assessment to address the nature, partiality, and biases of the information provided in videos. It was reassuring to find the strong overall scores on the CRAAP assessment. It is not surprising that the videos created by HCPs provided more accurate, objective, and evidence-based information in comparison to content by non-HCP rooted more often in personal experience. Prior studies have evaluated the quality and reliability of TikTok content focused on specific contraceptive methods, such as oral contraceptives and implants [20, 21]. Others have evaluated accuracy of contraception content on social media without examining quality or understandability of content [22]. This study analyzes both accuracy and understandability of a wide range of contraception content on TikTok created for and by people with varying levels of health literacy.
Our study has several implications related to TikTok’s utility in disseminating reliable contraception educational content. Given the platform’s popularity [4] and users’ high level of contraception video engagement [6, 23], TikTok has potential to strongly influence contraception decision making, particularly among younger users. Our cohort contained about 56% of videos created by HCPs, but these videos were quite popular, making up close to 80% of all video views. In an investigation of over 1,100 U.S. young people assigned female at birth aged 15–29, 82% said they prefer to get their information about birth control from medical professionals. Yet only 43% of that cohort had received contraception information from medical professionals in the past year [24]. Of note when considering our engagement metrics, we saw many of the HCP videos were made by a small handful of content creators who could be considered “influencers”. While it is beyond the scope of this paper to discuss engagement metrics or reasons behind the virality of these videos, the high viewership of these videos still indicates the reach of this content and the number of TikTok users, especially young women, who are exposed to the information provided.
Given this viewership coupled with peoples’ stated desire to get their birth control information from medical professionals and with our findings of greater accuracy, reliability and objectivity of those videos created by HCPs, we should consider using this platform more to reach and teach the public at large. In an age of long wait times for healthcare appointments and short appointment visit times due to overloaded clinic templates, medical professionals should consider public education through social media as part of their educational mission. Other studies of TikTok have demonstrated strong engagement with videos created by medical professionals for dermatologic conditions [4] and sex education [23]. Direct to consumer education and communication are common in modern healthcare with the rise of electronic patient chart messaging and direct to consumer medical advertising. Our findings demonstrate that even HCP-created videos score low on providing actionable recommendations. Resources have been developed to help guide medical professionals to create TikTok and other social media content [25, 26]. Given the large consumption of social media, this platform should not be overlooked or underestimated as a means of more effectively imparting our expertise and informing the public at large.
Strengths of this study include the high popularity of the videos with large engagement metrics and the use of two validated scoring metrics to assess both accessibility to users and the accuracy of information. This study was limited by the difficulties in adapting some of these validated questions to the short-form videos of TikTok. We attempted to minimize subjectivity of PEMAT and CRAAP assessment through the use of two coders and assessment of coder agreement. Of the original dataset from Stoddard et al. [6], only the videos that contained educational content could be assessed for understandability and accuracy with the PEMAT and CRAAP surveys, with another 4% of those excluded for being too short to be assessed using these instruments. This smaller cohort affects the conclusions that can be made from our results. Furthermore, since this is a secondary analysis of prior data collected in 2021, it is possible that changes have occurred in the content posted on TikTok. A next step in investigation will be to collect a new snapshot of contraception content on TikTok today and assess for changes between now and 2021 regarding content themes and accuracy. Lastly, our analysis is strictly on the content available and, to a limited degree what content TikTok users find appealing, as assessed through degree of views, likes, and shares. However, this line of inquiry does not assess what information they retain or what type of specific messaging they find interesting or helpful for contraceptive decision making.
Conclusion
In an era of more extreme U.S. abortion restrictions, contraceptive use as primary prevention for unintended pregnancy is particularly important. Contraception content on social media has the potential to inform, either positively or negatively, the contraceptive decision-making process. Our findings are overall reassuring, with the most viewed contraception videos on TikTok demonstrating understandable and accurate educational content. HCPs must be aware of the inaccurate content that their patients might be seeing and that may be informing their healthcare decisions. As informed reproductive health professionals, we should consider using this and other social media platforms ourselves to create and promote highly reliable contraception education to the public.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Acknowledgements
Not applicable.
Abbreviations
- PEMAT-A/V
Patient Education Materials Assessment Tool for Audiovisual Materials (PEMAT-A/V)
- CRAAP
Currency, relevance, authority, accuracy, and purpose
- HCP
Healthcare provider
- IUD
Intrauterine device
- IQR
Interquartile range
Author contributions
JKC contributed to study design and led data analysis and manuscript development; MEH-K contributed to data analysis and manuscript development; AP led data analysis and contributed to manuscript development ; RES led data collection and contributed to manuscript development; NRJ contributed to data analysis and manuscript development; JRB contributed to study design, data analysis and manuscript development; NYS contributed to study design and manuscript development; ASK contributed to study design and manuscript development; DB contributed to study design, data analysis and manuscript production. All authors approved the final version of the manuscript.
Funding
Not applicable.
Data availability
The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.
Declarations
Ethics approval and consent to participate
This study was deemed exempt by Mass General Brigham IRB as non-human research.
Consent for publication
Not applicable.
Competing interests
The authors declare no competing interests.
Footnotes
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
References
- 1.Yee L, Simon M. The role of the social network in contraceptive decision-making among young, African American and Latina women. J Adolesc Health. 2010;47:374–80. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 2.Kofinas JD, Varrey A, Sapra KJ, Kanj RV, Chervenak FA, Asfaw T. Adjunctive social media for more effective contraceptive counseling: a randomized controlled trial. Obstet Gynecol. 2014;123:763–70. [DOI] [PubMed] [Google Scholar]
- 3.Ketterer T, Sieke E, Min J, Quidgley-Martin M, Barral RL, Akers A, Adams A, Miller E, Miller MK, Mollen C. Contraception initiation in the Emergency Department: adolescent perspectives. J Adolesc Health. 2024;75(1):147–54. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 4.Villa-Ruiz C, Kassamali B, Mazori DR, Min M, Cobos G, LaChance A. Overview of TikTok’s most viewed dermatologic content and assessment of its reliability. J Am Acad Dermatol. 2021;85:273–4. [DOI] [PubMed] [Google Scholar]
- 5.Ceci L. TikTok: distribution of global audiences 2024, by age and gender. Statista 2024. Accessed April 20, 2024. https://www.statista.com/statistics/1299771/tiktok-global-user-age-distribution/
- 6.Stoddard RE, Pelletier A, Sundquist EN, et al. Popular contraception videos on TikTok: an assessment of content topics. Contraception. 2024;129:110300. [DOI] [PubMed] [Google Scholar]
- 7.Weber L, Malhi S. Women are getting off birth control amid misinformation explosion. The Washington Post 2024. Accessed April 17, 2024. https://www.washingtonpost.com/health/2024/03/21/stopping-birth-control-misinformation/
- 8.Pfender EJ, Devlin MM. What do Social Media influencers say about Birth Control? A content analysis of YouTube vlogs about Birth Control. Health Comm. 2023;38:3336–45. [DOI] [PubMed] [Google Scholar]
- 9.Vieth SJ, Hartmann-Boyce J, Maass N, Jani A. Survey of young women’s state of knowledge and perceptions about oral contraceptives in Germany. AJOG Glob Rep. 2022;2:100119. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 10.Kavanaugh ML, Friedrich-Karnik A. Has the fall of Roe changed contraceptive access and use? New research from four US states offers critical insights. Health Affairs Scholar. 2024;2:qxae016. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 11.Merz AA, Gutiérrez-Sacristán A, Bartz D, et al. Population attitudes toward contraceptive methods over time on a social media platform. Am J Obstet Gynecol. 2021;224(6):597.e1-597.e14. [DOI] [PubMed] [Google Scholar]
- 12.Shoemaker SJ, Wolf MS, Brach C. Development of the Patient Education materials Assessment Tool (PEMAT): a new measure of understandability and actionability for print and audiovisual patient information. Patient Educ Couns. 2014;96:395–403. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 13.Wu J, Trahair E, Happ M, Swartz J. TikTok, #IUD, and user experience with Intrauterine devices reported on Social Media. Obstet Gynecol. 2023;141(1):215–7. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 14.Lang JJ, Zane G, Hong S, Demeter J, El-Zawahry A, Puneet S, Obi E. Assessing vasectomy-related information on YouTube: an analysis of the quality, understandability, and actionability of information. Am J Mens Health. 2022;16:15579883221094716. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 15.Garcia M, Daugherty C, Ben Khallouq B, Maugans T. Critical assessment of pediatric neurosurgery patient/parent educational information obtained via the internet. J Neurosur Pediatr. 2018;21:535–41. [DOI] [PubMed] [Google Scholar]
- 16.Mehra M, Brody PA, Kollapaneni SS, Sakhalkar O, Rahimi S. Evaluating the quality, readability, and activity of Online Information on Brain Arteriovenous malformations. Cureus. 2023;15(9):e45984. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 17.Patterson B, Clark L, Sanchez-Birkhead AC, Martinez LI, Egger MJ. Developing a web-based toolkit for new mothers about postpartum pelvic floor health in collaboration with a professional medical association. J Med Libr Assoc. 2021;109(4):667–71. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 18.McNamee LA, Shakartzi H, Wasser TE, Li Y, Kim CS. Assessing the accuracy and quality of YouTube videos on early pregnancy loss. Heliyon. 2022;8(12):e11947. [DOI] [PMC free article] [PubMed]
- 19.Blakeslee S. The CRAAP test. LOEX Quarterly 2004;21:6–7. Accessed March 24, 2024. https://commons.emich.edu/loexquarterly/vol31/iss3/4
- 20.Shackleford M, Horvath A, Repetto M, Thi A, Twells R, Sanders M, Fernandez S, Netski D, Batra K, Gomez N, Free L. An analysis of oral contraceptive related videos on TikTok. AJOG Glob Rep. 2024;12(3):100364. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 21.Sütcüoğlu BM, Güler M. Social Media Videos on contraceptive implants: an Assessment of Video Quality and Reliability. J Pediatr Adolesc Gynecol. 2024;37(1):39–44. [DOI] [PubMed] [Google Scholar]
- 22.Gulifeiya A, Moustafa Q, Comfort Onyinye E, Jarboue SSA. Global Social Media Analysis of the Accuracy and Specific Characteristics of Information Shared about Family Planning and Contraception. Int J Med Net. 2023;1:40–9. [Google Scholar]
- 23.Fowler LR, Schoen L, Smith HS, Morain SR. Sex education on TikTok: a content analysis of themes. Health Promot Pract. 2021;23:739–41. [DOI] [PubMed] [Google Scholar]
- 24.Pleasants E, Whitfield B, Pleasure ZH, et al. Adolescents and young adults’ sources of Contraceptive Information. JAMA Netw Open. 2024;7(9):e2433310. [DOI] [PMC free article] [PubMed] [Google Scholar]
- 25.8 Tips for Medical Practices and Doctors on TikTok (Includes Guide to Getting Started). Insight Marketing Group. https://insightmg.com/8-tips-for-medical-practices-and-doctors-on-tiktok-includes-guide-to-getting-started/. Accessed October 7, 2024.
- 26.Khalil C, Van Deen W, Dupuy T, Bonthala N, Almario C, Spiegel B. Developing patient-centered inflammatory bowel disease–related Educational videos optimized for social media: qualitative research study. JMIR Med Educ. 2020;6:e21639. [DOI] [PMC free article] [PubMed] [Google Scholar]
Associated Data
This section collects any data citations, data availability statements, or supplementary materials included in this article.
Supplementary Materials
Data Availability Statement
The datasets used and/or analyzed during the current study are available from the corresponding author on reasonable request.