Skip to main content
AMIA Summits on Translational Science Proceedings logoLink to AMIA Summits on Translational Science Proceedings
. 2023 Jun 16;2023:261–270.

Generalizable Natural Language Processing Framework for Migraine Reporting from Social Media

Yuting Guo 1, Swati Rajwal 2, Sahithi Lakamana 1, Chia-Chun Chiang 3, Paul C Menell 3, Adnan H Shahid 3, Yi-Chieh Chen, Pharm D 4, Nikita Chhabra 5, Wan-Ju Chao 6, Chieh-Ju Chao 7, Todd J Schwedt 5, Imon Banerjee 8,9, Abeed Sarker 1
PMCID: PMC10283091  PMID: 37350878

Abstract

Migraine is a highly prevalent and disabling neurological disorder. However, information about migraine management in real-world settings is limited to traditional health information sources. In this paper, we (i) verify that there is substantial migraine-related chatter available on social media (Twitter and Reddit), self-reported by those with migraine; (ii) develop a platform-independent text classification system for automatically detecting self-reported migraine-related posts, and (iii) conduct analyses of the self-reported posts to assess the utility of social media for studying this problem. We manually annotated 5750 Twitter posts and 302 Reddit posts, and used them for training and evaluating supervised machine learning methods. Our best system achieved an F1 score of 0.90 on Twitter and 0.93 on Reddit. Analysis of information posted by our ‘migraine cohort’ revealed the presence of a plethora of relevant information about migraine therapies and sentiments associated with them. Our study forms the foundation for conducting an in-depth analysis of migraine-related information using social media data.

Introduction

Electronic health records (EHRs) encapsulate knowledge specific to patients’ health, and there is now widespread use of EHR data for studying targeted health-related factors. However, EHRs have limitations in capturing knowledge about factors influencing patients’ well-being, such as social determinants, behavior, lifestyle, personal interests, mental health, and self-management of health problems. This additional information is often captured in patient-generated data on social media, which have thus emerged as key sources of health information that captures daily information from real-world settings directly from patients.1 Data from social media has been used to supplement the information obtained from traditional sources including EHRs and knowledge obtained from clinical trials.24 Recent advances in data science and artificial intelligence have led to the increasing use of natural language processing (NLP) approaches for analyzing free-text, noisy, big data from social media platforms, such as Twitter and Reddit.

Given a health topic, the automatic detection and curation of relevant health data from a social network on the topic are only feasible if (i) there are sufficient numbers of subscribers who self-report their health condition/status associated with that topic, and (ii) these self-reports can be automatically distinguished from general posts on the topic. Once these posts related to a specific topic can be identified and curated, the data can be used to study population or cohort-level factors.5 Over recent years, several studies have attempted to build targeted cohorts from social media data and then study the information posted by the cohorts to understand health-related factors. Some of these cohort-level studies have utilized NLP to analyze information on Twitter and showed the potential to improve patient-centered outcomes. For example, a recent study from our group developed an NLP pipeline to identify patients who self-reported breast cancer on Twitter.6 Breast cancer-related tweets were first detected using keywords and hashtags, and a machine learning classifier was trained using manually annotated data to identify posts that represented personal experiences of breast cancer (i.e., self-reports) and separate them from the many generic posts (e.g., posts raising awareness or sharing news articles about breast cancer). Qualitative analysis of tweets extracted from users who self-reported breast cancer showed a wealth of information regarding mental health concerns, medication side effects, and medication adherence, demonstrating the tremendous potential of studying patient-centered outcomes that might not otherwise be available in EHRs. Similar approaches were applied to build cohorts of patients for whom comprehensive data may not be available from EHRs or populations excluded from traditional studies such as clinical trials. Such studies have focused on topics including but not limited to drug safety during pregnancy,7,8 substance use,9,10 intimate partner violence1113, and chronic stress.14,15 Since social media data are constantly generated, once the cohort-building methods are developed, they can be deployed to continuously grow the cohort, increase the volume of collected data, and conduct long-term studies. Recent studies have also shown that demographic characteristics of the cohorts (e.g., geolocation, race, gender) can be estimated fairly accurately, based on self-reported information, once large amounts of data are collected.16 Broadly speaking, these recent studies have helped establish social media as an important source of digital health data and have opened up opportunities to study various health topics using a different lens than traditional studies.

Building on our past work in this space, we focus on the topic of migraine in this paper. Migraine is a highly prevalent and disabling neurological disorder with a one-year prevalence of about 12% in the general population.17 A few recent studies have investigated migraine-related information on social media, including a study using NLP and sentiment analysis to study migraine-associated tweets classified as either “informative” or “expressive” tweets. They found that informative tweets were more likely to demonstrate extreme sentiment when they were positive tweets as opposed to negative tweets and that expressive tweets were more likely to demonstrate extreme sentiment when they were negative tweets as opposed to positive tweets.18 The study, however, did not explicitly attempt to filter out posts that were not based on personal experiences. From an informatics perspective, it is also unclear if the methods used in the study can be ported to other social media platforms. While social media presents an important source of migraine-related information, these research gaps need to be addressed to establish social media as a long-term and sustainable resource for migraine-related research. In this study, we aim to develop a platform-independent framework to study patient-reported outcomes related to migraine treatment posted on social media by developing a robust NLP pipeline. The study particularly focuses on the development and evaluation of the essential component of the pipeline, which is about identifying patients who self-report migraine on social media (Twitter and Reddit). Specifically, we first implemented and evaluated a supervised machine learning classifier for detecting migraine self-reports using data from Twitter. We then performed a thorough error and bias analysis of the model. We evaluated the portability of this model by evaluating its performance on manually-annotated data from the social network Reddit, which has substantially different characteristics compared to Twitter. We manually analyzed some of the data to assess the types of information posted by the cohort members, including but not limited to the mention of medications for migraine. As a use-case of this powerful pipeline, we demonstrate an automated analysis of sentiments regarding commonly used migraine medications from the tweets/posts of self-reported migraine sufferers. To the best of our knowledge, this is the first description of an NLP platform that analyzes social media posts regarding migraine from a self-identifying cohort.

Methods

Data collection

We collected publicly available English posts related to migraine and the user metadata from Twitter using its academic streaming application programming interface (API). We used the keyword “migraine” and the generic and brand names of migraine-specific medications, including sumatriptan (Imitrex), rizatriptan (Maxalt), naratriptan (Amerge), eletriptan (Relpax), zolmitriptan (Zomig), frovatriptan (Frova), almotriptan (Axert), erenumab (Aimovig), fremanezumab (Ajovy), galcanezumab (Emgality), eptinezumab (Vyepti), ubrogepant (Ubrelvy), rimegepant (Nurtec), atogepant (Qulipta) to select migraine-related tweets preliminarily. Since medication names are often misspelled on social media, we used an automatic misspelling generator to obtain common misspellings for the medication names.19 We collected the tweets between March 15, 2022, and June 28, 2022, accruing 343,559 migraine-related tweets. We also collected migraine-related posts from Reddit (r/migraine, r/NDPH, r/headache, r/headaches subreddit) using PRAW Reddit API, collecting a total of 3625 posts. The preprocessing of all posts included lowercasing, normalization of numbers, usernames, URLs, hashtags, and text smileys, and adding extra marks for capital words, hashtags, and repeated letters.

Annotation process

As the first step, 500 tweets were labeled by five clinician annotators to determine whether the tweets were self-reports about migraine from Twitter users. The initial mean agreement between any two annotators was 0.60 (Cohen’s kappa20). To optimize the agreement between annotators, we developed an annotation guideline to determine whether tweets should be labeled as self-reporting migraine. The guideline details are included in the supplementary file on Google Drive1, and the key rules are included in Table 1. After the guideline was implemented, we extracted an additional 250 tweets, which the same five clinicians annotated. The mean agreement improved to 0.78 between 4 annotators but was 0.66 when adding the annotation of the 5th annotator. The lead clinician investigator and headache specialist (CC) resolved the discrepancies according to the annotation guideline to ensure agreement. Finally, an additional 5000 tweets were labeled by the five annotators (1000 for each person). A total of 5750 tweets labeled by clinicians were used to train the NLP model, which was then applied to Reddit posts. For the annotation of Reddit posts, the same annotation guideline was applied. A total of 303 Reddit posts were annotated separately by the five clinician annotators.

Table 1.

Summary of Annotation Guideline for identifying self-reported migraine patients. Annotators were asked to label tweets or Reddit posts as Y or N.

Rules Examples tweets Label
Using first-person references such as “I” or “my” when referring to migraine-related experiences or intake of migraine-specific medications. “couldnt get all my work done yesterday because of a migraine so working extra hard today, pray for me” Y
Evidence of past or future, but not current migraine, as long as it is referring to the own experience of a user, should be labeled “Yes”. “The biggest health improvement since stopping hormonal birth control a year ago has been no more migraines. I used to get them monthly and they were particularly bad when I was on the arm implant.” Y
In some instances, the users might use “them” “they” and “that” to refer to migraine. If the sentence includes first-person references, it would be labeled “Yes” “Do you feel it when they give you migraines? I usually wake up with that on weekends.” Y
Emojis should be interpreted in this study. If a tweet includes a sentence or emoji referring to the user’s own feelings or experiences, those tweets should be labeled “Yes”. Note that links included in tweets should not be interpreted. “high on a rizatriptan and deep throated leftover sushi. feeling: 😵💫” Y
When a tweet does not include first-person references, as it could potentially refer to someone else’s (like their partner or close family member/friend) migraine, it should be labeled “No”. “Who else gets migraines? Do you ever get a weird “drunk-like” feeling in your eyes when its going away/lingering?” N
Advertisements, factual statements, or irrelevant statements should be labeled as “No”. “How is headache treated in emergency room in past decade? Opioids use reduced from more than 50% to just below 30%. Maybe use of migraine specific medications (triptans/ergots) has increased? Nope! Still less that 5% receive this therapy, our study shows <link>” N

Text classification model for identifying self-reporting of migraine

The primary objective of this module is to develop a generalizable free-text classification model that can distinguish between posts related to migraine self-reporting from the generic posts since the posts are retrieved only based on simple keywords (for Twitter) or the subreddit (for Reddit). We experimented with transformer-based models2 since they have achieved state-of-the-art performance on a wide range of NLP tasks.21,22 We investigated six transformer-based models to construct the migraine self-report classifier——RoBERTa,22 SciBERT,23 BioBERT,24 BioClinicalBERT,25 Clinical_KB_BERT,26 and BERTweet.27 These transformer-based models were pre-trained on large-scale text data from different domains. RoBERTa was pre-trained on the open domain, including English Wikipedia, news, books, Reddit comments, and story-like text from scratch; SciBERT was pretrained on scientific publications from scratch; BioBERT was initialized from BERT and pretrained on biomedical research papers; BioClinicalBERT was initialized from BioBERT and pretrained on clinical notes; Clinical_KB_BERT was initialized from BioBERT and pretrained on clinical notes with the knowledge base of the Unified Medical Language System (UMLS); BERTweet was pretrained on English tweets from scratch. We split the annotated Twitter data into the training, validation, and test sets at a 64-16-20 ratio. The model was trained on the training set for ten epochs and evaluated on the validation set for each epoch. The model checkpoint that achieved the best result on the validation set was selected for evaluation on the test set.

We applied the Twitter model to classify the Reddit posts without fine-tuning the model on the Reddit data. First, we removed duplicate posts from the Reddit dataset to get a unique cohort. We evaluated model performance by computing precision, recall, and F1 score for the positive class while comparing predictions against the expert annotations. Reddit posts are often longer than Twitter, so we split the posts at the sentence level and ran the classification model for each sentence. We finally combined the predictions and a post was considered to be positive if at least one sentence within the post was classified as self-reporting. We computed the 95% confidence intervals for the F1 scores via bootstrap resampling.

Post-classification analysis

Error analysis

To explore the potential reason for misclassifications of self-reporting, we conducted a thorough error analysis on the model that achieved the best performance. We manually analyzed the contents of false positives from both Twitter and Reddit and categorized the error types.

Bias analysis

In recent years, a number of studies have reported that deep learning methods can generate models that are biased towards specific populations.28 To explore whether there exists a bias in our model, we conducted a bias analysis by testing whether changing the gender-, race- or any identity-related words would change the model prediction. Specifically, for gender bias analysis, we changed gender-related words into words related to a different gender and vice versa. We checked if the change in the gender-related word would alter the classification result or significantly affect the word importance of the gender-related word in the tweet. Similarly, for race and identity bias analysis, we replaced relevant words with alternative possibilities.

Longitudinal data collection and analysis

To study longitudinal information posted by Twitter users who self-reported having migraine, we first used the best-performing classifier to automatically classify all migraine-related posts we initially collected. We then used the API to collect all the past posts of those users whose migraine-related posts were classified as positive. This provided us with many posts from people with migraine that we could analyze to identify the presence of other migraine-related information.

Sentiment analysis

From the timeline tweets of users who were detected to self-report migraine, we extracted tweets that included any of the migraine medication names mentioned in Table 2. We performed sentiment analysis using a publicly available sentiment analysis tool named VADER,29 which can assign a sentiment score ranging from -1 (most extreme negative) to 1 (most extreme positive) for a tweet. We analyzed the sentiment distributions of the migraine medications. To reduce the data noise, headache specialists (CC and TS) grouped the medications, and the details are shown in Table 2. We reported the frequency of occurrence of each medication group. For Twitter posts, if the same medication group is mentioned in multiple tweets/posts from the same user, the tweet/post that obtained the median sentiment score was chosen to represent the sentiment score of the user for that medication group. We also reported the mean, median, and standard deviation of the sentiment scores for each medication group. Additionally, we used kernel density estimate (KDE) plot to visualize the sentiment distribution. In the case of external validation using Reddit posts, we used the same tool and reported results in a similar manner. We performed sentiment analysis on the Reddit posts which were predicted as self-reporting (positive) and contained the medication keywords.

Table 2.

The expert-selected medications and their groupings.

Medication group Medications
Topiramate Topiramate (Topamax)
Beta Blockers Propranolol (Inderal), Atenolol (Tenormin), Metoprolol (Toprol)
Tricyclic antidepressants Amitriptyline (Elavil), Nortriptyline (Pamelor)
OnabotulinumtoxinA OnabotulinumtoxinA (Botox)
CGRP monoclonal antibodies Erenumab (Aimovig), Galcanezumab (Emgality), Fremanezumab (Ajovy), Eptinezumab (Vyepti)
Gepants Atogepant (Qulipta), Ubrogepant (Ubrelvy), Rimegepant (Nurtec)
Triptans Sumatriptan (Imitrex), Rizatriptan (Maxalt), Eletriptan (Relpax), Naratriptan (Amerge), Frovatriptan (Frova), Zolmitriptan (Zomig), Almotriptan (Axert)
Lasmiditan Lasmiditan (Reyvow)
Dihydroergotamine Dihydroergotamine (DHE), Migranal, Trudhesa

Results

Self-Reporting of Migraine Classification Results

Table 3 lists the statistics of the training, validation, and test sets for Twitter and of the external test data from Reddit. Table 3 shows the quantitative performance of transformer-based classification models. We observed that RoBERTa achieved the best recall score, BioBERT and BERTweet achieved the best precision score, and BERTweet achieved the best F1 score. In general, BERTweet outperformed other models on this task. It can be attributed to the fact that BERTweet was pre-trained on English tweets, which may have enabled the model to capture the nuances of social media language more efficiently. We applied the optimal models (RoBERTa and BERTweet) to the Reddit dataset and report the performances in Table 4.

Table 3.

The statistics of training, validation, and test set – Twitter and Reddit posts.

Dataset Positive cases Negative cases Total
Training 1034 2612 3646
Validation 272 647 919
Test 761 328 1089
External Test: Reddit 226 76 302

Table 4.

The classification results of 6 transformer-based models.

Model Precision Recall f1-score (95% CI)
Twitter Data
RoBERTa 0.84 0.95 0.89 (0.87-0.91)
SciBERT 0.87 0.89 0.88 (0.85-0.90)
BioBERT 0.88 0.89 0.88 (0.86-0.91)
BioClinicalBERT 0.85 0.91 0.88 (0.86-0.91)
BERTweet 0.88 0.91 0.90 (0.87-0.92)
Clinical_KB_BERT 0.86 0.91 0.88 (0.85-0.90)
External: Reddit data
RoBERTa 0.91 0.95 0.93 (0.91-0.95)
BERTweet 0.89 0.90 0.90 (0.87-0.93)

Self-Reporting of Migraine Error analysis

Due to a large number of Twitter posts available on the topic, precision is perhaps more important than recall for this work. Given that BERTweet achieved the best precision and F1 score compared with other models, we performed an error analysis on the false positives predicted by this model. We observed that one type of false positives occurred when the tweet was ambiguous, e.g. mentioned the word “migraine” without a clear indication if it was a self-reporting. For instance, “adulthood is preparing for migraines by taking ibuprofen the night before lmao”. Another type of false positive occurred when the tweet was incomplete or lacked context. For instance, These migraines Ain no ho”. According to the annotation guideline, both of the two instances should be labeled as N due to the absence and presence of a first-person reference. However, it is difficult to ascertain, even for a human annotator, whether the user had a migraine or not, which can be more difficult for the NLP model.

Self-Reporting of Migraine Bias analysis

We performed the manual analysis on 5% of all the tweets in our test set. Table 5 presents examples of this analysis. We observed that changes in gender/race/identity-related words slightly affected the word importance distributions, but the differences were trivial and did not alter the classification results. This provided further evidence that our model is not biased towards/against people with specific gender identities, race groups or other identities.

Table 5.

The examples for bias analysis on the predictions of self-reported and non-self-reported migraine posts. The pharmaceutical company name is presented as “<company>” in the tweet and corresponding figure. TP denotes true positive, which means that both the human annotation and the model prediction are self-report; TN denotes true negative, which means that both the human annotation and the model prediction are non-self-report.

Social Media Example TYPE Word Importance
Twitter I know nearly every girl says this, but I always feel sick and have a tummy ache and migraine and I bet ya a tenner more than anyone else TP Figure 2(a)
Twitter <company> Real-World Study Highlights Increased Healthcare Utilization Among Americans with Episodic Migraine having Higher Levels of Migraine-Related Disability TN Figure 2(b)
Reddit This is Asher. He likes to pet my face with his soft toe beans when I’m in bed with a migraine. I’ve had about 19 headache/migraine days in the last month. I just started Aimovig on Tuesday, so I’m hoping to reduce the number of migraine days, just not the number of cuddling days. My sweet boy. TP Figure 2(c)
Reddit My husband is a teacher. Yesterday, he turned on the lights in his classroom, to which a young woman in his class visibly flinched. He turned off the lights again right away and she breathed a sigh of relief. Upon asking if shes okay, she said she was absent the last couple of days due to migraines. After having seen me suffer, my SO now recognises some migraine signs. Id like to think his simple action of turning off the lights made this girls day a little more bearable. TN Figure 2(d)

Sentiment analysis

Table 6 shows the results of sentiment analysis for both Twitter and Reddit, and Figure 3 illustrates the sentiment distributions of the medications for which the frequency is greater than 40. On Twitter, the sentiment scores of onabotulinumtoxinA, triptans, topiramate, beta-blockers, and tricyclic antidepressants are concentrated near the 0 mean. This suggests that the sentiments for these medications tend to be neutral. In contrast, the sentiment distributions of CGRP monoclonal antibodies and gepants tend to be more positive. It is worth noting that the sentiment score was evaluated on the whole tweet rather than relative to the medication mentioned in the tweet. It is possible that the positive sentiment in the tweet is irrelevant to the medication. For example, “The woman behind me @<user_name> didn’t want her free chocolate chip cookie so she gave it to me! My day is looking up! On to my Botox injections! #EDS #lucky #winner #happy” and “Botox was approved for migraines!! Slowly but surely i’m making my symptoms manageable/not constant”. In the above examples, both tweets had positive sentiments, though only the sentiment of the second tweet refers directly to the medication itself. The sentiment analysis on Reddit posts shows clearer positive and negative trends. For example, beta-blockers have a positive sentiment score while topiramate posts have a mean negative sentiment. Note that conducting sentiment analysis at the medication level is a very challenging task, particularly in this setting where the data size is relatively small. To draw any reasonable clinical conclusions, further investigation with a larger dataset with equal presence of posts for each medication group is needed.

Table 6.

The sentiment distribution statistics for each medication group in the Twitter and Reddit datasets.

Medication Group Frequency Mean Median STDEV
Twitter
OnabotulinumtoxinA 521 0.05 0.5 0.14
Triptans 174 0 0.57 -0.05
CGRP monoclonal antibodies 103 0 0.54 -0.01
Gepants 87 0.16 0.5 0.14
Topiramate 55 0 0.5 -0.03
Beta Blockers 41 0 0.48 0.03
Reddit
Topiramate (Topamax) 32 -0.17 -0.36 0.71
Beta Blockers 18 0.39 0.63 0.64
Tricyclic Antidepressants 30 -0.3 -0.66 0.74
OnabotulinumtoxinA (botox) 41 0.01 0.18 0.75
CGRP monoclonal antibodies 41 -0.07 -0.3 0.77
Gepants 39 -0.04 0.25 0.78
Triptans 64 -0.03 0.06 0.8

Figure 3.

Figure 3.

The normalized sentiment distributions of the medications: (a) Twitter and (b) Reddit.

Conclusion

Monitoring of changes in migraine severity and responses to migraine treatments depends primarily on patient-reported feedback, information that is typically collected and documented in EHRs only during occasional doctor’s office visits. Since migraine typically affects individuals younger than 50 years old, people with migraine are adapted to social media (SM), such as Twitter, Facebook, Reddit. There is the potential of utilizing social media to fill the knowledge gap in the EHR. We proposed a robust NLP framework to perform a large-scale social media study of migraine. We first trained and evaluated a machine learning classifier to automatically detect self-reports of migraine from Twitter. Being trained on only the Twitter posts, the classification framework showed generalizability towards both shorter tweets as well as for the longer Reddit posts. Moreover, our pipeline displayed minimal bias towards gender, race, and identity-related words and may present a fair performance across the validation data. Finally, we presented a brief sentiment analysis of migraine medications as a possible use case for our system. We anticipate that our developed methods can be applied to collect information regarding migraine management in real-world settings and could be applied to other topics too.

Footnotes

2

The transformer-based models are open-source implemented by Hugging Face.

Figures & Table

Figure 1.

Figure 1.

The framework of our generalizable NLP system—model development and validation on Twitter data followed by additional evaluation on Reddit posts.

Figure 2.

Figure 2.

Examples from the bias analysis. The green color signifies positive attention to the words while red shows negative.

References

  • 1.Nittas V, Lun P, Ehrler F, Puhan MA, Mütsch M. Electronic Patient-Generated Health Data to Facilitate Disease Prevention and Health Promotion: Scoping Review. J Med Internet Res. 2019;21(10):e13320. doi: 10.2196/13320. doi:10.2196/13320. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 2.Conway M, Hu M, Chapman WW. Recent Advances in Using Natural Language Processing to Address Public Health Research Questions Using Social Media and ConsumerGenerated Data. Yearbook of medical informatics. 2019;28(1):208–217. doi: 10.1055/s-0039-1677918. doi:10.1055/s-0039-1677918. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Gonzalez-Hernandez G, Sarker A, O’Connor K, Savova G. Capturing the Patient’s Perspective: a Review of Advances in Natural Language Processing of Health-Related Text. Yearb Med Inform. 2017;26(1):214–227. doi: 10.15265/IY-2017-029. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Paul MJ, Sarker A, Brownstein JS, et al. Social Media Mining for Public Health Monitoring and Surveillance. Pacific Symposium on Biocomputing. World Scientific Publishing Co. Pte Ltd. 2016:468–479. doi:10.1142/9789814749411_0043. [Google Scholar]
  • 5.Ravindranath S, Zhao C, Tgavalekos K. Patient Status Indicator to Extract Key Temporal Changes in Continuous-Time Deterioration Risk Score. Critical Care Medicine. 2021;49(1) https://journals.lww.com/ccmjournal/Fulltext/2021/01001/374_Patient_Status_Indicator_to_Extract_Key.342.aspx . [Google Scholar]
  • 6.Al-Garadi MA, Yang YC, Lakamana S, et al. Michalowski M, Moskovitch R. Artificial Intelligence in Medicine. Springer International Publishing; 2020. Automatic Breast Cancer Cohort Detection from Social Media for Studying Factors Affecting Patient-Centered Outcomes; pp. 100–110. [Google Scholar]
  • 7.Golder S, Chiuve S, Weissenbacher D, et al. Pharmacoepidemiologic Evaluation of Birth Defects from Health-Related Postings in Social Media During Pregnancy. Drug Safety. 2019;42(3):389–400. doi: 10.1007/s40264-018-0731-6. doi:10.1007/s40264-018-0731-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Klein AZ, Gonzalez-Hernandez G. An Annotated Data Set for Identifying Women Reporting Adverse Pregnancy Outcomes on Twitter. Data in Brief. 2020;32:106249. doi: 10.1016/j.dib.2020.106249. doi:10.1016/J.DIB.2020.106249. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 9.Al-Garadi MA, Yang YC, Cai H, et al. Text Classification Models for the Automatic Detection of Nonmedical Prescription Medication Use From Social Media. BMC Medical Informatics and Decision Making. 2021;21(1):1–13. doi: 10.1186/s12911-021-01394-0. doi:10.1186/s12911-021-01394-0. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 10.Ghosh S, Misra J, Ghosh S, Podder S. Utilizing Social Media for Identifying Drug Addiction and Recovery Intervention. IEEE International Conference on Big Data (Big Data) 2020:3413–3422. doi:10.1109/BigData50022.2020.9378092. [Google Scholar]
  • 11.McCauley HL, Bonomi AE, Maas MK, Bogen KW, O’Malley TL. # MaybeHeDoesntHitYou: Social Media Underscore the Realities of Intimate Partner Violence. Journal of Women’s Health. 2018;27(7):885–891. doi: 10.1089/jwh.2017.6560. [DOI] [PubMed] [Google Scholar]
  • 12.Chu TH, Su Y, Kong H, Shi J, Wang X. Online Social Support for Intimate Partner Violence Victims in China: Quantitative and Automatic Content Analysis. Violence Against Women. 2021;27(3-4):339–358. doi: 10.1177/1077801220911452. doi:10.1177/1077801220911452. [DOI] [PubMed] [Google Scholar]
  • 13.Al-Garadi MA, Kim S, Guo Y, et al. Natural Language Model for Automatic Identification of Intimate Partner Violence Reports from Twitter. medRxiv. Published online 2021. [DOI] [PMC free article] [PubMed]
  • 14.Cao L, Zhang H, Li N, Wang X, Ri W, Feng L. Category-Aware Chronic Stress Detection on Microblogs. IEEE Journal of Biomedical and Health Informatics. 2022;26(2):852–864. doi: 10.1109/JBHI.2021.3090467. doi:10.1109/JBHI.2021.3090467. [DOI] [PubMed] [Google Scholar]
  • 15.Merolli M, Gray K, Martin-Sanchez F. Health outcomes and related effects of using social media in chronic disease management: A literature review and analysis of affordances. Journal of Biomedical Informatics. 2013;46(6):957–969. doi: 10.1016/j.jbi.2013.04.010. doi:https://doi.org/10.1016/j.jbi.2013.04.010. [DOI] [PubMed] [Google Scholar]
  • 16.Hinds J, Joinson AN. What demographic attributes do our digital footprints reveal? A systematic review. PLoS One. 2018;13(11):e0207112. doi: 10.1371/journal.pone.0207112. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 17.Pescador Ruschel MA, De Jesus O. Migraine Headache. Vol StatPearls [Internet]. Treasure Island (FL): StatPearls Publishing Accessed July 6; 2022. https://www.ncbi.nlm.nih.gov/books/NBK560787/ [PubMed] [Google Scholar]
  • 18.Deng H, Wang Q, Turner DP, et al. Sentiment analysis of real-world migraine tweets for population research. Cephalalgia Reports. 2020;3:2515816319898867. [Google Scholar]
  • 19.Sarker A, Gonzalez-Hernandez G. An unsupervised and customizable misspelling generator for mining noisy health-related text sources. Journal of Biomedical Informatics. 2018;88:98–107. doi: 10.1016/j.jbi.2018.11.007. doi:https://doi.org/10.1016/j.jbi.2018.11.007. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Cohen J. A Coefficient of Agreement for Nominal Scales. Educational and Psychological Measurement. 1960;20(1):37–46. doi:10.1177/001316446002000104. [Google Scholar]
  • 21.Devlin J, Chang MW, Lee K, Toutanova K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Proceedings of the 2019 North American Chapter of the Association for Computational Linguistics: Human Language Technologies. 2019 [Google Scholar]
  • 22.Liu Y, Ott M, Goyal N, et al. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv preprint arXiv:190711692. Published online 2019. [Google Scholar]
  • 23.Beltagy I, Lo K, Cohan A. Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conference on Natural Language Processing (EMNLP-IJCNLP) Association for Computational Linguistics; 2019. SciBERT: A Pretrained Language Model for Scientific Text; pp. 3615–3620. doi:10.18653/v1/D19-1371. [Google Scholar]
  • 24.Lee J, Yoon W, Kim S, et al. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics. 2020;36(4):1234–1240. doi: 10.1093/bioinformatics/btz682. doi:10.1093/bioinformatics/btz682. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Alsentzer E, Murphy J, Boag W, et al. 2nd Clinical Natural Language Processing Workshop. Association for Computational Linguistics; 2019. Publicly Available Clinical BERT Embeddings; pp. 72–78. doi:10.18653/v1/W19-1909. [Google Scholar]
  • 26.Hao B, Zhu H, Paschalidis I. Proceedings of the 28th International Conference on Computational Linguistics. International Committee on Computational Linguistics; 2020. Enhancing Clinical BERT Embedding using a Biomedical Knowledge Base; pp. 657–661. doi:10.18653/v1/2020.coling-main.57. [Google Scholar]
  • 27.Nguyen DQ, Vu T, Tuan Nguyen A. Empirical Methods in Natural Language Processing: System Demonstrations. Association for Computational Linguistics; 2020. BERTweet: A Pre-trained Language Model for English Tweets; pp. 9–14. doi:10.18653/v1/2020.emnlp-demos.2. [Google Scholar]
  • 28.Kurita K, Vyas N, Pareek A, Black AW, Tsvetkov Y. Proceedings of the First Workshop on Gender Bias in Natural Language Processing. Association for Computational Linguistics; 2019. Measuring Bias in Contextualized Word Representations; pp. 166–172. doi:10.18653/v1/W19-3823. [Google Scholar]
  • 29.Hutto C.J., Gilbert E. VADER: A Parsimonious Rule-based Model for. Eighth International AAAI Conference on Weblogs and Social Media. Published online 2014:18. [Google Scholar]

Articles from AMIA Summits on Translational Science Proceedings are provided here courtesy of American Medical Informatics Association

RESOURCES