Background
Social media has emerged as an important source of information generated by large segments of the population, which can be particularly valuable during infectious disease outbreaks. The recent outbreak of monkeypox led to an increase in discussions about the topic on social media, thus presenting the opportunity to conduct studies based on the generated data.
Objective
By analyzing posts from Twitter (subsequently rebranded X), we aimed to identify the topics of public discourse as well as knowledge and opinions about the monkeypox virus during the 2022 outbreak.
Methods
We collected data from Twitter focusing on English-language posts containing key phrases like “monkeypox,” “mpoxvirus,” and “monkey pox,” as well as their hashtag equivalents from August to October 2022. We preprocessed the data using natural language processing to remove duplicates and filter out noise. We then selected a random sample from the collected posts. Three annotators reviewed a sample of the posts and created a guideline for coding based on discussion. Finally, the annotators analyzed, coded, and manually categorized them first into topics and then into coarse-grained themes. Disagreements were resolved via discussion among all authors.
Results
A total of 128,615 posts were collected over a 3-month period, and 200 tweets were selected and included for manual analyses. The following 8 themes were generated from the Twitter posts: monkeypox doubts, media, monkeypox transmission, effect of monkeypox, knowledge of monkeypox, politics, monkeypox vaccine, and general comments. The most common themes from our study were monkeypox doubts and media, each accounting for 22% (44/200) of the posts. The posts represented a mix of useful information reflecting emerging knowledge on the topic as well as misinformation.
Conclusions
Social networks, such as Twitter, are useful sources of information in the early stages of outbreaks. Close to real-time identification and analyses of misinformation may help authorities take the necessary steps in a timely manner.
by
Ari Z. Klein;
Juan M. Banda;
Yuting Guo;
Ana Lucia Schmidt;
Dongfang Xu;
Jesus Ivan Flores Amaro;
Raul Rodriguez-Esteban;
Abeed Sarker;
Graciela Gonzalez-Hernandez
The aim of the Social Media Mining for Health Applications (#SMM4H) shared tasks is to take a community-driven approach to address the natural language processing and machine learning challenges inherent to utilizing social media data for health informatics. The eighth iteration of the #SMM4H shared tasks was hosted at the AMIA 2023 Annual Symposium and consisted of five tasks that represented various social media platforms (Twitter and Reddit), languages (English and Spanish), methods (binary classification, multi-class classification, extraction, and normalization), and topics (COVID-19, therapies, social anxiety disorder, and adverse drug events). In total, 29 teams registered, representing 18 countries. In this paper, we present the annotated corpora, a technical summary of the systems, and the performance results. In general, the top-performing systems used deep neural network architectures based on pre-trained transformer models. In particular, the top-performing systems for the classification tasks were based on single models that were pre-trained on social media corpora. To facilitate future work, the datasets—a total of 61,353 posts—will remain available by request, and the CodaLab sites will remain active for a post-evaluation phase.
Background
Headache frequency, defined as the number of days with any headache in a month (or four weeks), remains a key parameter in the evaluation of treatment response to migraine preventive medications. However, due to the variations and inconsistencies in documentation by clinicians, significant challenges exist to accurately extract headache frequency from the electronic health record (EHR) by traditional natural language processing (NLP) algorithms.
Methods
This was a retrospective cross-sectional study with human subjects identified from three tertiary headache referral centers- Mayo Clinic Arizona, Florida, and Rochester. All neurology consultation notes written by more than 10 headache specialists between 2012 to 2022 were extracted and 1915 notes were used for model fine-tuning (90%) and testing (10%). We employed four different NLP frameworks: (1) ClinicalBERT (Bidirectional Encoder Representations from Transformers) regression model (2) Generative Pre-Trained Transformer-2 (GPT-2) Question Answering (QA) Model zero-shot (3) GPT-2 QA model few-shot training fine-tuned on Mayo Clinic notes; and (4) GPT-2 generative model few-shot training fine-tuned on Mayo Clinic notes to generate the answer by considering the context of included text.
Results
The GPT-2 generative model was the best-performing model with an accuracy of 0.92[0.91 – 0.93] and R2 score of 0.89[0.87, 0.9], and all GPT2-based models outperformed the ClinicalBERT model in terms of the exact matching accuracy. Although the ClinicalBERT regression model had the lowest accuracy 0.27[0.26 – 0.28], it demonstrated a high R2 score 0.88[0.85, 0.89], suggesting the ClinicalBERT model can reasonably predict the headache frequency within a range of ≤ ± 3 days, and the R2 score was higher than the GPT-2 QA zero-shot model or GPT-2 QA model few-shot training fine-tuned model.
Conclusion
We developed a robust model based on a state-of-the-art large language model (LLM)- a GPT-2 generative model that can extract headache frequency from EHR free-text clinical notes with high accuracy and R2 score. It overcame several challenges related to different ways clinicians document headache frequency that were not easily achieved by traditional NLP models. We also showed that GPT2-based frameworks outperformed ClinicalBERT in terms of accuracy in extracting headache frequency from clinical notes. To facilitate research in the field, we released the GPT-2 generative model and inference code with open-source license of community use in GitHub.
Background
Substance use, including the non-medical use of prescription medications, is a global health problem resulting in hundreds of thousands of overdose deaths and other health problems. Social media has emerged as a potent source of information for studying substance use-related behaviours and their consequences. Mining large-scale social media data on the topic requires the development of natural language processing (NLP) and machine learning frameworks customized for this problem. Our objective in this research is to develop a framework for conducting a content analysis of Twitter chatter about the non-medical use of a set of prescription medications.
Methods
We collected Twitter data for four medications—fentanyl and morphine (opioids), alprazolam (benzodiazepine), and Adderall® (stimulant), and identified posts that indicated non-medical use using an automatic machine learning classifier. In our NLP framework, we applied supervised named entity recognition (NER) to identify other substances mentioned, symptoms, and adverse events. We applied unsupervised topic modelling to identify latent topics associated with the chatter for each medication.
Results
The quantitative analysis demonstrated the performance of the proposed NER approach in identifying substance-related entities from data with a high degree of accuracy compared to the baseline methods. The performance evaluation of the topic modelling was also notable. The qualitative analysis revealed knowledge about the use, non-medical use, and side effects of these medications in individuals and communities.
Conclusions
NLP-based analyses of Twitter chatter associated with prescription medications belonging to different categories provide multi-faceted insights about their use and consequences. Our developed framework can be applied to chatter about other substances. Further research can validate the predictive value of this information on the prevention, assessment, and management of these disorders.
Background:
Induction of buprenorphine, an evidence-based treatment for opioid use disorder (OUD), has been reported to be difficult for people with heavy use of fentanyl, the most prevalent opioid in many areas of the country. In this population, precipitated opioid withdrawal (POW) may occur even after individuals have completed a period of opioid abstinence prior to induction. Our objective was to study potential associations between fentanyl, buprenorphine induction, and POW, using social media data.
Methods:
This is a mixed methods study of data from seven opioid-related forums (subreddits) on Reddit. We retrieved publicly available data from the subreddits via an application programming interface, and applied natural language processing to identify subsets of posts relevant to buprenorphine induction, POW, and fentanyl and analogs (F&A). We computed mention frequencies for keywords/phrases of interest specified by our medical toxicology experts. We further conducted manual, qualitative, and thematic analyses of automatically identified posts to characterize the information presented.
Results:
In 267,136 retrieved posts, substantial increases in mentions of F&A (3 in 2013 to 3870 in 2020) and POW (2 in 2012 to 332 in 2020) were observed. F&A mentions from 2013 to 2021 were strongly correlated with mentions of POW (Spearman’s ρ: 0.882; p = .0016), and mentions of the Bernese method (BM), a microdosing induction strategy (Spearman’s ρ: 0.917; p = .0005). Manual review of 384 POW- and 106 BM-mentioning posts revealed that common discussion themes included “specific triggers of POW” (55.1%), “buprenorphine dosing strategies” (38.2%) and “experiences of OUD” (36.1%). Many reported experiencing POW despite prolonged opioid abstinence periods, and recommended induction via microdosing, including specifically via the BM.
Conclusions:
Reddit subscribers often associate POW with F&A use and describe self-managed buprenorphine induction strategies involving microdosing to avoid POW. Further objective studies in patients with fentanyl use and OUD initiating buprenorphine are needed to corroborate these findings.
Introduction
Decades of research have established the association between adverse childhood experiences (ACEs) and adult onset of chronic diseases, influenced by health behaviors and social determinants of health (SDoH). Machine Learning (ML) is a powerful tool for computing these complex associations and accurately predicting chronic health conditions.
Methods
Using the 2021 Behavioral Risk Factor Surveillance Survey, we developed several ML models—random forest, logistic regression, support vector machine, Naïve Bayes, and K-Nearest Neighbor—over data from a sample of 52,268 respondents. We predicted 13 chronic health conditions based on ACE history, health behaviors, SDoH, and demographics. We further assessed each variable’s importance in outcome prediction for model interpretability. We evaluated model performance via the Area Under the Curve (AUC) score.
Results
With the inclusion of data on ACEs, our models outperformed or demonstrated similar accuracies to existing models in the literature that used SDoH to predict health outcomes. The most accurate models predicted diabetes, pulmonary diseases, and heart attacks. The random forest model was the most effective for diabetes (AUC = 0.784) and heart attacks (AUC = 0.732), and the logistic regression model most accurately predicted pulmonary diseases (AUC = 0.753). The strongest predictors across models were age, ever monitored blood sugar or blood pressure, count of the monitoring behaviors for blood sugar or blood pressure, BMI, time of last cholesterol check, employment status, income, count of vaccines received, health insurance status, and total ACEs. A cumulative measure of ACEs was a stronger predictor than individual ACEs.
Discussion
Our models can provide an interpretable, trauma-informed framework to identify and intervene with at-risk individuals early to prevent chronic health conditions and address their inequalities in the U.S.
Background. The behaviors and emotions associated with and reasons for nonmedical prescription drug use (NMPDU) are not well-captured through traditional instruments such as surveys and insurance claims. Publicly available NMPDU-related posts on social media can potentially be leveraged to study these aspects unobtrusively and at scale.
Methods. We applied a machine learning classifier to detect self-reports of NMPDU on Twitter and extracted all public posts of the associated users. We analyzed approximately 137 million posts from 87,718 Twitter users in terms of expressed emotions, sentiments, concerns, and possible reasons for NMPDU via natural language processing.
Results. Users in the NMPDU group express more negative emotions and less positive emotions, more concerns about family, the past, and body, and less concerns related to work, leisure, home, money, religion, health, and achievement compared to a control group (i.e., users who never reported NMPDU). NMPDU posts tend to be highly polarized, indicating potential emotional triggers. Gender-specific analyses show that female users in the NMPDU group express more content related to positive emotions, anticipation, sadness, joy, concerns about family, friends, home, health, and the past, and less about anger than males. The findings are consistent across distinct prescription drug categories (opioids, benzodiazepines, stimulants, and polysubstance).
Conclusion. Our analyses of large-scale data show that substantial differences exist between the texts of the posts from users who self-report NMPDU on Twitter and those who do not, and between males and females who report NMPDU. Our findings can enrich our understanding of NMPDU and the population involved.
Introduction Prescription medication overdose is the fastest growing drug-related problem in the USA. The growing nature of this problem necessitates the implementation of improved monitoring strategies for investigating the prevalence and patterns of abuse of specific medications. Objectives Our primary aims were to assess the possibility of utilizing social media as a resource for automatic monitoring of prescription medication abuse and to devise an automatic classification technique that can identify potentially abuse-indicating user posts. Methods We collected Twitter user posts (tweets) associated with three commonly abused medications (Adderall®, oxycodone, and quetiapine). We manually annotated 6400 tweets mentioning these three medications and a control medication (metformin) that is not the subject of abuse due to its mechanism of action. We performed quantitative and qualitative analyses of the annotated data to determine whether posts on Twitter contain signals of prescription medication abuse. Finally, we designed an automatic supervised classification technique to distinguish posts containing signals of medication abuse from those that do not and assessed the utility of Twitter in investigating patterns of abuse over time. Results Our analyses show that clear signals of medication abuse can be drawn from Twitter posts and the percentage of tweets containing abuse signals are significantly higher for the three case medications (Adderall®: 23 %, quetiapine: 5.0 %, oxycodone: 12 %) than the proportion for the control medication (metformin: 0.3 %). Our automatic classification approach achieves 82 % accuracy overall (medication abuse class recall: 0.51, precision: 0.41, F measure: 0.46). To illustrate the utility of automatic classification, we show how the classification data can be used to analyze abuse patterns over time. Conclusion Our study indicates that social media can be a crucial resource for obtaining abuse-related information for medications, and that automatic approaches involving supervised classification and natural language processing hold promises for essential future monitoring and intervention tasks.
Background
Twitter is a potentially valuable tool for public health officials and state Medicaid programs in the United States, which provide public health insurance to 72 million Americans.
Objective
We aim to characterize how Medicaid agencies and managed care organization (MCO) health plans are using Twitter to communicate with the public.
Methods
Using Twitter’s public application programming interface, we collected 158,714 public posts (“tweets”) from active Twitter profiles of state Medicaid agencies and MCOs, spanning March 2014 through June 2019. Manual content analyses identified 5 broad categories of content, and these coded tweets were used to train supervised machine learning algorithms to classify all collected posts.
Results
We identified 15 state Medicaid agencies and 81 Medicaid MCOs on Twitter. The mean number of followers was 1784, the mean number of those followed was 542, and the mean number of posts was 2476. Approximately 39% of tweets came from just 10 accounts. Of all posts, 39.8% (63,168/158,714) were classified as general public health education and outreach; 23.5% (n=37,298) were about specific Medicaid policies, programs, services, or events; 18.4% (n=29,203) were organizational promotion of staff and activities; and 11.6% (n=18,411) contained general news and news links. Only 4.5% (n=7142) of posts were responses to specific questions, concerns, or complaints from the public.
Conclusions
Twitter has the potential to enhance community building, beneficiary engagement, and public health outreach, but appears to be underutilized by the Medicaid program.
Objective:
To mine Twitter and quantitatively analyze COVID-19 symptoms self-reported by users, compare symptom distributions across studies, and create a symptom lexicon for future research.
Materials and Methods:
We retrieved tweets using COVID-19-related keywords, and performed semiautomatic filtering to curate self-reports of positive-tested users. We extracted COVID-19-related symptoms mentioned by the users, mapped them to standard concept IDs in the Unified Medical Language System, and compared the distributions to those reported in early studies from clinical settings.
Results:
We identified 203 positive-tested users who reported 1002 symptoms using 668 unique expressions. The most frequently-reported symptoms were fever/pyrexia (66.1%), cough (57.9%), body ache/pain (42.7%), fatigue (42.1%), headache (37.4%), and dyspnea (36.3%) amongst users who reported at least 1 symptom. Mild symptoms, such as anosmia (28.7%) and ageusia (28.1%), were frequently reported on Twitter, but not in clinical studies.
Conclusion:
The spectrum of COVID-19 symptoms identified from Twitter may complement those identified in clinical settings.