Original Paper
Abstract
Background: Health misinformation, prevalent in social media, poses a significant threat to individuals, particularly those dealing with serious illnesses such as cancer. The current recommendations for users on how to avoid cancer misinformation are challenging because they require users to have research skills.
Objective: This study addresses this problem by identifying user-friendly characteristics of misinformation that could be easily observed by users to help them flag misinformation on social media.
Methods: Using a structured review of the literature on algorithmic misinformation detection across political, social, and computer science, we assembled linguistic characteristics associated with misinformation. We then collected datasets by mining X (previously known as Twitter) posts using keywords related to unproven cancer therapies and cancer center usernames. This search, coupled with manual labeling, allowed us to create a dataset with misinformation and 2 control datasets. We used natural language processing to model linguistic characteristics within these datasets. Two experiments with 2 control datasets used predictive modeling and Lasso regression to evaluate the effectiveness of linguistic characteristics in identifying misinformation.
Results: User-friendly linguistic characteristics were extracted from 88 papers. The short-listed characteristics did not yield optimal results in the first experiment but predicted misinformation with an accuracy of 73% in the second experiment, in which posts with misinformation were compared with posts from health care systems. The linguistic characteristics that consistently negatively predicted misinformation included tentative language, location, URLs, and hashtags, while numbers, absolute language, and certainty expressions consistently predicted misinformation positively.
Conclusions: This analysis resulted in user-friendly recommendations, such as exercising caution when encountering social media posts featuring unwavering assurances or specific numbers lacking references. Future studies should test the efficacy of the recommendations among information users.
doi:10.2196/62703
Keywords
Introduction
Approximately 16% of people reported using social media to inform their medical decisions [
]. This percentage, based on estimates from the National Cancer Center, equates to 37 million adults in the United States. A recent systematic review estimated that up to 40% of health-related social media posts contain misinformation [ ]. Misinformation could cause more harm to individuals with serious conditions such as cancer. Patients who believe in misinformation and use unproven therapies in parallel or in place of cancer treatment tend to be less adherent to evidence-based treatment [ - ]. Moreover, patients with cancer might choose to delay or reject evidence-based treatment and instead pursue unproven and potentially toxic therapies, which, for some patients, results in up to 2.5 times shorter life expectancy [ ]. Approximately 30% of cancer-related social media posts on Facebook, Reddit, Pinterest, and X (previously known as Twitter) contain misinformation, and a staggering 77% of these posts have the potential to encourage patients to pursue futile and toxic therapies, resulting in physical, psychological, and logistical burdens [ ]. Cancer misinformation persists across various cancer types and is more pervasive in more prevalent cancers. Across various social media platforms, two-thirds of the most shared posts about prostate cancer contain misinformation [ ]. Researchers identified misinformation in 59% of posts related to breast cancer prevention and treatment [ ] and 30% of posts related to gynecological cancer [ ]. When surveyed, 70% of patients with cancer reported encountering misinformation about cancer on social media, with 71% believing that some of this misinformation was accurate [ ].There is a growing need to protect health information users from misinformation, especially those who are affected by serious conditions such as cancer. Multiple recommendations have been developed to assist individuals in their search for reliable health information [
- ]. However, many of the recommendations are complex, as they require individuals to possess a certain level of scientific knowledge and skills. For instance, recommendations frequently suggest taking steps such as identifying authors and their credentials, evaluating potential conflicts of interest, understanding funding sources, and assessing the original sources of scientific information. Considering the time and expertise required, expecting individuals to perform these tasks routinely is unrealistic. Moreover, these guidelines often fall short when it comes to addressing the challenges posed by social media platforms. Those who post may not disclose their real names or sources of findings, which makes some recommended steps not possible.In this work, our goal is to identify user-friendly recommendations for addressing the high rate of misinformation on social media. We began by exploring literature on the algorithmic detection of misinformation. The algorithmic approach often involves the analysis of linguistic characteristics that differentiate between factual information and misinformation [
]. Linguistic characteristics describe a body of text in an abstract manner regardless of context and may include counts of words and word parts such as nouns, verbs, adjectives, and negations, as well as specific symbols such as URLs, hashtags, and question marks. An additional category of linguistic characteristics includes words associated with the psychological state of an author [ ], which includes words related to emotions, expressions of certainty, tentativeness, insight, persuasion, and gratitude. To date, linguistic characteristics have been used by algorithms only. However, some of these characteristics are observable and could be used by individuals when they need to evaluate the credibility of the text. While individuals are unlikely to count words in social media posts regularly, they may routinely note other linguistic characteristics, such as emotions, URLs, and a strong degree of certainty. Linguistic characteristics have been shown to be effective in distinguishing misinformation from factual information across multiple contexts. However, it is unknown (1) whether the linguistic characteristics are effective in cancer-related context and (2) which subset of user-friendly linguistic characteristics could effectively distinguish misinformation. In this work, we identify the linguistic characteristics specific to the context of cancer. These characteristics will be recommended as guidelines for health information users when browsing social media.Methods
Study Design
The main sequence of study procedures is illustrated in
and includes (1) a structured literature review, in which we assemble linguistic characteristics that were used in algorithms for distinguishing factual information and misinformation (phase 1); (2) data collection, which encompasses assembling cancer-related posts using the X application programming interface (API) and labeling them as misinformation and non-misinformation (phase 2); (3) identification of the linguistic characteristics in collected datasets using natural language processing tools (phase 3); and (4) conducting predictive modeling analysis to evaluate the effectiveness of linguistic characteristics in distinguishing social media posts with cancer misinformation (phase 4).![](https://asset.jmir.pub/assets/fe2e905e8a2baac2b0e5917dc91f0851.png)
Ethical Considerations
The study was institutional review board–approved by the University of North Carolina (IRB#21-2861). This was an analysis of publicly available data. As such, participants were not compensated and did not need to provide consent for the study, because the study did not involve any prospective data collection. To protect the confidentiality and anonymity of participants in this secondary data analysis, we reworded reported posts from X.
Structured Literature Review
To identify linguistic characteristics, we developed a literature review protocol that included the search strategy and keywords. This process was informed by a collaboration with a health sciences librarian (CBS), who suggested an initial set of keywords referenced in several relevant reviews [
- ]. She also created an expanded title, abstract, and keyword search strategies for each of the following concepts: (1) text as a unit of analysis, (2) misinformation, (3) algorithms, (4) internet, and (5) linguistic features or characteristics. After the search was peer reviewed by a second health sciences librarian (CB), 5 databases were searched: ProQuest Central (ProQuest), which includes the arXiv repository; Scopus (Elsevier); IEEE Xplore (Institute of Electrical and Electronics Engineers); ACM Digital Library (Association for Computing Machinery); and Communication & Mass Media Complete (EBSCOhost). The keywords and search strategies are reported in . Results were limited to citations published between January 2012 and December 2022. Within databases, results were limited to journal papers, conference proceedings, working papers, and book chapters.Two reviewers (IF and DB) independently coded titles and abstracts in Covidence software (Veritas Health Innovation) [
] and resolved conflict in codes during research meetings. Papers were included if they focused on detecting misinformation and contained a “Methods” section describing an approach for algorithmically detecting misinformation (eg, reviews and viewpoints were excluded). Examples of the algorithms included supervised and semisupervised machine learning (eg, Bidirectional Encoder Representations from Transformers [BERT] classification) that was built on linguistic characteristics. Papers were excluded if they did not report specific linguistic characteristics, focused on misinformation in any language other than English, or used human coding but not algorithms. The detailed inclusion-exclusion criteria and PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) diagram are reported in .![](https://asset.jmir.pub/assets/5abb374c8532cdc088449bdce5878ea2.png)
Identification of Linguistic Characteristics
Upon identifying eligible papers, 2 team members (IF and DB) reviewed the full text and extracted the linguistic characteristics. Around 11% (10/90) of papers underwent double-coding. After reviewers reached an agreement, we continued with single coding. The linguistic characteristics were extracted based on the following criteria: observability, applicability, and generalizability. The observability criterion was related to whether readers could easily observe the linguistic characteristics within the text; for example, positive emotions could be easily observed while morale or cognitive language styles may be difficult to distinguish. The applicability criterion distinguished linguistic characteristics that readers could easily apply while reading the text. For instance, common characteristics such as the number of words required substantial effort from readers to evaluate and, therefore, were deemed nonapplicable. In contrast, readers could easily use citations and hashtags in their post evaluations as the mere presence of these characteristics was determined to be helpful in identifying misinformation. The third criterion, generalizability, was chosen to ensure that linguistic characteristics were not related to a specific context but could be applied across various contexts. Thus, characteristics that were based on specific words such as “COVID-19,” or “cure” were excluded.
Data Collection: Unproven Therapy
Overview
To test how extracted linguistic characteristics could distinguish social media posts from misinformation and factual information, we collected social media posts from X. Misinformation was operationalized here as information that promoted cancer treatment that was known as ineffective or information that suggested cancer causes not supported by current scientific evidence [
]. This definition focused our investigation on misinformation that could be harmful to patients with cancer or cancer survivors. Based on this operationalization, we searched existing resources that summarized unproven cancer therapy, such as “List of unproven cancer therapy” [ ], a list of “Illegally sold cancer drugs” [ ], and previous literature [ , ]. We extracted keywords and constructed 176 queries associated with unproven cancer treatments ( ). Using these queries, we randomly selected up to 500 posts per query from social media. We used R software (R Foundation for Statistical Computing) to access the Academic X API. The data were manually evaluated to determine their relevance to the cancer context and unproven therapies. Queries were edited to ensure relevance. Upon corrections, the data collection was implemented on a schedule every other week between July 2022 and August 2023. After data collection was completed, the duplicate posts were removed.Data Labeling
To distinguish posts with misinformation from other discussions, 2 reviewers (IF and CR) double-coded a randomly chosen subset of 1064 posts, achieving an acceptable interrater agreement of 0.68 measured with Krippendorff a [
]. Since the agreement was rather on a lower bound, we followed the current recommendations [ ] and resolved disagreements between coders during research meetings, reaching consensus case by case. The initial criterion for coding misinformation was developed deductively based on the definition of misinformation used in this study. A post was coded as containing misinformation if it promoted an unproven therapy as a cancer-directed treatment. For example, a post claiming that an alkaline diet can eliminate cancer would be classified as misinformation: “Cure for cancer is an alkaline diet and lots of alkaline water.” As reviewers worked with the data, they developed additional criteria based on observed patterns. Specifically, posts were labeled as containing misinformation if they discussed unproven approaches to prevent cancer, for example, “Pygeum Bark is nature’s defense against prostate cancer.” Furthermore, if a post contained a combination of factual and false information it was labeled as “misinformation.”Posts that were labeled as non-misinformation fell into 4 distinct categories. First, posts mentioned complementary and alternative medicine but did not promote it as a cancer treatment, for example, “Acupuncture and acupressure seem to be helpful in reducing pain and anxiety in patients having surgery.” Second, posts that used sarcasm and actively debunked misinformation related to cancer were in the non-misinformation category, for example, “If what you stated is true, then Gerson treatment for cancer is false.” The third category included posts that discussed complementary and alternative therapies but not in the context of promotion of cancer treatment, for instance, “Grapes can help protect you from the sun! Who knew?” Finally, posts that presented information with ambiguity, lack of clarity, or insufficient context were categorized as non-misinformation, for instance, “As a pancreatic cancer patient providing myself with all the additional holistic care practices made all the difference.” The author did not specify whether his symptoms were alleviated or cancer progression was slowed down because of holistic practices. Therefore, the post was coded as non-misinformation.
Once a subset of the database was labeled by 2 reviewers (IF and RC), we applied an algorithm to populate labels to the entire database. We worked with BERT [
], a machine learning model for natural language processing. The BERT model was chosen because it (1) worked well with short, informal text [ ]; (2) was shown to be applicable to medical text extracted from X [ ]; and (3) was successfully used in previous research to identify misinformation on X [ ]. The BERT model was implemented with the programming language Python (Python Software Foundation). The manually prelabeled subset served as training data for the BERT model. Such semisupervised approaches are commonly used in similar classification tasks [ ]. After training, BERT used its understanding of the language and context learned from the large corpus it was originally trained on and the specific examples from the manually prelabeled dataset. BERT predicted labels for each post in the rest of the data (unlabeled dataset), determining whether each was likely to contain misinformation or not based on the patterns and features it learned from the manually coded dataset.After BERT algorithm assigned labels to the posts, a researcher (IF), blinded to the model’s results, manually coded a random subset of the posts (n=960) using the same “misinformation” and “non-misinformation” labels, adhering to the same criteria that were used to prelabel the data. When compared with manual coding, the algorithm identified misinformation with an accuracy of 83%, with a higher 86% specificity, and a slightly lower sensitivity of 82%. Upon labeling, 2 datasets were created and used in the first experiment: the misinformation dataset included only posts with misinformation, and control BERT dataset 1 included only posts with non-misinformation (
).Data Collection: Posts From Cancer Centers
Following the definition of misinformation as “information not supported by scientific evidence or expert consensus” [
] and the definition used for this research, we assumed that posts originating from cancer centers reflect scientific evidence and expert consensus. To collect posts with factual information, we retrieved X data posted by cancer centers. Cancer centers often shared internal announcements and organizational news on X. To make posts comparable between the dataset with misinformation and control datasets, we used the keywords “cancer,” “treatment,” “chemotherapy,” “healing,” and other words related to treating cancer or controlling cancer progress. With the help of R software, we sampled 300 posts per cancer center between June 2011 and November 2022. A researcher (IF) manually checked randomly chosen (n=100) posts. As expected, no misinformation was found in the posts originating from cancer centers. The dataset, therefore, was assumed to consist of non-misinformation posts from cancer centers and was designated as control dataset 2, which was used in the second experiment alongside the misinformation dataset.Linguistic Characteristics Modeling
Upon data collection and labeling, we used algorithmic approaches to model linguistic characteristics. First, we used an automated text search using regular expressions in Python [
] to capture digital numbers, hashtags, and URLs in the text.Second, we used the Linguistic Inquiry and Word Count (LIWC) software [
]. LIWC calculates the proportion of the words in the posts associated with distinct psychological dimensions [ ]. In this study, LIWC identified when authors of posts used certain, absolute, or tentative language.Third, we leveraged the natural language processing tool, Name Entity Recognition [
], which was trained on human-labeled datasets to extract names from unstructured text. Using Name Entity Recognition, we were able to identify which posts contained personal names, organizational names, or locations identified from text.Fourth, we experimented with several models for sentiment analysis and identified the DistilBERT algorithm as an optimal approach for its accuracy in handling health-related X data [
]. DistilBERT is a black-box algorithm that is trained on a large corpus of data and is based on multiple deep stack layers. The DistilBERT algorithm identified positive, negative, and neutral tones present in the posts. To evaluate the algorithm’s performance, we manually labeled 300 posts across the databases. On average, the DistilBERT algorithm achieved an 83% accuracy (82% for misinformation and 84% for the control database) in detecting the emotional tone within the posts.Linguistic Characteristics Testing: Prediction of Misinformation Labels
Identified linguistic characteristics were used in an algorithm to test whether these could distinguish misinformation in posts. As shown in
, we conducted 2 experiments using tidymodels package in R software [ ]. Using linguistic characteristics as predictors, we forecast the “misinformation” and “non-misinformation” labels in the datasets semimanually coded by researchers and BERT classification algorithm. Data were split 60:40 to enable evaluation of the predictive power of the model and reported the accuracy as a ratio of correctly classified posts to the total number of posts. We also reported area under the curve (AUC), which accounted for both false-positive and false-negative predictions. AUC value ranged from 0 to 1, where 0.5 indicated that the model performs no better than a random chance, and 1 was a perfect prediction. The model was built on the basis of Lasso (“Least Absolute Shrinkage and Selection Operator”) regression, which allowed variable selection by shrinking the coefficients of less important predictors to zero [ ]. Bootstrapping procedure was applied to optimize and stabilize the selection of variables [ ]. Lasso was chosen to address multicollinearity and overfitting issues in the regression model. More importantly, Lasso regression helped identify a set of linguistic characteristics that effectively distinguished posts containing misinformation. To evaluate the significance of specific linguistic characteristics, we computed importance scores, with higher scores indicating greater relevance in distinguishing posts containing misinformation. Importance scores, a common measure in predictive modeling, indicates to what extent individual predictors contribute to the overall model performance. The assessment involves permutating the characteristic values through shuffling and measuring the subsequent decline in model performance, effectively revealing the critical factors influencing predictions. Finally, we conducted a permutation statistical test (with 1000 permutations) to determine whether models with linguistic characteristics significantly outperformed random chance.Results
Structured Literature Review
A total of 5677 citations were initially identified across all databases. After removing 1598 duplicates, we screened 4070 unique citations in Covidence. Subsequently, 3605 were excluded during the title and abstract review phase, leaving 464 papers for full-text review. Ultimately, we extracted linguistic characteristics from 88 full-text papers. These papers featured algorithmic approaches for identifying misinformation through automated text analysis, spanning various contexts, including politics, social issues, and computer science. Exclusion reasons are detailed in
, and additional information about the included papers can be found in .Identified Linguistic Characteristics
The extracted linguistic characteristics and corresponding literature are detailed in
. Representative examples that contain each linguistic characteristic were chosen by selecting posts from the misinformation dataset. We used results from linguistic characteristic modeling to identify such posts. The first category of characteristics pertains to the sentiment and emotional expression in the text and includes positive, negative emotions, and neutral sentiments (absence of either). Some papers delved into more nuanced emotions such as anger, fear, surprise, and others. We excluded these emotions due to the potential difficulty for readers to detect nuanced emotions reliably in the text.The next category comprises linguistic characteristics that pertain to psychological concepts. It is worth noting that some psychological concepts consist of a combination of linguistic characteristics, such as social processes including references to family, friends, other people, and verbs indicating interactions. Although algorithms frequently use such combinations, we decided to exclude the following psychological concepts that consisted of combinations of linguistic characteristics such as cognitive, perceptual, social processes, and morality or deception. The rationale behind this exclusion is that users are unlikely able to observe and combine linguistic characteristics for evaluations of the posts. We also excluded characteristics mentioned in fewer than 4 studies, such as gratitude, insight, causation, and persuasion. Following our 3 criteria, we included negations, tentativeness, profanity (as a proxy of informality), and words associated with absolutes and certainty.
Other categories that met our inclusion criteria were linguistic characteristics such as names of individuals, locations, and organizations, as well as categories related to the presence of URLs, hashtags, personal pronouns, and numbers. Readers can identify these characteristics without additional efforts (observability criterion) and use them for evaluation of the text (applicability) because the presence of these characteristics in social media has historically been a distinguishable factor in detecting misinformation. Furthermore, these characteristics were not context-dependent and, therefore, satisfy the generalizability criterion.
Characteristics | Examples of linguistic characteristics and posts with misinformationa | Studies using characteristics for misinformation detection | |||
Sentimentb | [ | - ]||||
Negative emotions |
| ||||
Positive emotions |
| ||||
Neutral emotions |
| ||||
Psycholinguistic | |||||
Negation |
| [ | , , , , , , - ]|||
Tentativeness |
| [ | , , , , , , , , - ]|||
Absolute language or certainty |
| [ | , , , , , - ]|||
Profanity |
| [ | , , , , , , , , , , ]|||
Named entities | [ | , , , , , , , , - ]||||
Names |
| ||||
Location |
| ||||
Organization |
| ||||
URL |
| [45,51,52,54,55,62,69,78,79 ,86-88,92,93,98,99,101,104,107-117] | |||
Numeric data |
| [ | , , , , , , , , , , , , , , ]|||
Pronouns |
| [ | , , , , , , , , , , , , , - ]|||
Hashtag |
| [43,44,47,52-55,59,64,66,77-79,82,87, 92,96,98,101,104,107,108,111,115,119,122,123] |
aAll posts were paraphrased to protect the author’s anonymity.
bIn sentiment analysis, emotions are identified by a “black box” model (DistilBERT). While we report here examples and highlight “negative/positive” words in the sentence, we must acknowledge that the algorithm may or may not use these words for detecting emotions.
Collected Data From X
We collected a total of 45,791 posts related to unproven cancer therapies. Among these, 13,046 posts were labeled as misinformation (forming the misinformation dataset), while 32,745 posts were categorized as non-misinformation (comprising control dataset 1). Furthermore, we gathered 6782 posts from the profiles of comprehensive cancer centers, which were used as control dataset 2, as shown in
. The content description of both the misinformation dataset and the control dataset 1 is shown in . To illustrate the dataset in this study, we categorized the X posts into 9 distinct categories. The examples of the posts with misinformation are shown in .Categories of therapies | Total posts, n | Posts with misinformation, n (%)a | Examples of unproven cancer therapy |
Diet based | 5179 | 3069 (59) | Antioxidant, fasting, and alkaline diet |
Alternative health system | 7036 | 2250 (32) | Herbal therapy and ayurveda |
Plant- and fungus-based | 13,851 | 4386 (32) | Mushrooms |
Synthetic substances | 8471 | 2637 (31) | Antineoplastic Brudzinski and vitamin C |
Spiritual and mental healing | 2347 | 272 (12) | Meditation, praying, and tai chi |
Electromagnetic and energy-based | 2825 | 283 (10) | Polarity therapy and magnetic |
Physical procedures | 1144 | 49 (4) | Acupuncture |
Other | 4938 | 100 (2) | N/Ab |
Total | 45,791 | 13,046 (28) | N/A |
aOut of the total number of posts.
bN/A: not applicable.
Linguistic Characteristics Testing: Prediction of Misinformation Labels
As shown in
, experiment 1 demonstrated that linguistic characteristics predicted misinformation with 60% accuracy. In experiment 2, they exhibited even stronger predictive power, achieving an accuracy of 77%. The importance scores for each linguistic characteristic are shown in .Next, we selected linguistic characteristics with an impact score 0.05 and consistent predictive performance across experiments 1 and 2. These short-listed characteristics underwent further testing within the same experiments. In experiment 1, the short-listed characteristics achieved an accuracy rate of 50%, which did not significantly differ from random chance (P.90). However, in experiment 2, these characteristics predicted misinformation with an accuracy rate of 73% and an AUC of 83. This performance was significantly better than random chance (McNemar 21=5.7 ×107; P.001). The importance scores for the short-listed characteristics are shown in
. For a more detailed breakdown of the importance scores, we have summarized the percentage of posts containing these short-listed characteristics by dataset in and the complete list in .Name of the dataset | Total posts, n | Posts with misinformation, n | Accuracy, % |
Experiment 1: misinformation dataset and control dataset 1 | 45,791 | 13,046 | 60 |
Experiment 2: misinformation dataset and control dataset 2 | 19,828 | 13,046 | 77 |
Linguistic characteristics | Experiment with control group 1 | Experiment with control group 2 | Experiment with short-listed characteristics (control group 2) | |||
Predictors | Predictors | Predictors | ||||
Negative | Positive | Negative | Positive | Negative | Positive | |
Absolute language | —a | 0.11b | — | 0.69 | — | 0.84 |
Certainty | — | 0.21 | — | 1.13 | — | 1.02 |
First-person pronoun | 0.27 | — | — | 1.31 | — | — |
Hashtags | 0.56 | — | 1.55 | — | 1.6 | — |
Location | 0.27 | — | 0.27 | — | 0.46 | — |
Name | — | 0.08 | 0.91 | — | — | — |
Negation | 0.53 | — | — | 0.73 | — | — |
Negative emotions | 0.24 | — | 0 | — | — | — |
Neutral emotions | 0 | — | — | 0.07 | — | — |
Number | — | 0.17 | — | 0.29 | — | 0.28 |
Organization | — | 0.02 | 0.63 | — | — | — |
Positive emotions | — | 0.31 | 0.46 | — | — | — |
Profanity | 0.92 | — | — | 1.99 | — | — |
Second-person pronoun | 0.02 | — | 0.45 | — | — | — |
Tentativeness | 0.08 | — | 0.16 | — | 0.08 | — |
Third-person pronoun | 0 | — | 0.23 | — | — | — |
URL | 0.3 | — | 2.28 | — | 2.47 | — |
aNot applicable.
bItalicized values represent short-listed characteristics.
Linguistic characteristics | Misinformation dataset (n=13,046), n (%) | Control dataset 1 (n=32,745), n (%) | Control dataset 2 (n=6782), n (%) | ||||
Positive predictors | |||||||
Certainty | 1579 (12) | 3044 (9) | 208 (3) | ||||
Absolute | 2741 (21) | 7294 (22)a | 630 (9) | ||||
Number | 6358 (49) | 14,360 (44) | 2497 (37) | ||||
Negative predictors | |||||||
URL | 6978 (53) | 19,591 (60) | 6560 (97) | ||||
Hashtags | 2296 (18) | 8512 (26) | 4343 (64) | ||||
Location | 1212 (9) | 3373 (12) | 975 (14) | ||||
Tentativeness | 4154 (32) | 11,171 (34) | 1835 (27)a |
aValence of predictions is inferred from the model, which includes all characteristics simultaneously.
Discussion
Principal Findings
We have identified linguistic characteristics that can help people affected by cancer detect cancer misinformation on social media platforms such as X. Linguistic characteristics that were likely to be present in posts with misinformation were related to certain, absolute language, and numbers. Certain language included phrases that reflected a “degree of bravado” or “boasting of certainty.” Examples of certain languages could be “I really believe,” “it is definitely helpful,” and similar others [
]. The absolute language referred to phrases that reflect black-and-white thinking and included words such as “none,” “all,” “never,” and others [ ]. The number category encompassed any information reported with digits such as percentages, count of any units, years, and priorities. Notably, all 3 linguistic characteristics could be united under the umbrella of definite, confident language. Linguistic characteristics that were unlikely to be present in posts with misinformation encompassed URLs, hashtags, and location mentions. Each of these attributes could be considered as a form of citation or reference. URLs offered direct links to the original source or further information, hashtags connected posts to broader relevant discussions, while locations mentioned in posts provided context and a sense of origin to the information shared. Our findings are consistent with some of the suggestions provided by previous guidelines for identifying misinformation. For instance, the Food and Drug Administration recommends being vigilant if patients read confident statements such as a drug definitely “cures cancer” or “guarantees results” [ ]. Other guidelines encouraged users to search for references and original sources of health-related information [ - ].While consistent with previous recommendations, our findings make a unique contribution. Previous work has based the guidelines on theoretical assumptions, while our study is one of the first to provide some empirical evidence based on a large dataset to support the recommendations for users. Another contribution is that we outlined ineffective linguistic characteristics for detecting cancer misinformation. Despite a substantial body of research showing that social media posts with sentiments predicted fake news, we did not find these relationships. A potential explanation could be the algorithm’s limited efficiency in identifying emotions within cancer-related contexts. Furthermore, it is possible that authors express a limited range of emotions in cancer-related conversations, typically negative emotions toward cancer and both positive and negative emotions toward various treatments, including those that are unproven. These emotions may vary little across posts containing valid and nonvalid information, making emotions an unreliable factor for distinguishing misinformation.
Our work accumulates knowledge about misinformation detection from the literature covering a wide range of contexts—including political, social, and computer science—and translates this knowledge to the cancer context. The findings highlighted promising avenues for future research and could expedite the development of automated and augmented methods for identifying and verifying cancer-related misinformation on social media platforms. Finally, the robust labeled datasets developed by our research team are available to other researchers upon request to the corresponding author, thereby further supporting research on misinformation within the context of cancer and social media.
In practice, our work is at the forefront of customizing recommendations and contextualizing them for social network users. Our exploratory findings suggest a promising direction for studying linguistic characteristics that information users might apply when making quick judgments while scrolling through X feeds. Empowering users to stay vigilant in their initial evaluations could help reduce the spread of misinformation and the formation of erroneous beliefs. This is a crucial area for future research, which should explore how these findings apply in different cancer-related contexts and across various social networks.
Limitations
All the studies included in our analysis exclusively originate from peer-reviewed journals and conference proceedings; however, we must exercise caution when considering the potential for publication bias. Furthermore, in accordance with our selection criteria for linguistic characteristics, we included only those papers that focused on text and excluded other forms of social media content, such as videos and images. We recommend that future research comprehensively explore social media, including multimedia content, as it could potentially provide additional insights for user-friendly recommendations.
In selecting linguistic characteristics, we prioritized observability, applicability, and generalizability. However, alternative criteria may be considered when users are open to a more thorough exploration of a post’s validity. For example, future research should explore the use of metadata, link content analysis, and hashtag meanings. As misinformation evolves and its authors adjust to societal changes, the linguistic characteristics that identify misinformation may also shift. A longitudinal analysis is necessary to understand how linguistic characteristics perform in predicting misinformation over time.
Algorithms used in our analysis operate with a certain level of accuracy. Specifically, the accuracy of label identification in the dataset reached 83%, indicating that approximately 17% of posts were labeled incorrectly. This means that in experiment 1 some proportion of misinformation is included in the non-misinformation group and vice versa, making further exploration less accurate in experiment 1. This degree of uncertainty is common in algorithmic performance. Therefore, it is important to interpret our results in light of the inherent imperfections in algorithmic performance.
Furthermore, we encountered that the short-listed linguistic characteristics did not significantly outperform random chance in identifying misinformation in experiment 1. This outcome underscores a potential boundary condition of the effectiveness of the linguistic characteristics. Notably, experiment 1 encompassed more homogeneous data in contrast to experiment 2. Based on these findings, it becomes plausible to speculate that linguistic characteristics might provide limited help when a reader assesses posts within a closely knit community.
In experiment 2, the control dataset 2 consisted of posts shared by cancer centers and was compared with the misinformation dataset comprising random posts. To address this limitation, we collected posts from cancer centers that contain words related to cancer therapies. This step was taken to ensure a similar context of discussion as the posts with unproven therapy. Next, we exclude linguistic characteristics that are likely displayed differences between datasets due to the distinct nature of the information within control dataset 2. For example, linguistic traits such as “the use of profanity” or “first-person pronouns” were discarded. Furthermore, we decided to focus our analysis solely on the text within the posts and omitted other accompanying metainformation that users might observe, such as the user’s name, location of the author, and posting time. This approach allowed us to assume that posts shared by cancer centers might be perceived more broadly, for instance, as posts shared by researchers, physicians, administrators, and patient advocates. Because of these measures, we anticipate that the linguistic characteristics identified in this research may help differentiate between health misinformation and factual posts on social media, irrespective of their sources. Despite our precautionary measures, we cannot fully guarantee that identified linguistics characteristics certainly distinguish between posts with misinformation and non-misinformation versus posts produced by the general public and posts by health experts from health care systems. However, there are factors that support the first conclusion more than the second. First, our findings are consistent with the previous theoretical and practical recommendations for identifying misinformation [
- ]. Second, the associated with misinformation linguistic characteristics, such as numbers and assertive language, are expected to be used by health experts. For instance, providers use numbers more confidently than the general public [ ]. Professional guidelines for health providers encourage them to use numbers over verbal descriptions [ ] as well as the use of assertive language in communication with patients [ , ]. Yet, our study associated these characteristics with misinformation shared by the general public on social media, which suggests that we might be finding more than just a mere distinction between the general public language and the health professional language. One study in and of itself is not yet a comprehensive body of evidence. Our findings will need to be validated and built upon via additional studies—including those that use posts from other types of entities and comparison groups.Finally, our data were collected only on a single social network X. Many characteristics and customs of X are transferable to other social networks and our recommendations are likely to go beyond application on X, as demonstrated by the consistency of our recommendations with the recommendations of other researchers [
- ]. Given this limitation, our results need to be generalized cautiously, and further similar research is needed for different platforms (eg, Facebook, Pinterest, etc).Conclusions
Our structured review synthesized knowledge from studies that used algorithmic approaches for text analysis to detect misinformation in social media. From this literature, we identified user-friendly linguistic characteristics that can assist individuals in distinguishing misinformation when they seek health-related information on social media. The linguistic characteristics, such as certainty, absolute language, and numbers, were positively associated with misinformation, while characteristics such as URLs, hashtags, and location mentions were negatively predictive of misinformation. Based on these findings, we suggested that users should be cautious of social media posts containing confident promises or specific numbers without proper references to the original information. According to our analysis, we expect that this approach will allow users to filter out two-thirds of posts with cancer-related misinformation. Yet, before drawing a definitive conclusion, further testing with different datasets is required.
Acknowledgments
ChatGPT 3.5 (OpenAI) [
] was used to assist with the professional editing of the manuscript. This study was supported by North Carolina Translational Research and Clinical Science Institute, Pilot Award Spring 2022.Conflicts of Interest
None declared.
Strategy for literature review.
DOCX File , 29 KBList of unproven therapy.
DOCX File , 25 KBSummary of the literature.
DOCX File , 47 KBSummary of linguistic characteristics.
XLSX File (Microsoft Excel File), 10 KBReferences
- How much do you agree or disagree with the following statements? I use information from social media to make decisions about my health. HINTS. 2022. URL: https://hints.cancer.gov/view-questions/question-detail.aspx?PK_Cycle=14qid=1849 [accessed 2023-09-20]
- Suarez-Lledo V, Alvarez-Galvez J. Prevalence of health misinformation on social media: systematic review. J Med Internet Res. 2021;23(1):e17187. [FREE Full text] [CrossRef] [Medline]
- Johnson SB, Park HS, Gross CP, Yu JB. Use of alternative medicine for cancer and its impact on survival. J Natl Cancer Inst. 2017;110:121-124.
- Johnson SB, Park HS, Gross CP, Yu JB. Complementary medicine, refusal of conventional cancer therapy, and survival among patients with curable Cancers. JAMA Oncol. 2018;4(10):1375-1381. [FREE Full text] [CrossRef] [Medline]
- Weil CR, Hutten RJ, Barney BM, Fagerlin A, Gaffney DK, Gill DM, et al. Shifting perceptions of alternative therapies in cancer patients during the COVID-19 pandemic: results from the complementary and alternative medicine exposure in oncology (CAMEO) study. J. Clin. Oncol. 2022;40(16_suppl):e24130-e24130. [CrossRef]
- Johnson SB, Bylund CL. Identifying cancer treatment misinformation and strategies to mitigate its effects with improved radiation oncologist-patient communication. Pract Radiat Oncol. 2023;13(4):282-285. [CrossRef] [Medline]
- Johnson SB, Parsons M, Dorff T, Moran MS, Ward JH, Cohen SA, et al. Cancer misinformation and harmful information on facebook and other social media: a brief report. J Natl Cancer Inst. 2022;114(7):1036-1039. [FREE Full text] [CrossRef] [Medline]
- Alsyouf M, Stokes P, Hur D, Amasyali A, Ruckle H, Hu B. 'Fake News' in urology: evaluating the accuracy of articles shared on social media in genitourinary malignancies. BJU Int. 2019;124(4):701-706. [CrossRef] [Medline]
- Yussof I, Ab Muin NF, Mohd M, Hatah E, Mohd Tahir NA, Mohamed Shah N. Breast cancer prevention and treatment misinformation on Twitter: an analysis of two languages. Digit Health. 2023;9:20552076231205742. [FREE Full text] [CrossRef] [Medline]
- Chen L, Wang X, Peng T. Nature and diffusion of gynecologic cancer–related misinformation on social media: analysis of tweets. J Med Internet Res. 2018;20(10):e11515. [CrossRef] [Medline]
- Lazard AJ, Nicolla S, Vereen RN, Pendleton S, Charlot M, Tan H, et al. Exposure and reactions to cancer treatment misinformation and advice: survey study. JMIR Cancer. 2023;9:e43749. [FREE Full text] [CrossRef] [Medline]
- Health misinformation. The US Surgeon General's Advisory. 2021. URL: https://www.hhs.gov/surgeongeneral/priorities/health-misinformation/index.html [accessed 2023-09-05]
- Finding cancer information on the internet. American Cancer Society. 2020. URL: https://www.cancer.org/cancer/understand ing-cancer/cancer-information-on-the-internet.html [accessed 2023-07-20]
- How to find cancer resources you can trust. National Cancer Institute Internet. 2015. URL: https://www.cancer.gov/about -cancer/managing-care/using-trusted-resources [accessed 2022-06-28]
- Bondielli A, Marcelloni F. A survey on fake news and rumour detection techniques. Inf Sci. 2019;497:38-55. [CrossRef]
- Tausczik YR, Pennebaker JW. The psychological meaning of words: LIWC and computerized text analysis methods. J. Lang. Soc. Psychol. 2009;29(1):24-54. [CrossRef]
- Kapantai E, Christopoulou A, Berberidis C, Peristeras V. A systematic literature review on disinformation: toward a unified taxonomical framework. New Media Soc. 2020;23(5):1301-1326. [CrossRef]
- Zannettou S, Sirivianos M, Blackburn J, Kourtellis N. The web of false information. ACM J Data Inf Qual. 2019;11(3):1-37. [CrossRef]
- Sendra A, Farré J, Vaagan RW. Seeking, sharing and co-creating: a systematic review of the relation between social support theory, social media use and chronic diseases. Soc Theory Health. 2019;18(4):317-339. [CrossRef]
- Wang Y, McKee M, Torbica A, Stuckler D. Systematic literature review on the spread of health-related misinformation on social media. Soc Sci Med. 2019;240:112552. [FREE Full text] [CrossRef] [Medline]
- Molina MD, Sundar SS, Le T, Lee D. “Fake News” is not simply false information: a concept explication and taxonomy of online content. Am Behav Sci. 2019;65(2):180-212. [CrossRef]
- Babineau J. Product review: Covidence (systematic review software). J Can Health Libr Assoc. 2014;35:68-71.
- Ghenai A, Mejova Y. Fake cures. Proc ACM Hum Comput Interact. 2018;2:1-20. [CrossRef]
- The Free Encyclopedia. List of unproven and disproven cancer treatments. Wikipedia [Internet]. Wikimedia Foundation, Inc. 2013. URL: https://en.wikipedia.org/wiki/List_of_unproven_and_disproven_cancer_treatments [accessed 2023-07-21]
- Illegally sold cancer treatments. US Food and Drug Administration. 2017. URL: https://www.fda.gov/consumers/health-fraud -scams/illegally-sold-cancer-treatments [accessed 2023-09-01]
- White RW, Hassan A. Content bias in online health search. ACM Trans Web. 2014;8(4):1-33. [CrossRef]
- Marzi G, Balzano M, Marchiori D. K-Alpha Calculator-Krippendorff's Alpha Calculator: a user-friendly tool for computing Krippendorff's alpha inter-rater reliability coefficient. MethodsX. 2024;12:102545. [FREE Full text] [CrossRef] [Medline]
- Oortwijn Y, Ossenkoppele T, Betti A. Interrater disagreement resolution: a systematic procedure to reach consensus in annotation tasks. 2021. Presented at: Proceedings of the Workshop on Human Evaluation of NLP Systems (HumEval); 2021 April 01:131-141; Dublin, Ireland.
- Minaee S, Kalchbrenner N, Cambria E, Nikzad N, Chenaghlu M, Gao J. Deep learning--based text classification. ACM Comput Surv. 2021;54(3):1-40. [CrossRef]
- Qasim R, Bangyal WH, Alqarni MA, Ali Almazroi A. A fine-tuned BERT-based transfer learning approach for text classification. J Healthc Eng. 2022;2022:3498123. [FREE Full text] [CrossRef] [Medline]
- Roitero K, Bozzato C, Della MV, Mizzaro S, Serra G. Twitter goes to the doctor: detecting medical tweets using machine learning and BERT. 2020. URL: http://hdl.handle.net/11390/1189151 [accessed 2023-07-12]
- Devlin J, Chang MW, Lee K, Toutanova K. BERT: pre-training of deep bidirectional transformers for language understanding. arXiv. Preprint posted online on October 11, 2018. [CrossRef]
- Elroy O, Yosipof A. Semi-supervised learning classifier for misinformation related to earthquakes prediction on social media. 2023. Presented at: Artificial Neural Networks and Machine Learning – ICANN 2023: 32nd International Conference on Artificial Neural Networks; September 26-29, 2023; Heraklion, Greece. [CrossRef]
- Swire-Thompson B, Lazer D. Public health and online misinformation: challenges and recommendations. Annu Rev Public Health. 2020;41:433-451. [FREE Full text] [CrossRef] [Medline]
- Chapman C, Stolee K. Exploring regular expression usage and context in Python. 2016. Presented at: Proceedings of the 25th International Symposium on Software Testing and Analysis; July 18, 2016; New York, NY. [CrossRef]
- Boyd RL, Ashokkumar A, Seraj S, Pennebaker JW. The development and psychometric properties of LIWC-22. 2022. URL: https://www.liwc.app [accessed 2023-07-19]
- Tausczik YR, Pennebaker JW. The psychological meaning of words: LIWC and computerized text analysis methods. J Lang Soc Psychol. 2009;29(1):24-54. [CrossRef]
- Goyal A, Gupta V, Kumar M. Recent named entity recognition and classification techniques: a systematic review. Comput. Sci. Rev. 2018;29:21-43. [CrossRef]
- Ranganathan J, Tsahai T. Sentiment analysis of tweets using deep learning. In: Advanced Data Mining and Applications. Cham, Switzerland. Springer International Publishing AG; 2022:106-117.
- Kuhn M, Wickham H. Tidymodels: a collection of packages for modeling and machine learning using tidyverse principles. 2020. URL: https://www.tidymodels.org [accessed 2023-07-24]
- Bach F. Consistency of the group Lasso and multiple kernel learning. J Mach Learn Res. 2008;9:1179-1225.
- Bach FR. Bolasso: model consistent Lasso estimation through the bootstrap. 2008. Presented at: Proceedings of the 25th International Conference on Machine Learning; July 5, 2008:33-40; Helsinki, Finland. [CrossRef]
- Agarwal S, Samavedhi A. Profiling fake news: learning the semantics and characterisation of misinformation. In: Advanced Data Mining and Applications. ADMA 2022. Lecture Notes in Computer Science, vol 13087. Cham, Switzerland. Springer International Publishing; 2022:203-216.
- Agarwal R, Gupta S, Chatterjee N. Profiling fake news spreaders on Twitter: a clickbait and linguistic feature based scheme. In: Natural Language Processing and Information Systems. New York City, NY. Springer International Publishing; 2022:345-357.
- Alkhodair SA, Fung BCM, Ding SHH, Cheung WK, Huang S. Detecting high-engaging breaking news rumors in social media. ACM Trans Manage Inf Syst. 2020;12(1):1-16. [CrossRef]
- Garg S, Kumar Sharma D. Linguistic features based framework for automatic fake news detection. Comput Ind Eng. 2022;172(Part A):108432. [CrossRef]
- Agrawal C, Pandey A, Goyal S. Fake news detection system based on modified bi-directional long short term memory. Multimed Tools Appl. 2022;81(17):24199-24223. [CrossRef]
- Ahmad I, Yousaf M, Yousaf S, Ahmad MO. Fake news detection using machine learning ensemble methods. Complexity. 2020;2020:1-11. [FREE Full text] [CrossRef]
- Alsmadi I, O'Brien MJ. Rating news claims: feature selection and evaluation. Math Biosci Eng. 2019;17(3):1922-1939. [FREE Full text] [CrossRef] [Medline]
- Aneja S, Aneja N, Kumaraguru P. Predictive linguistic cues for fake news: a societal artificial intelligence problem. arXiv. Preprint posted online on November 26, 2022. [FREE Full text]
- Ebrahimi Fard A, Mohammadi M, Chen Y, van de Walle B. Computational rumor detection without non-rumor: a one-class classification approach. IEEE Trans Comput Soc Syst. 2019;6(5):830-846. [CrossRef]
- Sharma AK, Gangeya RA, Kumar H, Chaurasia S, Srivastava DK. Ensemble rumor text classification model applied to different tweet features. Int J Fuzzy Log Intell Syst. 2022;22(3):325-338. [CrossRef]
- Sharma D, Shrivastava P, Garg S. Utilizing word embedding and linguistic features for fake news detection. 2022. Presented at: 9th International Conference on Computing for Sustainable Global Development (INDIACom); March 23-25, 2022; New Delhi, India. [CrossRef]
- Shelke S, Attar V. Rumor detection in social network based on user, content and lexical features. Multimed Tools Appl. 2022;81(12):17347-17368. [FREE Full text] [CrossRef] [Medline]
- Shelke S, Attar V. Role of various features in identification of rumors in the social network. 2021. Presented at: 12th International Conference on Computing Communication and Networking Technologies (ICCCNT); July 06-08, 2021; Kharagpur, India. [CrossRef]
- Sicilia R, Giudice S, Pei Y, Pechenizkiy M, Soda P. Health-related rumour detection on Twitter. 2017. Presented at: IEEE International Conference on Bioinformatics and Biomedicine (BIBM); November 13-16, 2017; Kansas City, MO. [CrossRef]
- Sun QJ. A machine learning analysis of the features in deceptive and credible news. arXiv. Preprint posted online on October 5, 2019. [CrossRef]
- Tuarob S, Tucker CS, Salathe M, Ram N. An ensemble heterogeneous classification methodology for discovering health-related knowledge in social media messages. J Biomed Inform. 2014;49:255-268. [FREE Full text] [CrossRef] [Medline]
- van de Guchte L, Raaijmakers S, Meeuwissen E, Spenader J. Near real-time detection of misinformation on online social networks. In: Disinformation in Open Online Media. Cham, Switzerland. Springer International Publishing; 2020:240-260.
- Vijeev A, Mahapatra A, Shyamkrishna A, Murthy S. A hybrid approach to rumour detection in microblogging platforms. 2018. Presented at: International Conference on Advances in Computing, Communications and Informatics (ICACCI); September 19-22, 2018; Bangalore, India. [CrossRef]
- Volkova S, Jang JY. Misleading or falsification: inferring deceptive strategies and types in online news and social media. 2018. Presented at: Companion Proceedings of The Web Conference 2018; April 23, 2018:575-583; Lyon, France. [CrossRef]
- Xuan K, Xia R. Rumor stance classification via machine learning with text, user and propagation features. 2019. Presented at: International Conference on Data Mining Workshops (ICDMW); November 8-11, 2019; Beijing, China. [CrossRef]
- Zhou X, Jain A, Phoha VV, Zafarani R. Fake news early detection: an interdisciplinary study. arXiv. Preprint posted online on April 26, 2019. [Medline]
- Wang L, Wang Y, de Melo G, Weikum G. Five shades of untruth: finer-grained classification of fake news. 2018. Presented at: IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM); August 28-31, 2018; Barcelona, Spain. [CrossRef]
- Gautam A, Jerripothula KR. SGG: spinbot, grammarly and glove based fake news detection. 2020. Presented at: IEEE Sixth International Conference on Multimedia Big Data (BigMM); September 24-26, 2020; New Delhi, India. [CrossRef]
- Mahbub S, Pardede E, Kayes ASM. COVID-19 rumor detection using psycho-linguistic features. IEEE Access. 2022;10:117530-117543. [CrossRef]
- Niven T, Kao HY, Wang HY. Profiling spreaders of disinformation on Twitter: IKMLab and softbank submission. CLEF (Working Notes). 2020;2696:1-8. [FREE Full text]
- Mondal T, Pramanik P, Bhattacharya I, Boral N, Ghosh S. Analysis and early detection of rumors in a post disaster scenario. Inf Syst Front. 2018;20(5):961-979. [CrossRef]
- Basharat S, Ahmad M. Inferring trust from message features using linear regression and support vector machines. In: Communications in Computer and Information Science. Singapore. Springer Singapore; 2018:577-598.
- Butt S, Sharma S, Sharma R, Sidorov G, Gelbukh A. What goes on inside rumour and non-rumour tweets and their reactions: a psycholinguistic analyses. Comput Human Behav. 2022;135:107345. [CrossRef]
- Casillo M, Colace F, Conte D, de Santo M, Lombardi M, Mottola S. A multi-feature bayesian approach for fake news detection. In: Computational Data and Social Networks. Cham, Switzerland. Springer International Publishing; 2020:333-344.
- Castillo JM, Fadera KDF, Ladao AAA, Go JG, Tamayo MB, Octaviano MV. Fake news detection on English news article?s title. 2021. Presented at: 1st International Conference in Information and Computing Research (iCORE); December 11-12, 2021; Manila, Philippines. [CrossRef]
- Cerbin L, DeJesus J, Warnken J, Gokhale S. Unmasking the mask debate on social media. 2021. Presented at: IEEE 45th Annual Computers, Software, and Applications Conference (COMPSAC); July 12-16, 2021; Madrid, Spain. [CrossRef]
- Choudhary A, Arora A. Linguistic feature based learning model for fake news detection and classification. Expert Syst Appl. 2021;169:114171. [CrossRef]
- Choudhry A, Khatri I, Jain M, Vishwakarma DK. An emotion-aware multitask approach to fake news and rumor detection using transfer learning. IEEE Trans Comput Soc Syst. 2024;11(1):588-599. [CrossRef]
- Felber T. Constraint 2021: machine learning models for COVID-19 fake news detection shared task. arXiv. Preprint posted online on January 11, 2021 . [CrossRef]
- Ghanem B, Ponzetto SP, Rosso P. FacTweet: profiling fake news Twitter accounts. In: Statistical Language and Speech Processing. Cham, Switzerland. Springer International Publishing; 2020:35-45.
- Jogalekar NS, Attar V, Palshikar GK. Rumor detection on social networks: a sociological approach. 2020. Presented at: IEEE International Conference on Big Data (Big Data); December 10-13, 2020; Atlanta, GA. [CrossRef]
- Kumar A, Sangwan SR, Nayyar A. Rumour veracity detection on Twitter using particle swarm optimized shallow classifiers. Multimed Tools Appl. 2019;78(4):24083-24101. [CrossRef]
- Kumari R, Ashok N, Ghosal T, Ekbal A. A multitask learning approach for fake news detection: novelty, emotion, and sentiment lend a helping hand. 2021. Presented at: International Joint Conference on Neural Networks (IJCNN); July 18-22, 2021; Shenzhen, China. [CrossRef]
- Kwon S, Cha M, Jung K. Rumor detection over varying time windows. PLoS One. 2017;12(1):e0168344. [FREE Full text] [CrossRef] [Medline]
- Li C, Chun SA, Geller J. Stemming the tide of fake news about the COVID-19 pandemic. Int FLAIRS Conf Proc. 2022;35. [FREE Full text]
- Rajabi Z, Shehu A, Purohit H. User behavior modelling for fake information mitigation on social web. In: Social, Cultural, and Behavioral Modeling. Cham, Switzerland. Springer International Publishing; 2019.
- Sandrilla R, Devi MS. FNU-BiCNN: fake news and fake URL detection using Bi-CNN. Int J Adv Comput Sci Appl. 2022;13(2). [FREE Full text]
- Seddari N, Derhab A, Belaoued M, Halboob W, Al-Muhtadi J, Bouras A. A hybrid linguistic and knowledge-based analysis approach for fake news detection on social media. IEEE Access. 2022;10:62097-62109. [CrossRef]
- Singh JP, Kumar A, Rana NP, Dwivedi YK. Attention-based LSTM network for rumor veracity estimation of tweets. Inf Syst Front. 2020;24(2):459-474. [CrossRef]
- Singh JP, Rana NP, Dwivedi YK. Rumour veracity estimation with deep learning for Twitter. In: FIP Advances in Information and Communication Technology. Cham, Switzerland. Springer International Publishing; 2019.
- Slimi H, Bounhas I, Slimani Y. URL-based tweet credibility evaluation. 2019. Presented at: IEEE/ACS 16th International Conference on Computer Systems and Applications (AICCSA); November 3-7, 2019; Abu Dhabi, United Arab Emirates. [CrossRef]
- Sotirakou C, Karampela A, Mourlas C. Evaluating the role of news content and social media interactions for fake news detection. In: Bright J, Giachanou A, Spaiser V, Spezzano F, George A, Pavliuc A, editors. Disinformation in Open Online Media. MISDOOM 2021. Lecture Notes in Computer Science, vol 12887. Cham, Switzerland. Springer International Publishing; 2021.
- Verma PK, Agrawal P, Amorim I, Prodan R. WELFake: word embedding over linguistic features for fake news detection. IEEE Trans Comput Soc Syst. 2021;8(4):881-893. [CrossRef]
- Kula S, Choraś M, Kozik R, Ksieniewicz P, Woźniak M. Sentiment analysis for fake news detection by means of neural networks. In: Krzhizhanovskaya VV, Závodszky G, Lees MH, Dongarra JK, Sloot PMA, Brissos S, et al, editors. Lecture Notes in Computer Science. Cham, Switzerland. Springer International Publishing; 2020.
- Vadavalli A, Al E. Feed forward neural network based effective feature extraction technique for better classification accuracy. TURCOMAT. 2021;12(11):4576-4588.
- Saikh T, Anand A, Ekbal A, Bhattacharyya P. A novel approach towards fake news detection: Deep learning augmented with textual entailment features. In: Natural Language Processing and Information Systems. Cham, Switzerland. Springer International Publishing; 2019:345-358.
- Gravanis G, Vakali A, Diamantaras K, Karadais P. Behind the cues: a benchmarking study for fake news detection. Expert Syst Appl. 2019;128:201-213. [CrossRef]
- Hassan FM, Lee M. Political fake statement detection via multistage feature-assisted neural modeling. 2020. Presented at: IEEE International Conference on Intelligence and Security Informatics (ISI); November 9-10, 2020; Arlington, VA. [CrossRef]
- Patro J, Rathore PS. A sociolinguistic route to the characterizationdetection of the credibility of events on Twitter. 2020. Presented at: Proceedings of the 31st ACM Conference on Hypertext and Social Media; July 15, 2020; Virtual event, United States. [CrossRef]
- Kumar JM, Gopalani D, Kumar MY, Kumar R. Machine learning based fake news detection using linguistic features and word vector features. 2020. Presented at: 2020 IEEE 7th Uttar Pradesh Section International Conference on Electrical, Electronics and Computer Engineering (UPCON); November 27-29, 2020; Prayagraj, India. [CrossRef]
- Rastogi S, Bansal D. Disinformation detection on social media: an integrated approach. Multimed Tools Appl. 2022;81(28):40675-40707. [FREE Full text] [CrossRef] [Medline]
- Reddy H, Raj N, Gala M, Basava A. Text-mining-based fake news detection using ensemble methods. Int J Autom Comput. 2020;17(2):210-221. [CrossRef]
- Volkova S, Shaffer K, Jang JY, Hodas N. Separating facts from fiction: linguistic models to classify suspicious trusted news posts on Twitter. 2017. Presented at: Proceedings of the 55th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers); December 24, 2024; Stroudsburg, PA. [CrossRef]
- Lotfi S, Mirzarezaee M, Hosseinzadeh M, Seydi V. Rumor conversations detection in Twitter through extraction of structural features. Inf Technol Manag. 2021;22(4):265-279. [CrossRef]
- Khanam Z, Alwasel BN, Sirafi H, Rashid M. Fake news detection using machine learning approaches. IOP Conf Ser Mater Sci Eng. 2021;1099(1):012040. [CrossRef]
- Braşoveanu AMP, Andonie R. Integrating machine learning techniques in semantic fake news detection. Neural Process Lett. 2020;53(5):3055-3072. [CrossRef]
- Hamidian S, Diab MT. Rumor detection and classification for Twitter data. arXiv. Preprint posted online on November 25, 2019. [FREE Full text] [CrossRef]
- Iwendi C, Mohan S, Khan S, Ibeke E, Ahmadian A, Ciano T. Covid-19 fake news sentiment analysis. Comput Electr Eng. 2022;101:107967. [FREE Full text] [CrossRef] [Medline]
- Kasseropoulos D, Tjortjis C. An Approach Utilizing Linguistic Features for Fake News Detection. IFIP Advances in Information and Communication Technology. Cham. Springer International Publishing; 2021.
- Rezayi S, Soleymani S, Arabnia HR, Li S. Socially aware multimodal deep neural networks for fake news classification. 2021. Presented at: IEEE 4th International Conference on Multimedia Information Processing and Retrieval (MIPR). IEEE; 2021; September 10, 2021; Tokyo, Japan. [CrossRef]
- Anggrainingsih R, Hassan GM, Datta A. BERT based classification system for detecting rumours on Twitter. arXiv. Preprint posted online on September 7, 2021. [CrossRef]
- Tafannum F, Sharear Shopnil MN, Salsabil A, Ahmed N, Rabiul Alam MG, Tanzim Reza M. Demystifying black-box learning models of rumor detection from social media posts. 2021. Presented at: 2021 IEEE 12th Annual Ubiquitous Computing, Electronics Mobile Communication Conference (UEMCON). IEEE; 2021; December 4, 2021; New York, NY. [CrossRef]
- Al-Rakhami MS, Al-Amri AM. Lies kill, facts save: detecting COVID-19 misinformation in Twitter. IEEE Access. 2020;8:155961-155970. [CrossRef]
- Kalyanam J, Velupillai S, Doan S, Conway M, Lanckriet G. Facts and fabrications about ebola: a Twitter based study. arXiv. Preprint posted online on August 9, 2015. [CrossRef]
- Kwon S, Cha M, Jung K, Chen W, Wang Y. Prominent features of rumor propagation in online social media. 2013. Presented at: IEEE 13th International Conference on Data Mining; December 10, 2013; Dallas, TX. [CrossRef]
- Mazzeo V, Rapisarda A, Giuffrida G. Detection of fake news on COVID-19 on web search engines. Front Phys. 2021;9. [FREE Full text] [CrossRef]
- Puraivan E, Godoy E, Riquelme F, Salas R. Fake news detection on Twitter using a data mining framework based on explainable machine learning techniques. IET Conf Proc. 2021;2021(1):157-162. [CrossRef]
- Sahana VP, Pias AR, Shastri R, Mandloi S. Automatic detection of rumoured tweets and finding its origin. 2015. Presented at: International Conference on Computing and Network Communications (CoCoNet); December 16-19, 2015; Trivandrum, India. [CrossRef]
- Sailunaz K, Kawash J, Alhajj R. Tweet and user validation with supervised feature ranking and rumor classification. Multimed Tools Appl. 2022;81(22):31907-31927. [CrossRef]
- Sharma D, Singhal S. Detection of fake news on social media using classification data mining techniques. Int J Eng Adv Technol. 2019;9(1):3132-3138.
- Afroz S, Brennan M, Greenstadt R. Detecting hoaxes, frauds, and deception in writing style online. 2012. Presented at: IEEE Symposium on Security and Privacy. IEEE; May 23, 2012; San Francisco, CA. [CrossRef]
- Khan S, Hakak S, Deepa N, Prabadevi B, Dev K, Trelova S. Detecting COVID-19-related fake news using feature extraction. Front Public Health. 2021;9:788074. [FREE Full text] [CrossRef] [Medline]
- Levi O, Hosseini P, Diab M, Broniatowski D. Identifying nuances in fake news vs. satire: using semantic and linguistic cues. 2019. Presented at: Proceedings of the Second Workshop on Natural Language Processing for Internet Freedom: Censorship, Disinformation, and Propaganda; November 01, 2019; Hong Kong, China. [CrossRef]
- Argyris YA, Zhang N, Bashyal B, Tan PN. Using deep Learning to identify linguistic features that facilitate or inhibit the propagation of anti- and pro-vaccine content on social media. 2022. Presented at: 2022 IEEE International Conference on Digital Health (ICDH). IEEE; July 10-16, 2022:107-116; Barcelona, Spain. URL: https://europepmc.org/abstract/MED/37975063 [CrossRef]
- Maan M, Jain M, Trivedi S, Sharma R. Machine learning based rumor detection on Twitter data. In: Communications in Computer and Information Science. New York, NY. Springer International Publishing; 2022.
- Raj C, Meel P. Is dynamic rumor detection on social media viable? An unsupervised perspective. arXiv. Preprint posted online on November 23, 2021. [CrossRef]
- Office of the Commissioner. 6 Tip-offs to rip-offs: don't fall for health fraud scams. US Food and Drug Administration. 2021. URL: https://www.fda.gov/consumers/consumer-updates/6-tip-offs-rip-offs-dont-fall-health-fraud-scams [accessed 2023-07-20]
- Apter AJ, Paasche-Orlow MK, Remillard JT, Bennett IM, Ben-Joseph EP, Batista RM, et al. Numeracy and communication with patients: they are counting on us. J Gen Intern Med. 2008;23(12):2117-2124. [FREE Full text] [CrossRef] [Medline]
- Büchter RB, Fechtelpeter D, Knelangen M, Ehrlich M, Waltering A. Words or numbers? Communicating risk of adverse effects in written consumer health information: a systematic review and meta-analysis. BMC Med Inform Decis Mak. 2014;14:76. [FREE Full text] [CrossRef] [Medline]
- Omura M, Maguire J, Levett-Jones T, Stone TE. The effectiveness of assertiveness communication training programs for healthcare professionals and students: a systematic review. Int J Nurs Stud. 2017;76:120-128. [CrossRef] [Medline]
- Richard C, Lussier M, Millette B, Tanoubi I. Healthcare providers and patients: an essay on the importance of professional assertiveness in healthcare today. Med Educ Online. 2023;28(1):2200586. [FREE Full text] [CrossRef] [Medline]
- ChatGPT 3.5. OpenAI. URL: https://chatgpt.com/g/g-F00faAwkE-open-a-i-gpt-3-5 [accessed 2025-01-16]
Abbreviations
API: application programming interface |
AUC: area under the curve |
BERT: Bidirectional Encoder Representations from Transformers |
LIWC: Linguistic Inquiry and Word Count |
PRISMA: Preferred Reporting Items for Systematic Reviews and Meta-Analyses |
Edited by T Mackey; submitted 31.05.24; peer-reviewed by D Boatman, A King, W Ahmed; comments to author 05.07.24; revised version received 22.08.24; accepted 23.11.24; published 12.02.25.
Copyright©Ilona Fridman, Dahlia Boyles, Ria Chheda, Carrie Baldwin-SoRelle, Angela B Smith, Jennifer Elston Lafata. Originally published in JMIR Infodemiology (https://infodemiology.jmir.org), 12.02.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Infodemiology, is properly cited. The complete bibliographic information, a link to the original publication on https://infodemiology.jmir.org/, as well as this copyright and license information must be included.