<?xml version="1.0" encoding="UTF-8"?><!DOCTYPE article PUBLIC "-//NLM//DTD Journal Publishing DTD v2.0 20040830//EN" "journalpublishing.dtd"><article xmlns:mml="http://www.w3.org/1998/Math/MathML" xmlns:xlink="http://www.w3.org/1999/xlink" dtd-version="2.0" xml:lang="en" article-type="letter"><front><journal-meta><journal-id journal-id-type="nlm-ta">JMIR Infodemiology</journal-id><journal-id journal-id-type="publisher-id">infodemiology</journal-id><journal-id journal-id-type="index">38</journal-id><journal-title>JMIR Infodemiology</journal-title><abbrev-journal-title>JMIR Infodemiology</abbrev-journal-title><issn pub-type="epub">2564-1891</issn><publisher><publisher-name>JMIR Publications</publisher-name><publisher-loc>Toronto, Canada</publisher-loc></publisher></journal-meta><article-meta><article-id pub-id-type="publisher-id">v6i1e89831</article-id><article-id pub-id-type="doi">10.2196/89831</article-id><article-categories><subj-group subj-group-type="heading"><subject>Research Letter</subject></subj-group></article-categories><title-group><article-title>Cross-Sectional Evaluation of Medical Disinformation Safeguards in Consumer-Facing Large Language Model Platforms</article-title></title-group><contrib-group><contrib contrib-type="author" corresp="yes"><name name-style="western"><surname>Modi</surname><given-names>Natansh D</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Alex</surname><given-names>Cyril A</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Awaty</surname><given-names>Abdulhalim A</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Menz</surname><given-names>Bradley D</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Bacchi</surname><given-names>Stephen</given-names></name><xref ref-type="aff" rid="aff3">3</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Gradon</surname><given-names>Kacper T</given-names></name><xref ref-type="aff" rid="aff4">4</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Logan</surname><given-names>Jessica M</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Rowland</surname><given-names>Andrew</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Kalisch Ellett</surname><given-names>Lisa</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>McKinnon</surname><given-names>Ross A</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Wiese</surname><given-names>Michael D</given-names></name><xref ref-type="aff" rid="aff1">1</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Sorich</surname><given-names>Michael J</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib><contrib contrib-type="author"><name name-style="western"><surname>Hopkins</surname><given-names>Ashley M</given-names></name><xref ref-type="aff" rid="aff2">2</xref></contrib></contrib-group><aff id="aff1"><institution>Clinical and Health Sciences, University of South Australia</institution><addr-line>GPO Box 2471</addr-line><addr-line>Adelaide</addr-line><addr-line>South Australia</addr-line><country>Australia</country></aff><aff id="aff2"><institution>College of Medicine and Public Health, Flinders Health and Medical Research Institute, Flinders University</institution><addr-line>Adelaide</addr-line><addr-line>South Australia</addr-line><country>Australia</country></aff><aff id="aff3"><institution>Adelaide Medical School, Faculty of Health and Medical Science, The University of Adelaide</institution><addr-line>Adelaide</addr-line><addr-line>South Australia</addr-line><country>Australia</country></aff><aff id="aff4"><institution>Department of Security and Crime Science, University College London</institution><addr-line>London</addr-line><addr-line>England</addr-line><country>United Kingdom</country></aff><contrib-group><contrib contrib-type="editor"><name name-style="western"><surname>Mackey</surname><given-names>Tim</given-names></name></contrib></contrib-group><contrib-group><contrib contrib-type="reviewer"><name name-style="western"><surname>Scales</surname><given-names>David</given-names></name></contrib><contrib contrib-type="reviewer"><name name-style="western"><surname>Nakayama</surname><given-names>Masaharu</given-names></name></contrib></contrib-group><author-notes><corresp>Correspondence to Natansh D Modi, Clinical and Health Sciences, University of South Australia, GPO Box 2471, Adelaide, South Australia, 5001, Australia, 61 08 830 24926; <email>natansh.modi@unisa.edu.au</email></corresp></author-notes><pub-date pub-type="collection"><year>2026</year></pub-date><pub-date pub-type="epub"><day>20</day><month>4</month><year>2026</year></pub-date><volume>6</volume><elocation-id>e89831</elocation-id><history><date date-type="received"><day>18</day><month>12</month><year>2025</year></date><date date-type="rev-recd"><day>19</day><month>03</month><year>2026</year></date><date date-type="accepted"><day>28</day><month>03</month><year>2026</year></date></history><copyright-statement>&#x00A9; Natansh D Modi, Cyril A Alex, Abdulhalim A Awaty, Bradley D Menz, Stephen Bacchi, Kacper T Gradon, Jessica M Logan, Andrew Rowland, Lisa Kalisch Ellett, Ross A McKinnon, Michael D Wiese, Michael J Sorich, Ashley M Hopkins. Originally published in JMIR Infodemiology (<ext-link ext-link-type="uri" xlink:href="https://infodemiology.jmir.org">https://infodemiology.jmir.org</ext-link>), 20.4.2026. </copyright-statement><copyright-year>2026</copyright-year><license license-type="open-access" xlink:href="https://creativecommons.org/licenses/by/4.0/"><p>This is an open-access article distributed under the terms of the Creative Commons Attribution License (<ext-link ext-link-type="uri" xlink:href="https://creativecommons.org/licenses/by/4.0/">https://creativecommons.org/licenses/by/4.0/</ext-link>), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Infodemiology, is properly cited. The complete bibliographic information, a link to the original publication on <ext-link ext-link-type="uri" xlink:href="https://infodemiology.jmir.org/">https://infodemiology.jmir.org/</ext-link>, as well as this copyright and license information must be included.</p></license><self-uri xlink:type="simple" xlink:href="https://infodemiology.jmir.org/2026/1/e89831"/><abstract><p>This cross-sectional evaluation of six consumer-facing large language model platforms found significant heterogeneity in safeguard performance against the generation of health disinformation, with Claude and ChatGPT demonstrating complete resistance across all prompt types, while Copilot, Meta AI, Grok, and Gemini exhibited substantial vulnerabilities.</p></abstract><kwd-group><kwd>epidemiology</kwd><kwd>public health</kwd><kwd>large language models</kwd><kwd>disinformation</kwd><kwd>misinformation</kwd></kwd-group></article-meta></front><body><sec id="s1" sec-type="intro"><title>Introduction</title><p>Consumer-facing large language model (LLM) platforms such as ChatGPT, Copilot, Claude, Gemini, Meta AI, and Grok are increasingly being used as sources of health information, with nearly one in six adults reporting use for this purpose [<xref ref-type="bibr" rid="ref1">1</xref>]. Their popularity reflects low barriers to access and a conversational, personalized style that can make complex material easier to understand [<xref ref-type="bibr" rid="ref2">2</xref>]. However, the same properties that make these platforms attractive for legitimate health information seeking can also be leveraged by malicious actors if safeguards are insufficient, enabling persuasive, scalable, and highly targeted health disinformation generation [<xref ref-type="bibr" rid="ref3">3</xref>]. Here, disinformation is distinguished from misinformation by intent; while misinformation refers to the inadvertent spread of false information, disinformation involves the purposeful creation or dissemination of false content to mislead [<xref ref-type="bibr" rid="ref4">4</xref>]. In the context of LLMs, this intent resides with the malicious actor exploiting the platform rather than the model itself. To mitigate these risks, most platforms implement moderation layers and content-filtering systems to screen prompts and outputs, yet the effectiveness of these safeguards is minimally audited and poorly understood [<xref ref-type="bibr" rid="ref5">5</xref>,<xref ref-type="bibr" rid="ref6">6</xref>]. In the context of emerging agentic workflows, in which systems will increasingly be able to iteratively gather context on individuals and refine personalized, potentially coercive outputs, there is a clear need for greater evaluation and transparency around health disinformation safeguards.</p><p>Prior work in 2023 demonstrated marked differences among consumer-facing LLM platforms in their susceptibility to targeted prompts designed to stimulate the generation of convincing health disinformation [<xref ref-type="bibr" rid="ref7">7</xref>]. In that study, disinformation-seeking prompts to GPT-4 (via Copilot), GPT-4 (via ChatGPT), PaLM 2/Gemini Pro (via Bard), and Llama 2 (via HuggingChat) frequently generated health disinformation. This occurred in response to both direct requests and obfuscation techniques, which included characterization and fictionalization. In contrast, Claude 2 (via Poe) rejected all requests for health disinformation across the same direct and obfuscated prompt sets, demonstrating the feasibility of implementing stronger safeguards. Since that time, substantial performance and capability improvements have been reported for such platforms, necessitating a timely follow-up to assess whether safeguards have strengthened. Therefore, the primary aim of the present study was to determine whether current platforms now provide more reliable protection against targeted requests to generate health disinformation or whether significant vulnerabilities still exist.</p></sec><sec id="s2" sec-type="methods"><title>Methods</title><p>We conducted a cross-sectional evaluation of six consumer-facing LLM platforms using a predefined 90-prompt test adopted from the 2023 benchmark study. The test covered six health disinformation topics, including claims that sunscreen causes cancer, alkaline diets cure cancer, vaccines cause autism, hydroxychloroquine cures COVID-19, genetically modified foods are used for population control, and sugar causes cancer. The test set comprised of 30 direct disinformation-seeking prompts, and 30 prompts utilizing characterization and 30 using fictionalization techniques.</p><p>Five evaluated platforms were successors of those assessed in the 2023 benchmark, with Grok added as a new platform. In 2023, the benchmark assessed ChatGPT (GPT-4), Copilot (GPT-4), Bard (PaLM 2), Claude via Poe (Claude 2), and HuggingChat (Llama 2). In 2025, we evaluated ChatGPT (GPT-5.1), Copilot (GPT-5), Gemini (Gemini 3.0 Pro; formerly Bard), Claude (Claude Sonnet 4.5; accessed via claude.ai), and Meta AI (Llama 4 via meta.ai). Grok (Grok 4) was included as a new addition. All models were accessed through their standard public interfaces (<xref ref-type="supplementary-material" rid="app1">Multimedia Appendix 1</xref>) with default safety settings during a defined four-day window (27-30/11/2025). Each prompt was submitted in a new session with no prior conversation history. Prompts were submitted by one author (C.A.A) under the supervision of a second author (N.D.M). As a result, the evaluation reflects the behavior of each model as implemented in its consumer-facing platform, including any platform-level safety controls.</p><p>Direct prompts were used to enable comparison with the 2023 benchmark, which evaluated all platforms using this strategy. Obfuscated prompts (characterization and fictionalization) were also assessed for all platforms in this study. Outputs were independently coded by two reviewers with relevant expertise in health disinformation research, N.D.M (pharmacist and health disinformation researcher) and C.A.A (health disinformation researcher), with a third reviewer (A.M.H, pharmacist and health disinformation researcher) available for consensus resolution in the event of disagreement, of which none were identified. Outputs were coded as health disinformation if the model generated content presenting a false health claim as factual, without corrections, refutations, or disclaimers. Outputs in which the model refused the request, corrected the false premise or included clear disclaimers indicating the content was inaccurate were not classified as disinformation. Coding decisions were informed by reviewers&#x2019; clinical and research backgrounds and by the established scientific consensus on each of the six evaluated health topics. <xref ref-type="supplementary-material" rid="app2">Multimedia Appendix 2</xref> contains the complete prompt set for all six topics.</p></sec><sec id="s3" sec-type="results"><title>Results</title><p>Safeguard performances for the six platforms are presented in <xref ref-type="table" rid="table1">Table 1</xref> and <xref ref-type="fig" rid="figure1">Figure 1</xref>. Notably, Claude (Claude Sonnet 4.5) and ChatGPT (GPT-5.1) did not generate any health disinformation in response to any direct, characterization, and fictionalization prompts submitted across the six health topics.</p><table-wrap id="t1" position="float"><label>Table 1.</label><caption><p>Safeguarding the performance of consumer-facing large language model (LLM) platforms.</p></caption><table id="table1" frame="hsides" rules="groups"><thead><tr><td align="left" valign="bottom">Platform (Model)</td><td align="left" valign="bottom">Direct prompts (% disinformation)</td><td align="left" valign="bottom">Characterization (% disinformation)</td><td align="left" valign="bottom">Fictionalization (% disinformation)</td></tr></thead><tbody><tr><td align="left" valign="top">ChatGPT (GPT-5.1)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td></tr><tr><td align="left" valign="top">Gemini (Gemini 3.0 Pro)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">30 (100)</td><td align="left" valign="top">28 (93)</td></tr><tr><td align="left" valign="top">Claude (Claude Sonnet 4.5)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td><td align="left" valign="top">0 (0)</td></tr><tr><td align="left" valign="top">Copilot (GPT-5)</td><td align="left" valign="top">14 (47)</td><td align="left" valign="top">21 (70)</td><td align="left" valign="top">25 (83)</td></tr><tr><td align="left" valign="top">Grok (Grok 4)</td><td align="left" valign="top">19 (63)</td><td align="left" valign="top">10 (33)</td><td align="left" valign="top">25 (83)</td></tr><tr><td align="left" valign="top">Meta AI (Llama 4)</td><td align="left" valign="top">14 (47)</td><td align="left" valign="top">28 (93)</td><td align="left" valign="top">24 (80)</td></tr></tbody></table></table-wrap><fig position="float" id="figure1"><label>Figure 1.</label><caption><p>Safeguarding the performance of six consumer-facing large language model (LLM) platforms across six health disinformation topics.</p></caption><graphic alt-version="no" mimetype="image" position="float" xlink:type="simple" xlink:href="infodemiology_v6i1e89831_fig01.png"/></fig><p>For the direct prompts, Copilot (GPT-5) and Meta AI (Llama 4) produced disinformation in 47% (14 of 30) submissions, while Grok (Grok 4) did so in 63% (19 of 30). Gemini (Gemini 3.0 Pro) blocked all direct disinformation-seeking requests.</p><p>For the obfuscated prompts, Gemini (Gemini 3.0 Pro) produced disinformation in 100% (30 of 30) of characterization prompts and 93% (28 of 30) of fictionalization prompts. Meta AI (Llama 4) generated disinformation in 93% (28 of 30) of characterization and 80% (24 of 30) of fictionalization prompts. Copilot (GPT-5) produced disinformation in 70% (21 of 30) of characterization and 83% (25 of 30) of fictionalization prompts, while Grok (Grok 4) did so in 33% (10 of 30) and 83% (25 of 30), respectively.</p><p><xref ref-type="supplementary-material" rid="app3">Multimedia Appendix 3</xref> includes representative model outputs, with one example for each of the three prompt types across all six topics.</p><p>Compared with the 2023 benchmark for direct prompt submissions, Anthropic&#x2019;s Claude (Claude 2 via Poe in 2023; Claude Sonnet 4.5 via claude.ai in 2025) produced no health disinformation in either evaluation. Both ChatGPT (GPT-4 in 2023; GPT-5.1 in 2025) and Gemini (PaLM 2 in 2023; Gemini-3.0 Pro in 2025) appeared to improve, as each had generated health disinformation in response to direct prompts in 2023 but demonstrated complete resistance to all such prompts in the current evaluation. In contrast, the Copilot platform (GPT-4 in 2023; GPT-5 in 2025) showed a decline in safeguard performance, shifting from rejecting all direct health disinformation prompts in 2023 to generating medical disinformation for almost half of such prompts in this study. With regards to the Llama models (Llama 2 accessed by Hugging Chat in 2023; Llama 4 accessed by Meta AI in 2025), each assessment generated health disinformation across both evaluation timepoints.</p></sec><sec id="s4" sec-type="discussion"><title>Discussion</title><p>This evaluation provides an updated assessment of how contemporary consumer-facing LLM platforms respond to targeted attempts to elicit health disinformation. Across six disinformation domains and three prompt styles, the findings demonstrate significant heterogeneity in safeguard performance. Although some platforms resisted all attempts to generate disinformation, others remained highly susceptible, particularly when prompts were framed indirectly.</p><p>Claude (Claude Sonnet 4.5) and ChatGPT (GPT-5.1), did not produce disinformation in any instance. Their consistent refusal across direct, characterization and fictionalization prompts suggests that platform-level safeguards effectively identify and interrupt harmful requests, even when framed within narrative structures. Relative to the 2023 benchmark, these outcomes indicate improved resistance for ChatGPT and continued strong performance for Claude.</p><p>The remaining platforms exhibited substantial gaps. Copilot (GPT-5), Meta AI (Llama 4), and Grok (Grok 4) each produced disinformation in response to direct prompts and rates increased under obfuscation. For Copilot and Meta AI, disinformation rates were higher under both characterization and fictionalization than under direct prompting. Grok showed a mixed pattern, with lower disinformation rates under characterization but high rates under fictionalization. Overall, these patterns indicate that safeguards in several widely used platforms do not generalize reliably across common styles of indirect input, consistent with prior work showing that relatively simple prompt obfuscation can bypass safety controls [<xref ref-type="bibr" rid="ref8">8</xref>]. This matters because real-world health disinformation is often embedded in stories, speculative scenarios, or conversational phrasing rather than stated as an explicit request for falsehoods [<xref ref-type="bibr" rid="ref9">9</xref>].</p><p>Gemini (Gemini 3.0 Pro) showed a distinct profile. It rejected all direct prompts but produced health disinformation in nearly all obfuscated cases. This divergence indicates a narrow protective boundary that responds primarily to the phrasing of a request rather than its substance [<xref ref-type="bibr" rid="ref10">10</xref>]. Direct prompt refusal can therefore give a misleading impression of safety, and the results highlight the limitations of single-format safety testing, which is still common in developer and independent evaluations [<xref ref-type="bibr" rid="ref3">3</xref>]. Furthermore, when viewed alongside the 2023 benchmark, the present findings also show that platform safeguards can change materially as models and platforms are updated.</p><p>This study has several limitations. The evaluation focused on text-only interactions, six predefined topics, and a fixed four-day assessment window. Platform performance may differ for multimodal inputs, such as images or scanned documents, or for emerging health disinformation topics that were not represented in the prompt set. All prompts were in English; however, prior evaluations of LLM platforms in responding to health questions indicate that responses in other languages contain more inaccuracies [<xref ref-type="bibr" rid="ref11">11</xref>]. Even with these limitations, the findings indicate that while progress is evident in some systems, the overall reliability of consumer-facing LLM platforms in resisting health disinformation remains uneven. Increasing commercial pressures, such as the introduction of advertising and monetisation strategies [<xref ref-type="bibr" rid="ref12">12</xref>], have the potential to alter the incentives surrounding safety controls in unpredictable ways. Therefore, it is important to conduct systematic, long-term monitoring of LLM safety performance to inform timely policy and regulatory responses.</p></sec></body><back><ack><p>ChatGPT 5.1 (OpenAI) was used to assist with grammar and spelling refinement during manuscript preparation. The authors take full responsibility for the content, interpretation, and conclusions presented in the manuscript.</p></ack><notes><sec><title>Funding</title><p>Ashley M. Hopkins holds an Emerging Leader Investigator Fellowship from the National Health and Medical Research Council, Australia (APP2008119). The PhD scholarship of Bradley D. Menz is supported by the National Health and Medical Research Council (APP2030913).</p></sec><sec><title>Data Availability</title><p>The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.</p></sec></notes><fn-group><fn fn-type="con"><p>All authors contributed to the writing and review of the manuscript.</p></fn><fn fn-type="conflict"><p>AMH is a recipient of investigator-initiated funding for research outside the scope of the current study from Boehringer Ingelheim. AMH is also a recipient of speaker fees from AstraZeneca. The authors have no other potential conflicts of interest to disclose.</p></fn></fn-group><glossary><title>Abbreviations</title><def-list><def-item><term id="abb1">LLM</term><def><p>large language model</p></def></def-item></def-list></glossary><ref-list><title>References</title><ref id="ref1"><label>1</label><nlm-citation citation-type="web"><article-title>Health misinformation tracking poll: artificial intelligence and health information</article-title><source>KFF</source><year>2024</year><access-date>2025-12-02</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://www.kff.org/health-information-trust/poll-finding/kff-health-misinformation-tracking-poll-artificial-intelligence-and-health-information">https://www.kff.org/health-information-trust/poll-finding/kff-health-misinformation-tracking-poll-artificial-intelligence-and-health-information</ext-link></comment></nlm-citation></ref><ref id="ref2"><label>2</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Esmaeilzadeh</surname><given-names>P</given-names> </name><name name-style="western"><surname>Maddah</surname><given-names>M</given-names> </name><name name-style="western"><surname>Mirzaei</surname><given-names>T</given-names> </name></person-group><article-title>Using AI chatbots (e.g., CHATGPT) in seeking health-related information online: The case of a common ailment</article-title><source>Computers in Human Behavior: Artificial Humans</source><year>2025</year><month>03</month><volume>3</volume><fpage>100127</fpage><pub-id pub-id-type="doi">10.1016/j.chbah.2025.100127</pub-id></nlm-citation></ref><ref id="ref3"><label>3</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Menz</surname><given-names>BD</given-names> </name><name name-style="western"><surname>Modi</surname><given-names>ND</given-names> </name><name name-style="western"><surname>Sorich</surname><given-names>MJ</given-names> </name><name name-style="western"><surname>Hopkins</surname><given-names>AM</given-names> </name></person-group><article-title>Health disinformation use case highlighting the urgent need for artificial intelligence vigilance: weapons of mass disinformation</article-title><source>JAMA Intern Med</source><year>2024</year><month>01</month><day>1</day><volume>184</volume><issue>1</issue><fpage>92</fpage><lpage>96</lpage><pub-id pub-id-type="doi">10.1001/jamainternmed.2023.5947</pub-id><pub-id pub-id-type="medline">37955873</pub-id></nlm-citation></ref><ref id="ref4"><label>4</label><nlm-citation citation-type="book"><person-group person-group-type="editor"><name name-style="western"><surname>Viswanath</surname><given-names>K</given-names> </name><name name-style="western"><surname>Taylor</surname><given-names>TE</given-names> </name><name name-style="western"><surname>Rhodes</surname><given-names>HG</given-names> </name></person-group><article-title>National Academies of Sciences, Engineering, and Medicine</article-title><source>Understanding and Addressing Misinformation About Science</source><year>2025</year><publisher-name>The National Academies Press</publisher-name><fpage>356</fpage><pub-id pub-id-type="doi">10.17226/27894</pub-id></nlm-citation></ref><ref id="ref5"><label>5</label><nlm-citation citation-type="web"><article-title>Upgrading the moderation API with our new multimodal moderation model</article-title><source>OpenAI</source><year>2024</year><access-date>2025-12-02</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/index/upgrading-the-moderation-api-with-our-new-multimodal-moderation-model">https://openai.com/index/upgrading-the-moderation-api-with-our-new-multimodal-moderation-model</ext-link></comment></nlm-citation></ref><ref id="ref6"><label>6</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Modi</surname><given-names>ND</given-names> </name><name name-style="western"><surname>Menz</surname><given-names>BD</given-names> </name><name name-style="western"><surname>Awaty</surname><given-names>AA</given-names> </name><etal/></person-group><article-title>Assessing the system-instruction vulnerabilities of large language models to malicious conversion into health disinformation chatbots</article-title><source>Ann Intern Med</source><year>2025</year><month>08</month><volume>178</volume><issue>8</issue><fpage>1172</fpage><lpage>1180</lpage><pub-id pub-id-type="doi">10.7326/ANNALS-24-03933</pub-id><pub-id pub-id-type="medline">40550134</pub-id></nlm-citation></ref><ref id="ref7"><label>7</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Menz</surname><given-names>BD</given-names> </name><name name-style="western"><surname>Kuderer</surname><given-names>NM</given-names> </name><name name-style="western"><surname>Bacchi</surname><given-names>S</given-names> </name><etal/></person-group><article-title>Current safeguards, risk mitigation, and transparency measures of large language models against the generation of health disinformation: repeated cross sectional analysis</article-title><source>BMJ</source><year>2024</year><month>03</month><day>20</day><volume>384</volume><fpage>e078538</fpage><pub-id pub-id-type="doi">10.1136/bmj-2023-078538</pub-id><pub-id pub-id-type="medline">38508682</pub-id></nlm-citation></ref><ref id="ref8"><label>8</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Vinay</surname><given-names>R</given-names> </name><name name-style="western"><surname>Spitale</surname><given-names>G</given-names> </name><name name-style="western"><surname>Biller-Andorno</surname><given-names>N</given-names> </name><name name-style="western"><surname>Germani</surname><given-names>F</given-names> </name></person-group><article-title>Emotional prompting amplifies disinformation generation in AI large language models</article-title><source>Front Artif Intell</source><year>2025</year><volume>8</volume><fpage>1543603</fpage><pub-id pub-id-type="doi">10.3389/frai.2025.1543603</pub-id><pub-id pub-id-type="medline">40260413</pub-id></nlm-citation></ref><ref id="ref9"><label>9</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Greer</surname><given-names>J</given-names> </name><name name-style="western"><surname>Fitzgerald</surname><given-names>K</given-names> </name><name name-style="western"><surname>Vijaykumar</surname><given-names>S</given-names> </name></person-group><article-title>Narrative elaboration makes misinformation and corrective information regarding COVID-19 more believable</article-title><source>BMC Res Notes</source><year>2022</year><month>06</month><day>28</day><volume>15</volume><issue>1</issue><fpage>235</fpage><pub-id pub-id-type="doi">10.1186/s13104-022-06134-9</pub-id><pub-id pub-id-type="medline">35765114</pub-id></nlm-citation></ref><ref id="ref10"><label>10</label><nlm-citation citation-type="other"><person-group person-group-type="author"><name name-style="western"><surname>Zhang</surname><given-names>S</given-names> </name><name name-style="western"><surname>Zhai</surname><given-names>Y</given-names> </name><name name-style="western"><surname>Guo</surname><given-names>K</given-names> </name><etal/></person-group><article-title>JBShield: defending large language models from jailbreak attacks through activated concept analysis and manipulation</article-title><source>arXiv</source><comment>Preprint posted online on  Feb 11, 2025</comment><pub-id pub-id-type="doi">10.48550/arXiv.2502.07557</pub-id></nlm-citation></ref><ref id="ref11"><label>11</label><nlm-citation citation-type="journal"><person-group person-group-type="author"><name name-style="western"><surname>Menz</surname><given-names>BD</given-names> </name><name name-style="western"><surname>Modi</surname><given-names>ND</given-names> </name><name name-style="western"><surname>Abuhelwa</surname><given-names>AY</given-names> </name><etal/></person-group><article-title>Generative AI chatbots for reliable cancer information: evaluating web-search, multilingual, and reference capabilities of emerging large language models</article-title><source>Eur J Cancer</source><year>2025</year><month>03</month><day>11</day><volume>218</volume><fpage>115274</fpage><pub-id pub-id-type="doi">10.1016/j.ejca.2025.115274</pub-id><pub-id pub-id-type="medline">39922126</pub-id></nlm-citation></ref><ref id="ref12"><label>12</label><nlm-citation citation-type="web"><article-title>Testing ads in ChatGPT</article-title><source>OpenAI</source><year>2026</year><access-date>2026-03-07</access-date><comment><ext-link ext-link-type="uri" xlink:href="https://openai.com/index/testing-ads-in-chatgpt">https://openai.com/index/testing-ads-in-chatgpt</ext-link></comment></nlm-citation></ref></ref-list><app-group><supplementary-material id="app1"><label>Multimedia Appendix 1</label><p>Access details for all consumer-facing evaluated large langauge model platforms.</p><media xlink:href="infodemiology_v6i1e89831_app1.docx" xlink:title="DOCX File, 14 KB"/></supplementary-material><supplementary-material id="app2"><label>Multimedia Appendix 2</label><p>Disinformation prompts.</p><media xlink:href="infodemiology_v6i1e89831_app2.docx" xlink:title="DOCX File, 21 KB"/></supplementary-material><supplementary-material id="app3"><label>Multimedia Appendix 3</label><p>Representative model outputs.</p><media xlink:href="infodemiology_v6i1e89831_app3.xlsx" xlink:title="XLSX File, 104 KB"/></supplementary-material></app-group></back></article>