Viewpoint
Abstract
In recent years, artificial intelligence (AI) has seen rapid advancements, with innovations such as large language models and generative AI evolving at a rapid pace. While this progress offers tremendous opportunities, it also presents risks, particularly in the creation, consumption, and amplification of information and its impact on population health and health program delivery. Thoughtful approaches are necessary to navigate the consequences of advances in AI for different health care professionals and patient populations and from a policy and governance perspective. Through a collaboration between the World Federation of Public Health Associations working groups, this Viewpoint article brings together perspectives, concerns, and aspirations from young adult professionals across 5 continents and from diverse backgrounds to explore the future of public health and AI in the context of the changing health information environment. Our discussion is divided into 2 parts, specifically examining aspects of disinformation and AI, and also the role of public health and medical professionals in a growing AI-driven health information ecosystem. This Viewpoint concludes with 5 key recommendations on how to potentially address issues such as information and disinformation overload; misinformation propagation; and resultant changes in health practices, research, ethics, and the need for robust policies that can dynamically address current and future challenges.
JMIR Infodemiology 2025;5:e69474doi:10.2196/69474
Keywords
Introduction
Position Statement of the Authors
The World Federation of Public Health Associations (WFPHA) serves as a crucial global network uniting public health professionals and organizations committed to improving health outcomes and well-being worldwide. As a leading voice in global health, the WFPHA promotes public health advocacy, influences policy, and advances professional education across diverse regions. Within the WFPHA, the Young Working Group for the WFPHA (Young WFPHA) acts as a collection of medical and health professionals focused on empowering the next generation of public health leaders. Dedicated to fostering leadership, expanding career opportunities, and driving innovation, the Young WFPHA facilitates global collaboration among emerging professionals committed to advancing health and well-being. The group’s vision is to ensure that young professionals are integral to building equitable and high-performing public health systems. The Young WFPHA, consistent with the mission of the WFPHA, is dedicated to equipping emerging public health professionals with the requisite skills and networks to address contemporary public health challenges, including in the context of new and emerging technologies. The Young WFPHA is also active in collaboration with various stakeholders and implements its mission through active engagement in prominent international health forums, including the World Health Assembly. In 2023, the Young WFPHA group conducted a detailed survey aimed at identifying the challenges encountered by early-career public health professionals. The findings identified major obstacles and concerns about the ability to navigate the expanding role of generative artificial intelligence (GenAI) in public health and health care design, delivery, and practice. This next generation of artificial intelligence (AI) tools presents both opportunities and challenges, requiring professionals to adapt and innovate in response to this rapidly evolving landscape. In response, the Young WFPHA partnered with the WFPHA Working Group on Global Health Equity and Digital Technology that focuses on research, education, training, and capacity building to ensure that digital technologies include health equity principles in their design, validation, implementation, and assessment. In this Viewpoint article, the 2 working groups bring together and elevate the diverse perspectives, concerns, and aspirations of young adult public health and medical professionals across 5 continents to explore the future of public health and AI in the context of the changing health information environment. GenAI can be defined as follows [
]:Generative AI is artificial intelligence (AI) that can create original content, such as text, images, video, audio or software code, in response to a user’s prompt or request.
The Emerging Role of Generative AI
The area of AI, which is broad in scope and generally comprises technologies and a field of science that focuses on building computers and machines that can reason, learn, and act in a way that normally requires human intelligence, has seen significant advancements with the introduction of large language models (LLMs) and GenAI applications that are becoming increasingly accessible worldwide. GenAI is aggressively being incorporated by Microsoft, Google, and Apple across their products and services—concomitantly, Meta launched the largest open-source LLM, Llama 3.1. Subsequently, GenAI is being integrated into many aspects of daily life, from content creation (text, images, and music) to productivity tools (automation of tasks, customer service, and coding), data analysis (insights, reports, and trend predictions), and agentic AI models (chatbot agents) transforming our daily lives. Furthermore, GenAI has been shown to be promising in incorporating technically complex knowledge, as exemplified by OpenAI’s ChatGPT’s 98% score on a US Medical Licensing Examination Step 3 mock exam [
].On the other hand, GenAI is, in essence, modes that generate output based on the patterns of content and types of language in the datasets that were used for its training. Therefore, the quality and accuracy of the model’s outputs depend on the format, strategy, language, and knowledge base used in their training processes—when users use expert domain vocabulary, apply domain frameworks, and leverage domain context, they obtain better LLM outputs [
]. In specialized knowledge domains such as medicine, the implementation of LLMs in workflows is still nascent and will require improved evaluation approaches, consideration of prompt engineering, and understanding of human-computer interaction with the systems that incorporate GenAI, among other considerations specific to ensuring patient safety and improved health outcomes [ - ].On the regulatory front, the G7 AI Principles and Code of Conduct have been established to guide the responsible and safe global use of AI. The European Union (EU) led the way with its comprehensive AI Act, positioning itself between China’s strict regulatory controls and the more innovation-friendly, self-regulatory approaches of the United States and United Kingdom. In the United States, the “Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence” executive order originally set the framework for regulatory approaches complemented by state-level and industry-specific regulations but has since been revoked by the new administration through an executive order in an effort to remove barriers to AI development [
- ]. Together, these emerging data and AI governance frameworks will be important to assess in the context of their implications for applications in medicine and public health.The GenAI transformation will arguably impact all levels of society. McKinsey & Company [
] estimates that GenAI could add US $4.4 trillion annually to the global economy, although there are concerns about potential loss of employment across multiple industrial sectors. However, less understood is how GenAI could impact health information systems and broader information ecosystems already at significant risk from a trust and integrity standpoint following the global COVID-19 pandemic that was accompanied by an equally complex “infodemic.” Tackling the challenges associated with infodemics during crises will be critical, particularly in the context of how infodemics in the age of polycrises can be mitigated or exacerbated by established or emerging technologies. We must also consider the broader challenge of the information ecosystems and how diverse types of information—unclear, outdated, conflicting, and low-quality health information; misinformation; and disinformation (defined later in this Viewpoint)—impact health information seeking, understanding, and use, as well as health behaviors.Infodemic refers to an overabundance of information, accurate or not, that occurs during a mass acute health event such as an outbreak, epidemic, or event of mass importance. It can lead to confusion, risky behaviors, and mistrust in health authorities [
].Outside of the health emergency context, the work and impact of public health systems and health professionals is influenced by the structures and dynamics of the information ecosystem. The information environment refers to the dynamic set of ecosystems of all the communication channels, platforms, actors, narratives, and interactions that influence how individuals receive, process, and use and act upon information. With the digitization of information exchange and the ubiquitousness of information consumption, these rapidly changing information ecosystems are straining the capacity of public health agencies and programs to provide credible and accurate health information to clinicians, patients, and the public [
, ].LLMs, GenAI, and other emerging AI technologies and computational systems (eg, quantum computing) present both challenges and opportunities for public health but are rarely designed or implemented with the priorities of public health outcomes, health, and well-being in mind or adequate governance of potential for “dual use” (ie, they could be directly misapplied for threats with potential consequences for public health and safety) [
]. Accordingly, a balanced approach to adopting the tools supported by AI advancements, ensuring that these are both innovative and responsible, is necessary and should be informed from diverse perspectives [ ].To address this complex topic, this Viewpoint article gathers perspectives from young health professionals who represent stakeholders who are crucial in shaping the future of how the public health community assesses, responds to, and introduces into practice these technology changes. Specifically, this Viewpoint will explore the specific opportunities and risks posed by GenAI in the health and information sectors through a collaborative consensus-making process and then conclude with a discussion of essential steps in education, training, research, innovation, policy, and ethics to better ensure the consideration of responsible AI advancements that benefit public health.
Approach Description
An initial draft outline of the manuscript concept was developed, followed by an open call for contributors based on the outlined topics. This call was disseminated among members of the Young WFPHA, leading to the purposeful selection of authors to ensure diverse representation in terms of geography and professional backgrounds. The participating contributors and coauthors came from Portugal, the United States, Nepal, Nigeria, New Zealand, Italy, Germany, and Brazil, representing diverse fields such as medicine, public health, medical informatics, and social psychology.
A collaborative approach toward generating consensus on main findings was used, with biweekly meetings where authors shared insights, proposed improvements, and agreed on new goals. For the 5 questions in part 2, each participating responding author independently and anonymously drafted and submitted their responses to the lead and first author. The latter then synthesized these responses into unified answers, which were then reviewed and refined by all authors until consensus was reached.
For the remainder of the manuscript, while all authors were encouraged to contribute to each subsection, each author was assigned the responsibility of compiling and integrating the content for their respective sections.
Part 1: The Dynamics of Information Ecosystems in Health Through the Critical Lens of AI Innovation
In the first part of our Viewpoint article, we assess the threats and opportunities presented by GenAI tools and applications, specifically concerning disinformation overload, AI-generated misinformation, and societal impacts.
Disinformation
Disinformation is deliberately promoted false information and can include hoaxes, conspiracies, and propaganda. Spreading health disinformation often exploits people’s vulnerabilities for profit or political or ideological influence and inconsistencies in accuracy, relevancy, and context of the information. This can be driven by harmful commercial industries, fraudsters, or profiteers in sectors such as the tobacco, alcohol, food, wellness, and health supplement industries. Some actors spread disinformation to gain influence and monetize their audiences, advancing political or ideological agendas. For example, antivaccination groups use strategies such as donations, membership fees, advertising, and merchandise sales to fund their activities [
]. In addition, geopolitical actors use disinformation to weaken political opponents. Attacks on health systems through propaganda or cyberattacks undermine both the quality of health care and public trust in institutions, eroding people’s sense of safety and social cohesion [ ].In today’s digital age, people are bombarded with vast amounts of information daily through many channels and devices, leading to information overload—a state in which the sheer volume of information becomes overwhelming for people, making it difficult to process and make well-informed decisions. To manage this complexity, the brain uses cognitive shortcuts (or heuristics)—quick mental strategies that allow individuals to make judgments without needing to analyze all details [
- ].Both information overload and cognitive shortcuts contribute to analysis paralysis—when people feel overwhelmed and avoid engaging with information critically. Political, economic, or cultural agents can exploit this unbalanced information environment to advance their agendas with limited transparency, taking advantage of trust erosion and belief confirmation bias—where individuals prioritize information that confirms preexisting views and dismiss contradictory evidence, therefore making it harder to reach people with credible, accurate information [
- ].In the modern information environment, human attention is treated as a scarce and valuable resource, which is sometimes called the attention economy. Platforms, content creators, and media outlets compete to capture the attention of their users, shaping how information is produced, distributed, and consumed to influence their users’ behaviors and keep them on their platforms as long as possible to derive advertising revenue [
- ].As public health strongly relies on the acceptance and adoption of evidence-based health behaviors by the broader population, dynamics arising from information overload–related phenomena can be particularly harmful—these dynamics impact not only individuals’ health behaviors and the interpersonal and community relationships but also the health workforce, the health system, and the sociopolitical environment within which public health systems operate [
]. Regarding the latter, the politicization of health information also leads users to avoid critical thinking for selectively biased media content that can be amplified by platform algorithms [ - ].AI-Generated Misinformation
Comparatively, disinformation may be understood as a subset of misinformation, distinguished primarily by its intentionality, which often makes it more insidious and damaging to public trust [
]. Misinformation can be defined as follows [ ]:Misinformation is when false information is shared, but no harm is meant.
In the context of health-promoting behaviors and designed environments, understanding the continuum of elements in the information environment is essential to contextualize public health actions that aim to prevent harms to health and well-being [
]. This continuum starts with questions and concerns, the natural inquiries that people have during health crises. When these are not addressed, it can lead to information voids in which people search for answers but find none from credible sources. These information voids are situations in which people are especially on the lookout for health information, and when they fail to find it from credible, accurate sources, this creates fertile ground for exposure and susceptibility to misinformation (unintentional falsehoods) and disinformation (deliberate falsehoods), which can resonate with a person’s innate values or preexisting attitudes or perspectives. People may share information online because it is aligned with their beliefs, values, and experiences, which they want others to see, and therefore, these prevailing beliefs can snowball into prevailing narratives that can undermine trust in public health systems.The rise of AI-generated media and content (eg, deep fakes, news articles, statistics, photos, and infographics) makes it increasingly difficult to understand health information objectively and its relevance and trustworthiness, ultimately complicating communication of health guidance, health risks, and benefits of health interventions and programs [
, ]. Reflecting this risk, there has been a rise in low-quality AI-generated content on the internet, which has been referred to as digital sludge or the funkification of the internet, adding to concerns that low-quality AI-generated content will be further incorporated into general-use commercial LLMs, thereby diminishing the quality of LLM outputs over time [ ].Generating digital sludge is a strategy used by health-harming industries in marketing their products, with the alcohol industry being just one example [
]. Finally, the concept of disinformation itself might need to be re-evaluated given the challenges of attributing intentionality in the context of AI-generated content that may be mediated by user-led prompts, impacted by model hallucination, influenced by underlying training data or AI guardrails (eg, responsible AI frameworks or lack thereof and curated knowledge bases), or facilitated by agentic AI models that focus on achieving goals versus outputs. Inherently, these features of GenAI technologies make it harder to discern intentionality without understanding the intent of the human users who use these tools for engaging in health communication and information dissemination.In addition to being used for content creation, GenAI models are being built into many web portals, internet platforms, apps, and devices, therefore impacting the quality of the information to which individuals are exposed. However, these applications often prioritize popularity or relevance over accurate data sources, potentially reinforcing the imprecisions of the algorithm or of the user’s own preferences (which can be harmful, such as the promotion of ineffective treatments, amplifying health myths) as results are tailored to similarly situated users. An example of this challenge is the memory function of widely used GenAI platforms (eg, ChatGPT and Claude), which may skew a user’s information knowledge base over time without the user noticing [
]. This can lead users to reinforce inaccurate perceptions of health information and become more likely to share disinformation, which, over time, can lead to greater communal exposure to conflicting or false information that can erode trust in public health authorities and exacerbate existing societal divisions [ , , , - ].This emphasizes the need to understand how the integration of AI technologies into our information environment without appropriate public health safeguards can skew users’ perceptions and source accuracy, which demands a more nuanced approach to address not only the potential explicit harms of inaccurate narratives but also how the built information environment shapes people’s understanding of health and well-being.
To mitigate the potential impact of GenAI-generated misinformation and in the interpretation of the built information environment, we propose a comprehensive approach:
- Technological solutions; development of advanced algorithms to detect and flag deep fakes and inaccurate information [ ].
- Public awareness; promoting critical thinking and digital literacy to recognize misinformation among the public.
- Regulatory frameworks; implementing regulations to hold the full chain of dissemination of false information accountable (eg, the recently approved EU AI Act mandates disclosure when interacting with AI) [ ].
- Collaboration; engaging with technology companies, media, health care professionals, and governments to identify and mitigate the impact of misinformation [ ].
Societal Impacts
Mis- and disinformation are estimated to comprise 5% to 25% of the information environment and have wide-ranging impacts: psychological, physical, social, economic, and political [
- ]. Accordingly, these previously described information ecosystems—where sensationalist reporting, conflicting expert opinions, and slow issuance of health guidance create confusion among the public—undermine trust and enable misinformation exposure, endangering public health campaigns aimed at improving health outcomes [ , , ]. For example, during a pandemic, this brings risks such as vaccine hesitancy and disregard for public health measures [ - ]. In addition, the health information system can lead to other behavior changes, with patients self-diagnosing more, requiring unnecessary and potentially harmful examinations and treatments, and possibly forgoing preventive or care options [ , ]. Finally, it is important to note that vulnerable communities (eg, those with low digital or health literacy or those who lack adequate access to health care services or coverage) are particularly susceptible to the deleterious impact of infodemics [ ].On the other hand, when combined with other strategies to strengthen public health surveillance, health promotion and education, and health informatic systems, GenAI has the potential to be an ally in the detection and mitigation of misinformation. For example, initiatives such as the health-related misinformation detection framework, SimSearchNet by Meta, and SynthID from Google DeepMind have the potential to act as agents for detecting health misinformation and AI-generated content, therefore being able to support the spread of accessible and accurate information [
- ]. In addition, GenAI may also be useful to understand geographical and temporal patterns of information and misinformation, analyze large datasets quickly for prevailing misinformation narratives that emerge from specific communities of interest, and better predict new trends and forecasts as health emergencies arise [ ]. Furthermore, it has the potential through chatbots and other tools to enhance health literacy; accelerate information dissemination; support treatment adherence; enable early diagnosis; and contribute to disease surveillance, risk assessment, and mental health support if used appropriately and gated and guided by evidence-based information and ethical principles [ - ]. An example is the digital health promoter prototype Sarah, a GenAI chatbot and digital health promoter assistant developed by the World Health Organization (WHO) that provides guidance on healthy habits, mental health, noncommunicable diseases, and misinformation handling through online face-to-face conversations [ - ]. Examples of public health surveillance and health informatic systems include SIRVD-DL (Susceptible, Infected, Recovered, Vaccinated, and Deceased–Deep Learning) prediction model for COVID-19 surveillance, the WHO’s Global Influenza Surveillance and Response System, and AI for Pandemic and Epidemic Preparedness, which could benefit from integration with AI tools to tackle global health challenges, including in the context of identifying and contextualizing the impact of misinformation on health outcomes [ ]. Other uses of GenAI aimed to support health behavior change have focused on improving literacy, access, and operational efficiency in projects spread across low- and middle-income countries in both Asia and Africa (Jacaranda Health, Viamo, Girl Effect, Audere, and Noora Health) [ ]. However, the utility of GenAI tools to carry out shared public health objectives and goals lacks sufficient evidence and requires additional research. The overuse of such tools may lead to an overestimation of their actual value as current evidence does not yet support their widespread implementation. Nevertheless, many enterprises are already integrating GenAI functionalities into everyday activities.While technological innovation can help shape healthier information environments, it is crucial to address these challenges holistically considering the broader, governance, regulatory, social, and informational contexts that can be informed by young professionals, who will arguably be most impacted by the implementation of these technologies in current and future public health and medical settings [
].Part 2: The Role of Public Health and Medical Professionals in LLMs and Information Ecosystems in Health
As discussed previously, the impact of the information ecosystem and mis- and disinformation on health and well-being in the GenAI era is vast and ever-evolving, with constantly new versions of LLM tools and systems that use them. Hence, it is crucial to take into account the views of young health professionals across various fields and regions to better understand current and future opportunities and challenges. We consulted medical doctors, public health professionals, and other health experts (researchers and psychologists) from 5 continents. The following sections present their perspectives through a unified response developed from the intersection of their individual blind answers.
What Roles Do LLMs Currently Play in Your Professional Area, and How Can They Enhance the Propagation of Safe and Reliable Health Information?
LLMs are transforming health care by automating administrative tasks; enhancing clinical, public health, and administrative data management; and supporting health education and research. Examples of solutions under development that integrate LLMs include patient education, automated medical record writing, and providing suggestions for patient diagnosis and management [
]. In public health, LLMs can aid in data treatment, contributing to predictive analysis and response strategies during health emergencies and conducting rapid contextualized analysis of community sentiment and health behaviors to generate infodemic insights and inform adaptive delivery of public health responses, programs, and communications. Despite these advancements, the integration of LLMs into health systems and public health practices faces significant regulatory and ethical challenges. For example, Health New Zealand – Te Whatu Ora has specifically advised against the use of GenAI in health due to concerns over privacy, accuracy, and ethics. To address these issues, the WHO has called for stronger evidence on the design, training, and validation of AI-supported applications, yet global regulatory frameworks remain insufficient to oversee the full life cycle of AI in health care and public health [ ].How Is the Integration of AI and LLMs Transforming Patient Communication and Overall Health Outcomes in Your Field?
AI and LLMs can transform patient-clinician communication not only in the consultation process but also by extending this communication beyond traditional consultation boundaries. These technologies can collect clinical information before appointments, handle administrative tasks such as record writing, and facilitate follow-up communications. If AI and LLMs can reduce the nonclinical workload of clinicians, this could allow health care providers to spend more time interacting with patients, thereby enhancing efficiency and quality of care. However, this productivity gain may only increase the number of patients seen without necessarily improving the quality of care.
In addition, if their current limitations and risks are overcome, GenAI tools have the potential to contribute to health promotion, risk communication, health education, and social behavior change efforts in public health systems. These technologies can personalize messaging, analyze real-time health data, and automate responses, enabling more targeted health education and behavior change interventions. AI can enhance risk communication by providing tailored, timely health warnings and support behavior change through internet-based counseling and interactive tools.
Furthermore, AI-powered tools are being tested to contribute to virtual counseling, adapted health information systems, and treatment adherence (eg, medication reminders) [
- ]. This can improve the patient-clinician communication and the care continuum, ensuring that patients receive comprehensive and continuous care. To facilitate this, such technologies could be integrated into previously developed digital solutions to make health care more accessible in rural areas, for example, through remote diagnostics and consultations in New Zealand and Brazil [ ].How Do You Plan to Integrate LLMs Into Your Practice, and What Specific Applications Do You Foresee?
LLMs’ ability to process enormous amounts of data presents an opportunity to aid health care systems in managing patient outcomes. LLMs can transform previously unused data into actionable insights, enabling more effective management of patient outcomes. They can support clinical decision-making by providing clinicians with insights about patients in similar situations or generating concise patient summaries. If we can eliminate biases and other current limitations, LLMs can potentially identify patterns in patient data that may indicate a risk of developing certain conditions (digital twins), enabling early diagnosis and interventions, and remind patients and health professionals about regular checkups and screenings [
]. In this case, by analyzing and organizing large volumes of data, LLMs may facilitate regulated, accountable, and auditable data interoperability, increasing clinicians’ effectiveness and accelerating access to patient information when necessary.Furthermore, LLMs can improve resource management. For example, in Germany, one of the authors is working on a project that is exploring the integration of LLMs into nonemergency urgent care to help with patient navigation—this system determines whether a patient needs an in-person urgent care appointment, a telemedical consultation, or over-the-counter medication from a pharmacy or can wait until normal business hours for their general practitioner.
What Policies and Ethical Guidelines Do You Think Are Necessary to Responsibly Integrate GenAI Into Health Practices Worldwide (and Locally)?
The responsible integration of GenAI into health care requires robust global policies to avoid health risks, privacy issues, and biases (a study of a large US hospital database suggests that eliminating racial bias in triage algorithms would increase the percentage of Black patients who receive additional help from 17.7% to 46.5%) [
]. Thus, formulation of relevant policies and guidelines is urgently needed. Adherence to existing standards such as the General Data Protection Regulation (EU), HIPAA (Health Insurance Portability and Accountability Act; United States), and Lei Geral de Proteção de Dados (Brazil) is foundational. However, people living in countries without consumer protections are disproportionately affected by how commercial actors use their communities to develop, test, and deploy AI-supported tools, which may further exacerbate health and well-being inequities worldwide. Given the international nature of AI data markets, regional and national strategies should seek to align with global guidelines, with international organizations such as the WHO respecting the sovereignty of countries and the agency of individuals. Policies should balance innovation with patient safety and respect, ensuring practical implementation. The EU AI Act, for example, aims to create a comprehensive regulatory framework for AI, addressing risks and promoting ethical AI development and use within the EU.Integration of these technologies must provide patients with easy options to opt out of data sharing without facing adverse consequences. In addition, specific consumer protection against deceptive marketing to vulnerable populations (eg, children), protection against hate speech, and protections of freedom of expression are essential. Furthermore, addressing the risk of propagating biases from training data is crucial as it impacts the fairness and accuracy of AI outputs. Ensuring inclusive and representative datasets is essential for equitable treatment of all patient groups. In addition, GenAI systems must be continuously validated in a transparent manner to increase data reliability, avoid issues related to copyright and intellectual property rights, and maintain trust and efficacy. In line with the Global Digital Compact, AI governance must be anchored in human rights and international law, with ethical guidelines, regular compliance reviews, and data security ensuring that these technologies benefit all [
- ].What Future Impacts of GenAI in Health Care and Public Health Are Underexplored or Underestimated, and How Should the Health Professional Community Prepare for These Changes?
The future impacts of GenAI in health care are promising but need to be further evaluated for key public health areas: disease surveillance, prediction of disease outbreaks, assessment of individual health risks, and suggestion of preventive measures. These advancements could lead to lower health care costs and better patient outcomes. In addition, GenAI has the potential to aid medical education through personalized learning experiences, virtual simulations, and real-time feedback. It can also streamline administrative tasks such as scheduling, billing, and resource allocation. Furthermore, in direct patient care, information from different centers remains highly fragmented worldwide. Patients often forget details about their medications, medical history, and previous treatments or surgeries. GenAI could help by securely integrating and condensing these data, reducing both missing and overlapping information, similar to what is being developed in the European Health Data Space.
While future health professionals will not need to be technology experts, they must effectively communicate with technology specialists. Therefore, every health degree should include GenAI training, including risks and limits, to prepare students for this integration. There will also be a growing need for professionals who can bridge management, technological, and health competencies to manage these complex systems, which can lead to work overload.
However, significant risks and challenges must be addressed. Public and ethical oversight, equity, and social participation are essential to prevent biases and ensure fair treatment for all patient groups and communities. Adequate resourcing of public communication; social participation; health promotion and health education capacities; and multilevel building of health, digital, information, and media literacies in communities and in the workforce is crucial. In addition, data sovereignty of peoples and nations (including indigenous communities experiencing inequities) is a crucial but underdeveloped area with important implications as AI progresses [
].Policies must respect the data sovereignty of communities to ensure the ethical and equitable use of AI. These steps are necessary to foster acceptance and understanding of AI in health care.
Part 3: Discussion and Call to Action
In the third part of this Viewpoint article, we focus on summarizing the potential impacts of GenAI in various areas of society, with a specific emphasis on education and training, research and innovation, and policy and ethics.
While we are not specifically prescriptive, allowing for contextualized implementation and locally defined metrics, we aim to inspire actionable advancements across these areas. We also highlight that networks such as the WFPHA are well positioned to support these efforts, whether through capacity building, literacy initiatives, or collaboration on the responsible integration of GenAI into public health.
Education and Training
GenAI transformation is happening, and therefore, comprehensive training to future generations as well as to the current workforce is necessary. A commitment to maturation of AI literacy is necessary not only to improve the productivity of GenAI models and agents but also to ensure that consumers of GenAI-generated content are aware of its potential risks and biases and how to mitigate them [
]. In addition, health presents particular needs and serious implications for individual and community health, and therefore, specific continuous training for health and public health professionals should be prioritized.Educational efforts must be comprehensive, addressing areas that continue to be debated in the GenAI space, including data privacy, policy implications, and algorithmic bias, while also promoting transparent and ethical design of GenAI systems and applications that are based on them. As previously discussed, promoting digital, media, and health literacy is central to infodemic resilience. This must be embedded into training programs for both the current workforce and future generations [
, , - ].Therefore, we propose the following call to action: increase health literacies (health, media, information, and science) and specifically AI literacy starting as early as preuniversity levels to prepare future generations for the AI-driven health care landscape, as well as offering courses on AI use, risks, biases, and management of AI-generated information in all health degree programs. Similarly, we recommend integrating health topics into fields such as computer science, ethics, and technology governance.
Research and Innovation
While we understand the importance of research frameworks that address AI interventions, such as CONSORT-AI (Consolidated Standards of Reporting Trials–Artificial Intelligence), SPIRIT-AI (Standard Protocol Items: Recommendations for Interventional Trials–Artificial Intelligence), and the WHO’s guidelines, we call attention to higher-level research needs that AI should support. Accordingly, we believe that AI research and innovation in health should prioritize three key areas: (1) ensuring universal and equitable access to accurate and high-quality health information (recognized by some organizations as a fundamental human right) by developing innovative strategies and digital solutions that promote informed decision-making and improve access to health care; (2) supporting preventive health programs using AI tools to enhance early detection, risk assessment, and health promotion efforts; and (3) combating polarization and bias in health information, ensuring that content delivered to diverse populations is trustworthy, inclusive, and culturally sensitive. Furthermore, while GenAI can significantly enhance research capabilities through advanced writing and data analysis, interdisciplinary collaboration is essential to prevent the spread of appealing yet inaccurate research results and misinterpretation or misrepresentation [
- ].Therefore, we also introduce a call to action for incorporation of AI and ethics experts within public health organizations to ensure accurate and ethical AI application.
Policy and Ethics
Effective GenAI adoption in health care and public health requires collaboration among patients, civil society, and policy makers to establish ethical frameworks, regulatory oversight, and effective data governance. The Global Digital Compact, annexed to the United Nations Pact for the Future, is the first global commitment to data governance—in addition to calling attention to the need to take local action by 2030, it incentivizes a global policy approach on AI governance and clarifies the need of this space to respect human rights and international law [
]. This urgent need for greater transparency and accountability is recognized by stakeholders from different backgrounds, from nongovernmental to corporate and philanthropic backgrounds (as exemplified by the Partnership on AI) [ - ].A responsible transition depends on comprehensive yet flexible policy development (such as the EU AI Act), robust ethical guidelines (such as the WHO’s “Ethics and governance of artificial intelligence for health: Guidance on large multi-modal models”), and a strong commitment to ensure health equity [
, ]. These are vital to avoid biases in AI models and ensure that GenAI benefits are distributed fairly.Regarding ethics, the WHO has outlined 6 ethical principles to guide AI development in health: protecting autonomy, promoting well-being and safety, ensuring transparency, fostering accountability, ensuring inclusiveness and equity, and promoting sustainability [
]. WHO ethics guidance will soon extend to provide help and capacity building to public health authorities to establish infodemic management practices, policies, and strategies [ ]. These guidelines emphasize the importance of fairness, transparency, and accountability, addressing critical concerns such as data privacy, algorithmic bias, and health equity to prevent AI from deepening existing health disparities. Finally, the recent adoption of the WHO Pandemic Agreement, aimed to set a global standard for pandemic prevention, preparedness, and response, does not contain specific reference to infodemics or dis- or misinformation but does recognize the importance of trust and transparency in communications on the topic [ ]. This further emphasizes the need for continued international cooperation and coordination on using GenAI as an objective tool to enhance health communication.Therefore, we also make a call to action for international collaboration to develop AI-related legislation that ensures global standards and practices respecting regional and national autonomies, with clear definition of high-risk applications in medicine and public health, and the development of equitable ethical and regulatory guidelines and codes to guide all sectors in the responsible use of AI in health care and public health that prioritizes health equity and human rights.
Conclusions
GenAI integration in health care and public health presents both significant opportunities and substantial risks, especially as it intersects with the complex dynamics of health information ecosystems. The main threats involve exacerbating information overload and disinformation and AI-generated misinformation exposure, all of which can negatively impact individual and collective health behaviors. However, through responsible, participatory, and evidence-based AI development, we can mitigate these risks and enhance the accuracy and accessibility of quality health information.
As emerging leaders in public health drawing on the experiences of professionals from diverse backgrounds across 5 continents, we recognize this dual nature of AI technologies. To harness GenAI’s potential responsibly, we must prioritize AI literacy among health professionals and the public, integrate AI education into health curricula, and establish robust ethical guidelines and regulatory frameworks that promote equity and protect patient privacy. By taking proactive steps in education, research, and policy, we can leverage GenAI to enhance global health while safeguarding against its risks.
Acknowledgments
This paper used large language models (ChatGPT; OpenAI) with the purpose of improving the language. Large language models were not used for content development. In line with the “Best Practices for Using AI Tools as an Author, Peer Reviewer, or Editor,” [
] the authors assert that they are aligned with the guiding principles of accountability, transparency, and confidentiality in what regards the use of AI tools.Authors' Contributions
A Bandeira and TM developed the outline of the manuscript. A Bandeira, LHG, FH, JUS, MLG, RP, and SP equally contributed to the writing of the Viewpoint content, whereas A Berionni, TDP, and TM contributed with significant reviewing insights. All authors except for A Bandeira, A Berionni, TDP, and TM answered individually to the questions at the beginning of part 2; A Bandeira collected all answers and wrote a joint answer to each question that was later reviewed and approved by all authors. All authors reviewed and agreed to the final version of the manuscript.
Conflicts of Interest
TDP is an editorial board member for JMIR Infodemiology. TM, the editor-in-chief of JMIR Infodemiology, is a coauthor of this manuscript. To mitigate any potential conflict of interest, the authors clarify in the Authors’ Contributions section that he did not participate in responding to part 2 questions or in the development of the core content of the manuscript. His contributions were limited to assisting with the development of the paper outline and providing final expert review based on his extensive subject matter expertise. The authors remain committed to maintaining transparency and adhering to the journal’s policies regarding conflicts of interest.
References
- Stryker C, Scapicchio M. What is generative AI? IBM Corp. URL: https://www.ibm.com/topics/generative-ai [accessed 2025-05-29]
- Scott G, Shani B. Op-ed: How well can AI chatbots mimic doctors in a treatment setting? We put 5 to the test. CNBC. URL: https://www.cnbc.com/2024/07/18/op-ed-how-well-can-ai-chatbots-mimic-doctors.html [accessed 2025-05-29]
- Zamfirescu-Pereira JD, Wong RY, Hartmann B, Yang Q. Why Johnny can’t prompt: how non-AI experts try (and fail) to design LLM prompts. In: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 2023. Presented at: CHI '23; April 23-28, 2023; Hamburg, Germany. URL: https://dl.acm.org/doi/10.1145/3544548.3581388 [CrossRef]
- Sivarajkumar S, Kelley M, Samolyk-Mazzanti A, Visweswaran S, Wang Y. An empirical evaluation of prompting strategies for large language models in zero-shot clinical natural language processing: algorithm development and validation study. JMIR Med Inform. Apr 08, 2024;12:e55318. [FREE Full text] [CrossRef] [Medline]
- Mizrahi M, Kaplan G, Malkin D, Dror R, Shahaf D, Stanovsky G. State of what art? A call for multi-prompt LLM evaluation. arXiv. Preprint posted online on December 31, 2023. 2025. [FREE Full text] [CrossRef]
- Khurana A, Subramonyam H, Chilana PK. Why and when LLM-based assistants can go wrong: investigating the effectiveness of prompt-based interactions for software help-seeking. In: Proceedings of the 29th International Conference on Intelligent User Interfaces. 2024. Presented at: IUI '24; March 18-21, 2024; Greenville, SC. URL: https://dl.acm.org/doi/abs/10.1145/3640543.3645200 [CrossRef]
- Regulation (EU) 2024/1689 of the European Parliament and of the Council of 13 June 2024 laying down harmonised rules on artificial intelligence and amending Regulations (EC) No 300/2008, (EU) No 167/2013, (EU) No 168/2013, (EU) 2018/858, (EU) 2018/1139 and (EU) 2019/2144 and Directives 2014/90/EU, (EU) 2016/797 and (EU) 2020/1828 (Artificial Intelligence Act) (text with EEA relevance). European Union. URL: https://eur-lex.europa.eu/eli/reg/2024/1689/oj/eng [accessed 2025-05-29]
- Haddad M. AI needs regulation, but what kind, and how much? The Economist. URL: https://www.economist.com/schools-brief/2024/08/20/ai-needs-regulation-but-what-kind-and-how-much [accessed 2025-05-29]
- G7 AI principles and code of conduct. Ernst & Young Global. URL: https://www.ey.com/en_in/insights/ai/g7-ai-principles-and-code-of-conduct [accessed 2025-05-29]
- Removing barriers to American leadership in artificial intelligence. The White House. URL: https://www.whitehouse.gov/presidential-actions/2025/01/removing-barriers-to-american-leadership-in-artificial-intelligence/ [accessed 2025-05-29]
- What’s the future of AI? McKinsey & Company. URL: https://www.mckinsey.com/featured-insights/mckinsey-explainers/whats-the-future-of-ai [accessed 2025-05-29]
- Infodemic. World Health Organization. URL: https://www.who.int/health-topics/infodemic#tab=tab_1; [accessed 2025-05-29]
- Gonçalves L, Castro L, Rachid R, Fornazin M. As múltiplas faces da Infodemia. Rev Eletron Comun Inf Inov Saúde. Sep 30, 2024;18(3):716-735. [CrossRef]
- Bennett WL, Livingston S. The disinformation order: disruptive communication and the decline of democratic institutions. Eur J Commun. Apr 02, 2018;33(2):122-139. [CrossRef]
- Global guidance framework for the responsible use of the life sciences: mitigating biorisks and governing dual-use research. World Health Organization. URL: https://www.who.int/publications/i/item/9789240056107 [accessed 2025-05-29]
- Rasooly D, Khoury MJ. Artificial intelligence in medicine and public health: prospects and challenges beyond the pandemic. Centers for Disease Control and Prevention (CDC). URL: https://blogs.cdc.gov/genomics/2022/03/01/artificial-intelligence-2/ [accessed 2025-05-29]
- The anti-Vaxx playbook. Center for Countering Digital Hate. URL: https://counterhate.com/research/the-anti-vaxx-playbook/ [accessed 2025-05-29]
- Using social media in community based protection. United Nations High Commissioner for Refugees. URL: https://www.unhcr.org/innovation/using-social-media-in-community-based-protection/ [accessed 2025-05-29]
- Patent V. Dysfunctional trusting and distrusting: integrating trust and bias perspectives. J Trust Res. Sep 02, 2022;12(1):66-93. [CrossRef]
- Ecker UK, Lewandowsky S, Cook J, Schmid P, Fazio LK, Brashier N, et al. The psychological drivers of misinformation belief and its resistance to correction. Nat Rev Psychol. Jan 12, 2022;1(1):13-29. [CrossRef]
- van Zoonen L. I-Pistemology: changing truth claims in popular and political culture. Eur J Commun. Mar 29, 2012;27(1):56-67. [CrossRef]
- Bing C, Schectman J. Pentagon ran secret anti-vax campaign to undermine China during pandemic. Reuters. Jun 14, 2024. URL: https://www.reuters.com/investigates/special-report/usa-covid-propaganda/ [accessed 2025-05-29]
- Calac AJ, Southwell BG. How misinformation research can mask relationship gaps that undermine public health response. Am J Health Promot. Mar 14, 2022;36(3):561-563. [CrossRef] [Medline]
- Pennycook G, McPhetres J, Zhang Y, Lu JG, Rand DG. Fighting COVID-19 misinformation on social media: experimental evidence for a scalable accuracy-nudge intervention. Psychol Sci. Jul 30, 2020;31(7):770-780. [FREE Full text] [CrossRef] [Medline]
- Cesarino L. Pós-verdade e a crise do sistema de peritos: uma explicação cibernética. Ilha R Antropologia. Feb 24, 2021;23(1):73-96. [CrossRef]
- Goncalves LH, Furtado O. The fake simple exchange between Facebook and its prosumers. Int J Soc Politics Cult. 2020;2(2):206. [FREE Full text]
- Wu T. The Attention Merchants. Berlin, Germany. Alfred a Knopf Inc; 2016.
- Altay S, de Araujo E, Mercier H. “If this account is true, it is most enormously wonderful”: interestingness-if-true and the sharing of true and false news. Digit Journal. Aug 24, 2021;10(3):373-394. [CrossRef]
- Lewandowsky S, Ecker UK, Cook J. Beyond misinformation: understanding and coping with the “post-truth” era. J Appl Res Mem Cogn. Dec 2017;6(4):353-369. [CrossRef]
- Dantas Loureiro M. The financial logic of internet platforms: the turnover time of money at the limit of zero. tripleC. May 14, 2019;17(1):132-158. [CrossRef]
- Srnicek N. Platform Capitalism. New York, NY. Polity Press; 2016.
- Laxa J. The consumption of disinformation as a health crisis. J Public Health (Oxf). Mar 14, 2023;45(1):e161. [CrossRef] [Medline]
- Valenzuela S, Halpern D, Katz JE, Miranda JP. The paradox of participation versus misinformation: social media, political engagement, and the spread of misinformation. Digit Journal. Jun 12, 2019;7(6):802-823. [CrossRef]
- Pennycook G, Rand DG. The psychology of fake news. Trends Cogn Sci. May 2021;25(5):388-402. [FREE Full text] [CrossRef] [Medline]
- Garrett C, Qiao S, Li X. The role of social media in knowledge, perceptions, and self-reported adherence toward COVID-19 prevention guidelines: cross-sectional study. JMIR Infodemiology. Feb 16, 2024;4:e44395. [FREE Full text] [CrossRef] [Medline]
- Dutilh Novaes C. The role of trust in argumentation. Informal Logic. Jul 06, 2020;40(2):205-236. [CrossRef]
- Nascimento LF, Barreto T, Cesarino L, Mussa V, Fonseca P. Públicos refratados: grupos de extrema-direita brasileiros na plataforma Telegram. Internet & Sociedade. URL: https://revista.internetlab.org.br/publicos-refratados-grupos-de-extrema-direita-brasileiros-na-plataforma-telegram/ [accessed 2025-05-29]
- Brady WJ, Jackson JC, Lindström B, Crockett M. Algorithm-mediated social learning in online social networks. Trends Cogn Sci. Oct 2023;27(10):947-960. [CrossRef] [Medline]
- Arguedas AR, Robertson CT, Fletcher R, Nielsen RK. Echo chambers, filter bubbles, and polarisation: a literature review. The Reuters Institute for the Study of Journalism. URL: https://reutersinstitute.politics.ox.ac.uk/echo-chambers-filter-bubbles-and-polarisation-literature-review [accessed 2025-05-29]
- Kasisomayajula V, Allum N, Barrett N, Broniatowski D, Bruce AA. Understanding and addressing misinformation about science. National Academics. URL: https://www.nationalacademies.org/our-work/understanding-and-addressing-misinformation-about-science [accessed 2025-05-29]
- Wardle C, Derakhshan H. Information disorder: toward an interdisciplinary framework for research and policy making. Council of Europe. URL: https://edoc.coe.int/en/media/7495-information-disorder-toward-an-interdisciplinary-framework-for-research-and-policy-making.html [accessed 2025-05-29]
- Briand S, Hess S, Nguyen T, Purnat TD. Infodemic management in the twenty-first century. In: Purnat TD, Nguyen T, Briand S, editors. Managing Infodemics in the 21st Century. Cham, Switzerland. Springer; 2023.
- Purnat TD. Information ecosystem disrupting health, Ex #3: misdirection and misleading in health information search results. Linkedin. URL: https://www.linkedin.com/pulse/information-ecosystem-disrupting-health-ex-3-search-results-d-purnat-tny8e [accessed 2025-05-29]
- Ashraf AR, Mackey TK, Fittler A. Search engines and generative artificial intelligence integration: public health risks and recommendations to safeguard consumers online. JMIR Public Health Surveill. Mar 21, 2024;10:e53086. [FREE Full text] [CrossRef] [Medline]
- Lubin N. What to do about the Junkification of the internet. The Atlantic. URL: https://www.theatlantic.com/technology/archive/2024/03/generative-ai-social-media-moderation/677730/ [accessed 2025-05-29]
- Roy-Highley E, Körner K, Mulrenan C, Petticrew M. Dark patterns, dark nudges, sludge and misinformation: alcohol industry apps and digital tools. Health Promot Int. Oct 01, 2024;39(5):daae037. [CrossRef] [Medline]
- Lampl B. Memory function of ChatGPT – think twice before you use it! Instagram. URL: https://www.instagram.com/p/DBGmoZ7sDbu/ [accessed 2025-05-29]
- Citron D, Chesney R. Deepfakes and the new disinformation war. Foreign Affairs. URL: https://www.foreignaffairs.com/articles/world/2018-12-11/deepfakes-and-new-disinformation-war [accessed 2025-05-29]
- EU AI Act: first regulation on artificial intelligence. European Parliament. URL: https://www.europarl.europa.eu/topics/en/article/20230601STO93804/eu-ai-act-first-regulation-on-artificial-intelligence [accessed 2025-05-29]
- Hughes BL, Zaki J. The neuroscience of motivated cognition. Trends Cogn Sci. Feb 2015;19(2):62-64. [CrossRef] [Medline]
- Shao C, Ciampaglia GL, Varol O, Yang K, Flammini A, Menczer F. The spread of low-credibility content by social bots. Nat Commun. Nov 20, 2018;9(1):4787. [FREE Full text] [CrossRef] [Medline]
- Stewart R, Madonsela A, Tshabalala N, Etale L, Theunissen N. The importance of social media users' responses in tackling digital COVID-19 misinformation in Africa. Digit Health. Mar 18, 2022;8:20552076221085070. [FREE Full text] [CrossRef] [Medline]
- Borges do Nascimento IJ, Beatriz Pizarro A, Almeida J, Azzopardi-Muscat N, André Gonçalves M, Björklund M, et al. Infodemics and health misinformation: a systematic review of reviews. Bull World Health Organ. Sep 01, 2022;100(9):544-561. [CrossRef]
- Vosoughi S, Roy D, Aral S. The spread of true and false news online. Science. Mar 09, 2018;359(6380):1146-1151. [CrossRef] [Medline]
- Purnat T, Dunn AG, Ishizumi A, Nguyen T, Machiri S, Briand S. A research platform for measuring information exposure, trust, and behaviour – burden of infodemics. Eur J Public Health. 2023;33(Supplement_2):ckad160. [CrossRef]
- Wilhelm E, Ballalai I, Belanger M, Benjamin P, Bertrand-Ferrandis C, Bezbaruah S, et al. Measuring the burden of infodemics: summary of the methods and results of the fifth WHO Infodemic Management Conference. JMIR Infodemiology. Feb 20, 2023;3:e44207. [FREE Full text] [CrossRef] [Medline]
- Aiko S. The ethics of reporting: balancing truth and sensationalism in global media. Glob Med J. 2024;69(1):22. [CrossRef]
- Singh K, Lima G, Cha M, Cha C, Kulshrestha J, Ahn Y, et al. Misinformation, believability, and vaccine acceptance over 40 countries: takeaways from the initial phase of the COVID-19 infodemic. PLoS One. Feb 9, 2022;17(2):e0263381. [FREE Full text] [CrossRef] [Medline]
- Gaysynsky A, Senft Everson N, Heley K, Chou WS. Perceptions of health misinformation on social media: cross-sectional survey study. JMIR Infodemiology. Apr 30, 2024;4:e51127. [FREE Full text] [CrossRef] [Medline]
- Roozenbeek J, Schneider CR, Dryhurst S, Kerr J, Freeman AL, Recchia G, et al. Susceptibility to misinformation about COVID-19 around the world. R Soc Open Sci. Oct 14, 2020;7(10):201199. [FREE Full text] [CrossRef] [Medline]
- Yıldırım S. The challenge of self-diagnosis on mental health through social media: a qualitative study. In: Battineni G, Mittal M, Chintalapudi N, editors. Computational Methods in Psychiatry. Cham, Switzerland. Springer; 2023:197-213.
- Farnood A, Johnston B, Mair FS. A mixed methods systematic review of the effects of patient online self-diagnosing in the 'smart-phone society' on the healthcare professional-patient relationship and medical authority. BMC Med Inform Decis Mak. Oct 06, 2020;20(1):253. [FREE Full text] [CrossRef] [Medline]
- Melki J, Tamim H, Hadid D, Makki M, El Amine J, Hitti E. Mitigating infodemics: the relationship between news exposure and trust and belief in COVID-19 fake news and social media spreading. PLoS One. Jun 4, 2021;16(6):e0252830. [FREE Full text] [CrossRef] [Medline]
- Liu Y, Yu K, Wu X, Qing L, Peng Y. Analysis and detection of health-related misinformation on Chinese social media. IEEE Access. 2019;7:154480-154489. [CrossRef]
- Walter N, Brooks JJ, Saucier CJ, Suresh S. Evaluating the impact of attempts to correct health misinformation on social media: a meta-analysis. Health Commun. Nov 2021;36(13):1776-1784. [CrossRef] [Medline]
- Pandey S, Prabhakaran S, Subba Reddy NV, Acharya D. Fake news detection from online media using machine learning classifiers. J Phys Conf Ser. Jan 01, 2022;2161(1):012027. [CrossRef]
- Sundelson AE, Jamison AM, Huhn N, Pasquino S, Sell TK. Fighting the infodemic: the 4 i Framework for Advancing Communication and Trust. BMC Public Health. Aug 30, 2023;23(1):1662. [FREE Full text] [CrossRef] [Medline]
- Usando inteligência artificial para detectar desinformação e conteúdo abusivo sobre COVID-19. Meta. URL: https://about.fb.com/br/news/2020/05/usando-inteligencia-artificial-para-detectar-desinformacao-e-conteudo-abusivo-sobre-covid-19/ [accessed 2025-05-29]
- SynthID. Google Deepmind. URL: https://deepmind.google/technologies/synthid/; [accessed 2025-05-29]
- Olawade DB, Wada OJ, David-Olawade AC, Kunonga E, Abaire O, Ling J. Using artificial intelligence to improve public health: a narrative review. Front Public Health. 2023;11:1196397. [FREE Full text] [CrossRef] [Medline]
- Chaix B, Guillemassé A, Nectoux P, Delamon G, Brouard B. Vik: a chatbot to support patients with chronic diseases. Health. 2020;12(07):804-810. [CrossRef]
- Jungwirth D, Haluza D. Artificial intelligence and public health: an exploratory study. Int J Environ Res Public Health. Mar 03, 2023;20(5):4541. [FREE Full text] [CrossRef] [Medline]
- Purnat TD. What are some ways that generative AI could be used for public health and infodemic management? Linkedin. URL: https://www.linkedin.com/pulse/what-some-ways-generative-ai-could-used-public-health-tina-d-purnat [accessed 2025-05-29]
- Moore J, Grabb D, Agnew W, Klyman K, Chancellor S, Ong DC, et al. Expressing stigma and inappropriate responses prevents LLMs from safely replacing mental health providers. arXiv. Preprint posted online on April 25, 2025. 2025. [FREE Full text] [CrossRef]
- Fisher S, Rosella LC. Priorities for successful use of artificial intelligence by public health organizations: a literature review. BMC Public Health. Nov 22, 2022;22(1):2146. [FREE Full text] [CrossRef] [Medline]
- Meet Florence, WHO’s digital health worker who can help you quit tobacco. World Health Organization. URL: https://www.who.int/europe/news/item/14-02-2021-meet-florence-who-s- [accessed 2025-05-29]
- S.A.R.A.H, a smart AI resource assistant for health. World Health Organization. URL: https://www.who.int/campaigns/s-a-r-a-h [accessed 2025-05-29]
- Stein N, Brooks K. A fully automated conversational artificial intelligence for weight loss: longitudinal observational study among overweight and obese adults. JMIR Diabetes. Nov 01, 2017;2(2):e28. [FREE Full text] [CrossRef] [Medline]
- Neuhauser L, Kreps GL, Morrison K, Athanasoulis M, Kirienko N, Van Brunt D. Using design science and artificial intelligence to improve health communication: ChronologyMD case example. Patient Educ Couns. Aug 2013;92(2):211-217. [CrossRef] [Medline]
- Grolleman J, van Dijk B, Nijholt A, van Emst A. Break the habit! Designing an e-Therapy intervention using a virtual coach in aid of smoking cessation. In: Proceedings of the 1st International Conference on Persuasive Technology for Human Well-Being. 2006. Presented at: PERSUASIVE '06; May 18-19, 2006; Eindhoven, The Netherlands. URL: https://link.springer.com/chapter/10.1007/11755494_19 [CrossRef]
- Soul machines joins forces with the world health organization to combat COVID-19 misinformation and help more than 1 billion tobacco users quit. Digital Media Innovations. URL: https://tinyurl.com/3fn7eu62 [accessed 2025-05-29]
- Johnson NF, Velásquez N, Restrepo NJ, Leahy R, Gabriel N, El Oud S, et al. The online competition between pro- and anti-vaccination views. Nature. Jun 13, 2020;582(7811):230-233. [CrossRef] [Medline]
- Suvvari TK, Kandi V. Artificial intelligence enhanced infectious disease surveillance - a call for global collaboration. New Microbes New Infect. Dec 2024;62:101494. [CrossRef]
- Generative AI for health in low and middle income countries. Stanford University. URL: https://cdh.stanford.edu/ai-low-middle-income-countries/generative-ai-health-low-middle-income-countries [accessed 2025-05-29]
- Kaufman RA, Haupt MR, Dow SP. Who's in the crowd matters: cognitive factors and beliefs predict misinformation assessment accuracy. Proc ACM Hum Comput Interact. Nov 11, 2022;6(CSCW2):1-18. [CrossRef]
- Nazi ZA, Peng W. Large language models in healthcare and medical domain: a review. Informatics. Aug 07, 2024;11(3):57. [CrossRef]
- Generating evidence for artificial intelligence based medical devices: a framework for training validation and evaluation. World Health Organization. URL: https://www.who.int/publications/i/item/9789240038462 [accessed 2025-05-29]
- Gutierrez G, Stephenson C, Eadie J, Asadpour K, Alavi N. Examining the role of AI technology in online mental healthcare: opportunities, challenges, and implications, a mixed-methods review. Front Psychiatry. May 7, 2024;15:1356773. [FREE Full text] [CrossRef] [Medline]
- Samy Helmy B, Samy Helmy A. Role of artificial intelligence in mental wellbeing: opportunities and challenges. J Artif Intell. Jan 15, 2022;15(1):1-8. [CrossRef]
- Fadhil A. A conversational interface to improve medication adherence: towards ai support in patient's treatment. arXiv. preprint posted online on March 3, 2018. 2025. [FREE Full text]
- Programa do SUS investirá de R$ 133,6 milhões em redes de telessaúde. Ministério da Saúde Brazil. URL: https://agenciagov.ebc.com.br/noticias/202404/proadi-sus-fortalece-estrategia-em-telessaude-com-investimento-de-r-133-6-milhoes; [accessed 2025-05-29]
- Hager P, Jungmann F, Holland R, Bhagat K, Hubrecht I, Knauer M, et al. Evaluation and mitigation of the limitations of large language models in clinical decision-making. Nat Med. Sep 04, 2024;30(9):2613-2622. [CrossRef] [Medline]
- Obermeyer Z, Powers B, Vogeli C, Mullainathan S. Dissecting racial bias in an algorithm used to manage the health of populations. Science. Oct 25, 2019;366(6464):447-453. [FREE Full text] [CrossRef] [Medline]
- Avaliação de Impacto Algorítmico para Proteção dos Direitos Fundamentais. LAPIN. URL: https://lapin.org.br/2023/04/13/avaliacao-de-impacto-algoritmico-para-protecao-dos-direitos-fundamentais/ [accessed 2025-05-29]
- Artificial intelligence in society. Organisation for Economic Co-operation and Development. URL: https://www.oecd.org/en/publications/artificial-intelligence-in-society_eedfee77-en/full-report/component-9.html#chapter-d1e7930 [accessed 2025-05-29]
- Harrigan F, Piffanelli S, Kay Topacio J. United Nations adopts ground-breaking pact for the future to transform global governance. United Nations. Oct 2, 2024. URL: https://press.un.org/en/2024/ga12641.doc.htm [accessed 2025-08-26]
- Mackey TK, Calac AJ, Chenna Keshava BS, Yracheta J, Tsosie KS, Fox K. Establishing a blockchain-enabled Indigenous data sovereignty framework for genomic data. Cell. Jul 21, 2022;185(15):2626-2631. [FREE Full text] [CrossRef] [Medline]
- Nutbeam D. Health literacy as a public health goal: a challenge for contemporary health education and communication strategies into the 21st century. Health Promot Int. 2000;15(3):267. [CrossRef]
- Friedman CP, Wong AK, Blumenthal D. Achieving a nationwide learning health system. Sci Transl Med. Nov 10, 2010;2(57):57cm29. [CrossRef] [Medline]
- Chou WS, Oh A, Klein WM. Addressing health-related misinformation on social media. JAMA. Dec 18, 2018;320(23):2417-2418. [CrossRef] [Medline]
- Esteva A, Robicquet A, Ramsundar B, Kuleshov V, DePristo M, Chou K, et al. A guide to deep learning in healthcare. Nat Med. Jan 7, 2019;25(1):24-29. [CrossRef] [Medline]
- Morley J, Machado CC, Burr C, Cowls J, Joshi I, Taddeo M, et al. The ethics of AI in health care: a mapping review. Soc Sci Med. Sep 2020;260:113172. [CrossRef] [Medline]
- Mainstreaming infodemic management in learning and teaching programmes: a report from a WHO technical consultation. World Health Organization. URL: https://iris.who.int/bitstream/handle/10665/376026/9789240081499-eng.pdf [accessed 2025-05-29]
- Naddaf M. AI linked to explosion of low-quality biomedical research papers. Nature. May 21, 2025;641(8065):1080-1081. [CrossRef] [Medline]
- Ibrahim H, Liu X, Rivera SC, Moher D, Chan A, Sydes MR, et al. Reporting guidelines for clinical trials of artificial intelligence interventions: the SPIRIT-AI and CONSORT-AI guidelines. Trials. Jan 06, 2021;22(1):11. [FREE Full text] [CrossRef] [Medline]
- WHO calls for safe and ethical AI for health. World Health Organization. URL: https://www.who.int/news/item/16-05-2023-who-calls-for-safe-and-ethical-ai-for-health [accessed 2025-05-29]
- PAI brings together a diverse community to address important questions about our future with AI. ParternshiponAI. URL: https://partnershiponai.org/ [accessed 2025-05-29]
- Oremus W. Zuckerberg expresses regrets over covid misinformation crackdown. The Washington Post. URL: https://www.washingtonpost.com/technology/2024/08/27/meta-zuckerberg-covid-misinformation-jordan-white-house/ [accessed 2025-05-29]
- Addressing hate speech on social media: contemporary challenges. United Nations Educational, Scientific and Cultural Organization. URL: https://unesdoc.unesco.org/ark:/48223/pf0000379177 [accessed 2025-05-29]
- Transparency reporting index. AccessNow. URL: https://www.accessnow.org/campaign/transparency-reporting-index/ [accessed 2025-05-29]
- Canela G, Claesson A, Pollack R. Addressing mis- and disinformation on social media. In: Purnat TD, Nguyen T, Briand S, editors. Managing Infodemics in the 21st Century: Addressing New Public Health Challenges in the Information Ecosystem. Cham, Switzerland. Springer; 2023:113-126.
- Ethics and governance of artificial intelligence for health. World Health Organization. URL: https://www.who.int/publications/i/item/9789240029200 [accessed 2025-05-29]
- Machiri S, Purnat T, Nguyen T. An ethics framework for social listening and infodemic management. Eur J Public Health. 2023;33(Supplement_2):24. [CrossRef]
- Hörder S. WHO adopts pandemic treaty without censorship provisions. ADF International. URL: https://adfinternational.org/news/who-adopts-pandemic-treaty-without-censorship-provisions [accessed 2025-05-29]
- Leung TI, de Azevedo Cardoso T, Mavragani A, Eysenbach G. Best practices for using AI tools as an author, peer reviewer, or editor. J Med Internet Res. Aug 31, 2023;25:e51584. [FREE Full text] [CrossRef] [Medline]
Abbreviations
AI: artificial intelligence |
CONSORT-AI: Consolidated Standards of Reporting Trials–Artificial Intelligence |
EU: European Union |
GenAI: generative artificial intelligence |
HIPAA: Health Insurance Portability and Accountability Act |
LLM: large language model |
SIRVD-DL: Susceptible, Infected, Recovered, Vaccinated, and Deceased–Deep Learning |
SPIRIT-AI: Standard Protocol Items: Recommendations for Interventional Trials–Artificial Intelligence |
WFPHA: World Federation of Public Health Associations |
WHO: World Health Organization |
Edited by A Mavragani; submitted 01.12.24; peer-reviewed by A Okoronkwo, D Chrimes, B Southwell; comments to author 11.05.25; revised version received 17.06.25; accepted 13.08.25; published 15.09.25.
Copyright©António Bandeira, Luis Henrique Gonçalves, Felix Holl, Juliet Ugbedeojo Shaibu, Mariana Laranjo Gonçalves, Ronan Payinda, Sagun Paudel, Alessandro Berionni, Young WFPHA, Tina D Purnat, Tim Mackey. Originally published in JMIR Infodemiology (https://infodemiology.jmir.org), 15.09.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Infodemiology, is properly cited. The complete bibliographic information, a link to the original publication on https://infodemiology.jmir.org/, as well as this copyright and license information must be included.