Science Communication as a Collective Intelligence Endeavor: A Manifesto and Examples for Implementation

Effective science communication is challenging when scientific messages are informed by a continually updating evidence base and must often compete against misinformation. We argue that we need a new program of science communication as collective intelligence—a collaborative approach, supported by technology. This would have four key advantages over the typical model where scientists communicate as individuals: scientific messages would be informed by (a) a wider base of aggregated knowledge, (b) contributions from a diverse scientific community, (c) participatory input from stakeholders, and (d) better responsiveness to ongoing changes in the state of knowledge.

Many of the pressing challenges that societies face today, from climate change to global pandemics, require decisions informed by the best available scientific evidence.Ideally, citizens should have access to good quality scientific knowledge that they can trust.However, citizens may have difficulties accessing scientific information and grasping the technical terms used.Some of the difficulty can be mitigated by a better style of science communication, for example, using clearer and jargon-free language (Hanel & Mehler, 2019;Martínez & Mammola, 2021), more intuitive presentation formats (Sirota & Juanchich, 2019), effective graphics (Harold et al., 2016), and narratives that resonate with people (Freling et al., 2020).Similarly, there is a case for supporting people's competencies to critically engage with information (Brodsky et al., 2021;Hertwig & Grüne-Yanoff, 2017).While these aspects are important, it is also essential to consider the content of these messages: what is the best evidence and who is involved in generating it.Scientific knowledge is continually updating, and new evidence now emerges rapidly, with gaps, uncertainties, and ambiguities in the data and its interpretation.A new program of science communication is needed that can address these complexities and derive clear messages that (a) reflect the best available evidence and (b) are delivered in a way that maintains public trust.
Currently, individual scientists are incentivized to rapidly disseminate their findings, often at the expense of quality control (Higginson & Munafò, 2016).This can harm the reliability of scientific messages as well as public trust in them.Furthermore, scientific messages compete in a contested and complex online landscape that favors partisanship over reasoned debate (Lorenz-Spreen et al., 2020).Especially where evidence conflicts with political or commercial interests, organized efforts to misinform, sow public confusion, or advance conspiracy theories have distorted public discourse (Koehler, 2016), threatened evidence-based policy making (Vériter et al., 2020), and personally targeted individual prominent scientists (Mann, 2015).In this commentary, we argue that to combat the challenges of today's information landscape, science communication must go beyond "one-person reporting" and harness the collective knowledge and expertise of many scientists worldwide to provide high quality information and engage with stakeholders.In short, we propose to approach science communication as a collective intelligence process.
In its broadest form, "collective intelligence" can be seen as a collaborative approach to problem-solving, typically supported by technological tools, which allows for real-time co-ordination and mobilization of knowledge that is distributed among many individuals (Suran et al., 2021).To some extent, the scientific process already embeds collective intelligence, as scientific knowledge is informed by reasoned argument between scientists, generating better outputs through peer evaluation and debate (Mercier, 2016).Here, we focus on harnessing the most advantageous characteristics of existing collective intelligence systems that would benefit science communication (see, e.g., online Supplementary Table ).We explain why and how these characteristics could be an effective way to address specific obstacles present in the traditional, "one-person reporting" model of science communication.

Aggregating Distributed Knowledge
Collective intelligence can help science communication by aggregating knowledge that is distributed among individual scientists.First, aggregating data and evidence can build a more complete picture of the current state of scientific inquiry, leading to more confidence in the reliability of a scientific proposition.For example, distributed networks of laboratories can aggregate samples for an experimental protocol, spreading the time and labor costs of data collection and evidence syntheses (Coles et al., 2022).Monitoring and aggregating evidence can also increasingly be done in real time with new Artificial Intelligence (AI) tools, for example, using machine learning to screen databases for relevant evidence.
Second, aggregating independent expert judgments can mitigate bias in evidence interpretation and enhance accurate assessment.Furthermore, communicating judgments that fairly represent those of a collective avoids the false balance that may be presented if an audience only hears from a few, unrepresentative experts (Koehler, 2016).Showing the distribution of judgments can highlight when there is a consensus or, when judgments differ, it can illustrate the uncertainties involved in interpreting the available evidence and experts' level of confidence in the state of knowledge.Critically, technologically supported aggregation methods allow experts to add their judgments independently, reducing the risk of biases that can be introduced through group processes.
Third, aggregating expert discourse, that is, discussion of the evidence, can showcase how reasoned argument between scientists informs scientific knowledge.This can be as critical as the evidence itself, especially in crisis situations where action must be taken as evidence emerges.New digital tools for judgment aggregation in the civic participation sphere provide comprehensive packages for debating, proposing and voting on initiatives and data (e.g., Po.lis, PSi, Loomio, Consul, Decidim).These could be leveraged for communicating scientific discourse.
There are of course costs to setting up aggregation systems.To aggregate data and evidence, protocols must be developed and shared with participating researchers.Evidence quality must also be assessed to avoid undermining the accumulated knowledge base with the inclusion of unreliable data (Royal Society, 2018).When aggregating judgments and discourse, the expertise of those who are contributing needs to be verified and contributors should be representative of their collective field of research, to avoid those with vested interests gaming the power of scientific consensus (Cook et al., 2018).
Despite the costs, aggregation is highly beneficial.Communicating in terms of the "collective accumulated evidence" shifts the message toward what the best available evidence indicates.This can help resist arguments that science has not "proved" an effect (Oreskes & Conway, 2010).It is also harder for those interested in discrediting science to carry out ad hominem attacks on collective evidence from a group of scientists (Mann, 2015).Furthermore, accumulated evidence can make a scientific consensus more visible, which is important because well-communicated scientific consensus has influenced decision-making, shifted the public's attitudes, and strengthened calls for policy action across various domains (e.g., Bartoš et al., 2022;Budescu & Chen, 2014;Kerr & van der Linden, 2022), even for partisan individuals or those who tend to be predisposed toward rejecting scientific evidence (Lewandowsky et al., 2012).In areas where consensus has yet to form, aggregation can advance science by exposing areas in which further evidence is needed (Minas & Jorm, 2010).

Involving a More Diverse Group of Individuals
To optimize the quality of aggregated evidence and a scientific consensus, collective intelligence should increase the diversity of contributions.First, diversity in ideas (e.g., epistemic diversity) tends to invite greater scrutiny, increasing the robustness of scientific inquiry (Pesonen, 2022).Involving more diverse perspectives may help scientists challenge cognitive biases when seeking or interpreting evidence.Second, diversity in representation can boost the reach and effectiveness of science communication, especially when it comes to producing messages that the public trusts.Historically, a lack of diversity in science and research has perpetuated inequalities and contributed to the marginalization of voices from groups, such as women, minority groups, and citizens of countries in the Global South (Mertkan et al., 2017).This can undermine trust in science, especially among communities that experienced discrimination in the past (Woolf et al., 2021).
Diversity needs to be deliberately engineered because biases can easily be overlooked when values and norms are embedded into contemporary society.It is necessary to review processes, such as consensus-building, information gatekeeping, and sensemaking, and establish transparent frameworks to incorporate diversity in these processes (Thapar-Björkert & Farahani, 2019).For example, frameworks for inclusion can specify how experts will be invited or selected to contribute (e.g., by issuing invitations to all identified experts in the domain, regardless of their opinions on an issue).Although frameworks do not guarantee diversity, they make the lack of diverse representation more noticeable.A transparent framework for inclusion that discloses who the experts are and why they were chosen can also help verify expertise and avoid a "manufactured" collective scientific position from nonexperts (e.g., Cook et al., 2018).
Designing for diversity in the scientific collective also requires constructive spaces for deliberation, critique, and debate-discourse that is essential to knowledge-building-which support diverse participation.These spaces should be built around critiquing ideas rather than individuals, with recognized codes of conduct for respectful engagement.They should encourage scholars with opposing perspectives to collaborate rather than compete.Although there is no existing platform yet that promotes such behavior in online academic discourse, some researchers are considering how older methods to elicit, aggregate and discuss expert opinions could be harnessed as a model for shaping scientific discourse among diverse experts.Tools to scale up such processes could soon provide online infrastructure to visualize and convey the inputs to and outcomes of the consensus.

Increasing Public Participation
By definition, collective intelligence is participatory, leveraging the involvement of many individuals to produce outputs.Thus far, we have discussed the participatory input of experts in generating scientific knowledge that underpins science messages.However, science communication should also be informed by the people it will impact (Priest, 2018).Participatory input from citizens can help shape research to address the needs of those affected by it (Bruin & Bostrom, 2013).It can also generate interest and understanding from the public in how the research is conducted and evaluated (Bonney et al., 2015), thereby building trust in scientific messages (Bedessem et al., 2021).Increasingly, technological interfaces allow the public to participate in many ways.Participation can be active, for example, by acting as "citizen scientists" (Silvertown, 2009) or a mass monitoring system.The public can also passively inform scientists through their collective online discourse: such "social listening" has enabled science communicators to tackle misinformation outbreaks by targeting information provision to the public's needs (World Health Organization, 2021).
The accessibility of scientific findings is a precondition to harvest some of the benefits of public participation, such as a more knowledgeable citizenry.Accessibility can mean making research available.Researchers are increasingly doing so through "pre-prints," that is, draft-level papers submitted to a publicly accessible server.In theory, this gives the public early sight of findings, but pre-prints can be confused for scientific fact or weaponized to support a certain narrative (Bajak & Howeve, 2020).Hence, they should only be considered as emerging evidence in an aggregated system, and this needs to be clearly indicated on the pre-print platforms and papers.Accessibility also means making research comprehensible.Openly published articles (pre-prints or otherwise) often remain inaccessible to the public because of their technical language and general level of complexity, limiting informed discussion of these to scientists and small parts of the public (e.g., science journalists, think-tanks, and policymakers).Increasing accessibility could involve writing plain language summaries of papers (Stoll et al., 2022).It could involve supporting citizens' skills to engage with information, identify good quality evidence, and spot misleading argumentation (Brodsky et al., 2021;Roozenbeek et al., 2022).Scientific publications could even be augmented with technological tools that indicate how findings correspond to the broader literature or how samples should be structured for this kind of research.Accessibility could also be enhanced with collective projects to communicate the state of the evidence in comprehensible language.Ultimately, scientists have a duty to make research available and comprehensible to the public that provides them with funding and academic freedom (Greenwood & Riordan, 2001).

Improving Responsiveness
It can be difficult to identify relevant evidence and judge its quality at a given point in time when it can emerge rapidly, especially during a crisis situation where scientists may accelerate research production and dissemination (Fraser et al., 2021).Collective intelligence can leverage technology to enable real-time information monitoring, thereby enhancing the responsiveness of science communication to updates and changes.Traditional evidence syntheses are lengthy processes that often exclude the most recent studies that were not published by the time the research was conducted.In contrast, AI can enable a dynamic evidence synthesis, with some promising examples already emerging across different domains.In such systems, after having established the criteria for subsequent studies to be included, researchers can regularly monitor new publications and update their syntheses in real time.
Collective intelligence could also increase the responsiveness of evaluating new information.Emerging scientific papers typically undergo independent critique, or "peer review," but this process is notoriously slow.During the Covid-19 pandemic, researchers collectively responded by accelerating some peer review processes and, more commonly, openly sharing early-stage research as pre-prints.Not all rapid publication was helpful to the pandemic response, but some did provide valuable updates to inform decision-making (Fraser et al., 2021).Identifying and accelerating the review of better quality pre-prints could thus improve the responsiveness of science in times of crisis.A collective intelligence system could organize and support scientific evaluation of pre-prints, for example, by identifying potential reviewers through network analysis (Rodriguez & Bollen, 2008), or detecting information manipulation and erroneous statistical analyses (Henman, 2020).
However, AI cannot fully replace the human contributions needed for quality assurance.AI-supported tools to facilitate quicker pre-print or postpublication review by the scientific collective exist, but sustaining motivation to contribute collectively to this work over the longer term is difficult.This may in part be due to a lack of incentives.For example, academics often cite lack of time as the main reason for declining reviews, but it takes much less time to review a manuscript (typically hours) than to produce a new piece of research (typically months).Despite the critical contribution of peer review to the scientific process, it is not incentivised in the publication structure, nor by most employers.The same goes for maintaining contributions to consensus-building and communicating consensus.The recent Covid-19 crisis provided a glimpse of how a motivated scientific collective could produce, evaluate, and communicate research in a highly responsive fashion.However, this effort has been hard to sustain 2 years later.Harnessing the ability of collective intelligence in responding to crises and fast-paced research thus needs an overall structural change within the scientific community to better reward collective knowledge processes over individual efforts.

Implementing Collective Intelligence in Science Communication: An Example
This commentary is itself a product of our experience harnessing collective intelligence processes to create a "Manifesto for Science Communication as Collective Intelligence."We used group discussions and interactive online discourse via the tool pol.is to gather insights from attendees at an open virtual workshop on the topic.We then invited everyone to craft the manifesto, either as co-ordinating lead authors ("CLAs," n = 6) or contributing authors (n = 18).CLAs collectively voted on how to organize the points raised at the workshop.Each CLA then led a group of authors to draft a section of the manifesto.The CLAs condensed this draft into its key propositions and, using pol.is,all authors voted on which propositions from the draft were critical for the manifesto.Propositions with > 60% of votes were organized into the final Manifesto, which presented eight necessary features for science communication as collective intelligence.Altogether, we engaged a diverse group of researchers, captured and aggregated their judgments and discourse in an iterative fashion, and generated a consensus for communication.The full process is shared online as part of the Manifesto (https://scibeh.org/manifesto).

Conclusion
In this commentary, we highlighted the impetus for science communication to move away from a model where scientists disseminate individual findings and adopt a collective communication program that (a) develops messages from a wider base of aggregated evidence, judgments, and discourse, (b) is informed by a diverse community, (c) involves participation from stakeholders, and (d) is responsive to ongoing changes in the state of knowledge.In the online supplementary table, we provide examples that concretize how this new program would leverage collective processes, supported by participatory technology, in pursuit of a more collaborative form of science communication.While no single example (including our own) managed to harness all the advantages we describe in this commentary, they provide a glimpse of how collective processes are already enhancing the way in which scientists gather data, reach consensus, and communicate it.We hope that in the near future, more tools and examples will emerge to support a program of science communication as collective intelligence.
Angelo Fasce is a philosopher and postdoctoral researcher in the Faculty of Medicine of the University of Coimbra.He studies the cognitive basis of anti-scientific beliefs (e.g., alternative medicine and psychology, climate change denial, conspiracy theories, and vaccine hesitancy) and authoritarianism from a theoretical and empirical perspective.
Katy Tapper is a reader (associate professor) in psychology at City, University of London.She is interested in how we change health-related behaviors, particularly those linked to healthy eating and weight management.Her experimental research is aimed at identifying and understanding variables that influence behavior.This helps inform her applied work that focuses on the development and evaluation of health interventions for both adults and children.
Miso Demko worked at the Dynamic Decision-Making Laboratory at Carnegie Mellon University as social media manager and research communicator.In this role, he explored the metaphor of the lab as a place of "knowledge-production," by connecting the elements of theory, experimental procedure, and scientific discourse forums and innovating the standard research paper format to engage different communities (audiences) with the work of the lab.
Stephan Lewandowsky is a cognitive scientist at the University of Bristol.His research examines people's memory, decision-making, and knowledge structures, with a particular emphasis on how people update their memories if information they consider to be true turns out to be false.His research currently focuses on the persistence of misinformation and spread of "fake news" in society, including conspiracy theories.He is particularly interested in the variables that determine whether or not people accept scientific evidence, for example, surrounding vaccinations or climate science.Sameer Alladin is a doctoral student in the School of Psychological Science at the University of Bristol.He studies the interplay between gastric and neural physiology in disgust, using machine learning to align electrophysiological recordings from stomach and brain, and tests how these are impacted by pharmacological manipulations.He is also more broadly interested in decision-making, and music perception.

Ulrike
T. Sonia Boender is a health scientist and field epidemiologist at the Risk Communications group of the Robert Koch Institute, Germany's national public health institute.Her applied research focuses on managing the infodemic, bridging epidemiology, and communication.
Hendrik Bruns is a policy analyst at the European Commission Joint Research Center, Brussels.He works on aspects related to the environment, climate change, and misinformation from a behavioral science perspective.His work focuses on generating behavioral insights to support policymaking.
Helen Fischer is a postdoctoral researcher at the Leibniz Institut für Wissensmedien, Tübingen, Germany.Her research focuses on the metacognitive processes involved in forming beliefs and spreading information about politicized science, such as climate change.
Christian Gilde is a professor of business at the University of Montana Western.He conducts research in the areas of management and consumer studies.His research translated (and translates) into numerous journal articles and books on business and consumer science.His current projects center around decision-making, collective intelligence, and consumer sociology.
Paul H. P. Hanel is a lecturer (assistant professor) in psychology at the University of Essex, UK.A significant part of his empirical work includes human values (e.g., freedom, loyalty, security).Currently, he is especially interested in similarities between groups of people (e.g., women and men) as well as the effects of working remotely on well-being and productivity.
Stefan M. Herzog is a senior research scientist in the Adaptive Rationality Research Center at the Max Planck Institute for Human Development, Berlin.He studies how to boost human judgment and decision-making by understanding human and machine behavior and how humans themselves understand machines.He combines insights and methods from cognitive science, collective intelligence ("wisdom of crowds"), heuristics, and algorithms.He also works on applications in digital environments, medical decision-making, and meteorology.He co-leads an initiative on reconfiguring behavioral science for crisis knowledge management in response to COVID-19 and other, future disruptive events.
Astrid Kause is a psychologist and holds a junior professorship for Sustainability Science and Psychology at the Leuphana University Lüneburg.She studies perceptions and communications of risks and uncertainty related to climate and the environment, climate policies and health.Her aim is to provide empirical evidence for transparent communications as well as behavioral interventions contributing to a rapid societal transition toward sustainability.
Hahn holds a chair in Computational Modeling in the Department of Psychological Sciences, Birkbeck College, University of London, where she directs the Center for Cognition Computation and Modeling.Christoph M. Abels is a post-doctoral fellow at the Faculty of Human Sciences at the University of Potsdam.His research examines individuals' susceptibility and resilience to misinformation, and how this evidence can inform policy making.Ahmed Al-Rawi is an associate professor of News, Social Media, and Public Communication at the School of Communication at Simon Fraser University, Canada.He is the director of the Disinformation Project that empirically examines fake news discourses in Canada on social media and news media.His research expertise is related to social media, news, and global communication with emphasis on Canada and the Middle East.