Abstract
What should citizens understand about science to participate in democratic life? Against the prevailing approach, we argue that “what” a public understanding of science is about strongly depends on the specific epistemological nature of the science related issues considered in different contexts and circumstances. We identify three specific categories of such issues and show how, equally, specific models of public understanding are required to address them. Only by endorsing such an alternative approach will citizens arguably be able to form sound opinions about those very issues, as well as to discuss and deliberate rationally about them.
1. Introduction
In 2001, the American Association for the Advancement of Science published the Atlas of Science Literacy, a large volume that, using fifty linked maps, describes how students throughout 12th grade develop their understanding and skills to meet specific science-literacy goals. The Atlas was set up in the framework of an education reform initiative, the AAAS’s “Project 2061.” This program aimed to “promote literacy in science, mathematics, and technology to help people live interesting, responsible, and productive lives” (American Association for the Advancement of Science 1993, p. XI).
The expression “science literacy,” first popularized by Paul Hurd (1958), has been widely used over the last few decades (including by AAAS) to support the idea that, given the growing importance of scientific knowledge and expertise in our everyday lives, some elements of contemporary science should be mastered by ordinary citizens (e.g., Feinstein 2011; Slater et al. 2019; Huxster et al. 2018). This need for “mastering” some elements of the sciences can be justified at different levels (Shen 1975): practical science literacy helps people to make individual decisions in their everyday life, for instance about health or work (Shen 1975, p. 46); cultural science literacy makes people appreciate science as a great human achievement (1975, p. 49); and civic science literacy allows people to reach considered decisions about science related debates and issues (1975, p. 48). It is the latter that has recently taken center stage in the literature on scientific literacy, thanks also to a steady increase of social interest in science and technology debates (Miller 2004).
In the frame of civic science literacy one can legitimately ask, following, for instance, Arnon Keren: “what must laypersons understand about science to allow them to make sound decisions on science related issues?” (2018, p. 781). “Making sound decisions” in Keren’s context concerns, more than what laypeople want to do with scientific information, what laypeople come to believe when confronted by scientific information. He explicitly claims that good public understanding of science should clearly focus on the ability of laypeople “to be able to determine which scientific claims to accept” (Keren 2018, p. 788) or “which scientific claims to believe” (p. 799). In other words, the problem underlying civic science literacy as posed by Keren (among others, including Shen 1975) can be reframed in the following terms: what must laypersons understand about science to allow them to formulate sound opinions on science related issues?
Of course, figuring out what laypeople are to understand while addressing science related questions is crucial to finding an answer to the problem as just reframed.1 Indeed, this has been a concerning question for many of those engaged with public understanding of science over the last few decades. Phillips et al. (2018), for example, supported by the National Science Foundation and drawing from the Framework for Evaluating Impacts of Informal Science Education Projects (Friedman 2008), remark that understanding “science” can be applied to at least one of the following items: science content (with reference to “subject matter, i.e., facts or concepts”), science processes (“the methodologies that scientists use to conduct research,” for example the hypothetico-deductive method), and the nature of science (“the epistemological underpinnings of scientific knowledge and how it is generated”) (Phillips et al. 2018, p. 9).
Keren adds a further item of understanding. Citizens, he claims, are not required to use first-hand scientific evidence to form their beliefs (this is the type of understanding that professional scientists engage with), but to acquire a second order understanding based on “information about patterns of agreement and disagreement among purported experts and authorities” (Keren 2018, p. 785). In this sense Keren refers to a “division of scientific labour” between scientists and citizens. Other authors insist on the need for citizens to understand the “social structure of science,” or “science as a social enterprise” (Slater et al. 2019, p. 257), namely, to grasp in what sense and for what reasons scientific activity brings scientists to be at the same time competitive and cooperative with one another (Slater et al. 2019, p. 256).
Despite the multiplicity of these different approaches, all these normative models of public understanding of science share a central objective: the formulation of a unique, general answer to the civic science literacy problem, independently of the type of science issues that are at stake in the different domains and circumstances. Not much philosophical attention has been given to the epistemic diversity of the “science related issues” addressed in the literature on public understanding of science, and on how this very diversity might significantly affect the types of understanding of science citizens should acquire and pursue.
In the literature engaged with civic science literacy, issues as diverse as climate change, vaccines, evolution biology, and GMOs are generally considered equivalent instances of “science related issues.” For example, Miller (2010) puts a rather heterogeneous range of scientific issues on the same level within the “public policy agenda”: global climate change, the use of embryonic stem cells, the future of energy research, nuclear power, viral pandemics, genetically modified food (2010, p. 241). It appears to be the same in the case of Gerken’s study on the role of journalists in communicating scientific results: no distinction is made among debates on creationism vs. evolutionary biology, or on risks associated with GMO crops, vaccine-autism links, gun control (Gerken 2019, p. 5). Duncan, Chinn and Barzilai, in questioning “what students should understand about how experts work with evidence” and “how laypeople can use evidence reports themselves” (Duncan et al. 2018, p. 907), refer to “current social and political controversies about scientific claims regarding climate change, vaccination and evolution” as part of a same ensemble (2018, p. 930).
In this paper we argue that the different approaches to public understanding of science in current literature make sense, and can be best appreciated, if we start with making appropriate distinctions among the specific types of science related issues which they explicitly or tacitly address. In order to test our claim, we first articulate a tentative typology of “science related issues,” and then elucidate for each type which model of understanding appears more suitable. Our aim is to clarify what is at stake in some of the debates about what public understanding of science ought to be about, and possibly to allow such understanding to become a useful practical tool in the hands of citizens.
Our argument is organized as follows. In section 2 taking a cue from the civic science literacy problem as formulated by Keren, we will focus on some of the issues raised by this formulation, which we believe requires both specific attention and some qualification. In section 3 we put forward a three-dimensional typology of science related issues, which includes (i) an epistemological characterization of types of issues illustrated by concrete examples, and (ii) an indication of what model of public understanding seems most suitable to deal with each type. We finally conclude (section 4) by reflecting on the significance of promoting public understanding of science in the light of the diversity of the science related issues which populate public debate.
2. The Civic Science Literacy Problem: Some Preliminary Clarifications
The civic science literacy problem in Keren’s slightly reformulated form suggested in the previous section (i.e., what must laypersons understand about science to allow them to make sound opinions on science related issues) raises at least three questions in need of clarification: a) what counts as a sound opinion; b) what science related issues are targeted in the formulation above; and c) why citizens ought to acquire understanding of these issues.
An answer to the last question generally takes the form of a democratic ideal: one of the basic requirements of a democratic society is that citizens should “actively participate in the democratic process that weighs in on such [science related] issues” (Slater et al. 2019, p. 252). For the purpose of our discussion, we define “active participation” in the following, minimal, terms: citizens actively participate, or may be actively participative, not only when they form an opinion, but when they are able to defend it in a well-argued discourse that could be shared with, and gain support from, other citizens. A well-argued discourse is what specifically underwrites a sound opinion. In response to question a, an opinion is then sound in the sense that the arguments which are formulated in its support can be shared and discussed on rational bases. This entails that soundness depends on whether opinions are built, and then discussed, refined, or rejected by using a shared basic set of truth-related criteria. We will later show how these criteria are significantly dependent on the types of “science related issues” which are addressed in different discussions. This will constitute the core of our argument.
As to question b, it must first be emphasized that in the literature on science literacy or public understanding of science generally the science related issues referred to belong to a well identified category, namely they are issues debated in the public sphere because of their social, economic, or policy related importance. In other words, the “science related issues” focused on are those that can be labeled as politically relevant science related issues. Obviously, what counts as “politically relevant” might well change with time: for instance, at present, quantum mechanics does not immediately qualify as one of these issues, but it could become so in the future (let us think about the new developments of quantum informatics, for instance).
In the light of our answers to questions c and b we then suggest reframing the civic science literacy problem in the following way:
what ought laypersons to understand about science to allow them to actively participate in democratic life, that is to form and defend well-argued opinions about politically relevant science related issues?
Before we undertake our analysis, three clarifications are in order. First, the weight of our argument does not lay on the typology itself, but on how it shows that the nature of the issue at stake may influence the kind of understanding citizens need to actively participate in democratic life (in the sense suggested above). The classes of our typology, three among a much wider range of possibilities, were chosen having this objective in mind. Equally, these classes do not constitute, and should not be taken as, an exhaustive description of the epistemic diversity of the existing science related politically relevant issues.
Our second objective is to identify, for each class, a minimal requisite for public understanding of science, namely the kind of understanding which seems to be both necessary and sufficient for citizens to form sound opinions. Indeed, the fact that some kind of understanding is needed for a given class does not mean that it could not also be interesting or relevant (without being necessary) for other classes. Conversely, some kinds of understanding can be interesting or relevant for a given class, without being sufficient.
Third, the typology is organized around specific issues, and not around specific research domains. For instance, when considering the case of climate science, the science related issue we consider as characteristic of our first category is the anthropogenic origin of current global warming, not climate science as a research domain. It is this attribution of specificity that informs our argument.
3. A Working Typology of Science-Related Issues
We build our typology on three main classes of issues. A first class includes issues which are characterized by large consensus within the scientific community, for example, the benefits of vaccines or the reality of human-driven climate change. A second class refers to issues where intra-disciplinary disagreement exists among conflicting results, data, or approaches. Most often these disagreements reflect, and are fueled by, divergences of values. Suitable examples of these issues can be found in the field of chemical risks to the public and their management. A third class refers to issues that entail yet another type of disagreement, namely inter-disciplinary disagreement. In this case disagreements are due to clashes among different scientific disciplines or domains of inquiry dealing with the same issue or core of issues. For instance, the choice to promote or reject GMOs crops raises ecological, health, economic and social problems that often conflict with each other. The last two classes both deal with so-called “deep” disagreements, that is with disagreements which are epistemically grounded (in a sense that will be qualified below).
In the three following subsections we will describe each class more in detail and point out which among standard models of public understanding of science best addresses each class. At the same time we will highlight how none of them individually can effectively be used to address the typology in its totality. In fact, as pointed out earlier, none of these standard answers appear to take into specific account the effects that a variety of types of science related issues might produce on the possible answers themselves. We must not reason in terms of a “one size fits all” answer, but rather in terms of different answers for different types.
3.1. Type-One Issues: Wide Scientific Consensus
A first set of politically relevant science related issues is characterized by the existence of a large consensus among the scientific or expert community. Take for instance the case of human driven climate change: it is here acknowledged that “there is a broad expert consensus” about the anthropogenic drive on climate changes, and “most of the challenges to this claim come from interested parties outside the scientific community” (Oreskes 2018, p. 31). This does not mean that there is no disagreement about several aspects of climate change (e.g., how much and how far human impact accounts for such changes, or how to model/simulate/predict future changes), nor that all relevant aspects of climate, past and present, are thoroughly understood. However, on the basis of what the scientific community does know today—a combination of observational data, theoretical analysis, and computer simulation—we can rather confidently believe in (i.e., there are good scientific reasons to support the belief in) anthropogenic climate change. Not only some of the major international science institutions concord on this.2 If we also take a look at the scientific literature, there are almost no articles that question it. For instance, a survey made by Cook et al. of fourteen papers analyzing scientific consensus on human-driven climate change points out that consensus “is robust, with a range of 90%–100% depending on the exact question, timing and sampling methodology” (Cook et al. 2016, p. 11).
A similar story can be told about vaccine safety. On the basis of what is known today, the scientific reasons and evidence that support vaccine effectiveness is concurred upon. This does not mean that there are no risks associated with vaccine use, or recorded instances of vaccination mishandling. Yet, from such risks and cases no inference can be drawn to claiming fatal uncertainty for the science behind vaccines among the overwhelming majority of the scientific community (Kampourakis and McCain 2020, pp. 99–106). The degree of confidence in the use of vaccines is emphasized, for instance, by van der Linden (2016). He notes that “surveys of physicians and medical scientists have repeatedly indicated that over 90% of doctors agree that adults and children should receive all recommended vaccines. In other words, there is a strong medical consensus about vaccine safety that many patients may not be aware of” (Sturgis et al. 2021, p. 119).
These two examples (the anthropogenic origin of global warming and vaccine safety) are widely used as case studies within the literature on public understanding and communication of science. For instance, a large number of studies question the influence of communicating the mechanisms of anthropogenic global warming in changing people’s beliefs and attitudes (Ranney and Clark 2016; Guy et al. 2014; Bedford 2016; Johnson 2017). In the case of vaccines, it is often asked what kind of public understanding of science is needed for people to dismiss false results (notably, Wakefield’s 1998 retracted study about the link between vaccines and autism). The answers to these questions (e.g., Clarke et al. 2015; Gerken 2019; Duncan et al. 2018; Slater et al. 2019) are usually formulated in such a way as to include how to handle the rejection of scientific consensus. As described by Rutjens et al. (2018), a rejection attitude is determined by a mix of ideological, psychological, political, and educational features. It is not always clear how these elements interact with each other. For instance, a better level of education and science literacy has been reported to be a factor of polarization of opinions about climate change risks (Kahan et al. 2012). Clearly, understanding the science is only one dimension among many to consider when describing the distribution of opinions regarding these otherwise highly consensual issues among the scientific community. But if that is the case, what is the necessary and sufficient type of public understanding of science applied to consensus laden issues, able to respond effectively to the challenge posed by the civic science literacy problem as formulated above?
By looking at the available literature we suggest that Keren’s “scientific division of labour” (SDoL) model might provide the most appropriate answer. Keren contrasts SDoL with what he calls the “science content” model, taken to be the “dominant approach to the public understanding of science.” By science content Keren refers to “scientific concepts, theories, facts, and methods” belonging to science (Keren 2018, p. 782) By acknowledging SDoL instead, the public is not required to gain first-hand scientific evidence, and master scientific terminology, in order to form their beliefs. It rather needs to decide what and who to put trust on. Professional experts generate scientific knowledge through internal mechanisms of criticisms, debates and controversies, and scientifically literate people will require information and skills that allow them to assess “patterns of agreement and disagreement among purported experts and authorities” (Keren 2018, p. 785).
The SDoL model seems to be well-suited to solve the civic science literacy problem for issues that benefit from large scientific consensus. In these cases, being able to identify patterns of agreement indeed provide citizens with a rational tool which allows them to participate in democratic debate (that is, to defend one’s position in a well-argued discourse that could gain support from other citizens). As Keren himself acknowledges, “certain patterns of consensus among experts provide strong reasons to trust those experts.” But what happens for issues where such a consensus pattern is more difficult to identify? There are indeed cases where “scholars in a field fail to reach any substantial agreement on almost anything.” For such cases Keren points out that lack of consensus among experts “might be a reason not to trust them on such contested questions” (2018, p. 785).
Regarding such cases, advocates of SDoL do not offer solutions to the civic science literacy problem. Arguably, they could possibly be seen as subscribing to the view that for those kinds of politically relevant controversial issues the rational thing to do would be to (at least temporarily) withhold one’s judgement. However, withdrawing from forming a sound opinion is not an ideal position to be in vis-à-vis active participation in democratic life (in the sense given in section 2). This is made even worse by the fact that in practice the absence of clear scientific consensus is not a rare occurrence, and cases where scientific results are uncertain, and expertise not clearly identified, do matter when discussing the civic science literacy problems. In the next two sections we will focus on two kinds of politically relevant science related issues which are characterized by experts’ disagreement and yet, we will argue, ought not to end up necessarily in a position of suspended judgement on the part of laypeople.
3.2. Type-Two Issues: Deep Disagreements within Single Scientific Disciplines
A second ensemble of politically relevant science related issues is characterized by the existence of deep disagreement among scientists belonging to the same discipline. We borrow the concept of deep disagreement from Biddle (2018). He takes it from Lynch (2010) who defines it as a disagreement over “which methods are most reliable in a given domain.” To this characterization Biddle adds that often these types of disagreement don’t just concern the methods but also “what kinds of evidence are relevant to a given hypothesis” (p. 376). And what counts as evidence only partly emerges from disagreement over methods. It also originates from disagreement among other issues—for example, as we will see below, among values.3
Chemical risk evaluation offers a suitable area of discussion for deep disagreement. Let us consider as a first example the effects of the use of neonicotinoids on bees as discussed in Douglas (2017). Neonicotinoids are a class of pesticides widely used in agriculture to protect crops from insects (neonicotinoids act on insects’ central nervous system). These molecules are however suspected of being a cause of the decline in bees’ population. Because of widespread use over the last two decades, the issue of neonicotinoid toxicity has become a highly sensitive scientific issue for national and international institutions, (e.g., EFSA 2013 and DEFRA 2013 official reports), and the subject of a lively ongoing debate among environmental scientists specifically interested in negative effects on bees. The main aspect of controversy arises from how to interpret the difference in results between laboratory and field experiments. Studies made in laboratory show that sub-lethal concentrations of neonicotinoids negatively influence bee colony growth, with a significant reduction of bees’ ability to orient themselves when returning to the hive (Henry et al. 2012; Whitehorn et al. 2012; Gill et al. 2012). These results clash with evidence coming from various field studies which, by looking at not-controlled conditions, do not detect such an effect (Chauzat et al. 2009; Genersch et al. 2010; Sterk et al. 2016). One reason for such difference in results has been linked to the fact that concentrations of neonicotinoids under field conditions are much lower than those used in laboratory experiments (Blacquiere et al. 2012; Cresswell et al. 2012). However, more recent field studies suggest that neonicotinoids do have a measurable effect on bees’ performance also in real-world agricultural landscapes (Woodcock et al. 2017; Budge et al. 2015), but the extent of such an effect seems to vary depending on spatial location and bee species (Wintermantel et al. 2018). We are here typically in a situation of deep disagreement (in the sense qualified by Biddle): there is a debate on the methods to be used within a given scientific domain (lab vs. field research), and on what should count as reliable evidence.
A second example of chemical risk evaluation has to do with another group of chemicals known as endocrine disruptors (ED). EDs form a class of chemicals which are present in an increasingly large number of products (industrial chemicals, pesticides, cosmetics, pharmaceuticals). These chemicals are suspected to alter the functions of the hormonal system, with possible effects on human reproduction (Godfray et al. 2019). However, the scientific community is divided regarding both the reality of these effects on health, and the best regulatory approach to adopt (McIlroy-Young et al. 2021). More precisely, the core of the debate is how to rely on classical chemical risk assessment methods, which involve determining the probability of adverse effects for human health emerging from real world exposure (Lofstedt 2011). Some scientists claim that as long as exposure is below a certain threshold, EDs can be used safely (Lamb et al. 2014; Dietrich et al. 2013; Brescia 2020). Against this view, others argue that it is very difficult, and highly controversial, to determine in a reliable way where to set the bar for making an exposure threshold “safe” regarding EDs. In fact, they claim that classical risk assessment increases the chance of allowing the use of chemicals that are potentially harmful (Bergman et al. 2015; Gore et al. 2013), and therefore recommend the adoption of a precautionary principle in view of preserving safety. Again, this scientific debate raises deep disagreement about the reliability of risk assessment methods, as is often the case with the evaluation of chemical risks, of which neonicotinoids or EDs are only two instances. First, scientists must make methodological choices, e.g., regarding the level of statistical significance, the sample size, the time-lapse of the experiments, or the measured outputs. Second, assessing the toxicity of a chemical is a matter of balancing inductive risks, that is, the risks of making a mistake (Elliott and Richards 2017). False positives (judging that the chemical is dangerous whereas it is not) or false negatives (judging the chemical is safe whereas it is not) may have different kinds of consequences (for people’s health, the environment, life safety, economic growth, etc.).
As has been widely acknowledged, in the case of both methodological choice and inductive risk balance, non-epistemic values influence experts’ judgement and related policy decisions (Longino 1990; Dupré 2007; Douglas 2009; Elliott 2011; Biddle 2016). Scientists need, for example, to decide what uncertainties to emphasize in the collected evidence and what to leave aside; what methods to use in handling what they know and what they do not know; what data to select and how to interpret them; what constitutes sufficient warrant in particular cases; how to evaluate the risk, or the impact, of error outside laboratory research; etc. (see Montuschi 2017a discussing Douglas 2000 on scientists’ “judgement calls”).4 Decisions of this sort prove controversial for the scientists making them (and different scientists might “call upon” different decisions) and for the different choices that can be made at different junctures. Decisions and choices are not just dictated by facts. They often entail a wide range of evaluations—with ethical and social values playing a crucial role. Are the economic consequences on the market more or less important than human health safety in allowing the use of ED chemicals, and how do the two issues compare? What values should we appeal to in assessing the potential consequences of either under-estimating or over-estimating the risks associated with such use? As Biddle puts it, “there is always some possibility of being wrong (i.e., of accepting a false hypothesis or rejecting a true hypothesis) and being wrong brings different consequences (including moral consequences) for different parties” (2018, p. 363).
In the light of what is involved with this second class of science related issues, what kind of public understanding of science is then needed to allow citizens to participate actively (in the sense defined in section 2 above) in democratic life? Heather Douglas argues that in some cases of socially relevant scientific debates (such as the legislation regarding neonicotinoids) it is rational for a lay individual to follow the advice of experts who explicitly share the same values as the lay individuals themselves. This is not said in the sense that science should and can be used to support sides or for that matter any sides (2017, p. 94). It is rather because there are reasons to think that the experts with whom we (or other individuals) share values would manage inductive risks (that is, evaluate the consequences of making a mistake) in the same way as we (or identified others) would do, or that they address questions we believe ought to be addressed. In the case of the neonicotinoids debate, the scientists who are worried about bee health would be more trustworthy for us if bee health is our primary concern. If we are instead more concerned about farmers minimizing pest damage, then trusting scientists with similar concerns will be the rational thing to do. Let us note here that saying that it is “rational” to follow the advice of experts who share our values should be taken to mean that it is “justifiable publicly, i.e., [with] a reasoned basis that can be stated publicly” (Douglas 2017, p. 94n9). Therefore, following the advice of experts who appear scientifically legitimated to defend a given position amounts to following an advice based on an empirically grounded position (i.e., supported by evidence), which can also be legitimately defended as a justified choice of advice.
The kind of understanding which is needed here is then quite demanding: citizens should understand how their values enter the scientific process, and jointly with facts produce the kind of results they themselves defend as most relevant to guide policy making. This kind of understanding recalls what Phillips et al. label under the model “understanding the Nature of Science” (2018) in the specific sense of understanding the influence of social and cultural values as epistemological tenets of the scientific endeavor (see also Douglas 2017 on “Teaching the Nature of Science,” p. 85; Lederman et al. 2014 on the tenets of science).
On a more practical side, by granting her point Douglas draws the conclusion that it would be better for scientists to be as transparent and as open as possible about their non-epistemic values, especially when these offer clear guidance to how scientists carry out their research. To this claim we add two provisions. First, in order to achieve effective transparency and openness, scientists should undergo a radical transformation in their practice of presenting scientific results and be trained in such a way that they can recognize, and are willing to acknowledge, the values they endorse while preserving the objectivity of research. Second, citizens should be able to accept that scientific expertise often goes hand in hand with balancing inductive risks, and that non-epistemic values influence how this balance is achieved. Besides, they should be able to identify, in specific cases, the values at stake and the role they play (namely, the kind of risk they intend to avoid). In other words, a minimal requirement for public understanding of science, for this second category of politically relevant science related issues, would be an understanding of (i) the different kinds of inductive risks in the specific case at stake, and (ii) the non-epistemic values subscribed by the different scientists or experts who contribute to a specific debate. To go back to our examples, in the case of neonicotinoids citizens should be able to (i) understand that experts’ disagreement comes from the fact that there is a difference between laboratory experiments and real fields results, and (ii) understand that this ambiguity in results will be evaluated differently depending on whether the focus of research is on the harm on insects, or else on the damage to agricultural fields. Similarly, in the case of EDs, citizens should understand that there exists deep scientific disagreement about the reliability of classical risk assessment methods, so that they can take (and defend) a rational position depending on their own political or moral values.5
3.3. Type-Three Issues: Rational Disagreements across Disciplines and/or Domains
Our third category of politically relevant science related issues is constituted by a class where disagreement specifically cuts across different kinds of disciplinary fields and then feeds into the political debate.6 Let us start here from a paradigmatic example: the regulation of the use of GMOs in agriculture. As convincingly shown by Biddle, the debates on GMOs focus on different kinds of risks: “health and safety risks (e.g., allergenicity), risks to the environment (e.g., evolution of herbicide and/or pesticide-tolerant plants or animals), and legal and socio-economic risks (e.g., accidental flow of GM seeds onto organic farms, heavy-handed use of patent infringement lawsuits)” (Biddle 2018, p. 365). In other words, the regulation of the use of GMOs in agriculture is a multidisciplinary science related issue, which makes use of scientific results from biomedicine, plant biology and agronomy, ecology, sociology and economics.7 Issuing political decisions then depends on the way the problem is framed not by one disciplinary field alone, but across several fields, in combination and often in conflict with each other. Following Biddle, framing a problem in a certain way means choosing “which sort of evidence is relevant to that problem” (2018, p. 369). By “sort of evidence,” Biddle does not refer to the evidence produced by different methods within a given scientific discipline, but to the kinds of evidence made available by different scientific disciplines and domains of investigation. In other words, the way(s) a problem is framed determines the relative weight that a discipline is attributed vis-à-vis others. And this may “impact on the distribution of investigative resources, which in turn will impact on the ways in which those investigations balance the risks of false positives, false negatives, and failure to generate results at all” (Biddle 2018, p. 369).
Biddle illustrates his view about framing evidence by pointing out that “critics of GMOs, for example, tend to have a broad range of concerns, including health and safety risks (e.g., allergenicity), risks to the environment (e.g., the evolution of herbicide and/or pesticide tolerant plants or animals), and legal and socio-economic risks (e.g., the accidental flow of GM seeds onto organic farms, heavy-handed use of patent infringement lawsuits). On the contrary, “GMO proponents […] tend to adopt a much narrower conception of risk. Some argue—or, more commonly, assume—that health and safety is the only relevant criterion for evaluating GM crops” (Biddle 2018, p. 366). In making decisions regarding the regulation of GMOs, the framing of the problem is (or should be) thus a central object of debate. Indeed, framing is ultimately a political choice since it heavily depends on ideological or moral values. Let us note that, within each of the scientific disciplines which may offer a relevant specific expertise (in the GMOs case, biology, ecology, economics, sociology, agronomy), there is also room for the sort of rational disagreement we described previously. By comparison with our second class of science related issues, what is specific about this third class is the existence of a diversity of field perspectives, driven by different objectives, theoretical backgrounds, empirical methods, and rules of evidence production.
With this third type of politically relevant science related issues, which kind of understanding is needed? Looking at the example of GMOs, it appears that a minimum requirement for citizens to form a sound opinion is an ability to identify what the different disciplines involved in the debate are, and what the objects and aims of investigation each field selects in view of entering the debate itself. Understanding the different disciplines dealing with the issue at stake gives indeed citizens a more structured view of the debate. It allows them to situate the different disciplinary or domain-laden positions by referring to the sort of evidence each position gives priority to. Giving priority to one discipline over others is mostly a matter of values and interest. But this, once more, does not entail that it is not rational to form an opinion (and to make a political choice when needed) on the basis of an argument supplied by one particular scientific domain. What it does mean is that forming a sound opinion (in the sense, given previously, of an opinion that could be discussed in a political arena by using sound arguments) demands an understanding of the disciplinary structure of the debate—namely, identifying the scientific fields where the arguments so debated originate from.
This kind of understanding would find an adequate description by resorting to a mixed model, which includes reference to both the “Nature of Science” and the “Science process” models as described in Phillips et al. (2018). As noted in the previous section, understanding the “Nature of Science” entails, among other features, understanding the role of values in assessing scientific hypotheses. In the case of this third class of science related issues, values play a role in choosing the kind(s) of proofs, coming from one specific discipline, which ought to be favored in defending a particular solution to a given politically relevant science related issue. Nonetheless, public understanding in this case also refers to the diversity of methods, belonging to different disciplines, that can be used to solve a given issue. In that sense, it also amounts to understanding the “science processes” as described by Phillips et al.
It must finally be noted that scientists and experts, when communicating their results, should clearly state which disciplinary perspective they adopt in a debate, potentially exposing themselves to the type of deep disagreement concerning single disciplines (as discussed in the previous section). This, however, does not blur the distinction between the two types of disagreement. It rather strengthens the need to pay separate attention to the impact and the consequences each type produces on what is to count as sound public understanding.
4. Conclusion
So, what should citizens understand about science to participate in democratic life? Contrary to the background assumption underlying past and current literature on public understanding of science, this paper aims to demonstrate that there cannot exist one unique answer to this question. The politically relevant science related issues are indeed epistemologically diverse, and this diversity makes the search for one, “across the board” referent for public understanding of science irrelevant, besides being misguided. To defend this idea, we proposed a basic typology of science related issues and assessed the impact that individual types of issues produce on the features of understanding citizens should rely on. In Table 1 below we offer a summary of our position.
. | Type I issues . | Type II issues . | Type III issues . |
---|---|---|---|
Focal epistemological feature | Wide scientific consensus | Rational disagreement within one discipline | Rational disagreement between different disciplines |
Examples | ● Is current climate change due to human activities? | ● Is chemical A (resp. neonicotinoid/endocrine disruptors) dangerous for species B (resp. bees/human beings)? | ● Should GMOs culture be limited? |
● Are recommended vaccines safe? | |||
Referent of public understanding | Understanding of the pattern of (dis)agreement among experts | Understanding of the role of values in balancing inductive risks | Understanding of the cross-disciplinary structure of the debate as informed by specific fields of scientific inquiry |
Links to existing model(s) of public understanding of science | Scientific division of labour model (akin to Keren 2018) | Nature of science model (akin to Phillips et al. 2018) | Nature of science model and Science process model (akin to Phillips et al. 2018) |
. | Type I issues . | Type II issues . | Type III issues . |
---|---|---|---|
Focal epistemological feature | Wide scientific consensus | Rational disagreement within one discipline | Rational disagreement between different disciplines |
Examples | ● Is current climate change due to human activities? | ● Is chemical A (resp. neonicotinoid/endocrine disruptors) dangerous for species B (resp. bees/human beings)? | ● Should GMOs culture be limited? |
● Are recommended vaccines safe? | |||
Referent of public understanding | Understanding of the pattern of (dis)agreement among experts | Understanding of the role of values in balancing inductive risks | Understanding of the cross-disciplinary structure of the debate as informed by specific fields of scientific inquiry |
Links to existing model(s) of public understanding of science | Scientific division of labour model (akin to Keren 2018) | Nature of science model (akin to Phillips et al. 2018) | Nature of science model and Science process model (akin to Phillips et al. 2018) |
Our typology could by no means be modified and/or refined: the aim of creating such a typology—it is worth repeating here—is specifically to show how the nature of the issue at stake influences the kind of understanding that citizens need if they want to participate competently in democratic life.
Arguing in favor of a case by case understanding of science driven political problems is not the same as proposing practical solutions to facilitate the acquisition of relevant understanding. In this paper we are not dealing with the important practical issue of what institutional conditions allow citizens to access and identify sound scientific information. There is a well-established body of literature dealing with this issue—addressing either the role and the defining features of experts in the public debate or the tools that citizens can use to choose who can legitimately be considered an expert (see for instance Anderson 2011). A parallel debate concerns the way scientific information is communicated and received by citizens, notably in cases of deep scientific uncertainty (Gustafson and Rice 2019).
Instead, in this paper we raise a philosophical question on “what” public understanding or public knowledge of science is about, in the belief that any discussions concerning the civic science literacy problem can only make sense by elucidating ex ante what laypeople are asked to be “literate about.” Bringing conceptual clarity to what public understanding is about is preliminary to finding practical answers to the civic science literacy problem, and we take our discussion on the nature of the referents of such understanding to be a step in that direction.
Notes
This is not to say that figuring out what “understanding” itself amounts to is not equally crucial. And we are of course aware of the wide literature on the very concept of understanding in the philosophy of science. However the focus of this paper is not on the meaning of understanding but rather on its object or target.
IPCC (International Panel on Climate Change), NAS (National Academy of Sciences), American Meteorological Society, American Geophysical Union, American Association for the Advancement of Science, ECCP (European Climate Change Programme), etc.
Biddle contrasts deep disagreement (which he qualifies as “rational”) with another type that he calls dirty (or irrational): the latter is “disagreement that results from ignorance, bias, irrationality, or dishonesty on the part of one or more parties to the disagreement” (p. 376).
This is not to say that non-epistemic values are not at play in situations where consensus (and lack of deep uncertainty) is present, e.g., Montuschi (2017a), p. 71.
Citizens follow a “motivated” reasoning here, which is different from the model often described in some literature (e.g., Kahan et al. 2012, quoted by Douglas 2017). The motivation at stake does not intend to secure personal interest (which, left on their own, open the gate to confirmation bias, ideological conviction, and the like) to the detriment of legitimate and accountable scientific evidence.
In the (mainly sociological) literature, this kind of problems are sometimes labelled as “complex,” “uncertain”, or “wicked” (see Spruijt et al. 2014 for a review). We do not use here this terminology which is mainly designed to think about the role of experts as policy advisers.
Another interesting multidisciplinary and highly divisive case of evidence gathering consists of the scientific (and policy) debate on the causes of the spread of bovine tuberculosis in the United Kingdom. See Montuschi (2017b).
References
Author notes
This paper has been written thanks to the support of the ISEED research project (Grant Agreement No. 960366, European Union’s Horizon 2020 Research and Innovation Programme). It reflects only the author’s views, and the European Union is not liable for any use that may be made of the information contained therein. For their comments and advice we thank in particular Roberto Gronda, Pierluigi Barrotta, Stephanie Ruphie, and Elena Denia.