Hostname: page-component-586b7cd67f-rdxmf Total loading time: 0 Render date: 2024-11-30T06:14:21.291Z Has data issue: false hasContentIssue false

As We Like It: Did the UK’s 2016 EU Referendum Reveal the “Will of the People?”

Published online by Cambridge University Press:  10 August 2023

Kristoffer Ahlstrom-Vij
Affiliation:
Birkbeck, University of London, UK
William L. Allen
Affiliation:
University of Oxford and Nuffield College, UK
Rights & Permissions [Opens in a new window]

Extract

Rhetoric surrounding the United Kingdom’s 2016 referendum on continued European Union (EU) membership frequently has invoked the “will of the people.” Addressing the House of Commons in March 2019, then–Prime Minster Theresa May stated that “my sense of responsibility and duty has meant that I have kept working to ensure that we deliver on the result and the will of the people” (March 27, 2019).1 May’s successor, Boris Johnson, appealed to the same notion when suggesting in the Daily Telegraph (September 15, 2019) that opposition parties were “united in wanting to cancel the referendum result…and overturn the will of the people.” On the other side of the debate, Caroline Lucas (currently the sole Member of Parliament for the UK’s Green Party) stated that “[e]very recent opinion poll shows that the will of the people has changed since [the referendum]” (December 4, 2018).

Type
The Contemporary Politics of the United Kingdom: Brexit, Identity, and Democracy
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of the American Political Science Association

Rhetoric surrounding the United Kingdom’s 2016 referendum on continued European Union (EU) membership frequently has invoked the “will of the people.” Addressing the House of Commons in March 2019, then–Prime Minster Theresa May stated that “my sense of responsibility and duty has meant that I have kept working to ensure that we deliver on the result and the will of the people” (March 27, 2019).Footnote 1 May’s successor, Boris Johnson, appealed to the same notion when suggesting in the Daily Telegraph (September 15, 2019) that opposition parties were “united in wanting to cancel the referendum result…and overturn the will of the people.” On the other side of the debate, Caroline Lucas (currently the sole Member of Parliament for the UK’s Green Party) stated that “[e]very recent opinion poll shows that the will of the people has changed since [the referendum]” (December 4, 2018).

These statements raise a key question: What precisely is “the will of the people” in electoral contexts? This article contends that voters’ true preferences—and, in the aggregate, the “will of the people”—comprise the preferences that would have been reported had people been more fully informed on relevant matters (Ahlstrom-Vij Reference Ahlstrom-Vij2022a; Harsanyi Reference Harsanyi1997). As such, given widespread public ignorance on politically relevant matters (Achen and Bartels Reference Achen and Bartels2016; Delli Carpini and Keeter Reference Carpini, Michael and Keeter1996), there likely are gaps between the choices made by the electorate at the polls and what is in their collective “will.” Such gaps have been investigated in various electoral contexts under the banner of “information effects,” including in the United States (Ahlstrom-Vij Reference Ahlstrom-Vij2021; Althaus Reference Althaus2003; Bartels Reference Bartels1996; Delli Carpini and Keeter Reference Carpini, Michael and Keeter1996); Canada (Blais et al. Reference Blais, Gidengil, Fournier and Nevitte2008); Denmark (Hansen Reference Hansen2009); Sweden (Oscarsson Reference Oscarsson2007); and European Parliamentary elections (Bhatti Reference Bhatti2010). Crucially, reading across the results of this body of work does not suggest that more fully informed voters consistently prefer any particular ideological or party-political position.

Building on this research, we examined whether the results of the UK’s 2016 referendum on leaving the EU—51.9% for “Leave” and 48.1% for “Remain”—were potentially sensitive to differences in voters’ levels of political information. We did so by developing and applying two counterfactual models of how information influences vote choice to the 2017 British Election Study (BES) face-to-face survey wave (N=2,067)—that is, the wave following the referendum. In a more fully informed electorate, and across different sets of theoretical assumptions about how what people know is related to how they vote, we find that support for leaving the EU likely would have decreased by up to 10.6 percentage points, producing a 58.7%–41.3% split in favor of Remain. We view the modeled estimates as good reasons for taking the extent to which voters are informed more seriously in interpreting electoral outcomes. Moreover, given widespread misinformation about consequential political issues, we suggest that they offer cautionary lessons for politicians and policy makers who confidently interpret electoral results as reflecting their constituents’ collective and definitive will.Footnote 2

WHAT IS THE “WILL OF THE PEOPLE”?

The “will of the people” consists of the preferences that would have been reported had we been informed.Footnote 3 We consider the preferences of an individual first. Although the most straightforward way of finding out what people want is to ask them, interpreting their answers is difficult when they are mistaken about the nature or implications of the options in question. To illustrate, for example, consider two environmental policies. Policy A is disastrous for the environment and Policy B is good for the environment—yet, in all other ways, they are identical in line with their outcomes. As it happens, we have it the wrong way around: although we care deeply about the environment, we choose Policy A. However, this does not necessarily mean that Policy A is preferred over Policy B. Indeed, imagine that we realize our mistake after choosing Policy A. At that point, we likely would deny that we ever wanted Policy A: the only reason we chose it was because we thought it had the properties actually possessed by Policy B.

This is what underlies Goodin’s (Reference Goodin1995, 137) observation that, under circumstances of ignorance, “we can serve a person’s ‘real’ preferences only by censoring the misleading indication of his preferences that is revealed in his choices.” Harsanyi (Reference Harsanyi1997, 133) made a similar point when distinguishing “those choices of a person that really express his true preferences…from those choices of his that fail to do so because they are based on incorrect information.” Generalizing this point to the collective, Delli Carpini and Keeter (Reference Carpini, Michael and Keeter1996, 5) suggested that “the real interest of an individual—and by extension of a group and of the polity as a whole—are reflected in the choices one would make if he or she were fully informed about the consequences.” Similarly, we hold that the “will of the people” comprises the preferences its constituent individuals would hold if relevant yet false beliefs were corrected. Assuming a majoritarian electoral system, the will of a population thereby is indicated by what the majority prefers (if anything) under conditions of having full information.

When examining that will, we are squarely in the political context—a context in which mistaken beliefs are in fact rife (Achen and Bartels Reference Achen and Bartels2016; Friedman Reference Friedman1998).Footnote 4 To be sure, if the public reports wanting something, there usually is a strong case for government to act accordingly. However, in settings in which what voters say they want might be due to their not having had the time or opportunity to consider all relevant information—arguably, a central feature of the division of political labor involved in any representative democracy—directly reading the “will of the people” from electoral results becomes difficult. It would be convenient, then, if there were a way to identify diagnostically whether there were such gaps between what people say they want and what they would have said if they had possessed fuller information. The next section argues that work on “information effects” offers just such a way.

MEASURING THE SENSITIVITY OF ELECTORAL OUTCOMES TO INFORMATION

We argue that a long tradition of political scientists using statistical models to examine “information effects”—that is, gaps between actual outcomes or distributions of attitudes and the counterfactual outcomes or attitudes that we likely would have seen under full information—provides a way of diagnosing the presence of such gaps.

The Intuition Behind Counterfactually Modeling Information Effects

In electoral contexts, the procedure follows four steps. First, we construct a knowledge scale and place respondents on that scale. Such scales (see, e.g., Delli Carpini and Keeter Reference Carpini, Michael and Keeter1996) typically are built from questions that we might see on a civics exam—the roles that named people hold or the platforms of political parties—and are meant to indicate generalist political knowledge.Footnote 5 Second, using data on reported votes, we fit a model that estimates the probability of someone reporting having voted a particular way as a function of that person’s demographics and level of knowledge. Third, we increase each person’s placement on the knowledge scale to the maximum level—that is, the level at which they answer all of the questions correctly—while leaving everything else as is, and then use the fitted model to “predict” how each respondent would vote. This simulates how that person would have voted had they been “fully informed.” (If we want to relax this assumption of being fully informed, we can move individuals’ knowledge scores to a different threshold. Reporting and justifying this choice is good practice.) Fourth, by noting the difference between the actual electoral outcome and the model’s estimated outcome, we obtain a measure of the aggregate information effect. A substantial difference indicates high sensitivity to voters’ level of information and a greater risk that the result might be an artifact of mistaken beliefs to a non-trivial degree.

What counts as “substantial” in the context of elections? Bartels (Reference Bartels1996, 220) found information effects in the range of 2 to 5 percentage points in US presidential elections. Blais and colleagues (Reference Blais, Gidengil, Fournier and Nevitte2008) found an average information effect of 2.3 percentage points across parties in six Canadian elections. Oscarsson (Reference Oscarsson2007) found an average net gain of 2.7 percentage points for right parties in six Swedish elections. Bhatti (Reference Bhatti2010) modeled three European Parliament elections (i.e., in Denmark, Finland, and Sweden) with an average information effect of 3.5 percentage points across parties and elections. Collectively, these studies suggest that finding information effects of more than approximately 3 percentage points would be substantial.

Modeling a More Informed EU Referendum

In contrast to previous scholars’ focus on regularly scheduled elections, we considered whether, to what extent, and in which direction the UK’s EU referendum result may have displayed an information effect. This was motivated not only by observing politicians’ invocations of the “will of the people,” as described previously, but also by referenda representing a relatively rare event in British politics—and one in which information conveyed through campaign efforts has been shown experimentally to have significant effects on attitudes (Morisi Reference Morisi2018).

To this end, we used the 2017 BES face-to-face survey wave (N=2,194), which contains demographic information, responses to six questions measuring political knowledge, and respondents’ reported vote in the referendum (Fieldhouse et al. Reference Fieldhouse, Green, Evans, Schmitt, van der Eijk, Mellon and Prosser2018). Survey fieldwork occurred between June 26 and October 1, 2017. Omitting 127 observations that lacked survey weights left 2,067 observations, of which 2.6% had missing values that we imputed with multiple imputation using aregImpute in R ’s Hmisc package (Harrell et al. Reference Harrell and Dupont2019; R Core Team 2017).Footnote 6 By way of creating a knowledge scale, we then fitted a dichotomous item response theory (IRT) model using mirt (Chalmers Reference Chalmers2012) to estimate the underlying knowledge of respondents based on the six knowledge items (Ahlstrom-Vij and Allen Reference Ahlstrom-Vij and Allen2023).Footnote 7

Because we were looking to model a counterfactual—that is, how the electorate likely would have voted had we somehow been able to intervene on their level of knowledge by increasing its value—we used an explicitly causal model.Footnote 8 In terms of likely confounders (i.e., variables that likely have an effect on a person’s voting behavior and their level of knowledge), we controlled for gender (Plutzer Reference Plutzer, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020; vanHeerde-Hudson Reference vanHeerde-Hudson, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020), level of education (Hebbelstrup and Rasmussen Reference Hebbelstrup and Rasmussen2016), income (Plutzer Reference Plutzer, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020; Vowles Reference Vowles, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020), and age (Plutzer Reference Plutzer, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020). To reduce noise in the models, we also controlled for variables that likely influence voting behavior but not necessarily a person’s level of knowledge: ethnicity (Dawson Reference Dawson1994), religion and social class (Evans and Northmore-Ball Reference Evans, Northmore-Ball, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020), and marital status (Denver Reference Denver2008).

We viewed partisanship as a mediator because it likely is affected by political knowledge—specifically, knowledge of party and candidate positioning (Brader and Tucker Reference Brader and Tucker2018). Controlling for a mediator—that is, a causal node located on a direct or indirect pathway between political knowledge and political preference—would misestimate the relevant causal effect.Footnote 9 Nevertheless, in the interest of robustness, we fitted two multinomial logistic models: (1) a demographic model, containing all variables mentioned previously except partisanship; and (2) a partisanship model, containing all variables including partisanship. In both cases, an individual’s reported vote (i.e., “Leave,” “Remain,” or “Did Not Vote”) was the dependent variable. The models were fitted with R ’s nnet package (Venables and Ripley Reference Venables and Ripley2002) using “doubly robust” estimation of causal effects (Morgan and Winship Reference Morgan and Winship2015). For purposes of this estimation, we identified the knowledge score given by the IRT model as corresponding to having answered correctly all knowledge items. We then recoded the knowledge variable as a binary variable, with all observations meeting that threshold coded as 1 and everyone else as 0. That threshold occurred at a knowledge score of 1.036 and was met by 26% of the sample. IRT scores do not have any intrinsic meaning; however, because they can be interpreted as Z-scores, a knowledge score of 1.036 corresponds to someone at about one standard deviation above the estimated mean level of knowledge. We then estimated propensity scores using logistic regression via the informationeffects package (Ahlstrom-Vij Reference Ahlstrom-Vij2022b). In line with the assumptions about the causal determinants of political knowledge discussed previously, the model estimated the association between being fully informed (as given by the binary knowledge variable) and gender, age, income, and level of education. Diagnostic plots from the cobalt package (Greifer Reference Greifer2022) confirmed improved balance between the two groups (i.e., those who were and were not fully informed) across these covariates.Footnote 10 We then used these propensity scores as weights in fitting the models.Footnote 11

Finally, to estimate the distribution of support we likely would see for “Leave” and “Remain” in this counterfactually informed electorate, we compared the actual outcome with the distribution estimated by the two models. We did this after setting the (binary) knowledge variable for each respondent to 1 (which represents being fully informed, in the sense of answering correctly all of the knowledge items), with each observation weighted using the survey weights included with the dataset, to approximate representativeness.Footnote 12 Figure 1 reports the actual referendum outcome alongside the estimated outcomes on the two models.

Figure 1 Actual and Estimated Referendum Results Under Full Information Using Two Different Models, Including Survey Weights

On the partisanship model, the proportion in support for “Leave” decreased by 9.7 percentage points, from 51.9% to 42.2%. Using the demographic model, which assumes partisanship acting as a mediator, the “Leave” vote decreased by 10.6 percentage points to 41.3%. Reading across the models, not only does the aggregate result switch from “Leave” to “Remain,” the information effect also exceeds the standard set for a substantial information effect (i.e., 3 percentage points) by a fairly wide margin. This suggests that the results likely were sensitive to voters’ levels of information.

DISCUSSION

What does this exercise of simulating electoral results meaningfully contribute to the understanding of politics and political behavior? When supported by a robust set of causal models, this approach provides an empirical and replicable measure of the degree to which an outcome is sensitive to shifts in voters’ political knowledge. This should be of interest to political scientists who are concerned with the role of information in voters’ choices. Moreover, from a methodological point of view and as with causal modeling generally, this approach also forces researchers to be explicit and transparent about the mechanisms we think are at work when people vote, behave, and make sense of political issues. This invites a healthy skepticism of causal assumptions and enables us to pinpoint which assumptions matter. In our case, for example, conducting robustness checks involving partisanship—clearly, an important factor for political choices (De Vries, Hobolt, and Tilley Reference De Vries, Hobolt and Tilley2018)—enables us to gauge the sensitivity of information effects to different model specifications.

The outcomes of such models also should be of interest to policy makers who must translate electoral outcomes into political policy, especially because such modeling does not presume a priori that additional information benefits any particular side. This is a helpful and welcome feature in political contexts characterized by motivated reasoning and affective polarization. In the particular case of the 2016 UK referendum on EU membership, however, it might be objected that there actually was an abundance of information in the lead-up to the campaign: the problem was that this information, for whatever reason, did not influence people. Perhaps voters were sensitive to the partisan and ideological positions of either the messengers of factual information or the political elites who endorsed or dismissed those sources of evidence; studies about cues suggest that this happens in political domains (Aarøe Reference Aarøe2012; Darmofal Reference Darmofal2005). This explanation may be particularly relevant in the UK referendum context in which divisions between “Leave” and “Remain” voters seemingly have hardened into identities in their own right (Hobolt, Leeper, and Tilley Reference Hobolt, Leeper and Tilley2021).

Yet, whether this is true does not detract from what we attempt to do in this article. For example, it might be true that had we known X, we would not have believed Y or done Z. However, that counterfactual holding also is compatible with motivated reasoning preventing us from believing (and therefore knowing) X in the first place, and that someone telling us that X therefore will not have us relinquish our belief in Y or stop doing Z. For that reason, the point of the type of causal modeling conducted in this article is not to offer a recipe for attitude change. Rather, it is to isolate the effect of knowledge on political choices and attitudes in order to provide a diagnostic tool to stress-test claims about what is in the “will of the people.” In particular, having empirical evidence of substantial information effects should give policy makers pause in directly and conclusively inferring the will of the people from any result, including in contexts of relatively rare electoral events such as referenda in the United Kingdom. Specifically, to the extent that we believe that electoral outcomes must be sensitive to the will of the people to be legitimate—at least within reasonable liberal constraints on majority rule—substantial information effects offer valuable datapoints on whether acting on the relevant outcomes would be politically legitimate.

…evidence of substantial information effects should give policy makers pause in directly and conclusively inferring the will of the people from any result, including in contexts of relatively rare electoral events such as referenda in the United Kingdom.

Would applying such a diagnostic tool mean that policy makers pay attention only to the informed segment of the electorate and, as such, be elitist? On the contrary, it would be more egalitarian than simply catering to the stated preferences of the public, whether via traditional opinion polls, elections, or referenda. If our real preferences are those that we would have had if we had been fully informed (as we contended at the outset), then we only will tend to successfully express our real preferences insofar as we are politically informed. If we are not informed, by contrast, we are less likely to know what those preferences are. Given the correlation between privilege and political information, paying attention to stated preferences therefore will consistently—even if not deliberately—prioritize the preferences of the privileged and ignore those of the less privileged. By contrast, policy makers paying attention to information effects likely will be in a better position to listen to everyone’s true preferences—at least in the sense of being able to measure whether electoral outcomes likely mischaracterize the less privileged part of the population and, consequently, risk ignoring the injustices and concerns that disproportionally affect them.

ACKNOWLEDGMENT

William L. Allen acknowledges support from the British Academy (Grant No. PF21\210066).

DATA AVAILABILITY STATEMENT

Research documentation and data that support the findings of this study are openly available at the PS: Political Science & Politics Harvard Dataverse at https://doi.org/10.7910/DVN/Y06T2Y.

CONFLICTS OF INTEREST

The authors declare that there are no ethical issues or conflicts of interest in this research.

Footnotes

1. For this and all subsequent quotes from Members of Parliament, see https://hansard.parliament.uk, unless otherwise stated.

2. We are not suggesting that the “will of the people” is a myth (Weale Reference Weale2018) but instead that, at least in this case, the referendum may have failed to uncover it.

3. We provide a summary of our rationale here; however, see Ahlstrom-Vij (Reference Ahlstrom-Vij2022a) for a more detailed defense.

4. It has been suggested that the public is able to rely on cues and shortcuts to act as if informed (Popkin Reference Popkin1991) or that voters are able to vote retrospectively on minimal amounts of information (Fiorina Reference Fiorina1981; Key Reference Key1966). See Bartels (Reference Bartels1996) and Achen and Bartels (Reference Achen and Bartels2016) for critical discussions, as well as Ahlstrom-Vij (Reference Ahlstrom-Vij, Fricker, Graham, Henderson and Nikolaj2019) for a critical survey of suggestions for how democracies might cope with low levels of voter knowledge.

5. Some scholars object to this by arguing that competent voting does not require knowing answers to these types of survey items (e.g., Lupia Reference Lupia2006). We do not contest this point. Rather, we view these questions as diagnostic of whether people know things that are necessary for competent voting. This is for two reasons. First, these scales have high internal consistency and also correlate well with interviewer ratings of respondents’ political knowledge and behaviors associated with knowledge, such as participation (Delli Carpini and Keeter Reference Carpini, Michael and Keeter1996). Second, information effects based on these scales are consistently lower for groups that we have independent reason to believe will be more politically informed, such as those who are college educated (Hebbelstrup and Rasmussen Reference Hebbelstrup and Rasmussen2016), affluent (Plutzer Reference Plutzer, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020; Vowles Reference Vowles, Fisher, Fieldhouse, Franklin, Gibson, Cantijoch and Wlezien2020), and living in cities or suburbs (Althaus Reference Althaus2003).

6. Full reproducible R code for all subsequent analysis is available at https://github.com/ahlstromvij/informed_referendum.

7. The six (true/false) knowledge items were: (1) “Polling stations close at 10 pm on election day”; (2) “No one may stand for parliament unless they pay a deposit”; (3) “Only taxpayers are allowed to vote in a general election”; (4) “The Liberal Democrats favor a system of proportional representation”; (5) “MPs from different parties are on parliamentary committees”; and (6) “The number of Members of Parliament is about 100.” A parallel analysis suggested that the items might form two dimensions, but a confirmatory factor analysis using only one dimension exhibited very good fit, indicating unidimensionality. Both a two- and a three-parameter model (accounting for any guessing) were fitted to the items, but a likelihood ratio test suggested no significant difference among the models; therefore, the simpler two-parameter model was used. The test-information function suggested good precision (with a peak around mean ability), and tests indicated local independence (by Yen’s Q3) and good model fit (evaluated through a plot of observed-versus-expected values). For more details on these diagnostics, see DeMars (Reference DeMars2010).

8. Pearl (Reference Pearl2009) is the central text here, but see also Keele, Stevenson, and Elwert (Reference Keele, Stevenson and Elwert2020) for an overview of good causal inferential practices in political science.

9. Even if partisanship is not a mediator, controlling for it in this context likely is unnecessary. Socialization is centered around group-identity considerations relating to religion, ethnicity, gender, social class, and the like. All of these factors shape individuals’ conceptions of who they are and, consequently, also about which positions “people like us” take in politics (Green, Palmquist, and Schickler Reference Green, Palmquist and Schickler2002). Consequently, controlling for such group-level variables, in the manner described here, already would account for partisanship.

10. One of the observations ultimately was assigned a very high propensity score of 120, with the second highest 48, and the mean score 2. On manual inspection, the observation was deemed to exhibit an unusual response pattern but not so unusual that it was clear that it had been miscoded. Nevertheless, to avoid that observation having a disproportionate influence on subsequent modeling, its propensity score was trimmed to 48.

11. The McFadden value for the partisanship model was 0.187 and the coefficients for the knowledge variable 0.330 for “Leave” (p=0.004) and 0.896 for “Remain”’ (p=0.000), with “Did Not Vote” as the reference category. The McFadden value for the demographic model was 0.139 and the coefficients 0.406 for “Leave” (p=0.023) and 0.960 for “Remain” (p=0.000), again with “Did Not Vote” as the reference category. The variance inflation factor for each predictor did not diverge substantially from 1 on either model and in no case was greater than 5, which suggests an absence of multicollinearity.

12. For each observation, the models ascribed a probability of voting “Remain,” voting “Leave,” or not voting, respectively, given full information. The informed proportion of “Remain” and “Leave,” respectively, was calculated as the weighted mean of the corresponding probabilities, with the survey weights as weights.

References

REFERENCES

Aarøe, Lene. 2012. “When Citizens Go Against Elite Directions: Partisan Cues and Contrast Effects on Citizens’ Attitudes.” Party Politics 18 (2): 215–33.10.1177/1354068810380093CrossRefGoogle Scholar
Achen, Christopher L., and Bartels, Larry M.. 2016. Democracy for Realists: Why Elections Do Not Produce Responsive Government. Princeton, NJ: Princeton University Press.10.1515/9781400882731CrossRefGoogle Scholar
Ahlstrom-Vij, Kristoffer, and Allen, William L.. 2023. “Replication Data for ‘As We Like It: Did the UK’s 2016 EU Referendum Reveal the “Will of the People”?’PS: Political Science & Politics. DOI: 10.7910/DVN/Y06T2Y.Google Scholar
Ahlstrom-Vij, Kristoffer. 2022a. “The Case for Modelled Democracy.” Episteme 19 (1): 89110.10.1017/epi.2020.10CrossRefGoogle Scholar
Ahlstrom-Vij, Kristoffer. 2022b. “informationeffects: A Complete Pipeline for Calculating Information Effects.” R Package Version 0.1.0. https://github.com/ahlstromvij/informationeffects.Google Scholar
Ahlstrom-Vij, Kristoffer. 2021. “Do We Live in a ‘Post-Truth’ Era?Political Studies 71 (2). https://doi.org/10.1177/00323217211026427.Google Scholar
Ahlstrom-Vij, Kristoffer. 2019. “The Epistemic Benefits of Democracy: A Critical Assessment.” The Routledge Handbook of Social Epistemology, ed. Fricker, Miranda, Graham, Peter J., Henderson, David, and Nikolaj, J. L. L. Pedersen, 406–14. New York: Routledge.10.4324/9781315717937-40CrossRefGoogle Scholar
Althaus, Scott. 2003. Collective Preferences in Democratic Politics. Cambridge, UK: Cambridge University Press.10.1017/CBO9780511610042CrossRefGoogle Scholar
Bartels, Larry M. 1996. “Uninformed Voters: Information Effects in Presidential Elections.” American Journal of Political Science 40 (1): 194230.10.2307/2111700CrossRefGoogle Scholar
Bhatti, Yosef. 2010. “What Would Happen If We Were Better Informed? Simulating Increased Knowledge in European Parliament Elections.” Representation 46 (4): 391410.10.1080/00344893.2010.518078CrossRefGoogle Scholar
Blais, André, Gidengil, Elisabeth, Fournier, Patrick, and Nevitte, Neil. 2008. “Information, Visibility and Elections: Why Electoral Outcomes Differ When Voters Are Better Informed.” European Journal of Political Research 48:256–80.10.1111/j.1475-6765.2008.00835.xCrossRefGoogle Scholar
Brader, Ted, and Tucker, Joshua A.. 2018. “Unreflective Partisans? Policy Information and Evaluation in the Development of Partisanship.” Advances in Political Psychology 39 (Suppl. 1): 137–57.10.1111/pops.12480CrossRefGoogle Scholar
Chalmers, R. Philip. 2012. “mirt: A Multidimensional Item Response Theory Package for the R Environment.” Journal of Statistical Software 48 (6): 129.10.18637/jss.v048.i06CrossRefGoogle Scholar
Darmofal, David. 2005. “Elite Cues and Citizen Disagreement with Expert Opinion.” Political Research Quarterly 58 (3): 381–95.10.1177/106591290505800302CrossRefGoogle Scholar
Dawson, Michael C. 1994. Behind the Mule: Race and Class in African American Politics. Princeton, NJ: Princeton University Press.Google Scholar
Carpini, Delli, Michael, X., and Keeter, Scott. 1996. What Americans Know about Politics and Why It Matters. New Haven, CT: Yale University Press.Google Scholar
DeMars, Christine. 2010. Item Response Theory. Oxford, UK: Oxford University Press.10.1093/acprof:oso/9780195377033.001.0001CrossRefGoogle Scholar
Denver, David. 2008. “Another Reason to Support Marriage? Turnout and the Decline of Marriage in Britain.” British Journal of Politics and International Relations 10 (4): 666–80.10.1111/j.1467-856x.2008.00329.xCrossRefGoogle Scholar
De Vries, Catherine E., Hobolt, Sara B., and Tilley, James. 2018. “Facing Up to the Facts: What Causes Economic Perceptions?Electoral Studies 51 (February): 115–22.10.1016/j.electstud.2017.09.006CrossRefGoogle Scholar
Evans, Geoffrey, and Northmore-Ball, Ksenia. 2020. “Long-Term Factors: Class and Religious Cleavages.” In The Routledge Handbook of Elections, Voting Behavior and Public Opinion, ed. Fisher, Justin, Fieldhouse, Edward, Franklin, Mark N., Gibson, Rachel, Cantijoch, Marta, and Wlezien, Christopher, 123–35. New York: Routledge.Google Scholar
Fieldhouse, Edward, Green, Jane, Evans, Geoffrey, Schmitt, Hermann, van der Eijk, Cees, Mellon, Jonathan, and Prosser, Christopher. 2018. British Election Study, 2017: Face-to-Face Post-Election Survey. http://doi.org/10.5255/UKDA-SN-8418-1.CrossRefGoogle Scholar
Fiorina, Morris. 1981. Retrospective Voting in American National Elections. New Haven, CT: Yale University Press.Google Scholar
Friedman, Jeffrey. 1998. “Introduction: Public Ignorance and Democratic Theory.” Critical Review 12 (4): 397411.10.1080/08913819808443510CrossRefGoogle Scholar
Goodin, Robert E. 1995. Utilitarianism as Public Policy. Cambridge, UK: Cambridge University Press.Google Scholar
Green, Donald P., Palmquist, Bradley, and Schickler, Eric. 2002. Partisan Hearts and Minds: Political Parties and the Social Identities of Voters. New Haven, CT: Yale University Press.Google Scholar
Greifer, Noah. 2022. “cobalt: Covariate Balance Tables and Plots.” R Package Version 4.3.2. https://CRAN.R-project.org/package=cobalt.Google Scholar
Hansen, Kasper M. 2009. “Changing Patterns in the Impact of Information on Party Choice in a Multiparty System.” International Journal of Public Opinion Research 21 (4): 525–46.10.1093/ijpor/edp042CrossRefGoogle Scholar
Harrell, Frank E., Dupont, Charles, and Collaborators. 2019. “Hmisc: Harrell Miscellaneous.” R Package Version 4.2-0. https://CRAN.R-project.org/package=Hmisc.Google Scholar
Harsanyi, John C. 1997. ‘Utilities, Preferences, and Substantive Goods.” Social Choice and Welfare 14:129–45.10.1007/s003550050057CrossRefGoogle Scholar
Hebbelstrup, Stig, and Rasmussen, Rye. 2016. “Education or Personality Traits and Intelligence as Determinants of Political Knowledge?Political Studies 64 (4): 1036–54.Google Scholar
Hobolt, Sara B., Leeper, Thomas J., and Tilley, James. 2021. “Divided by the Vote: Affective Polarization in the Wake of the Brexit Referendum.” British Journal of Political Science 51 (4): 1476–93.10.1017/S0007123420000125CrossRefGoogle Scholar
Keele, Luke, Stevenson, Randolph T., and Elwert, Felix. 2020. “The Causal Interpretation of Estimated Associations in Regression Models.” Political Science Research and Methods 8 (1): 113.10.1017/psrm.2019.31CrossRefGoogle Scholar
Key, V. O. 1966. The Responsible Electorate: Rationality in Presidential Voting 1936–1960. Cambridge, MA: Belknap Press.10.4159/harvard.9780674497764CrossRefGoogle Scholar
Lupia, Arthur. 2006. “How Elitism Undermines the Study of Voter Competence.” Critical Review 18 (3): 217–32.10.1080/08913810608443658CrossRefGoogle Scholar
Morgan, Stephen L., and Winship, Christopher. 2015. Counterfactuals and Causal Inference: Methods and Principles for Social Research. Second edition. New York: Cambridge University Press.Google Scholar
Morisi, Davide. 2018. “Choosing the Risky Option: Information and Risk Propensity in Referendum Campaigns.” Public Opinion Quarterly 82 (3): 447–69.10.1093/poq/nfy033CrossRefGoogle Scholar
Oscarsson, Henrik. 2007. “A Matter of Fact? Knowledge Effects on the Vote in Swedish General Elections, 1985–2002.” Scandinavian Political Studies 30 (3): 301–22.10.1111/j.1467-9477.2007.00182.xCrossRefGoogle Scholar
Pearl, Judea. 2009. Causality: Models, Reasoning, and Inference. Second edition. Cambridge, UK: Cambridge University Press.10.1017/CBO9780511803161CrossRefGoogle Scholar
Plutzer, Eric. 2020. “Demographics and the Social Bases of Voter Turnout.” In The Routledge Handbook of Elections, Voting Behavior and Public Opinion, ed. Fisher, Justin, Fieldhouse, Edward, Franklin, Mark N., Gibson, Rachel, Cantijoch, Marta, and Wlezien, Christopher, 6982. New York: Routledge.Google Scholar
Popkin, Samuel L. 1991. The Reasoning Voter: Communication and Persuasion in Presidential Elections. Chicago: University of Chicago Press.10.7208/chicago/9780226772875.001.0001CrossRefGoogle Scholar
R Core Team. 2017. R: A Language and Environment for Statistical Computing. Vienna, Austria: R Foundation for Statistical Computing.Google Scholar
vanHeerde-Hudson, Jennifer. 2020. “Political Knowledge: Measurement, Misinformation and Turnout.” In The Routledge Handbook of Elections, Voting Behavior and Public Opinion, ed. Fisher, Justin, Fieldhouse, Edward, Franklin, Mark N., Gibson, Rachel, Cantijoch, Marta, and Wlezien, Christopher, 369–82. New York: Routledge.Google Scholar
Venables, William N., and Ripley, Brian D.. 2002. Modern Applied Statistics with S. Fourth edition. New York: Springer Publishing.10.1007/978-0-387-21706-2CrossRefGoogle Scholar
Vowles, Jack. 2020. “The Big Picture: Turnout at the Macro-Level.” In The Routledge Handbook of Elections, Voting Behavior and Public Opinion, ed. Fisher, Justin, Fieldhouse, Edward, Franklin, Mark N., Gibson, Rachel, Cantijoch, Marta, and Wlezien, Christopher, 5768. New York: Routledge.Google Scholar
Weale, Albert. 2018. The Will of the People: A Modern Myth. Cambridge, UK: Polity Press.Google Scholar
Figure 0

Figure 1 Actual and Estimated Referendum Results Under Full Information Using Two Different Models, Including Survey Weights

Supplementary material: Link

Ahlstrom-Vij and Allen Dataset

Link