Guess - Et Al - 2019 - Prevalence and Predictors of Fake News Dissemination On Facebook
Guess - Et Al - 2019 - Prevalence and Predictors of Fake News Dissemination On Facebook
INTRODUCTION sults hold when alternate lists are used, such as that used by peer-
One of the most discussed phenomena in the aftermath of the 2016 U.S. reviewed studies (2).
Fig. 1. Distribution of total and fake news shares. (Left) Histogram of the total number of links to articles on the web shared by respondents in the sample who iden-
tified as Democrats, Republicans, or independents. (Right) Stacked histogram of the number of fake news articles shared by respondents who identified as Democrats,
identifying as “very conservative,” shared the most articles from fake association is also robust to controlling for party, as the various
news domains. On average, a conservative respondent shared 0.75 alternative specifications provided in the Supplementary Materials
such stories [95% confidence interval (CI), 0.537 to 0.969], and a very illustrate. In column 2, the coefficient on “Age: over 65” implies that
conservative respondent shared 1.0 (95% CI, 0.775 to 1.225). This is being in the oldest age group was associated with sharing nearly seven
consistent with the pro-Trump slant of most fake news articles pro- times as many articles from fake news domains on Facebook as those
duced during the 2016 campaign, and of the tendency of respon- in the youngest age group, or about 2.3 times as many as those in the
dents to share articles they agree with, and thus might not represent next-oldest age group, holding the effect of ideology, education, and the
a greater tendency of conservatives to share fake news than liberals total number of web links shared constant (e1.9 ≈ 6.69, e1.9−1.079 ≈ 2.27).
conditional on being exposed to it (3). This association is also found in the specifications using the alternate
Figure 2D shows that, if anything, those who share the most content peer-reviewed measure (2) as a dependent variable in columns 3 and 4,
in general were less likely to share articles from fake news–spreading with those over 65 sharing between three and four times as many
domains to their friends. Thus, it is not the case that what explains fake news links as those in the youngest age group.
fake news sharing is simply that some respondents “will share any- Aside from the overall rarity of the practice, our most robust and
thing.” These data are consistent with the hypothesis that people who consistent finding is that older Americans were more likely to share
share many links are more familiar with what they are seeing and articles from fake news domains. This relationship holds even when
are able to distinguish fake news from real news. (We note that we we condition on other factors, such as education, party affiliation,
have no measure as to whether or not respondents know that what ideological self-placement, and overall posting activity. It is robust
they are sharing is fake news.) Turning to a key demographic char- to a wide range of strategies for measuring fake news (see Materials
acteristic of respondents, a notable finding in Fig. 2B is the clear as- and Methods). Further, none of the other demographics variables in
sociation between age group and the average number of articles from our model—sex, race, education, and income—have anywhere close
fake news domains shared on Facebook. Those over 65 shared an average to a robust predictive effect on sharing fake news. We subject our
of 0.75 fake news articles (95% CI, 0.515 to 0.977), more than twice findings to a battery of robustness tests in the Supplementary Materials.
as many as those in the second-oldest age group (0.26 articles; 95% CI, Among them, we show that model specification, other predictors such
0.206 to 0.314). Of course, age is correlated with other characteristics, as political knowledge, and distributional assumptions about the de-
including political predispositions. Thus, we turn to a multivariate anal- pendent variable do not appear to be driving our results (tables S1 to
ysis to examine the marginal impact of individual characteristics. S8 and S13). Last, we show in table S14 that, when we try to explain
Table 2 shows that the age effect remains statistically significant patterns of hard news sharing behavior using the same approach, the
when controlling for ideology and other demographic attributes. The predictors are more varied and do not include age.
DISCUSSION pected. More puzzling is the independent role of age: Holding con-
Using unique behavioral data on Facebook activity linked to individual- stant ideology, party identification, or both, respondents in each age
level survey data, we find, first, that sharing fake news was quite rare category were more likely to share fake news than respondents in the
during the 2016 U.S. election campaign. This is important context next-youngest group, and the gap in the rate of fake news sharing
given the prominence of fake news in post-election narratives about between those in our oldest category (over 65) and youngest category
the role of social media disinformation campaigns. Aside from the is large and notable.
relatively low prevalence, we document that both ideology and age These findings pose a challenge and an opportunity for social sci-
were associated with that sharing activity. Given the overwhelming entists. Political scientists tend to favor explanations based on sta-
pro-Trump orientation in both the supply and consumption of fake ble, deeply held partisan or ideological predispositions (10, 11). The
news during that period, including via social pathways on Facebook predictive power of demographic traits evaporates when subjected
(3), the finding that more conservative respondents were more likely to multiple regression analyses that control for other characteristics cor-
to share articles from fake news–spreading domains is perhaps ex- related with those demographics. Yet, when an empirical relationship
Table 2. Determinants of fake news sharing on Facebook. Quasi-Poisson models with YouGov’s sample-matching weights applied. Dependent variables are
counts of fake news articles shared using measures derived from (7) (columns 1 and 2) and (2) (columns 3 and 4). The reference category for ideology is “Not
sure.” “Number of links shared” refers to the number of Facebook posts by each respondent that includes a link to an external URL. A&G, Allcott and Gentzkow.
Number of stories shared Number of stories shared (A&G)
(1) (2) (3) (4)
Very liberal 0.487 0.387 1.634* 1.485*
(1.238) (1.209) (0.876) (0.800)
Liberal −1.127 −1.141 0.873 0.812
(1.439) (1.404) (0.886) (0.809)
Moderate 0.333 0.392 0.748 0.824
(1.186) (1.157) (0.875) (0.799)
Conservative 2.187* 2.248** 1.736** 1.800**
(1.155) (1.128) (0.868) (0.794)
Very conservative 2.366** 2.297** 2.231** 2.087***
(1.158) (1.132) (0.869) (0.795)
Age: 30–44 0.772 0.742 0.253 0.172
such as the one documented here emerges, we are challenged to view and beyond, lacks the level of digital media literacy necessary to reli-
demographic traits not as controls to be ignored but as central ex- ably determine the trustworthiness of news encountered online (13, 14).
planatory factors above and beyond the constructs standard in the There is a well-established research literature on media literacy and
literature (12). This is especially the case with age, as the largest gener- its importance for navigating new media technologies (15). Build-
ation in America enters retirement at a time of sweeping demographic ing on existing work (16, 17), researchers should further develop
and technological change. Below, we suggest possible avenues for further competency-based measures of digital media literacy that encompass
research incorporating insights from multiple disciplines. the kinds of skills needed to identify and avoid dubious content de-
Given the general lack of attention paid to the oldest generations signed to maximize engagement. Research on age and digital media
in the study of political behavior thus far, more research is needed literacy often focuses on youth skills acquisition and the divide be-
to better understand and contextualize the interaction of age and tween “digital natives” and “digital immigrants” (18), but our results
online political content. Two potential explanations warrant further suggest renewed focus on the oldest age cohorts.
investigation. First, following research in sociology and media stud- Within this cohort, lower levels of digital literacy could be com-
ies, it is possible that an entire cohort of Americans, now in their 60s pounded by the tendency to use social endorsements as credibility
cues (19). If true, this would imply a growing impact as more Amer- sibility. Similarly, it may be the case that the composition of older
icans from older age groups join online social communities. A second Facebook users’ News Feeds differs in systematically important ways
possibility, drawn from cognitive and social psychology, suggests a from that of younger users; while we lack the data in the present work
general effect of aging on memory. Under this account, memory de- to test this proposition, future research with access to these data could
teriorates with age in a way that particularly undermines resistance prove illuminating. These concerns aside, the evidence we have pre-
to “illusions of truth” and other effects related to belief persistence sented is strongly suggestive of an emerging relationship between not
and the availability heuristic, especially in relation to source cues (20–22). only ideological affiliation but also age and the still-rare tendency to
The severity of these effects would theoretically increase with the spread misinformation to one’s friends on social media platforms.
complexity of the information environment and the prevalence of If the association with age holds in future studies, there are a whole
misinformation. host of questions about the causes, mechanisms, and ramifications
We cannot definitively rule out the possibility that there is an omit- of the relationship that researchers should explore. First, how much
ted variable biasing our estimates, although we have included con- of the effect can be attributed to lack of digital or media literacy as
trols for many individual-level characteristics theoretically related to opposed to explanations rooted in cognitive abilities and memory?
acceptance of misinformation and willingness to share content online. Answering this question will require developing measures of digital
Even if our models are correctly specified, we use observational data literacy that can be behaviorally validated. Second, what is the role
that cannot provide causal evidence on the determinants of fake news of the (currently) unobserved News Feed and social network envi-
sharing. This study takes advantage of a novel and powerful new dataset ronment on people’s tendency to see, believe, and spread dubious
combining survey responses and digital trace data that overcomes content? How are consumption and spreading, if at all, related? How
well-known biases in sample selection and self-reports of online be- does social trust over networks mediate the relationship between age
havior (8, 9). However, we are still limited in our ability to collect and the sharing of misinformation?
*Column 2 summarizes characteristics of respondents who said in the survey that they have a Facebook account (i.e., they selected “Facebook” from the list of
response options to the question “Do you have accounts on any of the following social media sites?”). †Column 3 subsets to respondents (regardless of
their answer in the previous question) who consented to share Facebook profile information with the researchers. ‡P values are computed from t tests of
the difference in means between the sample of respondents who reported having a Facebook account and those who consented to provide access to their
profile data. §The final column subsets to those who shared any Facebook data at all that we were able to link back to the survey.
cooperation between academic researchers and the platforms them- genuine—posts that our respondents did not feel compelled to re-
selves (24). move at a later time. There may be an additional concern that some
types of people were more likely to delete fake news articles that they
posted, leading us to biased inferences. However, to the extent that
MATERIALS AND METHODS these characteristics are negatively correlated with the characteristics
Survey data that predict posting in the first place, such deletion activity (which
We designed and conducted a panel survey (fielded by online polling is likely very rare) should reduce noise in the data that would other-
firm YouGov) during the 2016 U.S. presidential election to under- wise be generated by split-second sharing decisions that are imme-
stand how social media use affects the ways that people learn about diately retracted.
politics during a campaign. In addition to including a rich battery of
individual-level covariates describing media use and social media use, Defining fake news
we were able to match many respondents to data on their actual The term fake news can be used to refer to a variety of different
Facebook behavior (see below). The survey had three waves. Wave 1 phenomena. Here, we largely adopted the use suggested in (25) of
was fielded 9 April to 1 May 2016 (3500 respondents), wave 2 was knowingly false or misleading content created largely for the purpose
fielded 9 September to 9 October 2016 (2635 respondents), and wave 3 of generating ad revenue. Given the difficulty of establishing a com-
was fielded 25 October to 7 November 2016 (2628 respondents). monly accepted ground-truth standard for what constitutes fake news,
our approach was to build on the work of both journalists and academ-
Facebook profile data ics who worked to document the prevalence of this content over the
We were able to obtain private Facebook profile data from a substan- course of the 2016 election campaign. In particular, we used a list of
tial subset of our survey respondents. Starting 16 November 2016, fake news domains assembled by Craig Silverman of BuzzFeed News,
2. H. Allcott, M. Gentzkow, Social media and fake news in the 2016 election. 22. B. Swire, U. K. H. Ecker, S. Lewandowsky, The role of familiarity in correcting inaccurate
J. Econ. Perspect. 31, 211–236 (2017). information. J. Exp. Psychol. Learn. Mem. Cogn. 43, 1948–1961 (2017).
3. A. Guess, B. Nyhan, J. Reifler, “Selective exposure to misinformation: Evidence from the 23. E. Bakshy, S. Messing, L. A. Adamic, Exposure to ideologically diverse news and opinion
consumption of fake news during the 2016 U.S. presidential campaign” (2016); on facebook. Science 348, 1130–1132 (2015).
http://www-personal.umich.edu/~bnyhan/fake-news-2016.pdf. 24. G. King, N. Persily, “A new model for industry-academic partnerships” (2018);
4. G. Pennycook, T. D. Cannon, D. G. Rand, Prior exposure increases perceived accuracy of https://gking.harvard.edu/files/gking/files/partnerships.pdf.
fake news. J. Exp. Psychol. Gen., in press. 25. J. A. Tucker, A. Guess, P. Barberá, C. Vaccari, A. Siegel, S. Sanovich, D. Stukal, B. Nyhan,
5. E. Porter, T. J. Wood, D. Kirby, Sex trafficking, Russian infiltration, birth certificates, and Social Media, Political Polarization, and Political Disinformation: A Review of the Scientific
pedophilia: A survey experiment correcting fake news. J. Exp. Polit. Sci. 5, 159–164 Literature (William and Flora Hewlett Foundation, 2018).
(2018). 26. C. Silverman, Here are 50 of the biggest fake news hits on facebook from 2016 (BuzzFeed,
6. S. Vosoughi, D. Roy, S. Aral, The spread of true and false news online. Science 359, 2016); www.buzzfeed.com/craigsilverman/top-fake-news-of-2016.
1146–1151 (2018). 27. D. Rivers, Sample Matching: Representative Sampling from Internet Panels (Polimetrix
7. C. Silverman, “This analysis shows how fake election news stories outperformed real White Paper Series, 2006).
news on facebook” (BuzzFeed, 2016); www.buzzfeed.com/craigsilverman/viral-fake-
election-news-outperformed-real-news-on-facebook.
Acknowledgments: This study has been approved by the New York University Institutional
8. A. M. Guess, Measure for measure: An experimental test of online political media
Review Board (IRB-12-9058 and IRB-FY2017-150). We thank S. Luks at YouGov for facilitating
exposure. Polit. Anal. 23, 59–75 (2015).
survey data collection and J. Ronen for development and technical assistance. Funding: This
9. A. Guess, K. Munger, J. Nagler, J. Tucker, How accurate are survey responses on social
research was supported by the INSPIRE program of the NSF (Award SES-1248077). Author
media and politics? Polit. Commun. 1–18 (2018).
contributions: A.G., J.N., and J.T. designed the research. J.N. and J.T. oversaw data collection.
10. A. Campbell, P. E. Converse, W. E. Miller, The American Voter (Wiley, 1960).
A.G. analyzed the data and produced the figures. A.G. drafted the paper, and all authors
11. D. Green, B. Palmquist, E. Schickler, Partisan Hearts and Minds (Yale Univ. Press, 2002).
contributed to revisions. Competing interests: J.T. received a small fee from Facebook to
12. C. A. Klofstad, The sociological bases of political preferences and behavior, in The Future
compensate him for time spent in organizing a 1-day conference for approximately 30
of Political Science: 100 Perspectives (Routledge, 2009).
academic researchers and a dozen Facebook product managers and data scientists that was
13. B. Schäffer, The digital literacy of seniors. Res. Comp. Int. Educ. 2, 29–42 (2007).
held at NYU in the summer of 2017 to discuss research related to civic engagement. He did not
14. B. B. Neves, F. Amaro, Too old for technology? How the elderly of lisbon use and perceive