CHAPTER 10
The Public and Its Problems
“Fake News” and the Battle for Hearts and Minds
Susan Maret

Following the major political events of 2016–2017 and the Facebook/ Cambridge Analytica revelations, phrases such as “alternative facts,” “post-fact,” “post-truth,” and “fake news” have deluged global channels of communication. Of these terms, the use of “fake news” is now so commonplace—and vulgarized—that it has been included on the annual “List of Words Banished from the Queen’s English for Misuse, Over-use and General Uselessness” as of January 1, 2018.1 These fuzzy terms point to larger social problems that not only concern the authority, credibility, and believability of information, but its very manipulation.2

Under the guise of facts, shades of false, unvetted information, plagiarized stories, and clickbait from nation-states, contractors, advertisers, social media, news conglomerates, and hidden actors flood public communication spaces, usurping the traditional 24-hour news cycle. The technological ability to influence information choices, as well as to predict, persuade, and engineer behavior through algorithms (a “recipe” or set of instructions carried out by computer) and bots (applications that run automated, repetitive tasks) now blurs the line between potentially meaningful information and micro/targeted messaging.3

Several research studies illustrate the deleterious effects of false information on public communication.4 For example, a survey by Pew Research Center and Elon University’s Imagining the Internet Center found that the “fake news ecosystem preys on some of our deepest human instincts.”5 A Pew Research Center Journalism and Media survey of 1,002 US adults revealed two-in-three Americans (64 percent) find that “fabricated news stories” create confusion over current issues and events and believe that this confusion is “shared widely across incomes, education levels, partisan affiliations, and most other demographic characteristics.”6 This survey also revealed that “16% of US adults say they have shared fake political news inadvertently, only discovering later that it was entirely made up.”7 Still other research uncovered the influence of fake news, propagated by algorithms and bots, on the 2016 US election8 and Cambridge Analytica’s role in the Brexit and Leave.EU movements.9

But there is something more telling about the influence of fake news, which pertains to global perceptions of the media and trust in their sources. An investigation of approximately 18,000 individuals across the United States, United Kingdom, Ireland, Spain, Germany, Denmark, Australia, France, and Greece found reduced trust in the media due to “bias, spin, and agendas.”10 That is, a “significant proportion of the public feel that powerful people are using the media to push their own political or economic interests”; moreover, “attempts to be even-handed also get the BBC and other public broadcasters into trouble. By presenting both sides of an issue side by side, this can give the impression of false equivalence.”11

In this chapter, I briefly outline the ways fake news is characterized in the research literature. I then discuss how fake news is manufactured and the global entities responsible for its propagation. I close the chapter by reporting on ongoing technological and educational initiatives and suggest several avenues in which to explore and confront this controversial, geopolitical social problem.

FAKE NEWS: AN UNSTABLE CONCEPT

The term “fake news” was reported as early as the 6th century CE, and persisted into the 18th century as a means of “diffusing nasty news . . . about public figures.”12 Merriam-Webster, however, situates fake news as “seeing general use at the end of the 19th century,” and defines it as news (“material reported in a newspaper or news periodical or on a newscast”) that is fake (“false, counterfeit”).13 The term “fake news” is “frequently used to describe a political story, which is seen as damaging to an agency, entity, or person . . . [I]t is by no means restricted to politics, and seems to have currency in terms of general news.”14 This expanded description by the esteemed dictionary allows for falsehood to be viewed within the sphere of harm, fallout, and consequence.

In addition to Merriam-Webster’s definition, scholars have divided “fake news” into categories such as commercially-driven sensational content, nation state–sponsored misinformation, highlypartisan news sites, social media, news satire, news parody, fabrication, manipulation, advertising, and propaganda.15 The use of the term “fake news” is now so prevalent that it is considered a “catchall phrase to refer to everything from news articles that are factually incorrect to opinion pieces, parodies and sarcasm, hoaxes, rumors, memes, online abuse, and factual misstatements by public figures that are reported in otherwise accurate news pieces.”16

Fake news is often framed as misinformation, disinformation, and propaganda in mainstream and scholarly literature, which complicates collective understanding of what constitutes “counterfeit” information and the actors behind its manufacture. Misinformation, for example, is defined as “information that is initially assumed to be valid but is later corrected or retracted. . . [though it] often has an ongoing effect on people’s memory and reasoning.”17 Misinformation is either intentional (e.g., manufactured for some sort of gain) or unintentional (e.g., incomplete fact-checking, “hasty reporting,” or sources who were misinformed or lying).18 Misinformation has been identified with outbreaks of violence.19 Misinformation is also described in relation to disinformation as “contentious information reflecting disagreement, whereas disinformation is more problematic, as it involves the deliberate alienation or disempowerment of other people.”20

Disinformation, as characterized by the European Commission’s High Level Expert Group (HLEG), “goes well beyond the term fake news.”21 Similar to Merriam-Webster’s definition of “fake news” and its emphasis on damage and resulting harm, the HLEG likens fake news to disinformation to include “all forms of false, inaccurate, or misleading information designed, presented and promoted to intentionally cause public harm or for profit.”22

To further muddy the waters, in an analysis of “mainstream and social media coverage” during the 2016 US election, disinformation was linked to “propaganda as the ‘intentional use of communications to influence attitudes and behavior in the target population’”; disinformation is the “communication of propaganda consisting of materially misleading information.”23 But such characterizations are not as seamless as they at first appear. A specific category of propaganda called black propaganda is described as “virtually indistinguishable” from disinformation and “hinges on absolute secrecy . . . usually supported by false documents.”24 These particular definitions perhaps supplement longstanding descriptions of propaganda as a “consistent, enduring effort to create or shape events to influence the relations of the public to an enterprise, idea or group” and as a “set of methods employed by an organized group that wants to bring about the active or passive participation in its actions of a mass of individuals, psychologically unified through psychological manipulations and incorporated in an organization.”25

Perhaps the most dramatic evolution in the study of fake news as it applies to Internet society is the proposal by Samuel C. Woolley and Philip N. Howard, who offer the term computational propaganda as the “assemblage of social media platforms, autonomous agents, and big data tasked with the manipulation of public opinion.”26 Computational propaganda—a mashup of technology, ideology, and political influence, coupled with deception and secrecy—results in calculated messaging designed to influence opinion and beliefs. Here it is critical to interject that some researchers find the term “fake news” “difficult to operationalize” and instead suggest in its place junk news, or content consisting of “various forms of propaganda and ideologically extreme, hyper-partisan, or conspiratorial political news and information.”27

Under certain conditions, “fake news” may be infused with noise or classed as information distortion. Described as “undoubtedly the most damaging to the clarification of a catastrophe,”28 information distortion occurred, for example, during the Las Vegas shootings when Google prominently displayed search results derived from 4chan, Twitter, and Facebook as “news.”29 Distortion also featured in the media coverage of Devin Patrick Kelley, who opened fire on 26 people in a Sutherland Springs, Texas, church on November 5, 2017. During the early hours of the disaster, Kelley was described as both a “liberal, Antifa communist working with ISIS” and a supporter of Bernie Sanders, creating “noise”—a conspiratorial, inaccurate, rumor-fed, politically-charged news stream run amok.30 A similar scenario occurred during the aftermath of the Parkland shootings. Researchers found that individuals

outraged by the conspiracy helped to promote it—in some cases far more than the supporters of the story. And algorithms—apparently absent the necessary “sentiment sensitivity” that is needed to tell the context of a piece of content and assess whether it is being shared positively or negatively—see all that noise the same.31

Sidestepping well-worn terms such as misinformation, disinformation, and propaganda, truth decay was recently suggested as a vital model for understanding the “increasing disagreement about facts and analytical interpretations of facts and data” and diminishing trust in sources of factual information.32

From this overview, it is apparent that “fake news” and related conditions of information suffer from epistemic fluidity. On one hand, recent attempts to flesh out fake news in relation to misinformation, disinformation, and propaganda potentially advance our previouslyheld notions. On the other hand, the lack of universally-established definitions and refinement of concepts creates Babel-like complexity in the journalism and research communities, resulting in a kitchensink approach to the study of falsehood and the disordered information-communication ecosystem where it thrives.

Falsehood—fake news—as misinformation, disinformation, and/ or propaganda, is a crisis of collective knowledge. Whatever definitions or concepts with which we choose to categorize the phenomenon, the reality is that falsehood contributes to a cacophonous, polluted information-sharing environment. The infosphere now resembles a light-pollution map. In confronting fake news, we now face the challenge described by Karl Mannheim in his Ideology and Utopia: “not how to deal with a kind of knowledge which shall be ‘truth in itself,’ but rather how man deals with his problems of knowing.”33

HOW FAKE NEWS HAPPENS

It is critical to acknowledge that fake news as counterfeit information is created by humans and/or humans tasking technologies to moderate information and/or disrupt communications. Often this moderation and disruption is conducted in secret by anonymous actors. For example, patented algorithms construct a search engine results page (SERP)’s appearance and content (e.g., ad words, knowledge panels, news boxes, rating stars, reviews, snippets), thus directing an individual’s information-seeking and research gaze; hackers or “pranksters” manipulate search results through the practice of “Google bombing”;34 and algorithms and bots fabricated by troll factories and/or nation-states actively engage in the shaping of information in order to sow confusion and discord. In the following section, I briefly illustrate how fake news is promoted by these search features and global entities.

The Invisible Hand

One product of a SERP, the snippet, is “extracted programmatically” from web content to include a summary of an answer based on a search query. Google describes the snippet as reflecting “the views or opinion of the site” from which it is extracted.35 The DuckDuckGo search engine also supplies snippets in the form of “Instant Answers” or “high-quality answers” situated above search results and ads. Instant Answers are pulled from news sources such as the BBC, but also include Wikipedia and “over 400 more community built answers.”36 The Russian web search engine Yandex also provides “interactive snippets,” or an “island,” to supply individuals with a “solution rather than an answer.”37 Although the snippet is not entirely “counterfeit information,” it is designed to confederate knowledge from across the web to make it appear “more like objective fact than algorithmically curated and unverified third-party content.”38

More to the point, third-party advertising networks connect advertisers with website creators to attract paid clicks by producing often misleading, sensational, tabloid-like information or ads that mimic news stories.39 To stem the rising tide of fake news monetization, Google disabled a total of 112 million ads that “trick to click,” with Facebook following suit with new guidelines that informed “creators and publishers” that they must have an “authentic, established presence.”40 However, Facebook’s “Content Guidelines for Monetization” do not include a category for false or deceptive information that would disqualify this type of content from monetization.41 In the end, “deciding what’s fake news can be subjective, and ad tech tends to shrug and take a ‘who am I to decide’” stance.42

At the Halifax International Security Forum in November 2017, Eric Schmidt, former executive chairman of Alphabet, Google’s parent company, disclosed that “it was easier for Google’s algorithm to handle false or unreliable information when there is greater consensus, but it’s more challenging to separate truth from misinformation when views are diametrically opposed.”43 To address this technological quandary, Google altered its secret patented algorithm to de-rank “fake news,” relegating it to a lower position on subsequent search result pages. The search giant also revised its “Search Quality Evaluator Guidelines” to assist its legion of search quality raters with “more detailed examples of low-quality webpages for raters to appropriately flag, which can include misleading information, unexpected offensive results, hoaxes and unsupported conspiracy theories.”44 Referring to Google’s multi-pronged approach, Schmidt stated that he is “strongly not in favour of censorship. I am very strongly in favour of ranking. It’s what we do.”45 In direct response to reports of Russia’s alleged election interference, Google de-ranked content from Russia Today (RT) and Sputnik.46 Contesting Google’s actions, an op-ed published on RT argued “there is nothing good or noble about de-ranking RT. It’s not a war against ‘fake news’ or misinformation. It’s a war of ideas on a much, much wider scale.”47

Google’s techno-fix may have artificially repaired one problem while posing others. First, research on information-seeking behavior suggests that individuals tend to focus on the first page of search results, often not venturing to “bottom” layers or pages that may potentially contain meaningful links to sites representing a diversity of views.48 Secondly, it is not only reactionary censorship we must guard against in the “war of ideas.” It is gatekeeping as well.49 In allowing behind-the-curtain algorithmic technology—the invisible hand—to distinguish “true” from “false” information without agreement, nuance, or context, the open pursuit of knowledge and ideas without boundaries is challenged on a fundamental level.50 In this regard, “the human right to impart information and ideas is not limited to ‘correct’ statements.” This information right also “protects information and ideas that may shock, offend and disturb.”51

Information Warfare and Information Operations

In addition to behind-the-curtain assemblages of information by search engines and algorithmic determination of search results, “fake news” is produced by way of information warfare and information operations (also known as influence operations).52 Information warfare (IW) is the “conflict or struggle between two or more groups in the information environment”;53 information operations (IO) are described in a Facebook security report as

actions taken by governments or organized non-state actors to distort domestic or foreign political sentiment, most frequently to achieve a strategic and/or geopolitical outcome. These operations can use a combination of methods, such as false news, disinformation, or networks of fake accounts (false amplifiers) aimed at manipulating public opinion.54

With its emphasis on damage, it is possible to apply MerriamWebster’s definition of “fake news” (and possibly the High Level Expert Group’s as well) to IW and IO, as they are conducted by cyber troops, or “government, military or political-party teams committed to manipulating public opinion.”55 These often anonymous actors from around the globe—invisible to oversight and regulation—are involved in delivering weaponized falsehoods to unwitting consumers of information via fake accounts.56 In these cases, deception is a key element in the manipulation of information.57

The earliest reports of organized social media manipulation came to light in 2010, with revelations in 2017 that there are “offensive” organizations in 28 countries.58 Less is known about corporations, such as Facebook’s global government and politics team, and private actors that moderate information and troll critics in order to influence public perceptions.59 With concealed geographic locations and secret allegiances, cyber troops engineer information machines to produce an embattled Internet. For instance, paid disinformation agents at the Internet Research Agency in St. Petersburg operated the “Jenna Abrams” Twitter account, which had 70,000 followers.60 Coined “Russia’s Clown Troll Princess,” “Abrams” produced tweets that were covered by the BBC, Breitbart News, CNN, InfoWars, the New York Times, and the Washington Post.61 The Internet Research Agency is also implicated in more than 400 fake Twitter accounts used to influence UK citizens regarding Brexit.62 In addition, the Guardian and BuzzFeed reported that the Macedonian town of Veles was the registered home of approximately 100 pro-Trump websites, many reporting fake news; “ample traffic was rewarded handsomely by automated advertising engines, like Google’s AdSense.”63 And to support the “infowar,” Alex Jones, founder of the alternative media sites infowars.com, newswars.com, prisonplanet.com, and europe. infowars.com, likened purchases from his online store to purchasing “war bonds, an act of resisting the enemy to support us and buy the great products.”64

Both IW and IO are useful in framing widespread allegations of “fake news” and micro/targeted messaging during major political events, including the 2016 US election,65 the French election coverage in 2016–17,66 the 2017 UK and German elections,67 and the 2018 Italian election.68 IW and IO potentially fueled “purported” fake news posted on the Qatar News Agency’s website (which inflamed relations between Iran and Israel),69 the spread of a false story about a $110 billion weapons deal between the US and Saudi Arabia (reported as true in such outlets as the New York Times and CNBC),70 and “saberrattling” between Israel and Pakistan after a false report circulated that Israel had threatened Pakistan with nuclear weapons.71

The conceptual framework of IW/IO has profound implications for the transgressions of Cambridge Analytica, whose techniques influenced the US and Nigerian elections.72 Information warfare and/or information operations as high-octane opposition research certainly describes Donald Trump’s former campaign manager Paul Manafort’s alleged involvement in a covert media operation. This operation included revision of Wikipedia entries “to smear a key opponent” of former Ukrainian president Viktor Yanukovych and activating a “social media blitz” aimed at European and US audiences.73 Ahead of the May 25, 2018 vote to repeal the 8th amendment to the Irish Constitution on a woman’s right to terminate her pregnancy, journalist Rossalyn Warren wrote that “Facebook and the public have focused almost solely on politics and Russian interference in the United States election. What they haven’t addressed is the vast amount of misinformation and unevidenced stories about reproductive rights, science, and health.”74

THE PUBLIC AND ITS PROBLEMS: ADDRESSING FAKE NEWS

In 1927, philosopher John Dewey remarked that, “until secrecy, prejudice, bias, misrepresentation, and propaganda as well as sheer ignorance are replaced by inquiry and publicity, we have no way of telling how apt for judgment of social policies the existing intelligence of the masses might be.”75 In the age of “post-truth” and “fake news,” Dewey’s comments might be construed as patronizing or perhaps elitist. But embedded in his remarks is respect for democratic values, which recognize the power of openness and the essential role of education in transforming social action. Below I report on ongoing technological and educational initiatives and suggest several interconnected steps to address the fake news challenge.

Educational Initiatives

Echoing John Dewey’s confidence in the formative power of education and literacy, several approaches originating from higher education can be employed across educational settings to address fake news. Since 1976, Project Censored’s faculty–student partnerships have developed the Validated Independent News (VIN) model, subjecting independent news stories to intense evaluation in order to validate them as “important, timely, fact-based, well documented, and under-reported in the corporate media.”76 Supporting multiple literacies, the Association of College & Research Libraries (ACRL)’s “Framework for Information Literacy for Higher Education” offers tools for librarians and faculty to emphasize “conceptual understandings that organize many other concepts and ideas about information, research, and scholarship into a coherent whole.”77 Both Project Censored’s VIN model and the ACRL’s “Framework” encourage the development of metaliteracies and metacognition, leading to increased awareness and skill regarding the creation of collective knowledge and the reporting of facts.78

Government Response to Fake News

Governmental response to the fake news problem varies internationally. In the US, for instance, the State Department’s Global Engagement Center (GEC), established during the Obama administration, aims to “counter the messaging and diminish the influence of international terrorist organizations and other violent extremists abroad.”79 The 2017 National Defense Authorization Act expanded the GEC to

identify current and emerging trends in foreign propaganda and disinformation in order to coordinate and shape the development of tactics, techniques, and procedures to expose and refute foreign misinformation and disinformation and proactively promote fact-based narratives and policies to audiences outside the United States.80

The US Congress also convened several high-profile hearings on combatting fake news and foreign disinformation.81

On a state level, the Internet: Social Media: False Information Strategic Plans (SB-1424), introduced into the California State Legislature in February 2018, would require “any person who operates a social media, as defined, Internet Web site with a physical presence in California” to create a strategic plan to “mitigate” the spread of false news. The proposed legislation, which duplicates numerous ongoing educational-literacy programs and NGO activities that address “fake news,” mandates the use of fact-checkers to verify news stories and outreach to social media users regarding stories that contain “false information” (which the proposed bill does not define), and requires social media platforms to place a warning on news that contains false information.82

European governmental agency responses to counterfeit news include legislation and regulatory stopgap measures. For example, Croatia, Ireland, Malaysia, and South Korea proposed legislation to counter fake news; legislation in France includes an emergency procedure that allows a judge to delete web content, close a user’s account, or block access to a website altogether.83 During its 2018 elections, the Italian government launched an online “red button” system for people to report “fake news” to the postal police, the federal agency responsible for monitoring online crime.84 Sweden created a “psychological defence” authority to counter fake news and disinformation,85 while the German Ministry of the Interior proposed a “Center of Defense Against Disinformation.”86 Linking fake news with national security concerns, and subtly with IO/IW, the United Kingdom is in the process of creating a national security communications unit to battle “disinformation by state actors and others.”87 In early February 2018, the UK Parliament’s 11-member Committee on Digital, Culture, Media and Sport traveled to Washington, DC, to convene an “evidence session” with Facebook, Google, and Twitter executives on the subject of fake news, and to meet with members of the US Senate Intelligence Committee regarding Russian influence on social media during the 2016 election.88

Going several steps further, Chinese President Xi Jinping appeared to suggest increasing control and censorship of China’s digital communications system when he remarked that “without Web security there’s no national security, there’s no economic and social stability, and it’s difficult to ensure the interests of the broader masses.”89 The Russian Federation’s Security Council recommended an alternate Domain Name System (DNS) for BRICS countries (Brazil, the Russian Federation, India, China, and South Africa), citing “increased capabilities of western nations to conduct offensive operations in the informational space.”90 If implemented, the new DNS would essentially create another layer of a walled-off, compartmentalized Internet in these countries.

NGOs and Social Media Companies

Nongovernmental organizations (NGOs) and social media companies have responded to the spread of counterfeit information by forming fact-checking programs and initiatives. Fact Tank, the International Fact-Checking Network at Poynter, Media Bias/Fact Check, Metabunk, Mozilla Information Trust Initiative, PolitiFact, and Snopes.com analyze specific, often memetic, news stories, while Google News and Google’s search engine now include a “fact check” tag that alerts information consumers that a particular slice of information was scrutinized by organizations and publishers. Google is now partnering with the Trust Project, which—much like food ingredient labels for fat and sugar content—developed a “core set” of indicators for news and media organizations that range from best practices to the use of citations, references, and research methodologies, to having a “commitment to bringing in diverse perspectives.”91 Facebook now requires political ads to reveal their funding sources, and the creation of a nongovernmental, voluntary accreditation system to “distinguish reliable journalism from disinformation” is another proposal toward stemming the creation and dissemination of fake news.92

Technological Solutions

Technological approaches to curbing fake news include the formation of the World Wide Web Consortium (W3C)’s Credible Web Community Group, which investigates “credibility indicators” as these apply to structured data formats. MisinfoCon is a “global movement focused on building solutions to online trust, verification, fact checking, and reader experience in the interest of addressing misinformation in all of its forms.”93 Browser add-ons, such as Project FiB, a Chrome-based extension, detect fake news on Facebook, Reddit, and Twitter, while the controversial PropOrNot browser plugin misreports websites as delivering “Russian propaganda” targeted toward Americans. The Hamilton 68 Dashboard, a project of the German Marshall Fund’s Alliance for Securing Democracy, monitors 600 Twitter accounts linked to Russian influence efforts online, “but not all of the accounts are directly controlled by Russia. The method is focused on understanding the behavior of the aggregate network rather than the behavior of individual users.”94

In the following discussion, I suggest additional interconnected steps to bring clarity to public and academic investigations of false information and to combat fake news on a local, national, and global scale.

Defend Journalistic Integrity and Protect Journalists

The preponderance of fake news traversing public channels of communication challenges established editorial practices, information and media ethics, transparency, publicity, and the public right to know. In addition to the proliferation of fake news, newsbots created and employed by media organizations have shifted human authorship to algorithmic or “robo-journalism.”95 The rise of elected officials claiming that the media spreads “fake news” by “lying” to the public increases “the risk of threats and violence against journalists” and undermines the public’s trust and confidence in members of the Fourth and Fifth Estates.96

Global Public/Information Policy

In the US, Congress must reconstitute the Office of Technology Assessment (OTA). For 23 years, OTA policy analysis included research on adult literacy, information security, emerging technologies, and US information infrastructure.97 On an international level, a nonpartisan body of stakeholders (e.g., the Tallinn Manual Process) should be formed to address the policy implications of tech platforms, algorithms, bots, and their use in weaponizing information.98 Above all, as Marietje Schaake suggests, “regulators need to be able to assess the workings of algorithms. This can be done in a confidential manner, by empowering telecommunications and competition regulators.”99 This proposed international body might also address the global implications of algorithmic accountability and algorithmic justice in order to confront “dissemination of knowingly or recklessly false statements by official or State actors” within the established framework of international human rights.100 Both OTA v.2 and the proposed international body suggested above would address the influence of fake news and other information conditions that impact the right to communicate and freedom of expression.

Re-Imagine Education and the Curriculum

Institutional support and funding for inter/multi/transdisciplinary courses and initiatives that address intermeshed literacies across the curriculum, continuing education, and community is imperative. Studies of how editorial practices and ethics contribute to the construction of knowledge in crowdsourced and academic reference works would be an essential part of this proposed curriculum. Such initiatives can only be realized through global partnerships of “intermediaries, media outlets, civil society, and academia.”101

Research as a Necessity and Global Social Good

Dewey observed that “tools of social inquiry will be clumsy as long as they are forged in places and under conditions remote from contemporary events.”102 Using Dewey as our North Star, theoretical forays into how “fake news” differs from various information conditions (e.g., misinformation, disinformation, or propaganda) serve to build a common language for deeper studies. Evidence-based research on those factors that influence the public’s information-seeking and media habits is essential. This research would go a long way toward explaining how and why individuals adopt certain ideas and are influenced by specific information. For example, research suggests that “the greater one’s knowledge about the news media . . . the less likely one will fall prey to conspiracy theories.”103 There is also a need for intensive qualitative and quantitative investigations into those actors who manipulate information for sociopolitical gain and to cause damage; one study indicates that “robots accelerated the spread of true and false news at the same rate, implying that false news spreads more than the truth because humans, not robots, are more likely to spread it.”104

Support Local Media Through Citizen/Community Journalism

One solution to the problem of fake news, media conglomeration, and prepackaged news is community-based journalism. In forging educational partnerships with communities (see above), universities, libraries, and organizations have the opportunity to cultivate a climate focused on freedom of information and skill building. These partnerships allow citizens, including student journalists, to investigate local perspectives, stories, and events that national, corporate media ignore, marginalize, or abbreviate. According to one account, citizen journalism “ranks low on revenues and readers. It ranks high on perceived value and impact. While it aspires to report on community, it aspires even more to build community.”105

In his book The Public and Its Problems, Dewey established that knowledge is social. As such, knowledge is a function of communication and association, which depend “upon tradition, upon tools and methods socially transmitted, developed, and sanctioned.”106As discussed in this chapter, “fake news” as falsehood disrupts established ways of knowing and communication. In its wake, fake news destabilizes the very trust in information required to sustain relationships across the social world.


SUSAN MARET, PHD, is a lecturer in the School of Information at San José State University, where she teaches courses on information secrecy and information ethics. Maret is the editor/author of several works on government secrecy and a founding member of the peer-reviewed online journal Secrecy and Society.

Notes
  1. “List of Words Banished from the Queen’s English for Mis-use, Over-use and General Uselessness,” Lake Superior State University, 2018, https://www.lssu.edu/banished-words-list/.
  2. From its roots, informatio: in (within) forma (form, shape, pattern) tio (action, process). Information can be thought of as giving form to meaning, ideas, knowledge, and action; this powerful thing called information is crucial to decision-making, self-determination, and social trust. Based on this “special nature” of information, unique moral dilemmas arise in its use. For a discussion of information ethics, see Richard O. Mason, Florence M. Mason, and Mary J. Culnan, Ethics of Information Management (Thousand Oaks, CA: SAGE Publications, 1995); and Luciano Floridi, “Information Ethics, Its Nature and Scope,” ACM SIGCAS Computers and Society 36, No. 3 (September 2006): 21–36, http://pages.uoregon.edu/koopman/courses_readings/phil123-net/intro/floridi_info-ethics_nature_scope.pdf.
  3. None of my discussion in this chapter is meant to suggest a “golden age” of manipulation-and bias-free communications. Numerous theorists and journalists have documented media bias and tampering: C. Wright Mills, Marshall McLuhan, Carl Bernstein, Noam Chomsky and Edward S. Herman, Michael and Christian Parenti, Robert McChesney, Chris Hedges, Carl Jensen, Project Censored, and the Senate Select Committee to Study Governmental Operations with Respect to Intelligence Activities (Church Committee).
  4. Susan Maret, “Intellectual Freedom and Government Secrecy,” in The Library Juice Press Handbook of Intellectual Freedom: Concepts, Cases, and Theories, eds. Mark Alfino and Laura Koltutsky (Sacramento, CA: Library Juice Press, 2014), 247–81.
  5. Janna Anderson and Lee Rainie, “The Future of Truth and Misinformation Online,” Pew Research Center, October 19, 2017, http://www.pewinternet.org/2017/10/19/the-future-of-truth-and-misinformation-online/.
  6. Michael Barthel, Amy Mitchell, and Jesse Holcomb, “Many Americans Believe Fake News is Sowing Confusion,” Pew Research Center, December 15, 2016, http://www.journalism.org/2016/12/15/many-americans-believe-fake-news-is-sowing-confusion/.
  7. Ibid.
  8. For example, Yochai Benkler, Robert Faris, Hal Roberts, and Ethan Zuckerman, “Breitbart-Led Right-Wing Media Ecosystem Altered Broader Media Agenda,” Columbia Journalism Review, March 3, 2017, https://www.cjr.org/analysis/breitbart-media-trump-harvard-study.php; Rob Faris, Hal Roberts, Bruce Etling, Nikki Bourassa, Ethan Zuckerman, and Yochai Benkler, “Partisanship, Propaganda, and Disinformation: Online Media and the 2016 U.S. Presidential Election,” Berkman Klein Center for Internet & Society at Harvard University, August 16, 2017, https://cyber.harvard.edu/publications/2017/08/mediacloud; and Chengcheng Shao, Giovanni Luca Ciampaglia, Onur Varol, Kaicheng Yang, Alessandro Flammini, and Filippo Menczer, “The Spread of Low-Credibility Content by Social Bots,” arXiv.org, May 24, 2018, https://arxiv.org/pdf/1707.07592.pdf. See also Maura Rocio Tellez and Olivia Jones, with Kenn Burrows and Rob Williams, “Big Data and Dark Money behind the 2016 Election,” in Censored 2018: Press Freedoms in a “Post-Truth” World, eds. Andy Lee Roth and Mickey Huff with Project Censored (New York: Seven Stories, 2017), 51–54.
  9. See University of Essex lecturer Dr. Emma L. Briant’s research, which unearthed how “Leave. EU sought to create an impression of ‘democracy’ and a campaign channeling public will, while creating deliberately ‘provocative’ communications to subvert it and win by channeling hateful propaganda.” Emma L. Briant, “Three Explanatory Essays Giving Context and Analysis to Submitted Evidence,” UK Parliament’s Digital, Culture, Media and Sport Committee: Evidence, April 16, 2018, https://www.parliament.uk/documents/commons-committees/culture-media-and-sport/Dr%20Emma%20Briant%20Explanatory%20Essays.pdf. See also Tellez and Jones, with Burrows and Williams, “Big Data and Dark Money,” in Censored 2018, 53.
  10. Nic Newman and Richard Fletcher, “Bias, Bullshit and Lies: Audience Perspectives on Low Trust in the Media,” Digital News Project, 2017, 5, https://reutersinstitute.politics.ox.ac.uk/ our-research/bias-bullshit-and-lies-audience-perspectives-low-trust-media.
  11. Ibid.
  12. Robert Darnton, “The True History of Fake News,” New York Review of Books, February 13, 2017, https://www.nybooks.com/daily/2017/02/13/the-true-history-of-fake-news/. See also David Uberti, “The Real History of Fake News,” Columbia Journalism Review, December 15, 2016, https://www.cjr.org/special_report/fake_news_history.php.
  13. “The Real Story of ‘Fake News,’” Merriam-Webster, March 23, 2017, https://www.merriam-webster.com/words-at-play/the-real-story-of-fake-news.
  14. Ibid.
  15. James Titcomb and James Carson, “Fake News: What Exactly is It—and Can It Really Swing an Election?” The Telegraph, November 14, 2017, http://www.telegraph.co.uk/technology/0/fake-news-exactly-has-really-had-influence/ [updated, and renamed “Fake News: What Exactly is It—and How Can You Spot It?” on May 30, 2018]; and Edson C. Tandoc Jr., Zheng Wei Lim, and Richard Ling, “Defining ‘Fake News’: A Typology of Scholarly Definitions,” Digital Journalism 6, No. 2 (2018): 137–53.
  16. Jen Weedon, William Nuland, and Alex Stamos, “Information Operations and Facebook,” Facebook, April 27, 2017, v.1.0, 4, https://www.mm.dk/wp-content/uploads/2017/05/facebook-and-information-operations-v1.pdf.
  17. Ullrich K.H. Ecker, Stephan Lewandowsky, Olivia Fenton, and Kelsey Martin, “Do People Keep Believing Because They Want To? Preexisting Attitudes and the Continued Influence of Misinformation,” Memory & Cognition 42, No. 2 (February 2014): 292–304, 292.
  18. See Melanie C. Green and John K. Donahue, “The Effects of False Information in News Stories,” in Misinformation and Mass Audiences, eds. Brian G. Southwell, Emily A. Thorson, and Laura Sheble (Austin: University of Texas Press, 2018), 109–23. Information design and organization may also influence the integrity and veracity of information. This important detail goes unmentioned in most discussions.
  19. See, for example, Amanda Taub and Max Fisher, “Where Countries are Tinderboxes and Facebook is a Match: False Rumors Set Buddhist against Muslim in Sri Lanka, the Most Recent in a Global Spate of Violence Fanned by Social Media,” New York Times, April 21, 2018, https:// www.nytimes.com/2018/04/21/world/asia/facebook-sri-lanka-riots.html.
  20. Brian G. Southwell, Emily A. Thorson, and Laura Sheble, “Misinformation among Mass Audiences as a Focus for Inquiry,” in Misinformation and Mass Audiences, eds. Brian G. Southwell, Emily A. Thorson, and Laura Sheble (Austin: University of Texas Press, 2018), 1–11, 2.
  21. According to the HLEG, as a term, “fake news” “has been appropriated and used misleadingly by powerful actors to dismiss coverage that is simply found disagreeable.” High Level Expert Group, “Final Report of the High Level Expert Group on Fake News and Online Disinformation,” European Commission, March 12, 2018, 5, https://ec.europa.eu/digital-single-market/en/news/final-report-high-level-expert-group-fake-news-and-online-disinformation.
  22. Ibid.
  23. Faris et al., “Partisanship, Propaganda, and Disinformation,” 19.
  24. Victor Marchetti and John D. Marks, The CIA and the Cult of Intelligence (New York: Alfred A. Knopf, 1974), 165.
  25. Edward L. Bernays, Propaganda (New York: H. Liveright, 1928), 25; and Jacques Ellul, Propaganda: The Formation of Men’s Attitudes, trs. Konrad Kellen and Jean Lerner (New York: Vintage Books, 1965 [French orig. 1962]), 61.
  26. Samuel C. Woolley and Philip N. Howard, “Political Communication, Computational Propaganda, and Autonomous Agents,” International Journal of Communication 10 (2016): 4882–90, 4886.
  27. Philip N. Howard, Samantha Bradshaw, Bence Kollanyi, Clementine Desigaud, and Gillian Bolsover, “Junk News and Bots during the French Presidential Election: What are French Voters Sharing over Twitter?” COMPROP Data Memo, April 22, 2017, 3, http://comprop.oii.ox.ac.uk/wp-content/uploads/sites/93/2017/04/What-Are-French-Voters-Sharing-Over-Twitter-v10-1.pdf.
  28. José Vicente García-Santamaría, “The Crisis of Investigative Journalism in Spain: The Journalism Practice in the Spanair Accident,” Revista Latina de Comunicación Social 65 (2010), http://www.revistalatinacs.org/10/art3/916_UC3M/38_Santamaria.html [author translation]; see also Claude E. Shannon, “Communication in the Presence of Noise,” Proceedings of the IEEE 86, No. 2 (1949/1998), 447–57.
  29. Alfred Ng, “How Social Media Trolls Capitalized on the Las Vegas Shooting,” CNET, October 2, 2017, https://www.cnet.com/news/las-vegas-shooting-fake-news-hoax-social-media-facebook-twitter/.
  30. Dana Liebelson and Paul Blumenthal, “The Texas Shooter was Called a Liberal, Antifa Communist Working with ISIS—Before Anyone Knew Anything,” Huffington Post, November 7, 2017, updated November 8, 2017, https://www.huffingtonpost.com/entry/texas-shooter-conspiracy-theories_us_5a01ed2be4b092053058499e; and Christal Hayes, “Texas Church Shooter was a Bernie Sanders Supporter, Alt-Right Claims in Latest Lie,” Newsweek, November 16, 2017, http://www.newsweek.com/texas-church-shooter-was-bernie-sanders-supporter-altright-claims-latest-lie-713677.
  31. Molly McKew, “How Liberals Amped Up a Parkland Shooting Conspiracy Theory,” Wired, February 27, 2018, https://www.wired.com/story/how-liberals-amped-up-a-parkland-shootingconspiracy-theory/.
  32. Jennifer Kavanagh and Michael D. Rich, “Truth Decay: An Initial Exploration of the Diminishing Role of Facts and Analysis in American Public Life,” RAND Corporation, 2018, https://www.rand.org/pubs/research_reports/RR2314.html.
  33. Karl Mannheim, Ideology and Utopia: Introduction to the Sociology of Knowledge (New York: Harcourt, Brace & Co, 1954), 168.
  34. Techniques that artificially increase a site’s page rank. See Karen Wickre, “Google Bombs are Our New Normal,” Wired, October 11, 2017, https://www.wired.com/story/google-bombs-are-our-new-normal/.
  35. “Featured Snippets in Search,” Google, undated, https://support.google.com/webmasters/ answer/6229325?hl=en.
  36. Adam S. Cochran, answer to the question “What Search Algorithm Does DuckDuckGo Run in Order to Search? Is It Similar to PageRank?” Quora, May 20, 2015, https://www.quora.com/What-search-algorithm-does-DuckDuckGo-run-in-order-to-search-Is-it-similar-to-PageRank.
  37. “Yandex Takes Leap with Its New Interactive Search Platform—First in Turkey,” Yandex, June 5, 2013, https://yandex.com/company/blog/22/.
  38. Jamie Condliffe, “Google’s Algorithms May Feed You Fake News and Opinion as Fact,” MIT Technology Review, March 6, 2017, https://www.technologyreview.com/s/603796/googles-algorithms-may-feed-you-fake-news-and-opinion-as-fact/.
  39. Joshua Gillin, “The More Outrageous, the Better: How Clickbait Ads Make Money for Fake News Sites,” PunditFact (PolitiFact), October 4, 2017, http://www.politifact.com/punditfact/article/2017/oct/04/more-outrageous-better-how-clickbait-ads-make-mone/.
  40. Scott Spencer, “How We Fought Bad Ads, Sites and Scammers in 2016,” Google, January 25, 2017, https://blog.google/topics/ads/how-we-fought-bad-ads-sites-and-scammers-2016/; and Nick Grudin, “Standards and Guidelines for Earning Money from Your Content on Facebook,” Facebook, September 13, 2017, https://media.fb.com/2017/09/13/standards-and-guidelines-for-earning-money-from-your-content-on-facebook/. Monetization is a source of major controversy for established brands that find their ads on certain YouTube channels. A CNN investigation discovered that ads from approximately 300 companies (e.g., Adidas, Amazon, Cisco, Facebook, LinkedIn, Netflix, Nordstrom) and organizations (e.g., the US Department of Transportation, Centers for Disease Control and Prevention) “may have unknowingly helped finance” channels promoting white nationalism, Nazis ideology, “conspiracy theories,” and pedophilia. See Paul P. Murphy, Kaya Yurieff, and Gianluca Mezzofiore, “YouTube Ran Ads from Hundreds of Brands on Extremist Channels,” CNN Tech, April 20, 2018, http://money.cnn.com/2018/04/19/technology/youtube-ads-extreme-content-investigation/.
  41. “Content Guidelines for Monetization,” Facebook, undated, https://www.facebook.com/facebookmedia/get-started/monetization_contentguidelines.
  42. Lucia Moses, “‘The Underbelly of the Internet’: How Content Ad Networks Fund Fake News,” Digiday, November 28, 2016, https://digiday.com/media/underbelly-internet-fake-news-getsfunded/.
  43. Liam Tung, “Google Alphabet’s Schmidt: Here’s Why We Can’t Keep Fake News out of Search Results,” ZDNet, November 23, 2017, http://www.zdnet.com/article/google-alphabetsschmidt-heres-why-we-cant-keep-fake-news-out-of-search-results/.
  44. Ben Gomes, “Our Latest Quality Improvements for Search,” The Keyword (Google), April 25, 2017, https://blog.google/products/search/our-latest-quality-improvements-search/; and Mark Bergen, “Google Rewrites Its Powerful Search Rankings to Bury Fake News,” Bloomberg, April 25, 2017, https://www.bloomberg.com/news/articles/2017-04-25/google-rewritesits-powerful-search-rankings-to-bury-fake-news.
  45. Tung, “Google Alphabet’s Schmidt.”
  46. In its report, the Office of the Director of National Intelligence labeled RT as the “Kremlin’s principal international propaganda outlet” and cited its involvement with WikiLeaks in influencing the 2016 US election. See “Background to ‘Assessing Russian Activities and Intentions in Recent US Elections’: The Analytic Process and Cyber Incident Attribution,” Office of the Director of National Intelligence, January 6, 2017, 3, https://www.dni.gov/files/documents/ICA_2017_01.pdf.
  47. Danielle Ryan, “Google’s De-Ranking of RT in Search Results is a Form of Censorship and Blatant Propaganda,” RT, November 26, 2017, https://www.rt.com/op-edge/410981-google-rtcensorship-propaganda/.
  48. Bing Pan, Helene Hembrooke, Thorsten Joachims, Lori Lorigo, Geri Gay, and Laura Granka, “In Google We Trust: Users’ Decisions on Rank, Position, and Relevance,” Journal of Computer-Mediated Communication 12, No. 3 (April 2007): 801–23. In addition, a report by Amy Gesenhues notes that a study by SurveyMonkey found that some individuals appear to trust search results labeled “Google” rather than “Bing,” even when search results were swapped out. Taken together, research on individuals not scanning past the first page of search results, their preference for Google’s brand, and Google’s decision to de-rank results raises many questions regarding the depth of web users’ information/knowledge-seeking habits. See Amy Gesenhues, “Study: Many Searchers Choose Google Over Bing Even When Google’s Name is on Bing’s Results,” Search Engine Land, April 15, 2013, https://searchengineland.com/usersprefer-google-even-when-155682.
  49. Engin Bozdag, “Bias in Algorithmic Filtering and Personalization,” Ethics and Information Technology 15, No. 3 (September 2013): 209–27.
  50. For example, Facebook came under attack for tagging as “child pornography” Nick Ut’s awardwinning 1972 Vietnam War photograph of then nine-year-old Phan Thi Kim Phuc fleeing from a napalm attack. This points to the inability of algorithms to clearly discriminate between types of information. See Mark Scott and Mike Isaac, “Facebook Restores Iconic Vietnam War Photo It Censored for Nudity,” New York Times, September 9, 2016, https://www.nytimes.com/2016/09/10/technology/facebook-vietnam-war-photo-nudity.html. For another example, see Dutch Parliament member Marietje Schaake’s video, which mentioned the word “torture” and was subsequently flagged by a Google algorithm as “spam” and removed from YouTube. Marietje Schaake, “Algorithms Have Become So Powerful We Need a Robust, Europe-Wide Response,” The Guardian, April 4, 2018, https://www.theguardian.com/commentisfree/2018/apr/04/algorithms-powerful-europe-response-social-media.
  51. Organization for Security and Co-operation in Europe, “Joint Declaration on Freedom of Expression and ‘Fake News’, Disinformation and Propaganda,” March 3, 2017, 1, https://www.osce.org/fom/302796.
  52. Martin C. Libicki writes that information warfare, “as a separate technique of waging war, does not exist.” Libicki organizes seven distinct categories of information warfare, “each laying claim to the larger concept”: command-and-control warfare, intelligence-based warfare, electronic warfare (radio-electronic or cryptographic techniques), psychological warfare (in which information is used to change the minds of friends, neutrals, and foes), “hacker” warfare (in which computer systems are attacked), economic information warfare (blocking information or channeling it to pursue economic dominance), and cyberwarfare (“a grab bag of futuristic scenarios”). See his What is Information Warfare? ([Washington, DC]: Center for Advanced Concepts and Technology, Institute for National Strategic Studies, National Defense University, 1995), http://www.dtic.mil/get-tr-doc/pdf?AD=ADA367662.
  53. Isaac R. Porche III, Christopher Paul, Michael York, Chad C. Serena, Jerry M. Sollinger, Elliot Axelband, Endy M. Daehner, and Bruce Held, Redefining Information Warfare Boundaries for an Army in a Wireless World (Santa Monica, CA: RAND Corporation, 2013), xv, https://www.rand.org/pubs/monographs/MG1113.html.
  54. Weedon, Nuland, and Stamos, “Information Operations and Facebook,” 5.
  55. Samantha Bradshaw and Philip N. Howard, “Troops, Trolls and Troublemakers: A Global Inventory of Organized Social Media Manipulation,” Working Paper No. 2017.12, Computational Propaganda Research Project, University of Oxford, 2017, 4, http://comprop.oii.ox.ac.uk/wp-content/uploads/sites/89/2017/07/Troops-Trolls-and-Troublemakers.pdf.
  56. Ibid.; Emerson T. Brooking and P.W. Singer, “War Goes Viral: How Social Media is being Weaponized Across the World,” The Atlantic, November 2016, https://www.theatlantic.com/magazine/archive/2016/11/war-goes-viral/501125/; and Sara Fischer, “How Bots and Fake Accounts Work,” Axios, October 31, 2017, https://www.axios.com/how-bots-and-fakeaccounts-work-2503972339.html.
  57. In her classic account of deception, Sissela Bok wrote that “deceptive messages, whether or not they are lies,” are affected by self-deception, error, and “variations in the actual intent to deceive.” Further, and this is central to the matter at hand, Bok identified three “filters”—thickness, distortion, and color—that alter how a message is experienced by both the deceived and the deceivers. Importantly, Bok stated that “someone who intends to deceive can work with these filters and manipulate them; he can play with the biases of some persons, the imagination of others, and on errors and confusion throughout the system.” See her Lying: Moral Choice in Public and Private Life (New York: Vintage Books, 1978/1989), 15.
  58. Bradshaw and Howard, “Troops, Trolls and Troublemakers,” 3.
  59. Lauren Etter, Vernon Silver, and Sarah Frier, “How Facebook’s Political Unit Enables the Dark Art of Digital Propaganda,” Bloomberg, December 21, 2017, https://www.bloomberg.com/ews/features/2017-12-21/inside-the-facebook-team-helping-regimes-that-reach-out-andcrack-down. See also Facebook CEO Mark Zuckerberg’s April 10, 2018 revelations before the Senate Committee on the Judiciary that Facebook not only collects data on nonusers for “security reasons,” but doesn’t provide any opt-out for non-Facebook subscribers. “Transcript of Zuckerberg’s Appearance before House Committee,” Washington Post, April 11, 2018, https://www.washingtonpost.com/news/the-switch/wp/2018/04/11/transcript-of-zuckerbergsappearance-before-house-committee/.
  60. See Ben Collins and Joseph Cox, “Jenna Abrams, Russia’s Clown Troll Princess, Duped the Mainstream Media and the World,” Daily Beast, November 2, 2017, https://www.thedailybeast.com/jenna-abrams-russias-clown-troll-princess-duped-the-mainstream-media-and-the-world; and Maya Kosoff, “The Russian Troll Farm that Weaponized Facebook Had American Boots on the Ground: The Shadowy Internet Research Agency Duped American Activists into Holding Protests and Self-Defense Classes,” Vanity Fair, October 18, 2017, https://www.vanityfair.com/news/2017/10/the-russian-troll-farm-that-weaponized-facebook-had-american-boots-on-theground.
  61. Collins and Cox, “Jenna Abrams.”
  62. Robert Booth, Matthew Weaver, Alex Hern, Stacee Smith, and Shaun Walker, “Russia Used Hundreds of Fake Accounts to Tweet about Brexit, Data Shows,” The Guardian, November 14, 2017, https://www.theguardian.com/world/2017/nov/14/how-400-russia-run-fake-accountsposted-bogus-brexit-tweets.
  63. Samanth Subramanian, “Inside the Macedonian Fake-News Complex,” Wired, February 15, 2017, https://www.wired.com/2017/02/veles-macedonia-fake-news/.
  64. Alex Jones, The Alex Jones Show (Infowars), episode of November 29, 2017 (podcast on iTunes), http://rss.infowars.com/20171129_Wed_Alex.mp3, at 1:40:34. See also Cyberwar, S02e01, titled “The Great Meme War,” Viceland, October 3, 2017, https://www.viceland.com/en_us/video/cyberwar-the-great-meme-war/595f95afd978e31b73496a7e, where alt-right blogger and founder of GotNews Charles C. (Chuck) Johnson calls for the creation of a “citizen troll” army to make “journalism indistinguishable from trolling.” Perhaps we can add “culture war” to the scope of IW and IO.
  65. Office of the Director of National Intelligence, “Background to ‘Assessing Russian Activities’”; Senate Intelligence Committee, “Facebook, Google, and Twitter Executives on Russia Election Interference,” C-SPAN video of the hearing in Washington, DC, November 1, 2017, https://www.c-span.org/video/?436360-1/facebook-google-twitter-executives-testify-russias-influence-2016-election&live [the text of which is also available as Senate Report 105-1]; and Kent Walker and Richard Salgado, “Security and Disinformation in the U.S. 2016 Election,” The Keyword (Google), October 20, 2017, https://www.blog.google/topics/public-policy/security-and-disinformation-us2016-election/.
  66. “Fake News: Five French Election Stories Debunked,” BBC News, March 15, 2017, http://www.bbc.com/news/world-europe-39265777.
  67. See, for example, Horand Knaup and Gerald Traufetter, “Innenministerium will Abwehrzentrum gegen Falschmeldungen einrichten” (“Ministry of the Interior Wants to Set Up a Defense Center against False Reports”), Der Spiegel, December 23, 2016, http://www.spiegel.de/netzwelt/netzpolitik/fake-news-bundesinnenministerium-will-abwehrzentrum-einrichten-a-1127174.html; and Peter Walker, “New National Security Unit Set Up to Tackle Fake News in UK,” The Guardian, January 23, 2018, https://www.theguardian.com/politics/2018/jan/23/new-national-security-unit-will-tackle-spread-of-fake-news-in-uk.
  68. Yasmeen Serhan, “Italy Scrambles to Fight Misinformation Ahead of Its Elections,” The Atlantic, February 24, 2018, https://www.theatlantic.com/international/archive/2018/02/europe-fake-news/551972/.
  69. Jon Gambrell, “Hack, Fake Story Expose Real Tensions Between Qatar, Gulf,” Associated Press, May 24, 2017, https://apnews.com/f5da3293be18401a954d48249f75394e.
  70. Bruce Riedel, “The $110 Billion Arms Deal to Saudi Arabia is Fake News,” Brookings Institution, June 5, 2017, https://www.brookings.edu/blog/markaz/2017/06/05/the-110-billion-arms-deal-to-saudi-arabia-is-fake-news/.
  71. Russell Goldman, “Reading Fake News, Pakistani Minister Directs Nuclear Threat at Israel,” New York Times, December 24, 2016, https://www.nytimes.com/2016/12/24/world/asia/pakistan-israel-khawaja-asif-fake-news-nuclear.html.
  72. Carole Cadwalladr, “Revealed: Graphic Video Used by Cambridge Analytica to Influence Nigerian Election,” The Guardian, April 4, 2018, https://www.theguardian.com/uk-news/2018/apr/04/cambridge-analytica-used-violent-video-to-try-to-influence-nigerian-election.
  73. Luke Harding, “Former Trump Aide Approved ‘Black Ops’ to Help Ukraine President,” The Guardian, April 5, 2018, https://www.theguardian.com/us-news/2018/apr/05/ex-trump-aidepaul-manafort-approved-black-ops-to-help-ukraine-president.
  74. Rossalyn Warren, “Facebook is Ignoring Anti-Abortion Fake News,” New York Times, November 10, 2017, https://www.nytimes.com/2017/11/10/opinion/facebook-fake-news-abortion.html. See also Laura Hazard Owen, “Can Facebook Beat Back the Fake News in Ireland’s Upcoming Vote on Abortion?” NiemanLab, April 20, 2018, http://www.niemanlab.org/2018/04/can-facebook-beat-back-the-fake-news-in-irelands-upcoming-vote-on-abortion/.
  75. John Dewey, The Public and Its Problems: An Essay in Political Inquiry (Chicago: Gateway Books, 1927/1946), 209.
  76. “Validated Independent News,” Project Censored, undated, https://projectcensored.org/category/validated-independent-news/. See also Andy Lee Roth and Project Censored, “Breaking the Corporate News Frame through Validated Independent News Online,” in Media Education for a Digital Generation, eds. Julie Frechette and Rob Williams (New York: Routledge, 2016), 173–86.
  77. The “Framework” is organized as “a cluster of interconnected core concepts” (or “frames”) that include authority as constructed and contextual, information creation as a process, information as valuable, research as inquiry, scholarship as conversation, and searching as strategic exploration. See “Framework for Information Literacy for Higher Education,” Association of College & Research Libraries (American Library Association), January 11, 2016, http://www.ala.org/acrl/standards/ilframework.
  78. Ibid.
  79. Barack Obama, “Executive Order 13721: Developing an Integrated Global Engagement Center to Support Government-Wide Counterterrorism Communications Activities Directed Abroad and Revoking Executive Order 13584,” Federal Register, March 14, 2016, published March 17, 2016, https://www.federalregister.gov/documents/2016/03/17/2016-06250/developing-anintegrated-global-engagement-center-to-support-government-wide-counterterrorism.
  80. It is not clear how the Global Engagement Center counters fake news on the Internet. See the Senate Committee on Armed Services, National Defense Authorization Act for Fiscal Year 2017, S. 2943, § 1287: Global Engagement Center, PL 114-328, December 23, 2016, Congress.gov, last updated December 23, 2016, https://www.congress.gov/bill/114th-congress/senate-bill/2943/text#toc-H533D0AE113D24D90A6BD9F8A0B9D679C. See also Samuel Mathias Ditlinger and Tom Field, with Andy Lee Roth and Mickey Huff, “US Quietly Established New ‘AntiPropaganda’ Center,” in Censored 2018: Press Freedoms in a “Post-Truth” World, eds. Andy Lee Roth and Mickey Huff with Project Censored (New York: Seven Stories Press, 2017), 71–73.
  81. US House of Representatives Permanent Select Committee on Intelligence, “Russia Investigative Task Force Hearing with Former Secretary of Homeland Security Jeh Johnson,” C-SPAN video of the hearing in Washington, DC, June 21, 2017, https://www.c-span.org/video/?c4674447/trump-unwitting-agent-russia-2016-election; and the November 1, 2017 hearing by the Senate Intelligence Committee, “Facebook, Google, and Twitter Executives on Russia Election Interference.” The latter hearing investigated “incendiary” Russian ads. See Nicholas Fandos, Cecilia Kang, and Mike Isaac, “House Intelligence Committee Releases Incendiary Russian Social Media Ads,” New York Times, November 1, 2017, https://www.nytimes.com/2017/11/01/us/politics/russia-technology-facebook.html.

    The 2015 House Committee on Foreign Affairs, Confronting Russia’s Weaponization of Information: Hearing Before the Committee on Foreign Affairs, House of Representatives, 114-1, April 15, 2015 (Washington, DC: US Government Publishing Office, 2015), online at House.gov, https://docs.house.gov/meetings/FA/FA00/20150415/103320/HHRG-114-FA00-Transcript-20150415.pdf, undeservedly received less media scrutiny.

  82. SB-1424 Internet: Social Media: False Information Strategic Plans, February 16, 2018, amended in the California State Senate, March 22, 2018, published on California Legislative Information on May 25, 2018, https://leginfo.legislature.ca.gov/faces/billTextClient.xhtml?bill_id=201720180SB1424.
  83. Reuters Staff, “Macron Plans Law to Fight ‘Fake News’ in 2018,” Reuters, January 3, 2018, https://www.reuters.com/article/us-france-macron/macron-plans-law-to-fight-fake-news-in-2018-idUSKBN1ES1LJ.
  84. Angela Giuffrida, “Italians Asked to Report Fake News to Police in Run-Up to Election,” The Guardian, January 19, 2018, https://www.theguardian.com/world/2018/jan/19/italians-asked-report-fake-news-police-run-up-election.
  85. “Sweden to Create New Authority Tasked with Countering Disinformation,” The Local, January 15, 2018, https://www.thelocal.se/20180115/sweden-to-create-new-authority-tasked-with-countering-disinformation.
  86. Knaup and Traufetter, “Innenministerium will Abwehrzentrum gegen Falschmeldungen einrichten.”
  87. Walker, “New National Security Unit Set Up.”
  88. As this chapter goes to publication in May 2018, there is little public information available on this meeting.
  89. Adam Jourdan, “China’s Xi Says Internet Control Key to Stability,” Reuters, April 21, 2018, https://www.reuters.com/article/us-china-internet/chinas-xi-says-internet-control-key-to-stability-idUSKBN1HS0BG.
  90. Patrick Tucker, “Russia will Build Its Own Internet Directory, Citing US Information Warfare,” Defense One, November 28, 2017, http://www.defenseone.com/technology/2017/11/russia-will-build-its-own-internet-directory-citing-us-information-warfare/142822/.
  91. “Who We Are, What We Do,” The Trust Project, 2017, https://thetrustproject.org.
  92. Anna Gonzalez and David Schulz, “Helping Truth with Its Boots: Accreditation as an Antidote to Fake News,” Yale Law Journal Forum, October 9, 2017: 315–36, https://www.yalelawjournal.org/pdf/GonzalezandSchulz_b6fvqdro.pdf.
  93. “About,” MisinfoCon, undated, https://misinfocon.com/about. One fly in the ointment here is that “misinformation” does not distinguish between disinformation, propaganda, and other conditions of information.
  94. Hamilton 68 monitors “accounts likely controlled by Russian government influence operations,” “‘patriotic’ pro-Russia users that are loosely connected or unconnected to the Russian government, but which amplify themes promoted by Russian government media,” and “users who have been influenced by the first two groups and who are extremely active in amplifying Russian media themes. These users may or may not understand themselves to be part of a pro-Russian social network.” For additional details on methods, see “The Methodology of the Hamilton 68 Dashboard,” Alliance for Securing Democracy, August 7, 2017, http://securingdemocracy.gmfus.org/publications/methodology-hamilton-68-dashboard.
  95. Shelley Podolny, “If an Algorithm Wrote This, How Would You Even Know?” New York Times, March 7, 2015, https://www.nytimes.com/2015/03/08/opinion/sunday/if-an-algorithm-wrote-this-how-would-you-even-know.html; and Joe Keohane, “What News-Writing Bots Mean for the Future of Journalism,” Wired, February 16, 2017, https://www.wired.com/2017/02/robots-wrote-this-story/.
  96. Organization for Security and Co-operation in Europe, “Joint Declaration.”
  97. “The OTA Legacy,” Office of Technology Assessment, Princeton University, undated, https://www.princeton.edu/ota/.
  98. Michael N. Schmitt, ed., Tallinn Manual 2.0 on the International Law Applicable to Cyber Operations, prepared by the International Groups of Experts at the invitation of the NATO Cooperative Cyber Defence Centre of Excellence (New York: Cambridge University Press, 2017).
  99. Schaake, “Algorithms Have Become So Powerful.”
  100. “Algorithmic Accountability: Applying the Concept to Different Country Contexts,” World Wide Web Foundation, July 2017, https://webfoundation.org/docs/2017/07/Algorithms_ Report_WF.pdf; and Organization for Security and Co-operation in Europe, “Joint Declaration,” 1.
  101. Organization for Security and Co-operation in Europe, “Joint Declaration,” 5.
  102. Dewey, The Public and Its Problems, 181.
  103. Stephanie Craft, Seth Ashley, and Adam Maksl, “News Media Literacy and Conspiracy Theory Endorsement,” Communication and the Public 2, No. 4 (2017): 388–401, 388, 395.
  104. Soroush Vosoughi, Deb Roy, and Sinan Aral, “The Spread of True and False News Online,” Science 359, No. 6380 (2018): 1146–51, 1146.
  105. Jan Schaffer, “Citizen Media: Fad or the Future of News?—The Rise and Prospects of Hyperlocal Journalism,” J-Lab: The Institute for Interactive Journalism (Philip Merrill College of Journalism, University of Maryland, College Park), 2007, 7, http://www.j-lab.org/wpcontent/pdfs/citizen-media-report.pdf.
  106. Dewey, The Public and Its Problems, 158.