Skip to main content
SearchLogin or Signup

4    An Encyclopedia with Breaking News

Published onOct 15, 2020
4    An Encyclopedia with Breaking News
·

Wikipedia’s response to the September 11 attacks profoundly shaped its rules and identity and illuminated a new strategy for growing the project by coupling the supply and demand for information about news. Wikipedia’s breaking news collaborations offer lessons for hardening other online platforms against polarization, disinformation, and other sociotechnical sludge.

The web was a very different place for news in the United States between 2001 and 2006. The hanging chads from the 2000 presidential election, the spectacular calamity of 9/11, the unrepentant lies around Operation Iraqi Freedom, and the campy reality television featuring Donald Trump were all from this time. The burst of the dot-com bubble and corporate malfeasance of companies like Enron dampened entrepreneurial spirits, news publishers were optimistically sharing their stories online without paywalls, and blogging was heralded as the future of technology-mediated accountability and participatory democracy. “You” was Time Magazine’s Person of the Year in 2006 because “Web 2.0” platforms like YouTube, MySpace, and Second Life had become tools for “bringing together the small contributions of millions of people and making them matter.”1

Wikipedia was a part of this primordial soup, predating news-feed-mediated engagement, recommender-driven polarization, politicized content moderation, and geopolitical disinformation campaigns. From very early in its history, Wikipedia leveraged the supply and demand for information about breaking news and current events into strategies that continue to sustain this radical experiment in online peer production. This chapter will explore Wikipedia’s earliest efforts to cover breaking news events, common features of these unique collaborations, and how these features may serve as a model for other social platforms grappling with problems like disinformation.

I first encountered Wikipedia as an undergraduate student around 2004. My introduction to Wikipedia was likely a product of the sociotechnical coupling between Google and Wikipedia during this era. Google helped Wikipedia because Google’s ranking algorithms privileged Wikipedia’s highly interlinked articles, which brought an influx of users, some (tiny) fraction of whom became contributing editors like myself. Wikipedia also helped Google because Wikipedia could reliably generate both general interest and up-to-date content that satisfied its users’ information-seeking needs, which brought users back to Google rather than its competitors. The aftermath of natural disaster, the death of a celebrity, or a new pop culture sensation are all occasions for people to seek out background information to help them make sense of these events. Traditional journalistic offerings provide incremental updates about the immediate subject but often lack context or background: Why are there earthquakes in Indonesia? Who is Saddam Hussein? What is Eurovision? The availability and timeliness of Wikipedia content around topics of general interest would prove to be critical for its own sustainability in addition to complementing other platforms’ need to serve relevant and up-to-date content.

Wikipedia also entered the popular awareness of undergraduates like myself through the pitiless warnings from instructors and librarians about its lack of reliability as a citation. While these anxieties were largely reversed through empirical research and changes in professional culture, they also missed the forest for the trees: the value and authority of Wikipedia was not in any single article’s quality but in its network of hyperlinked articles. More than synthesizing knowledge as a tertiary source like traditional encyclopedias, Wikipedia’s hyperlink network invited users to follow their interests, dive deeper into topics, introduce missing connections, and create new articles where none existed. Where the decentralized web created a fragmented user experience requiring directories (e.g., Yahoo!) and search engines (e.g., Google) for navigation, Wikipedia’s hyperlinked articles foreshadowed an era of centralized web platforms that sustain user engagement with a consistent experience and “bottomless” content to consume and engage.

There are many ways to promote Wikipedia articles to its front page. Immediately to the right of “From today’s featured article” is the “In the news” (ITN) box featuring “articles that have been substantially updated to reflect recent or current events of wide interest.”2 The presence of news-like content in an encyclopedia is uncanny. On the one hand, encyclopedias are supposed to be stable references of historical knowledge rather than dynamic accounts of current events. On the other hand, there is a long history of encyclopedia editors grappling with how to incorporate new knowledge and encyclopedia publishers competing to be the most up-to-date.3 Wikipedia’s choice to privilege content related to current events via the ITN is also shrewd: it simultaneously is a shortcut to content users may already be searching for, it showcases the dynamism and quality of Wikipedia articles, and it invites users to consume and contribute to content outside of their primary interests.

September 11 and Wikipedia

To understand how Wikipedia’s “ITN” template and its broader culture of breaking news collaborations came about, we have to return to the immediate aftermath of the September 11, 2001, attacks. Wikipedia was ten months old at the time of the attacks, and while it already surpassed its elder sibling Nupedia in the number of articles, it was far from certain that the project would ever reach a sustainable level of activity. Although a comprehensive accounting of the editing activity in the immediate aftermath of the events has been lost to a server migration, snapshots from the Internet Archive’s Wayback Machine along with listserv discussions document the extent to which the Wikipedia community at the time went into overdrive in response to the attacks.4 Far from being an idiosyncratic case of online collaboration, the decisions made by editors at the time to use Wikipedia’s unique collaborative capacities to deeply cover the September 11 attacks would fundamentally change the direction, scope, and culture of Wikipedia as a project to the present day.

A Wayback Machine snapshot of the “September 11, 2001 terrorist attack” article from October 9, 2011, captures the remarkable breadth and depth of topics that were authored and organized together about the attacks.5 There were timelines; documentation of closings and cancellations; lists of casualties; links to donating blood and money; articles on political and economic effects; and newly created articles about the buildings, cities, flights, and perpetrators as well as topics like “terrorism,” “box-cutter knife,” and “collective trauma.” Approximately one hundred September 11–related articles were created in total (at a time when Wikipedia as a whole had only thirteen thousand articles) but Wikipedia’s content attracted links from other prominent web gateways like Yahoo! that brought in an influx of desperately-needed new users to the project.

The list of casualties enumerating each of the nearly three thousand victims (sorted by name and location and categorized by civilian or first responder) became a source of tension in the weeks following the attacks. Some editors argued this level of detailed coverage was unbecoming of the traditional encyclopedia Wikipedia was trying to emulate stylistically. Supporters referenced the rule that “Wikipedia is not paper” to justify a goal of writing biographies for thousands of victims, survivors, and leaders. As the trauma-induced altruism continued to fade, Wikipedia editors continued to raise concerns about the quality, notability, and importance of these memorialization efforts given the other demands of writing an encyclopedia. By September 2002 the community reached a consensus decision to move the September 11–related recollections and nonnotable pages to a “memorial wiki.” The launch of the memorial wiki led to heated discussions about which September 11–related articles would get to stay on Wikipedia and which would be relegated to the memorial wiki. The memorial wiki ultimately failed to thrive: its stagnant content and lack of editing activity led to accumulating vandalism, and it was effectively shuttered by September 2006. The creation, rejection, and disappearance of the September 11 memorial wiki’s content remains an underappreciated cautionary tale about the presumed durability of peer-produced knowledge: this content only persists when it remains integrated to the larger common project rather than being relegated to a smaller and more specialized project. Wikipedia’s peer production model is not immune from “rich get richer” mechanisms.

The Wikipedia community’s overreaction to the September 11 attacks and the discussions about the memorial content led to reflexive rule making about news that persists today. The “What Wikipedia is not” (WP:NOT) policy predates the attacks and enumerates that Wikipedia is not a dictionary, manual, directory, or a variety of other reference genres. In the midst of the debates in 2002 about what to do with the September 11 memorial content, the WP:NOT policy was expanded to assert that Wikipedia is not “a news report.” The revised policy attempted to thread the needle between the channeling of collaborative energy following current events against diluting the mission of writing an encyclopedia. The policy emphasized that “Wikipedia should not offer news reports on breaking stories” but conceded “creating encyclopedia articles on topics currently in the news is an excellent idea”6 as long as current events articles are written in an encyclopedic style. This “NOT NEWS” policy has persisted to the present, and the policy now emphasizes that “Wikipedia should not offer first-hand news reports on breaking stories” and “newsworthy events do not [automatically] qualify for inclusion … breaking news should not be emphasized or treated differently from other information.”7 Another change in identity that emerged as a result of the September 11 memorial content was the addition of “Memorials” to the WP:NOT policy. The policy, revised in 2004, now emphasizes that “Wikipedia is not the place to memorialize deceased friends, relatives, acquaintances, or others who do not meet such requirements.”8 These normative guardrails remain in place today to channel the outpouring of pro-social collaborative energy and sensemaking in the aftermath of traumatic events.

Features of Breaking News Collaborations

Even as an extremely active Wikipedia editor who made hundreds of revisions per month, I was always disappointed that I was never the first to create or update an article about a major current event. Wikipedia’s editors had remarkable alacrity in revising content in response to current events: articles about deceased celebrities, political scandals, and natural disasters were all updated or created seemingly within minutes of the news breaking. My disappointment at being unable to author the first revisions shifted into curiosity, and I began to explore the revision histories of these breaking news articles.9 These explorations raised more questions about the emergent social behaviors, and I switched my dissertation research project to exploring these breaking news collaborations. I was not alone in this inquiry. The Wikipedia model of a single, central account is much more legible to search engines like Google that boosted these articles’ authority and that drove the virtuous feedback loops of more traffic, more contributors, more updates, and better content. In 2009, then Google Vice President Marissa Mayer imagined a new web-oriented form of journalism where news stories did not compete against each other for authority or search engine results:

How [might] the authoritativeness of news articles grow if an evolving story were published under a permanent, single URL as a living, changing, updating entity?10

It is hard to imagine that Ms. Mayer’s vision of the future of journalism was not influenced by the enormous volumes of traffic her search engine was referring to Wikipedia in the aftermath of current events. More than a decade later, Wikipedia’s collaborations around breaking news continues to be a generative research context for myself and other researchers.11 Several general patterns have consistently emerged from my research over the past decade into Wikipedia’s breaking news collaborations.

First, the contributors to breaking news articles are drawn from editors across the Wikipedia community, not introduced by a small set of “ambulance-chasing” editors who had specialized roles and routines of breaking news editing. This suggests the motivation and ability for editors to engage in breaking news collaborations is widely shared. This distributed collaborative capacity proved to be important throughout Wikipedia’s history for mobilizing when multiple major events happened simultaneously. In March 2011 while the events of the Arab Spring demanded complex revisions across articles related to Tunisia, Egypt, Libya, and Syria, a 9.0-magnitude earthquake off the coast of Japan’s Tōhoku province triggered a massive tsunami that ultimately killed more than twenty thousand people and led to the most serious nuclear disaster since Chernobyl. Because of Wikipedians’ distributed collaborative capacity, editors were able to process these major historical events in parallel when each event itself required a massive undertaking of synthesizing, coordinating, and deliberating across dozens of articles, talk pages, administrative processes, and language editions. Moreover, the contributors to breaking news article collaborations have diverse repertoires and roles on the project: an editor specializing in editing articles about Japanese boy bands shifted their focus to updating infrastructure damaged by the 2011 tsunami while another editor migrated their dispute resolution experience from Harry Potter articles to the Fukushima nuclear disaster article.12 Other topical areas that are proximate to breaking news have developed specialized routines for managing common coordination problems. When a new storm happens, members of the WikiProject Tropical Cyclones shift to editing these articles and bring a wealth of experience for structure, style, references, and multimedia about storms to structure these collaborations. Pro-social responses in the aftermath of disaster and catastrophe are ubiquitous,13 but Wikipedia uniquely channels this energy into producing enduring and highly networked knowledge artifacts.

Second, breaking news events are sites of large, rapid, and temporary collaborations that were otherwise rare on Wikipedia. The average Wikipedia article has accumulated fewer than ten unique editors and revisions over a span of years while breaking news articles can have hundreds of editors and revisions over a span of days. Examining the archival “zeitgeist” statistics for the English Wikipedia articles,14 the most actively revised articles in any given month tend to be related to breaking news events or people in the news. In 2004, the articles with the most unique editors in a month included the “2004 Madrid train bombings” (112 editors in March), “Ronald Reagan” (114 editors in June), “2004 Summer Olympics” (92 editors in August), “Timeline of the 2004 United States Presidential election” (154 editors in November), and “2004 Indian Ocean earthquake and tsunami” (345 editors in December). The number and frequency of revisions to these articles was also extremely high: on major events, multiple revisions can be made in the same minute, complicating efforts for longer-form writing or copyediting.

The MediaWiki software on which Wikipedia runs did not anticipate this kind of synchronous editing behavior, so editors revert to strategies for working around the limitations of the software such as making smaller and more frequent edits, merging in changes from a sandbox, or requesting an administrative lock on the article to incorporate requested changes. These collaborations are often temporary, involving editors with disparate expertise and interests to come together to collaborate, with most of them never having worked together before and with no expectations of collaborating again in the future. In the absence of social relationships to shape these emergent collaborations, editors are guided by common interests and shared values around writing an encyclopedia. Even if most participants in breaking news collaborations return to editing their usual topics afterward, these breaking news collaborations play a crucial role as “watering holes” where different groups’ norms are reaffirmed and best practices are synthesized and then diffused back out through the rest of the project. Breaking news collaborations arguably play an important role in the viability of the broader Wikipedia project by engaging editors in challenging experiences, validating the investments of volunteer editors, and circulating innovations throughout the project.

Finally, breaking news articles are exceptionally high quality when compared with the median Wikipedia article: they tended to be longer; have more links to other Wikipedia articles; have more references and citations; and have more images, maps, and multimedia. Recent events make more “raw” material available in the form of reporting and social media content than historical events requiring archival research skills, providing a richer set of inputs to generate better articles. But breaking news articles also benefit from “Linus’s Law”15 where a large number of diverse editors can accomplish tasks that would seem only possible for a small group of experts to accomplish. These articles also have a complex life cycle of different cohorts of editors cycling through the collaboration over the course of days, weeks, and years. Biographical articles about the recently deceased often go through a major rewrite to incorporate information from obituaries as well as a general reappraisal and standardization of structure and style rather than simply changing verb tenses and adding in the relevant information about the subject’s death. Anniversaries have also become occasions for readers and editors to revisit an article and make new contributions. Wikipedia articles about current events provide a unique commons for emergent communities to gather, not only to document and reappraise our understanding of the causes, contexts, and consequences of major and often traumatic events but also to support others’ information seeking and sensemaking as well.

All of these patterns reinforce the idiom that “Wikipedia works in practice, not in theory.” Who are these editors that rapidly self-select and self-organize themselves in the absence of any formal coordination or delegation? Why have breaking news collaborations continued to employ generalists rather than develop a class of specialists? How did dozens of users synchronously edit a shared document using an asynchronous tool with none of features we take for granted in something like Google Docs? These remain open and vital questions for researchers twenty years after Wikipedia’s launch.

Wikipedia in the Age of Disinformation

Despite being the “encyclopedia that anyone can edit” and one of the most trafficked websites in the world, Wikipedia did not show the same susceptibility to the coordinated disinformation campaigns that plagued social platforms like Facebook, YouTube, and Twitter around 2016. Although these platforms have made massive investments in human and automated moderation to improve users’ experience, allay advertisers’ concerns, and head off regulator scrutiny, disinformation, harassment, and other sociotechnical sludge remain endemic.16 Provocateurs, outrage-mongers, and outright fascists have flocked to these “virality engine” platforms to actively recommend fringe ideas and compensate their creators while distributing them to audiences of millions. Platforms’ attempts at commonsense content moderation by removing or “demonetizing” the most egregious examples of hate speech and harassment have in turn led to accusations of their threats to “free speech” and “anti-conservative bias.” What explains Wikipedia’s apparent resilience to the sociotechnical sludge polluting other platforms?

The most obvious hypothesis is the difference in incentives between the user experience of advertising-driven engagement maximization and commons-based peer-production models. Facebook, YouTube, and other popular social platforms generate billions of dollars in revenue by injecting personalized advertising alongside bottomless recommendations and news feeds managed by expensive engineers and infrastructures to engage users’ attention. Every user’s Facebook News Feed is personalized in response to their relationships, interests, and behavior. Content featuring novelty, humor, and outrage receives greater “engagement,” so publishers and advertisers are locked in an arms race to produce ever more attention-grabbing content and target it for users’ personalized feeds. Wikipedia has no newsfeed,17 runs no advertising, and has a comparatively minuscule operating budget. But an overlooked and critical difference between Wikipedia and other social platforms is the absence of personalization in the user experience. Every English Wikipedia user’s “Abraham Lincoln” article is the same regardless of their geography, gender, browsing history, or social graph. This common experience concentrates collective scrutiny and deliberative capacity rather than diffusing these accountability mechanisms across inscrutable and incommensurable personalized news feeds. Linus’s Law—“given enough eyes, all bugs are shallow”—evidently holds for preserving the integrity of social information feeds.

A second hypothesis explaining Wikipedia’s resilience to sociotechnical sludge is the absence of algorithmic amplification. The background above illustrates how Wikipedia articles can “trend” in response to current events and popular culture. However, Wikipedia’s editors exercise considerable “human in the loop” editorial discretion over both the substance of trending content as well as its amplification mechanisms, unlike the algorithms driving news-feed-centered platforms like Facebook and YouTube that can be manipulated into privileging viral and outrageous content. The most common user experience of Wikipedia is arriving from a search engine and navigating to related articles via hyperlinks or follow-on searches rather than navigating in from a news feed or home page. To the extent Wikipedia has mechanisms for amplifying content to users, they exist on the homepage as “From today’s featured article,” “In the news,” “Did you know,” and “On this day.” These mechanisms are all explicitly vetted by human editors following documented public policies and consensus-driven deliberation that still have remarkable alacrity in responding to current events. The responsiveness of Wikipedia editors to current events also provides an important counterfactual to claims from engineering culture that human-in-the-loop systems lack the scalability, speed, and accuracy of automated systems, despite accumulating evidence of automated systems’ multiple liabilities. Because the oversight and capacity to intervene in Wikipedia’s attention amplification mechanisms is delegated across hundreds of administrators and/or thousands of editors, they are substantially harder to compromise than algorithmic systems operating under “security through obscurity” strategies.

Social platforms confronting the limitations of their current engagement and moderation models are turning to Wikipedia. In October 2017, Facebook announced that it would provide “contextual information” about articles in the news feed that would include links to Wikipedia.18 In March 2018, YouTube Chief Executive Officer Susan Wojcicki outlined a strategy wherein YouTube would connect videos containing conspiracies to corresponding Wikipedia articles in an effort to combat the spread of disinformation.19 YouTube’s decision, in particular, came as a surprise to the Wikipedia community and the Wikimedia Foundation, who were given no forewarning that they would be indirectly policing YouTube’s toxic content. The fundamental risk was that the same dynamics that converted information-seeking Google search users into Wikipedia editors could also convert the conspiracists, ideologues, and culture warriors on these platforms into Wikipedia editors. These decisions to outsource content moderation to Wikipedia were deeply irresponsible: either YouTube failed to comprehend the obvious risks of swamping the smaller volunteer project with their content moderation problems or they did not care.

Facebook’s and YouTube’s conduct in this case is a classic problem of governing what economists call “common goods” and its corresponding “tragedy of the commons.” The knowledge produced—and more importantly, governed—by Wikipedia is “nonexcludable,” which means that it can still be used by people who have not contributed to it. However, the governance of this knowledge exhibits patterns of “rivalrousness” in which consumption by one actor reduces availability for others. In this case, Facebook and YouTube contributed nothing to Wikipedia’s governance but could still benefit from the credible content generated and governed by the Wikipedia community (nonexcludability). But in outsourcing content moderation to Wikipedia editors and administrators, Facebook and YouTube were potentially reducing Wikipedia editors’ capacity to attend to other content generation and moderation demands (rivalrousness). Facebook and YouTube were effectively “overfishing” the capacity of Wikipedia editors and administrators to handle sociotechnical sludge by requiring the volunteer Wikipedia community to do more of all of this work on behalf of a corporation who profits from not having to moderate its own content. But commons do not inevitably end up as tragedies; the research of Elinor Ostrom (which culminated in her 2009 Nobel Memorial Prize in Economic Sciences) charts out strategies for designing institutions for sustaining commons in the face of threats like overuse. Her 2006 edited volume with Charlotte Hess, Understanding Knowledge as a Commons, charts prescient strategies for communities like Wikipedia to pursue to “define, protect, and build the knowledge commons in the digital age.”20

The case of Wikipedia content being redeployed by unscrupulous platforms for their content moderation needs illustrates the risks associated with the “interoperability” of online platforms: content from Platform A can be plugged into Platform B, but these connections can also cause blowback as the bad behavior from Platform B moves to Platform A. Wikipedia’s content is reused in both visible and invisible ways across platforms: Google serves up Wikipedia content alongside its search results, Facebook uses it to populate information for its pages, and Apple’s Siri or Amazon’s Alexa will read summaries of articles. Wikipedia’s content is also used in more invisible ways to train algorithms used for translation, image recognition, and concept similarity. These interoperable connections increase the prominence of Wikipedia’s content, recruit new users to contribute, and highlight the need to preserve this commons, but every new interoperable link also introduces new threats. If a malicious actor wanted to undermine trust in these other major platforms, an under-realized vector can subtly compromise the quality of information from the Wikipedia and Wikidata content that they ingest.

Wikipedia’s resilience to the strategic disinformation campaigns from 2016 should not be interpreted as intrinsic immunity to information manipulation: Wikipedia’s most active editors are not representative of the population at large, which creates both biases in its content and blind spots in its responses, which are then ingested and amplified through the web of interoperable dependencies outlined above. Wikipedia administrators botched its response to the Gamergate controversy in 2015 by acquiescing to a manipulative influence campaign and banning five editors who had been fending off extremist content:21 this case illustrated how Wikipedia’s administrative procedures can be hijacked by bad-faith actors to target good-faith editors. On a lighter note, another illustration of the threats of interoperability is a case from October 2017. When users of Apple Siri asked “What is the national anthem of Bulgaria?,” they were served “Despacito,” a 2017 reggaeton pop hit, rather than the nineteenth-century hymn “Mila Rodino.”22 Somewhere deep in Apple’s knowledge graph, much of which is likely trained on Wikipedia and Wikidata, this erroneous pairing was introduced and never validated before being pushed out to millions of users.

Wikipedia and its increasingly important sister project Wikidata have been able to resist disinformation efforts because of the ability to match its supply of human-in-the-loop governance with demand for information: oversight follows the action. While it might be hard to embed disinformation into articles about candidates for an upcoming election because of this superabundance of editorial attention, it might be trivial to persistently embed disinformation into provincial articles about distant historical events, specialized scientific topics, or marginal trivia about national anthems that lack sustained editorial oversight. While Wikipedia’s unique editorial model has shown greater resistance to the disinformation, harassment, and manipulation plaguing other social platforms to the point that its content is serving as a front-line defense, there are nevertheless growing precedents that Wikipedia’s content and governance has very real vulnerabilities that could easily and quickly propagate throughout a complex technical stack of interoperable technologies.

Conclusion

Encyclopedists have always struggled with the limitations of synthesizing knowledge into paper documents because when the knowledge changes, so must the paper. Wikipedia was not the first encyclopedia to use the online medium to rapidly and inexpensively revise content in response to changes, but its unique “anyone can edit” model had the effect of entangling current events with the viability of the project.

The September 11 attacks were a critical moment in Wikipedia’s history. The events brought in an influx of new editors motivated to document the events, perpetrators, victims, and contexts, and the outpouring of collaborative effort in the aftermath of the September 11 attacks validated an underappreciated strategy for growing the project. By simultaneously tapping into editors’ pro-social motivations following traumatic events as well as showcasing the quality and timeliness of the project’s content in a time of acute information seeking and sensemaking, Wikipedia could convert the large numbers of information-seeking users into new contributors as well as increase popular trust in its radical editorial model. However, Wikipedia editors’ overzealous creation of September 11–related content also required the development of new rules and identities as guardrails that persist today about what the encyclopedia is and is not.

Wikipedia editors continue to invest enormous amounts of effort in covering breaking news and current events within the confines of these guardrails. Articles about the recently deceased, natural disasters, conflicts, and popular culture are sites of large and extremely dynamic collaborations involving dozens of editors making hundreds of revisions within hours. While Wikipedia’s MediaWiki software was not designed with this use case in mind, these high-tempo collaborations continue to serve crucial roles in sustaining the health of the broader project, close to twenty years after the early precedent of the September 11 attacks: they bring in new users to the project, provide opportunities to disparate subcommunities to temporarily congregate, disseminate innovations and best practices into the rest of the community, and produce high-quality content hyperlinked to other relevant background.

Wikipedia remains a product of a particular historical moment from the early 2000s, in terms of not only its adorably dated interface but also the absence of advertising and engagement, news feeds and recommendation systems, and virality and polarization as central features that define so much of the user experience on other social platforms. Wikipedia’s resilience to the disinformation that plagued Facebook, YouTube, and Google in 2016 would suggest this archaic user experience provided an important defense against actors who weaponized these attention amplification mechanisms on other platforms to malicious ends. But this story overlooks other explanations for Wikipedia’s apparent resilience: Wikipedia users and editors’ attention is shared around common articles rather than distributed across personalized news feeds.

Does Wikipedia’s success in covering breaking news and current events chart a path for other platforms to follow? Information seeking and sensemaking about current events drive enormous flows of online collective attention, which explains why “News feeds” and “Trending” topics are ubiquitous on social platforms. Whether and how Wikipedia can channel this demand for information likewise has been central to its ongoing identity, relevance, and sustainability. Wikipedia remains a valuable counterfactual for the potential of designing around information commons, human-in-the-loop decision making, and strong editorial stances in the face of the Silicon Valley consensus emphasizing content personalization, automated moderation, and editorial indifference. The differences in how Wikipedia handles current event information may have insulated it from manipulation, but as platforms increasingly turn to Wikipedia for providing and moderating content, Wikipedia’s very real vulnerabilities risk becoming a target.

Comments
0
comment

No comments here