āTrump Accuses Google of Burying Conservative News in Search Results,ā reads an August 28 New York Times headline. The piece features a bombastic president, a string of bitter tweets, and accusations of censorship. āAlgorithmsā are mentioned, but not until the twelfth paragraph.
Trumpālike so many other politicians and punditsāhas found search and social media companies to be convenient targets in the debate over free speech and censorship online. āThey have it RIGGED, for me & others, so that almost all stories & news is BAD,ā the president recently tweeted. He added: āThey are controlling what we can & cannot see. This is a very serious situationāwill be addressed!ā Trump is partly right: They are controlling what we can and cannot see. But ātheyā arenāt the executives leading Google, Facebook, and other technology companies. āTheyā are the opaque, influential algorithms that determine what content billions of internet users read, watch, and share next.
These algorithms are invisible, but they have an outsized impact on shaping individualsā experience online and society at large. Indeed, YouTubeās video-recommendation algorithm inspires 700,000,000 hours of watch time per dayāand can spread misinformation, disrupt elections, and incite violence. Algorithms like this need fixing.
But in this moment, the conversation we should be havingāhow can we fix the algorithms?āis instead being co-opted and twisted by politicians and pundits howling about censorship and miscasting content moderation as the demise of free speech online. It would be good to remind them that free speech does not mean free reach. There is no right to algorithmic amplification. In fact, thatās the very problem that needs fixing.
TO SEE HOW this algorithm amplification works, simply look to RT, or Russia Today, a Russian state-owned propaganda outlet thatās also among the most popular YouTube presences. RT has amassed more than 6 billion views across 22 channels, more than MSNBC and Fox News combined. According to YouTube chief product officer Neal Mohan, 70 percent of views on YouTube are from recommendationsāso the siteās algorithms are largely responsible for amplifying RTās propaganda hundreds of millions of times.
How? Most RT viewers donāt set out in search of Russian propaganda. The videos that rack up the views are RTās clickbait-y, gateway content: videos of towering tsunamis, meteors striking buildings, shark attacks, amusement park accidents, some that are years old but have comments from within an hour ago. This disaster porn is highly engaging; the videos have been viewed tens of millions of times and are likely watched until the end. As a result, YouTubeās algorithm likely believes other RT content is worth suggesting to the viewers of that contentāand so, quickly, an American YouTube user looking for news finds themselves watching Russiaās take on Hillary Clinton, immigration, and current events. These videos are served up in autoplay playlists alongside content from legitimate news organizations, giving RT itself increased legitimacy by association.
The social internet is mediated by algorithms: recommendation engines, search, trending, autocomplete, and other mechanisms that predict what we want to see next. The algorithms donāt understand what is propaganda and what isnāt, or what is āfake newsā and what is fact-checked. Their job is to surface relevant content (relevant to the user, of course), and they do it exceedingly well. So well, in fact, that the engineers who built these algorithms are sometimes baffled: āEven the creators donāt always understand why it recommends one video instead of another,ā says Guillaume Chaslot, an ex-YouTube engineer who worked on the siteās algorithm.
These opaque algorithms with their singular purposeāākeep watchingāācoupled with billions of users is a dangerous recipe. In recent years, weāve seen how dire the consequences can be. Propaganda like RT content is circulated far and wide to disinform and worsen polarization, especially during democratic elections. YouTubeās algorithms can also radicalize by suggesting āwhite supremacist rants, Holocaust denials, and other disturbing content,ā Zeynep Tufekci recently wrote in the Times. āYouTube may be one of the most powerful radicalizing instruments of the 21st century.ā
The problem extends beyond YouTube, though. On Google search, dangerous anti-vaccine misinformation can commandeer the top results. And on Facebook, hate speech can thrive and fuel genocide. A United Nations report about the genocide in Myanmar reads: āThe role of social media is significant. Facebook has been a useful instrument for those seeking to spread hate, in a context where for most users Facebook is the Internet ⦠The extent to which Facebook posts and messages have led to real-world discrimination and violence must be independently and thoroughly examined.ā
So what can we do about it? The solution isnāt to outlaw algorithmic ranking or make noise about legislating what results Google can return. Algorithms are an invaluable tool for making sense of the immense universe of information online. Thereās an overwhelming amount of content available to fill any given personās feed or search query; sorting and ranking is a necessity, and there has never been evidence indicating that the results display systemic partisan bias. That said, unconscious bias is a concern in any algorithm; this is why tech companies have investigated conservative claims of bias since the Facebook Trending News debacle of 2016. There hasnāt been any credible evidence. But there is a trust problem, and a lack of understanding of how rankings and feeds work, and that allows bad-faith politicking to gain traction. The best solution to that is to increase transparency and internet literacy, enabling users to have a better understanding of why they see what they seeāand to build these powerful curatorial systems with a sense of responsibility for what they return.
There have been positive steps in this direction. The examples of harms mentioned above have sparked congressional investigations aimed at understanding how tech platforms shape our conversations and our media consumption. In an upcoming Senate hearing next week, the Senate Intelligence Committee will ask Jack Dorsey of Twitter and Sheryl Sandberg of Facebook to provide an accounting of how, specifically, they are taking steps to address computational propaganda.
Itās imperative that we focus on solutions, not politics. We need to build on those initial investigations. We need more nuanced conversations and education about algorithmic curation, its strange incentives, and its occasionally unfortunate outcomes. We need to hold tech companies accountableāfor irresponsible tech, not evidence-free allegations of censorshipāand demand transparency into how their algorithms and moderation policies work. By focusing on the real problem here, we can begin addressing the real issues that are disrupting the internetāand democracy.
You copied advertisements into the body text of your post š
There is no question that there is room for improvement in the recommendation algorithm, but the claim that Big Tech is being (unfairly) harsh on conservative information is worth considering. However, this is partly because conservative information on the Internet contains a lot of wrong information.
Hey. Social media decided to implement fact check, and conservatives happen to lie a lot.