Posted inMedia / Politics / Social Control Media / ToMl

It’s time to get rid of the Facebook “news feed,” because it’s not news

In the wake of the US election, critics have blamed Facebook for bringing about—at least in part—Trump’s surprise win. A BuzzFeed report showed that Facebook users interacted far more with “fake news” stories about both candidates than they did with mainstream news outlets before the election. This wouldn’t seem like such a big deal if it weren’t for a Pew Research Center survey showing that 44% of Americans rely on Facebook to get their news.

But proving whether fake news influenced the election more than the usual political propaganda is impossible. What’s certain is that fake news on Facebook is a symptom of a larger problem: the company is trying to play contradictory roles as both trustworthy news publisher and fun social media platform for personal sharing. The problem is that it cannot be both—at least not without making some changes.

Facebook shapes people’s perceptions

When you log into your Facebook account, your default page is dominated by a cascading “news feed,” automatically selected for your pleasure, which consists of whatever your friends have shared. The company uses a mix of secret-sauce algorithms to choose which pieces of news you see. Some items are displayed based on what you’ve responded to before. For example, if you always like or reply to news from Trevor but ignore news from Mike, you’re going to see more Trevor and less Mike.

Other news that Facebook thinks you probably want to see is fed to you based on your profile. The Wall Street Journal has an incredible infographic on this, showing how Democrats on Facebook see liberal-leaning stories and Republicans see conservative-leaning ones.

Mark Zuckerberg has protested since the election that it’s preposterous to believe that Facebook could change anyone’s minds based on their news feed—yet the company behaves as if it does. Given that Facebook’s main goal is to serve you ads and get you to buy things, their number-one priority is keeping you glued to your feed. If you see a bunch of things you hate in your feed, you’re going to stop looking at it and take your clicks elsewhere. Common sense dictates that Facebook should avoid showing you news that will upset you or make you angry.

But Facebook’s decision goes beyond common sense. It’s based on real data the company gathered in a 2012 experiment in which algorithms fed users positive and negative stories to see whether they would affect people’s moods. Sure enough, the people fed positive news responded with more indicators of happiness. To the extent that people at Facebook believe their own data analysis, they know that the news feed affects people’s emotions and shapes their perceptions of the world. Their business depends on it.

Though there’s something ineffably creepy about Facebook manipulating our emotions, the site is no different than any other ad-driven business. Television series toy with your feelings to keep you coming back for more, even after you’ve seen the same stupid ad on Hulu nine times. Hollywood pumps out sequels to get you to pay $16 for a ticket. Facebook’s big innovation was the discovery that it could sell ads against people’s friendship networks. We consume each other’s posts on Facebook the same way we consume new episodes of Mr. Robot and with the same result. Our eyeballs translate into ratings, which translate into ad dollars.

The trouble with “news feed”

The problem is that Facebook decided to go beyond friendship networks. For several years now, the company has courted professional news outlets, ranging from the New York Times and BuzzFeed to Breitbart and FOX News, promising them prominent placement in users’ feeds. This courting intensified with the creation of the Facebook Instant service in 2015, which allows media companies to publish stories directly on Facebook and share ad revenues.

But then, earlier this year, Facebook announced that it would no longer be giving special weight to news coming from pages that belong to media outlets. From that point on, they would be weighting news from “friends and family” more heavily than from news outlets. Your news feed would not discriminate between stuff your friends passed along and news sources.

This rule had one exception. To the right of the news feed on your default Facebook page is the “trending” module. This module surfaces news that Facebook readers are discussing, so it’s never sourced to friends and family. For a while, Facebook had a team of human editors who curated trending news from a long list of reputable sources. But after several disgruntled ex-employees accused the news team of censoring right-wing stories, Facebook fired them and set up an algorithm to take their places.

With no human checks on it, that algorithm immediately started posting fake news.

Facebook’s algorithms are great at keeping people glued to their screens, but they are terrible at distinguishing real news from fake. Yet the most prominent feature of Facebook is called a “news feed.” Given the company’s almost hilarious inability to identify news, this feature clearly has a misleading name.

“But what is news?” you might ask. “Are you talking about liberal media?” No. I am talking about stories published by professional media organizations—organizations that take legal and ethical responsibility for what they publish. They pledge to print news that is the truth. They may define truth differently, depending on whether they’re Mother Jones or the National Review. Nevertheless, they come up with a definition and attempt to stick to it.

Plus, if those publications post articles that are libelous, infringing, obscene, or otherwise unlawful, they can be held liable in a court of law for what they’ve published. In short, news comes from organizations that stand by what they post as the truth. That’s why the public can trust news from a professional media organization more than that rant Uncle Tommy posted about chemtrails on Facebook. Unfortunately, however, Uncle Tommy’s rant is classified as part of the same “news feed” that contains headlines from your most trusted professional media sources.

The legal issues

Our concerns with Facebook’s news feed aren’t just a matter of truthy semantics, either. If you want to understand what’s at stake, think about two rules that the US legal system uses to regulate the news media. One is the First Amendment, which has been interpreted to mean that a publisher has the right to publish lawful information without interference from the government. The other is a rather obscure section of the Communications Decency Act known as “CDA 230.”

Section 230 offers immunity from legal prosecution to “intermediaries” or “interactive computer services” who publish things created by other people. The classic example of a person protected by CDA 230 is a blogger who allows open comments on their stories. Under CDA 230, the blogger cannot be sued if a commenter says something libelous or posts an infringing movie. The blogger is merely an “intermediary” and thus does not bear responsibility for what commenters say.

Facebook enjoys the protection of both the First Amendment and CDA 230, but CDA 230 is key to its survival. It means that Facebook cannot be held liable for anything posted on Facebook, whether written by Uncle Tommy or a BuzzFeed journalist. Without CDA 230, President-Elect Trump could sue Facebook for libel when users post fake news about him. Angry consumers could sue Facebook for fraud when quack doctors post false claims about herbal remedies. The government could sue Facebook for obscenity when a naughty photo goes viral. CDA 230 is what’s keeping Facebook from being sued out of existence.

But CDA 230 may not protect Facebook if the company continues to use algorithms to shape people’s experience in the news feed. We’re entering a legal gray area with CDA 230, where courts are still undecided on what actions boot an organization out of the protected category. If Facebook starts acting too much like a publisher, or what CDA 230 calls an “information content provider,” it may open itself up to liability.

Given that one of the traditional roles of publishers is to choose which stories to show their readers, a court might possibly decide that Facebook is actually a publisher. After all, if I’m following 100 people, and Facebook only shows me “news” from 20 of them, is that not taking an editorial role?

As attorney and Stanford law researcher Morgan Weiland told Ars via email:

Facebook’s highly personalized algorithmic curation of its users’ newsfeeds falls in a legal gray area with respect to CDA 230. As you know, CDA 230 provides immunity for “interactive computer services,” drawing a line between that category and “information content providers.” But it’s not entirely clear when the former becomes the latter; in other words, it’s not clear when an intermediary engages in enough editing of third-party content that it becomes an “information content provider” and loses CDA 230 immunity. So the question becomes: does the editing that Facebook is doing here—editing what information is displayed in a users’ newsfeed through algorithmic curation —transform Facebook into an “information content provider” for the purposes of the law? Lawyers would argue that one either way.

Courts may decide it either way, too. It’s not out of the realm of possibility that a few important court cases could redefine Facebook as an information content provider. That would force the company into taking responsibility for fake news in the most dramatic possible way. It would also destroy their business.
Put up or shut up, Facebook

Facebook wants to have it both ways, both ethically and legally. The company wants to be seen as a source of news. The company offers media outlets a chance to use Facebook Instant, because having native content from the New York Times enhances Facebook’s credibility. And yet it also wants to retreat into the safe harbor of being a mere intermediary that does not distinguish between Uncle Tommy’s rants and a Washington Post investigation. That’s why Zuckerberg has responded to criticisms of fake news by saying Facebook’s job is not to determine what’s true.

Facebook needs to to stop playing both sides, but it keeps trying to walk the line. Vice correspondent Nellie Bowles said on Twitter, “On the phone with Facebook PR and they literally ask me ‘what is truth’.” As Bowles pointed out in an article later, Facebook already has an answer to this question: the company just officially banned fake news sources from buying ads. Clearly, Zuckerberg et al know how to distinguish real from fake, but are choosing to feign ignorance when it suits them.

Facebook, stop it. Stop calling it a “news feed.” Call it a “friend feed.” Call it a “fun feed.” Call it whatever you want—just don’t call it news, because it’s a dangerous misrepresentation for those 44% of Americans who choose to look to Facebook for actual news.

There are several ways forward from here. Facebook could get out of the news business entirely and go back to being a pure social network. It could abandon media partnership initiatives like Facebook Instant and Facebook Live. It could tune its algorithms to feed people only fun updates from friends, and enjoy the benefits of being one of the world’s biggest entertainment companies.

Or the company could go in a more radical direction and declare itself an actual news provider. It could hire a legitimate team of news writers and editors and bring readers the news they seem to crave. But this scenario seems unlikely, given what I said earlier about how Facebook’s business model depends on CDA 230 immunity. Almost nothing would get posted to Facebook if the company had to take full legal responsibility for everything in the feed.

Or, as a third option, Facebook could keep walking the line, but in a much more honest fashion. It might do this by creating a separate news section, curated in part by humans and perhaps by better algorithms. This section could draw from a vast array of professional media organizations, the same way Google News does. It would also be separate from the friend feed, to make it easy for users to understand when they’re reading something from an accountable organization and when they’re reading goofball crap.

No matter what Facebook does, it needs to stop pretending that “news” can be anything from anyone, granting equal weight to the trivial and the truly important. Most people think that news means truth and accountability. We know the public depends on the company to help show them what’s important in the world. Facebook needs to take full responsibility for posting news or get out the game. That’s not just a matter of ethics—it’s good business.

— source arstechnica.com By Annalee Newitz

Leave a Reply

Your email address will not be published. Required fields are marked *