Debate has been waging over whether the government should change Section 230 of the Communications Decency Act, which shields platforms from liability for content posted by their users. But public officials who want to curb the spread of inaccurate information and growth of violent movements over these sites could be more impactful by taking on the way platforms’ algorithms amplify content and recommend connecting with extremist users on the sites, said researchers, journalists and legal experts speaking at the Summit.
“You have these platforms … who love to make sure we keep talking about this as a free speech versus censorship problem because that benefits them, because then they don’t really have to discuss the harder issues,” said panelist and Future of Democracy Fellow Yaël Eisenstat, who previously headed the Global Elections Integrity team of Facebook’s Business Integrity unit during 2018.
Should someone attempt to sue Facebook on charges of complicity in the Jan. 6 Capitol attack, Einestat said, “I suspect Facebook will try to use the Section 230 argument to say ‘We are not responsible for the conspiracy theories these users are posting.’… Yes, but did your recommendation engines, your targeting tools, your curation, did it do anything of these things that actually helped facilitate this crime?”
An April 27 Congressional hearing is slated to confront this question of algorithms, and MIT panelists recently offered their own views into whether and how the government can intervene in social media regulation, and the potential challenges and ramifications.
What’s at stake
Social media platforms didn’t invent falsehoods, but they have enabled incorrect information due to posters’ mistakes (misinformation) and deliberate lies (disinformation) to proliferate rapidly, and common social media features like tailored newsfeeds can mean users live inside echo chambers in which false information is reiterated until normalized.
The platforms were designed to encourage high user engagement — not promote considered public discourse — and the algorithms that determine what items appear in users’ newsfeeds do not distinguish between facts and falsehood, several panelists said. The result is that users may see inaccurate stories repeated without contesting views mixed in, and individuals can struggle to be aware of — let alone dispute — news items not presented to them in their own newsfeeds.
Ali Velshi, NBC correspondent and host of MSNBC’s “Velshi,” said the result is that residents often cannot agree on what is real, so debates about political and social issues become hung up on efforts to establish basic facts. Conversation cannot progress to productive debate over how to respond to realities described by those facts.
“You never get to the discussion of, ‘What should good policing look like?’… [or] ’How should we deal with the application of health care throughout this country?’” Velshi said.
The Section 230 question
Government has a variety of tools at its disposal for combating the spread of harmful inaccuracies over social media.
The platforms — like all private companies — have no obligation to provide free expression, and current law already bans certain speech, said Richard Stengel, former U.S. under secretary of state for Public Diplomacy and Public Affairs. He noted that these sites actively monitor content to remove child pornography and content that violates copyrights. Governments simply need to update or repeal Section 230 or pass new legislation to make removing conspiracy theories or other harmful content similarly incentivized.
“Facebook can make any [content] law it wants, but unless you give them more liability, they won’t take content down,” Stengel said.
That added liability could have unintended consequences, however, said Jeff Kosseff, United States Naval Academy cybersecurity law professor and author of a history of Section 230. Social media platforms’ legal advisers are likely to recommend platforms remove even genuine news items that prove controversial or provoke complaints, to minimize the risk of lawsuit, he said.
Part of the complication is that some content is clearly problematic while other posts are more subjective, and platforms are wary of drawing the wrong conclusions. Stengel said that government regulation can spare platforms from that risk by offloading that decision-making and responsibility to public policymakers.
“They [platforms] want to be regulated more — they don’t like being in the gray area of subjective decisions; they want to be able to say, ‘Well, the government made me do this,’” Stengel said.
Speech vs. amplification
Flagging posts for removal and banning users who breach policies is often an uphill battle given how fast content is created and spread. Summit host and MIT Sloan School of Management professor Sinan Aral pointed to findings from a 2018 study he had co-authored that revealed inaccurate stories are retweeted faster than true ones and reach 1,500 users six times sooner. Real humans played a greater role than bots did in spreading the falsehoods.
This may point to the heart of the problem — not that some users and bots are posting disinformation, but that the platforms’ algorithms are then proactively recommending the posts to a wide audience and encouraging reposting.
“There’s always been this division between your right to speak and your right to have a megaphone that reaches hundreds of millions of people,” said Renée DiResta, research manager at Stanford Internet Observatory. “There’s value to you being able to post your … views but [the] platform doesn’t need to boost it.”
There may already be tools on the books to tackle this, with Eisenstat subscribing to the view that Section 230 is most accurately interpreted as protecting platforms from liability only for user’s speech, not for the actions of the platform’s own engines.
A variety of approaches
Platforms can rework their designs and strategies to slow disinformation — if public regulations push them to.
Social media companies could deploy fact checkers on standby to flag or remove inaccurate posts in advance of major events like elections, DiResta suggested, and platforms can redesign the user experience so account holders are prompted to reconsider whether they’ve fully read content or know it is accurate before reposting. Social media companies could also be required to fact check paid political ads that users might reasonably assume have been verified by the platforms, Eisenstat said.
Stengel also proposed reconsidering speech regulations in a larger way, suggesting state and local governments tackle certain harmful content by passing ordinances against hate speech.
An ideal strategy may not quickly present itself, but Eisenstat said the government should focus on passing something that at least makes things better.
“Everyone thinks we need to have the end-all, be-all, silver bullet, one piece of legislation that suddenly makes social media a lovely, healthy place for democracy. That’s not going to happen,” she said. “But we can all agree, the status quo cannot continue.”
Government Technology is a sister site to Governing. Both are divisions of e.Republic.