Supreme Court will decide on First Amendment protections for social media

The court has the opportunity to reexamine the nature of content moderation and the constitutional limitations of the government regarding speech on platforms such as Facebook, X, TikTok, and Instagram.

SHARE Supreme Court will decide on First Amendment protections for social media
The Supreme Court building in Washington, D.C. on Sept. 25. The court is tackling the question of First Amendment rights and social media platforms.

The Supreme Court building in Washington, D.C. on Sept. 25. The court is tackling the question of First Amendment rights on social media platforms.

J. Scott Applewhite/AP

The First Amendment does not protect messages posted on social media platforms.

The companies that own the platforms can, and do, remove, promote or limit the distribution of any posts according to corporate policies. But all that might soon change.

The Supreme Court has agreed to hear five cases during this current term, ending June 2024, that collectively give the court the opportunity to reexamine the nature of content moderation — the rules governing discussions on social media platforms such as Facebook and X, formerly known as Twitter — and the constitutional limitations on the government to affect speech on the platforms.

Content moderation, whether done manually by company employees or automatically by a platform’s software and algorithms, affects what viewers can see on a digital media page. Messages that are promoted garner greater viewership and interaction; those that are deprioritized or removed will obviously receive less attention. Content moderation policies reflect decisions by digital platforms about the relative value of posted messages.

As an attorney, professor and author of a book about the boundaries of the First Amendment, I believe that the constitutional challenges presented by these cases will give the court the occasion to advise government, corporations and users of interactive technologies of their rights and responsibilities as communications technologies continue to evolve.

Opinion bug

Opinion

In late October, the Supreme Court heard oral arguments on two related cases in which both sets of plaintiffs argued that elected officials who use their social media accounts either exclusively or partially to promote their politics and policies cannot constitutionally block constituents from posting comments on the officials’ pages.

In one of those cases, O’Connor-Radcliff v. Garnier, two school board members from the Poway Unified School District in California blocked a set of parents — who frequently posted repetitive and critical comments on the board members’ Facebook and Twitter accounts — from viewing the board members’ accounts.

In the other case, Lindke v. Freed, the city manager of Port Huron, Michigan, apparently angered by critical comments about a posted picture, blocked a constituent from viewing or posting on the manager’s Facebook page.

Courts have long held that public spaces, like parks and sidewalks, are public forums, which must remain open to free and robust conversation and debate, subject only to neutral rules unrelated to the content of the speech. The silenced constituents in the two cases insisted that in a world where a lot of public discussion is conducted on social media, digital spaces used by government representatives to communicate with constituents are also public forums and should be subject to the same First Amendment rules as their physical counterparts.

If the Supreme Court rules that public forums can be both physical and virtual, government officials will not be able to arbitrarily block users from viewing and responding to their content or remove constituent comments with which they disagree. But if the Supreme Court rejects the plaintiffs’ argument, the only recourse for frustrated constituents will be to create competing social media spaces where they can criticize and argue at will.

Moderation as editorial choice

Two other cases — NetChoice LLC v. Paxton and Moody v. NetChoice LLC — also relate to the question of how the government should regulate online discussions. Florida and Texas have both passed laws that modify the internal policies and algorithms of large social media platforms by regulating how the platforms can promote, demote or remove posts.

NetChoice, a tech industry trade group representing a wide range of social media platforms and online businesses, including Meta, Amazon, Airbnb and TikTok, contends that the platforms are not public forums, and that the Florida and Texas legislation unconstitutionally restricts the social media companies’ First Amendment right to make their own editorial choices about what appears on their sites.

In addition, NetChoice alleges that by limiting Facebook’s or X’s ability to rank, repress or even remove speech — whether manually or with algorithms — the Texas and Florida laws amount to government requirements that the platforms host speech they didn’t want to, which is also unconstitutional.

NetChoice is asking the Supreme Court to rule the laws unconstitutional so that the platforms remain free to make their own independent choices regarding when, how and whether posts will remain available for view and comment.

To reduce harmful speech that proliferates across the internet — speech that supports criminal and terrorist activity as well as misinformation and disinformation — the federal government has engaged in wide-ranging discussions with internet companies about their content moderation policies.

The Biden administration has regularly advised — some say strong-armed — social media platforms to deprioritize or remove posts the government flagged as misleading, false or harmful. Some posts related to misinformation about COVID-19 vaccines or promoted human trafficking. On several occasions, the officials would suggest that platform companies ban a user who posted the material from making further posts. Sometimes, the company corporate representatives themselves would ask the government what to do with a particular post.

While the public might be generally aware that content moderation policies exist, people are not always aware of how those policies affect the information to which they are exposed. Specifically, audiences have no way to measure how content moderation policies affect the marketplace of ideas or influence debate and discussion about public issues.

In Missouri v. Biden, the plaintiffs argue that government efforts to persuade social media platforms to publish or remove posts were so relentless and invasive that the moderation policies no longer reflected the companies’ own editorial choices. They argue that the policies were, in reality, government directives that effectively silenced — and unconstitutionally censored — speakers with whom the government disagreed.

The court’s decision in this case could have wide-ranging effects on the manner and methods of government efforts to influence the information that guides public debates and decisions.

Lynn Greenky is professor emeritus of Communication and Rhetorical Studies, Syracuse University.

The Sun-Times welcomes letters to the editor and op-eds. See our guidelines.

The views and opinions expressed by contributors are their own and do not necessarily reflect those of the Chicago Sun-Times or any of its affiliates.

This article was originally published on theconversation.com

The Latest
The police action began around daybreak Tuesday as campus officers surrounded the university’s main quadrangle. According to live reports from the scene, the police were keeping students from entering the quad.
Joel Edgerton, Jennifer Connelly keep morphing their characters in visually stunning tale of different Chicagos in the multiverse.
Barbara Glusak warned two of Washington Federal Bank’s board members about loan fraud, but the two did nothing to stop it — and were later convicted of participating in it. Glusak is now scheduled to testify at the sentencing hearing for Janice Weston and George Kozdemba.
DEI programs help eliminate hiring barriers for otherwise qualified candidates who don’t have “the inside track” for jobs, a nonprofit leader writes. Without such programs, Gen Z could lose faith in their career prospects.
A rougher allergy season and the prevalence of wildfires will make the summer especially tough for people with asthma.