By Brett Pinkus, Partner, Wick Phillips – Intellectual Property Litigation Group

Published in the Spring 2021 Issue of Accessible Law, UNT Dallas College of Law, April 27, 2021

The public seems to have a fundamental misunderstanding about the true extent of “freedom of speech” under the First Amendment. Who can or cannot restrict free speech? What type of speech can be restricted? And how does this apply to speech restrictions on social media platforms which have become so prevalent?

Lawsuits alleging free speech violations against social media companies are routinely dismissed.  The primary grounds for these dismissals are that social media companies are not state actors and their platforms are not public forums, and therefore they are not subject to the free speech protections of the First Amendment. Consequently, those who post on social media platforms do not have the right to free speech on these social media platforms. This article will attempt to explain the relationship between social media and free speech so that we can understand why.

Who Can Restrict Free Speech – State v. Private Actors

The overarching principle of free speech under the First Amendment is that its reach is limited to protections against restrictions on speech made by the government.1 The text of the First Amendment itself only prevents Congress (i.e., U.S. Congress) from making laws that restrict the freedom of speech. This protection is extended to the states, and to local governments, through the State Action Doctrine and the Due Process Clause of the Fourteenth Amendment.2 However, under the State Action Doctrine, First Amendment restrictions traditionally do not extend to private parties, such as individuals or private companies.3 In other words, a private person or private company (such as a social media company) cannot violate your constitutional free speech rights, only the government can do so. That is, unless the private party attempting to restrict speech qualifies for one of the three exceptions to the State Action Doctrine.

The first exception is when an action to restrict speech by a private party involves a function that is traditionally and exclusively reserved for the State, which is known as the Exclusive Public Function Doctrine.4 The Exclusive Public Function Doctrine is limited to extreme situations where a private party has stood in the shoes of the state. For example, when a private company has been given control of a previously public sidewalk or park, it has been found that the private company is performing municipal powers exclusively performed by the state.5 Courts have repeatedly refused efforts to characterize the provision of a news website or social media platform as a public function that was traditionally and exclusively performed by the government.6

The second and third exceptions, which are related to each other, are the entanglement and entwinement exceptions. The entanglement exception applies when an action to restrict speech by a private party is such that the state has significantly involved, or entangled itself, with the private action.7 This occurs when the “power, property, and prestige” of the government is behind the private action, and where there is evidence of the overt, significant assistance of state officials.8 The entwinement exception applies when an action of a private party can be treated as though the action were of the government itself (i.e., overlapping identities).9 These exceptions are rarely used in free speech cases and apply in very limited situations, typically in cases involving the Equal Protection or Establishment Clauses which are not relevant in most social media contexts.

Where Can Speech Be Restricted – Public v. Private Forums

When speech takes place in a public forum, that speech can qualify for protection of speech under the First Amendment.10 This is known as the Public Forum Doctrine. While there is no constitutional right for a person to express their views in a private facility (such as a shopping center),11 speech that takes place in a traditional or designated public forum for expressive activity (such as a sidewalk or park on government property) is protected and only limited restrictions of speech are allowed.12 A designated public forum can only be created when the government intentionally opens a nontraditional forum for public discourse.13 A private forum (such as a grocery store or comedy club), however, does not perform a public function by merely inviting public discourse on its property.14

Social media platforms are often characterized as a digital public square. Yet, courts have repeatedly refused arguments that social media platforms are public forums subject to the First Amendment.15 This reasoning is justified because their networks are private, and merely hosting speech by others does not convert a private platform to a public forum.16 Only in limited cases have social media sites been found by courts to qualify as a public forum. For example, in a recent case, an appellate court held that the official Twitter page operated by then President Donald Trump was a designated public forum. As a result, government officials could not engage in viewpoint discrimination by blocking individuals from posting comments with critical views of the President and his policies.17 In contrast, a private person or organization’s social media page is not a public forum and is not protected by the First Amendment.

Social media platforms may also be analogized to newspapers when they attempt to exercise editorial control and judgment over the publishing of users’ posts. In this scenario, the Supreme Court has held that newspapers exercise the freedom of the press protected by the First Amendment and cannot be forced to print content they would not otherwise include.18 This is due to a newspaper’s ability to exercise editorial control and judgment, including making decisions on the size and content of the paper, along with treatment of public issues and public officials (whether such treatment is fair or unfair). This leads us to next examine what protections are afforded to social medial companies for content posted by their users on their platforms.

Social Media’s Immunity for User Content – 47 U.S.C. § 230(c)

Section 230 of the Communications Decency Act (“CDA”), codified as 47 U.S.C. § 230, was enacted in response to a court decision ruling that an internet service provider, Prodigy, was considered a “publisher” of defamatory statements that a third party had posted on a bulletin board hosted and moderated by Prodigy, and Prodigy could therefore be subject to a civil lawsuit for libel.19 Sec. 230(c)(1) remedies this by providing immunity to internet service providers from lawsuits that attempt to make them liable for the user content posted on their sites.20 Social media companies, which are currently considered to be service providers under Sec. 230(c)(1), are broadly protected from responsibility for what users say while using their social media platforms.21

The next question that logically follows is whether a social media company can restrict or exercise editorial control over content on its platform. Sec. 230(c)(2) of the CDA answers this, by precluding liability for decisions to remove or restrict access to content that the provider deem “obscene, lewd, lascivious, filthy, excessively violent, harassing, or otherwise objectionable.”22 Social media platforms therefore set their policies and Terms and Conditions to state that they can remove violent, obscene, or offensive content and can ban users who post or promote such content. For example, Facebook, Twitter, and YouTube have banned terrorist groups that post material promoting violence or violent extremism, and have also banned ISIS, Al Qaeda, and Hezbollah solely because of their status as U.S.-designated foreign terrorist organizations. As was recently seen following the 2020 Presidential election, Facebook, Twitter, Snapchat, YouTube (Google), Reddit, and Twitch (Amazon) also justified their suspension of the accounts of President Trump and some of his supporters under Sec. 230(c)(2) for continuing to post misinformation, hate speech, and inflammatory content about the election.

What are Permissible Restrictions on Speech

As discussed above, if a social media company chooses to remove content from its platform in accordance with its designated policies, that removal does not raise a First Amendment issue and there is no civil liability as a result of Sec. 230 of the CDA. But what if precedent was to be reversed, and a social media platform was declared a state actor or a public forum such that the First Amendment would apply to them? Or what if Sec. 230 was repealed to make social media companies liable for their users’ posts when they attempt to moderate the content? If either were to happen, the type of speech being restricted would play a significant role in its permissibility.

Restrictions of speech in a public forum are permissible if they are appropriately limited in time, place, and manner.23 Speech can be restricted under a less demanding standard when it is done without regard to the content of the speech or the speaker’s point of view.24 A content-neutral restriction on speech, for example, would be prohibiting all picketing within 150 feet of any school building while classes are in session without regard to their message, whereas a content-based restriction would be one that allows picketing only if the school is involved in a labor dispute.25 Other reasonable content-neutral regulations include regulating noise by limiting decibels, or the hours and place of public discussion.26 It is unlikely that content-neutral restrictions could be implemented to effectively regulate violent, obscene, or offensive content on social media platforms, which leaves us with content-based restrictions that would be subjected to heightened scrutiny. Content-based restrictions in a public forum require that there must be a compelling government interest in the restriction and the least restrictive means are employed to further that interest.27

It is important to emphasize that the First Amendment “does not guarantee the right to communicate one’s views at all times and places or in any manner that may be desired.”28 For that reason, if there is an alternative channel of communication for the desired speech, it may be a suitable alternative even if it is not a perfect substitute for the preferred forum that has been denied.29 For example, if a user were blocked from posting on a social media platform, alternative channels to make the desired speech might include other social media platforms or different forms of media. Other possibilities might include remedial steps for regaining posting privileges, such as imposing temporary posting suspensions that can be lifted over time or requiring the poster to agree to specific posting restraints before regaining unrestricted access.

What Types of Content-Based Restrictions are Permitted

It is also worthwhile to review the types of protected and unprotected content-based speech to understand the extent of the speech protected by the First Amendment, particularly in view of the recent unrest reflected on social media following the 2020 election. Content-based restrictions on speech have been permitted within a few traditionally recognized categories of expression.30

Misinformation, Defamation, Fraud, Perjury, Government Officials

Misinformation is defined as false or inaccurate information. False statements of fact about a public concern or public officials are protected from censorship under the First Amendment, unless the statement is made with knowledge or reckless disregard that it was a false statement made and/or made with intent to harm.31 It is not safe to assume that false statements can be made on social media platforms without impunity. There can be civil liability imposed for defamatory statements, which are knowingly false statements of fact published without authorization that damage others’ reputations (e.g., libel if written and slander if spoken), and for fraud, which is a false statement of fact made with the intent to cause the hearer to alter their position.32 At the time of this writing, statements pushing claims of election fraud following the 2020 election made by various public figures and news commentators on television and social media are being pursued for defamation by electronic voting machine manufacturers Dominion Voting Systems and Smartmatic.

Hate Speech and Speech that Incites Imminent Lawless Action

The First Amendment generally protects even hate or racist speech from government censorship. However, speech advocating the use of force is unprotected when it incites or is likely to incite imminent lawless action.33 Likewise, speech that is considered an incitement to riot, which creates a clear and present danger of causing a disturbance of the peace, is also not protected by the First Amendment.34 “Fighting words” which “by their very utterance inflict injury or tend to incite an immediate breach of the peace” are unprotected and may be punished or prohibited.35

Harassment and True Threats of Violence

Harassment refers to unwanted behavior that makes someone feel degraded, humiliated, or offended. Harassing someone for the purpose of irritating or tormenting them is protected from censorship by the First Amendment. However, harassment that goes so far as to present a “true threat of violence,” is an exception not protected by the First Amendment and is banned by all social media platforms. True threats of violence directed at a person or group of persons that have “the intent of placing the target at risk of bodily harm or death” are unprotected, regardless of whether the speaker actually intends to carry out the threat.36 Intimidation “is a type of true threat,” and would likewise be unprotected by the First Amendment.37


Advertising, which is a type of commercial speech, receives only limited protection under the First Amendment.38 If an advertisement is shown to be misleading or unlawful, a restriction on that speech is permissible.39 A website or social medial platform, much like a newspaper, cannot be forced to print advertisements in contravention of their right of editorial control.40


Current legal precedent conclusively establishes that social media users do not have a right to free speech on private social media platforms. Social media platforms are allowed to remove offending content when done in accordance with their stated policies as permitted by Sec. 230 of the CDA, and that removal does not raise a justiciable First Amendment issue or a real risk of civil liability. The users, on the other hand, put themselves at risk of being banned for making violent, obscene, or offensive content on social media, and may even expose themselves to civil liability for making false, misleading, or violence-inciting statements.

About Wick Phillips’ Intellectual Property Practice

The attorneys in Wick Phillips’ Intellectual Property team has represented clients in the protection of their inventions, designs, trade names, and other original works. Our team of intellectual property attorneys balance their legal experience with business insight to provide comprehensive and strategic counseling for every type of intellectual property matter – from the procurement of intellectual property rights to litigation. To learn more about the practice, click here.

Sources listed in PDF version of this article.

View PDF