Justice Thomas jumped into the #Section230 debate to embrace GOP arguments for narrowing protections for content moderation. He might think differently in a case where the issues he raised were actually briefed by both sides—unlike this very narrow case
Thomas often issues such statements when SCOTUS decides not to take a case—to vent his frustrations about the state of the law
But this is the first time SCOTUS has ever considered taking a case involving #Section230. The briefs here did not even address the issues Thomas raises
Justice Thomas is free to call for fuller briefing on Section 230’s meaning in, as he says, “an appropriate case,” but this is not that case. Justice Thomas had no need to express his own views, in extensive dicta, without the benefit of the briefing he acknowledges is needed.
The Malwarebytes case was about content filtering tools. The Ninth Circuit imported a "good faith" requirement into #Section230 (c)(2)(B), which protects those who offer content filtering tools to others.
Ironically, Thomas objected to courts’ ‘rel[ying] on purpose and policy’ when interpreting Section 230, yet that is precisely what the Ninth Circuit did; it’s why the Supreme Court should have taken this case. The 9th Circuit read into the statute words that are not there.
It makes sense to require sites to prove ‘good faith’ when claiming (c)(2)(A)’s immunity for their content moderation decisions
It makes no sense when a filtering tool developer claims (c)(2)(B)’s immunity for providing its tool to others to make their own decisions
Letting the Ninth Circuit’s decision stand invites litigation against the makers of malware software, parental controls and other tools that empower users to filter content online. That liability will cause many small developers to exit the market even before they are sued.
The Malwarebytes case involved only the interplay between (c)(2)(A) & (c)(2)(B), not the interplay between (c)(2)(A) & (c)(1) or the meaning of (c)(1)
Thomas argues that (c)(1) protects only decisions to leave content up and only (c)(2)(A) protects decisions to take content down
Justice Thomas’s opinion tracks political talking points advanced by the White House about #Section230. Those theories are framed in textualist terms, but they quickly break down upon close examination—as Thomas himself might ultimately agree if he waited to hear from both sides
Across the board, #Section230 protects the same thing the First Amendment does: editorial discretion. That’s why Congress said website operators cannot be held liable ‘as publishers’ for content they in no way created.
Refusing to carry content one finds objectionable is a core function of any publisher. The courts have interpreted the statute correctly — as a way to short-circuit expensive litigation and thus avoid what one appeals court called ‘death by ten thousand duck-bites.’
Instead of being able to resolve lawsuits over content moderation with a motion to dismiss, Thomas would force websites to litigate lawsuits through discovery—which, on average, accounts for up to 90% of litigation costs
That, in turn, will only discourage content moderation
But #Section230’s central purpose was to avoid the Moderator’s Dilemma: Congress wanted to ensure that websites weren’t discouraged from trying to clean up harmful or illegal content
If, as Thomas argues, 230(c)(1) doesn’t protect websites from being held liable as distributors for content they knew, or should have known, was illegal, this liability will create a perverse incentive not to monitor user content—another version of the Moderator’s Dilemma
Holding sites liable for content they edit in any way, as Thomas proposes, could, conversely, discourage them from attempting to make hard calls, such as blotting out objectionable words (eg racial sluts) while leaving other content up
They may simply take down content entirely
So, in summary...
Thomas's opinion has no legal weight, isn't based on the kind of briefing the Court relies on in actual decisions
He should have limited himself to urging plaintiffs to bring more #Section230 cases so these issues could be fully briefed
We explained why the Court should have taken, and reversed, the Ninth Circuit's decision in Malwarebytes when we filed our amicus brief back in June: techfreedom.org/scotus-should-…
#Section230 protects “tech platforms” just as it protects National Review’s site, or a user’s (ahem, Trump's) retweet of someone else’s defamatory statements
230 DOES Internet media differently from other print & broadcasting, because they ARE different
Traditional publishers review content pre-publication but Internet media just can’t b/c:
—SCALE: billions of pieces of content created daily
—SPEED: much content is real-time
WEBSITES AREN’T PUBLIC FORA: Supreme Court jurisprudence and case law DOES NOT support the EO’s claims that they are. The EO cites two cases that don’t apply to social media platforms:
1) Pruneyard (1980) was limited to shopping malls (very different from websites) and definitely wouldn’t be upheld by the Court today anyway, as made clear in Johnson v Twitter (2018) 2) Packingham (2017) is about restrictions on Internet imposed by STATE LAW, not private actors
1/6 @LindseyGrahamSC's #EARNIT Act would give AG Barr a blank check—via #Section230—to crack down on Internet services, effectively ban secure #encryption & impose other de facto mandates that could never get through Congress as legislation
2/6 AG Barr could use Graham’s bill to force Apple to give law enforcement a backdoor on iMessage, iCloud or even iPhones—effectively banning end-to-end encryption
3/6 #Section230 has never shielded child porn (CSAM) traffickers from federal prosecution, but Graham's bill would create vast new legal liability for websites—then use that liability to force them to do whatever the AG commands
Trump’s order would transform the FCC & FTC from consumer protection agencies into regulators of online speech
Rs complain about ‘censorship’ by private companies. But this order would mean REAL censorship, empowering regulators to decide what kinds of speech are allowed online
That the GOP, after decades of fighting government meddling in broadcasting, now wants its own Fairness Doctrine for the Internet is staggeringly hypocritical
We value the balance between freely exchanging ideas, fostering innovation, & limiting harmful speech. Because this is an exceptionally delicate balance, #Section230 reform poses a substantial risk of failing to address policymakers’ concerns and harming the Internet overall.
Principle #1: Content creators bear primary responsibility for their speech and actions.
TRUMP: I would work with Democrats to regulate social media platforms to rein in their power
Guys, c'mon. You can't possibly be dumb enough to fall for this. He's been on the warpath about social media companies "censoring conservatives" (without actual evidence) for months
1/7
You REALLY want to give Trump broad power to crack down on a website just because user content on the platform pisses him off?
Jeff Sessions has been rallying state AGs to do just this
So that's why Trump is trying to get new legislation passed. The First Amendment SHOULD protect websites' exercise of editorial discretion, but there's a lot the President could do to circumvent its protections and harass websites—if given enough power by Congress
3/7