This week, the White House released its antimonopoly for the tech sector. It's important, because a free, fair and open internet is a necessary precondition for organizing all our other fights about human rights, equity, labor, the climate, and racial and gender justice. 1/
If you'd like an essay-formatted version of this thread to read or share, here's a link to it on pluralistic.net, my surveillance-free, ad-free, tracker-free blog:
The White House plan is a mixed bag. They set out six action points, each of them amorphous enough that they could all be summarized as "the devil is in the details" - that is, depending on how these are handled, they could be great, or terrible. 3/
But one point stands out as especially fraught, controversial and dangerous: a vague promise of "fundamental reforms to Section 230," which is incorrectly characterized as "special legal protections for large tech platforms." 4/
I'm going to go through all the points below, and in the end I'll get into detail on 230 - it's one of the worst-understood areas of internet law, a favored punching bag of the right and the left, and getting this one wrong could deliver permanent dominance to Big Tech. 5/
I. "Promote competition in the technology sector": This covers both meat-and-potatoes trustbusting (breakups, merger scrutiny) and modern, tech-specific tactics, like interoperability mandates and bans on self preferencing. 6/
This is generally great stuff, but there are three important pitfalls to avoid:
i. Interop mandates that expose users to risk through hasty action. The EU's Digital Markets Act unwisely kicked off by mandating interop in messaging tools on an unrealistically short timeline. 7/
Maintaining the security of encrypted messengers is extremely important; failures in messaging encryption are a source of existential risk to human rights workers, journalists and marginalized people all around the world. 8/
Recall that Jamal Khashoggi was lured to his slaughter by the Saudi government after they broke into his peers' encrypted messages using a cyberweapon produced by the NSO Group.
ii. Must-carry rules that force platforms to carry speech. Big online platforms have become our new public square, except that they aren't public - they're private. 10/
Their choices about which speech to block and which speech to carry are enormously consequential for our civics and politics. But rules that allow regulators to force providers to carry speech they disagree with set a dangerous precedent. 11/
Even if you think that the Biden admin's compelled speech will be fine (say, a rule requiring warnings alongside vaccine disinformation), imagine how this power will be handled by President Marjorie Taylor Green's administration. 12/
The platforms' moderation choices are a danger because the platforms dominate our discourse. Allowing the platforms to corner the market for online speech has profound First Amendment implications:
But the answer isn't to turn the platforms into an arm of the state - it's to make their moderation choices less consequential for all of us, by devolving control over community norms to the communities themselves:
iii. Self-preferencing bans are *very* hard to administer. If Apple puts its own weather app at the top of the app-store listings, or if Google shows you an infobox with its weather prediction at the top of a search, that might feel like self-preferencing. 15/
But maybe Apple *really* believes that it has the best weather app. There isn't an objective standard for "best weather app." Unless you've got a front-row seat for the wall of Plato's Cave, distinguishing self-preferencing from good-faith curation is often impossible. 16/
That's not to say that we should tolerate self-preferencing, nor is it to say that we can't *ever* detect and punish self-preferencing. 17/
Sometimes, tech companies actually *document* the fact that they're self-preferencing, as Google did when its engineers emailed their bosses to complain about being forced to put Google's inferior results ahead of rivals:
But we can't rely on Big Tech tripping over its own dick every time it does a bit of nefarious self-preferencing. 19/
The *real* remedy for self-preferencing is #StructuralSeparation: banning platform operators from competing with platform users. Referees shouldn't own one of the teams on the field, period.
II. "Provide robust federal protections for Americans’ privacy." A no-brainer. 21/
The US needs a federal privacy law, with a private right of action that allows individuals (and human rights groups) to sue firms that violate it, rather than waiting for a prosecutor to take up their cause. Do it. 22/
I'm entirely unsympathetic to the argument that "targeted ads" are better than "untargetted ads" because they are "more relevant" to users. Users *fucking hate* targeted ads. Ad-blockers are the largest boycott in human history. 23/
When users are given the change to opt out of targeted ads, they do so in such overwhelming numbers that the holdouts are likely to be people who accidentally clicked the wrong button:
III. "Protect our kids by putting in place even stronger privacy and online protections for them, including prioritizing safety by design standards and practices for online platforms, products, and services."
Sounds good. As a dad, I like the idea. 25/
But there's so many ways it can go wrong. California's version of this rule was so vaguely worded that it's effectively impossible to comply with.
It's not just that this could result in kids being banned from using any online service - it's also that all online services might institute invasive verification procedures (like storing - and, inevitably, leaking - government IDs to prove that none of their users are kids). 27/
But the difficulties here don't mean we have to be nihilists. We can demand that platforms that target kids - that market themselves as services for children - eschew advertising, minimize data collection, and take other steps to protect kids from commercial predation. 28/
V. "Increase transparency about platform’s algorithms and content moderation decisions." Opponents of this one will claim that telling people how you moderate is a gift to trolls and griefers. 29/
I'm unsympathetic to the idea that there is "security through obscurity":
There's a lot of room for debate about how the "civil justice" system of big platforms should operate. 30/
One thing is clear: automated judgments about user speech can't be balanced by human review. The former happens at scale and near-instantaneously. The latter will either be deliberative and too slow to matter, or rapid and too quick to make sense of nuance. 31/
One intriguing idea is to structure moderation review as a "systemic" matter, which can address "immoderation" (content that *isn't* moderated) as well as moderation. Note: no one has tried this yet, so while it sounds great, it's also a gamble:
VI. "Stop discriminatory algorithmic decision-making." This one is also maddeningly vague. If they're talking about ensuring that machine learning classifiers don't discriminate on the basis of speech, it's going to be very hard to make work. 33/
Algorithmic moderation often operates on the *context* of speech as much as the *content* - if many seemingly coordinated users all post harassmentm that speech might get labelled or deleted. The exact same speech, posted by one person, once, might be left alone. 34/
But there's another kind of algorithmic discrimination, whose most obvious case is the algorithms that target predatory financial products to Black users, or exclude women and racial minorities from being shown good jobs on employment sites. 35/
This is illegal - and we don't need new laws to prosecute it. But we *do* need new enforcement powers and resources for existing regulators to tackle it. 36/
All right, that's the five least controversial points in the White House plan. But I left out point IV: "Remove special legal protections for large tech platforms." 37/
Here, the White House is talking about Section 230 of the Communications Decency Act, AKA "The 26 Words That Made the Internet."
CDA230 is a rule that says that if a user's speech violates federal law, legal responsibility for that speech falls on the speaker, not the intermediary that brought you that speech. It's a rule that makes hosting user speech possible, period. 39/
It's how we get Facebook and Twitter, sure - but also how we get blog comments, Mastodon instances, and other independent platforms. 40/
It's also how we get the infrastructure that makes it possible for individuals, nonprofits, private groups and co-ops to create their own speech forums. 41/
CDA230 means that a hosting company doesn't need to review all its customers' users' speech before hosting them (imagine if every web-page had to be vetted by your host before you could make it live - and then every change also had to go through legal review). 42/
This is important in a competitive market, but it's even more important in our current, monopolized world, where getting kicked off of a platform might doom a speech forum. 43/
Again, if you're comfortable with this being used to nuke forums that the politicians you agree with get rid of, imagine which forums President DeSantis will target. 44/
Any gun on the mantlepiece in Act I is sure to go off by Act III. If we hand any aggrieved party the right to remove speech without a trial, we can be sure that this facility will be abused by the worst people in the worst ways. 45/
We know this because we've got decades of experience with the "notice-and-takedown" system for copyright enforcement, which allows anyone claiming to be a rightsholder to get almost anything taken down from almost anywhere, irrespective of whether an infringement took place. 46/
To see that in action, check out Eliminalia, which sells fraudulent copyright takedowns to launder the reputations of dictators, torturers, murderers and rapists, getting articles and personal accounts of their victims and survivors removed::
In Germany, @sonymusic is attempting to force @Quad9DNS, a public DNS provider, to block the records of websites whose users have allegedly posted links to *other* websites where infringing copies of Sony's copyrighted works can be found:
Sony is a serial abuser of its ability to moderate speech. They routinely, wantonly deletes independent performances of classical compositions by falsely claiming they violate Sony's copyrights. Sony is a music pirate on an *unimaginable* scale:
Evidence for what a post-CDA230 world would look like isn't limited to the copyright wars - for a more recent, more direct look at what happens when you make intermediaries responsible for their users' speech, look at the aftermath of SESTA/FOSTA.
SESTA/FOSTA is a (nominal) anti-sex-trafficking rule that creates criminal liability for companies whose services are used in connection with the heinous crime of sex trafficking. 51/
The immediate impact of SESTA/FOSTA was the mass, internet-wide removal of sites that sex workers used to keep themselves safe. 52/
SESTA/FOSTA pushed sex workers back onto the streets, deprived them of the forums where they shared information about dangerous clients, and created a renaissance in pimping, as sex workers were forced to turn to third parties for their protection.
Curbing CDA230 is especially dangerous in light of the calls for a "fairness doctrine" for online platforms. 54/
One of the activities that CDA230 protects is moderation, allowing online hosts to remove harassing, hateful, threatening or otherwise odious speech without worrying that this requires that they remove *every* such instance. 55/
This allows moderators to distinguish between a racist who calls another user by a slur, and a user who says, "Can you believe that racist called me :slur:?" 56/
Before 230 was enacted, courts took the position that once a service moderated *any* speech, it took on the duty to moderate *all* speech, creating the perverse incentive to ignore bad speech. 57/
Some say that CDA230 protects Big Tech platforms only to the extent that it protects *all* online speech forums, including independent ones. But this is wrong. 58/
CDA230 protects small platforms *more* than it protects large ones - because large ones are better situated to hire the armies of lawyers and moderators to pore over and comma-fuck everything their users post. 59/
That's why Mark Zuckerberg supports eliminating CDA230. As he is fond of pointing out, Facebook's budget for human moderators exceeds Twitter's total *revenue*.
He understands that if you need to be as big as Facebook to compete with Facebook that: 60/
a) No company will ever compete with Facebook, and
b) No government will ever make Facebook any smaller.
It's not just Zuck that hates 230 - it's also Donald Trump. 61/
Trump understands that removing legal protections for intermediaries will make them less able to stand up to rich and powerful people who can hire vicious attack lawyers who pride themselves on suppressing speech:
Trump loves the kinds of lawyers who kept #MeToo at bay for decades, not just by threatening the survivors of abuse, but by scaring anyone who might host their testimony into removing it. 63/
The fact that Zuck and Trump think killing CDA230 is a great idea should at least give its progressive opponents a moment's pause.
I'm Kickstarting a DRM-free audiobook of Chokepoint Capitalism, the book @rgibli and I wrote about how tech- and entertainment monopolies rip off artists, and how to unrig those terrible labor markets:
My fellow Americans, I regret to inform you that our beloved health insurance industry has done a major fuckery. I know this is hard to believe, given the probity and honor we associate with our fine insurance companies, but the evidence is incontrovertible. 1/
Back in 2019, the Trump administration ordered insurers and hospitals to start disclosing their prices, despite tens of thousands of comments filed by employers, insurers and hospitals objecting to the proposal. 2/
This is one of those pox-on-all-your-houses/you-can't-get-there-from-here situations. The Trump admin wanted to continue the fiction that the blame for America's worst-in-class health care was the result of bad market dynamics. 3/
I'm Kickstarting a DRM-free audiobook of Chokepoint Capitalism, the book @rgibli and I wrote about how tech- and entertainment monopolies rip off artists, and how to unrig those terrible labor markets: