In today's #vatniksoup, I'll talk briefly about the Community Notes system and why it doesn't work. I've previously stated that the Community Notes mechanism is a "mob rule" and can be played easily by big accounts and troll farms.
1/15
Community Notes is a community-driven content moderation program, intended to provide informative context based on a crowd-sourced voting system. As of Nov 2023, this system had over 130 000 contributors.
2/15
The idea of a crowd-sourced system as a moderation tool did not come from Elon - it was announced already back in 2020 when it was called Birdwatch. Musk later rebranded the system as Community Notes and sold it to the platform as something new.
3/15
Vitalik Buterin (@VitalikButerin) has made a very extensive (and technical) analysis on the tool and the Community Notes algorithm as a whole.
I disagree with him on some points, but I really suggest to everyone to read it:
Also, focusing the analysis only on the algorithm and the technical aspects is simplifying the concept, as it rules out the most important variable: human factor.
People are prone to bias and disinformation tends to spread much more aggressively than the truth.
5/15
Twitter's former head of safety, Yoel Roth, has stated that the system was never intended to replace the curation team, but to complement it. But all this of course changed after Elon sacked everyone from Twitter's Trust and Safety in order to save money.
6/15
These sackings have resulted in long response times on reports on hate speech - X's attempts to deal with hateful direct messages has slowed down by 70%.
As of today, the company doesn't have any full-time staff singularly dedicated to hateful conduct issues globally.
7/15
Some Community Notes contributors (who are also NAFO activists) have claimed that the system is riddled with coordinated manipulation, infighting and lack of oversight from the platform. Also, many contributors engage in conspiracy-fueled discussions.
8/15
The Notes system also has a huge problem with its scalability. During events like the 7 Oct 2023 Hamas terrorist attack, the amount of disinformation grows so large, that it's simply impossible for the small community to keep up and check factuality of said content.
9/15
Analysis by NewsGuard showed that the most popular disinformation posts related to the Israel-Hamas war (not so surprisingly originating from serial liars like @jacksonhinklle, @drloupis and @ShaykhSulaiman) failed to receive Community Notes 68% of the time.
10/15
These big accounts also have the ability to fight against the Notes they've received by mobilizing people who support their views. In the most tragicomic instance, @elonmusk claimed, without any evidence, that a Community Note on his post was "gamed by state actors".
11/15
Other than being humiliated and ridiculed, getting Community Noted doesn't really have any major downsides. Noted posts don't provide you income and advertisers can decide if they want to show ads on accounts like @dom_lucre's, but most of these...
12/15
...so-called superspreader accounts make most of their income through other means, namely through the X's subscription system. Also, many of them, including @stillgray and (allegedly) @jacksonhinklle are employed by state actors like Russia and the CCP.
13/15
With accounts that post tens or hundreds posts a day, the Notes are also inefficient - while the Community is trying to put a note on a post that's clearly disinformation, there are already 10 or 20 new ones to replace it in the algorithm.
14/15
To conclude, Community Notes are a non-functional and slow mechanism that's desperately trying to replace the Trust and Safety team. They work on a "mob rule" basis and big enough accounts (including the owner of the platform) can play around the system.
In today’s Vatnik Soup, we’ll explain the context of the upcoming Budapest Blunder, and how it follows the infamous Alaska Fiasco from two months ago and Trump’s absurd delaying of serious aid to Ukraine and effective sanctions on Russia for the past nine months.
1/20
Two months ago, Trump embarrassed the United States by rolling out the red carpet for war criminal dictator Putin and overall acting like a pathetic servant eager to meet his master. Of course, the Alaska Fiasco didn’t bring peace any closer.
Worse, the main outcome of the humiliation was to delay serious sanctions, which the US Congress, in rare bipartisan unity against Russia, was on the verge of passing. Two weeks by two weeks, Trump Always Chickens Out, postponing any real pressure on Putin for 9 months now.
In today’s Vatnik Soup, I’ll introduce American propagandist Alexandra Jost, aka “Sasha” (@sashameetsrus). She’s best known for being paid by the Russian state to spread pro-Kremlin propaganda — and for doing it with a big smile.
1/23
Alexandra, now 26, was born in Hong Kong. Her father is from Texas, and her mother is from Siberia. According to her, she has “dreamed of living in Russia since childhood.” Sasha's mother runs a dance studio in Moscow and her younger brother is avoiding mobilization.
2/23
Since the beginning of her creator career, Sasha has been adamant about one thing: that she’s “never had to be paid” to speak of her “love” for Russia. But, as always with Russia, this turned out to be nothing but vranyo — a Russian “tactical lie.”
In today’s Vatnik Soup REBREW, I’ll reintroduce an American political commentator and pro-Kremlin propagandist, Tucker Carlson (@TuckerCarlson). He’s best known for his promotion of crazy conspiracy theories and for his support of authoritarian regimes around the world.
1/24
Tucker’s career spans decades, but he’s also been very active in recent years, so a lot has happened since our previous soup on him, which can be found here:
Once described as “the most powerful conservative in America”, Tucker has now fully transformed into a grifting conspiracy theorist and propagandist willing to work for whoever pays him the most. It’s unclear whether Tucker truly believes his endless conspiracy theories or…
3/24
In today’s Vatnik Soup, I’ll cover the agenda-setting and flood of disinformation that spread on X and other platforms right after Charlie Kirk’s assassination. It’s far from the first or last time a tragedy has been weaponized for political purposes.
1/18
Every major political event, especially those involving violence, attracts massive attention. In the immediate aftermath, reliable information is scarce, making it highly vulnerable to both coordinated and improvised disinformation campaigns.
2/18
As I’ve mentioned in my previous soups and lectures, in disinformation campaigns, being first with a narrative is crucial, as people often remember the first version best — psychology studies show it sets the mental schema, and later updates rarely overwrite it.
In today’s Vatnik Soup, I’ll introduce American social media personality David Freeman, AKA Gunther Eagleman™ (@GuntherEagleman). He’s best known for spreading political disinformation on X and shamelessly sucking up to Trump, Putin, and other authoritarian leaders.
1/22
David is a textbook example of someone profiting from MAGA grievance politics. He uses extreme, provocative language to farm engagement on X and never hesitates to flatter anyone who might give him more exposure — or money.
2/22
But David wasn’t always like this. At some point, in his mid-40s, he even tried a real job: he trained to become a cop. He spent three years with the Metro Transit PD, but after that he either got fired or quit, and never looked back.