In today's #vatniksoup, I'll talk about a phenomenon called information laundering. It refers to an activity where false or deceitful information is legimitised through a network of intermediaries, such as fake news blogs and/or social media networks.
1/20
When for example the Kremlin wants to spread a narrative, they spread it through various channels to legitimatize it. As basically no one trusts the Russian TASS anymore, Russian propagandists must get creative and use various backchannels to legimitise the information...
2/20
...in order to inject into the mainstream. To do this, they'll have to "pass the information" around some other news outlets, preferably that the people (at least some) trusts. The groundwork for these outlets has been created a long time ago - for example a well-known...
3/20
...conspiracy theory website, Veteran's Today, was established already back in 2003, and the pro-Kremlin financial blog and "news aggregator" Zero Hedge was launched in 2009. Some of these may have even started as a legitimate info sources, but at some point they...
4/20
...turned into information laundering platforms. Also, many of the Facebook groups that were later exposed to be troll farms were established already around 2013.
Like money launderers who use shell companies, info launderers rely on certain social media accounts or fake news blogs that can then mask the original source and its intent.
6/20
Social media can also help with the legitimisation process, as people often perceive stories that are liked and shared as something important. In addition, a study has shown that fake news tends to spread much, much faster than factual news.
7/20
Troll farms can also be used to spread these messages even faster,and since Elon's takeover, these farms have regained their power to spread disinfo on Twitter.Both Russia & the CCP also use diplomat and embassy accounts actively to spread disinformation and false narratives.8/20
Social media platforms can also be the original source which is then spread throughout the fake news blogs. For example photoshopped images (Ukrainians with Nazi flags, doctored documents, etc.) or old videos with wrong context often start spreading on social media, ...
9/20
...after which the fake news blogs publish articles on them, hoping for the mainstream to take the bait.
The goal with all this is to "layer" the disinformation so, that it spreads from its point of origin to more credible sources.
10/20
This credibility can then be increased with reposts, likes and shares on social media. For example, the hacked Podesta e-mails in 2016 were spread through various middle-men, including WikiLeaks. These seemingly authentic platforms provided credibility and the ...
11/20
...hack-and-leak operation wasn't immediately identified as a Russian intelligence operation. The illusion of legitimacy of fake news blogs and/or individual "journalists" can also be increased by awards and nominations. One example of this is the "Serena Shim Award for...
12/20
...Uncompromised Integrity in Journalism", that's been awarded to totalitarian regime bootlickers like Max Blumenthal, Aaron Maté and Jackson "Z" Hinkle.
Both Russia and China also like to use Western public figures to propagate their agenda and strengthen it in their own..13/20
...news outlets. Outlets like TASS and Global Times often interview MEP's like Clare Daly, Mick Wallace or Maximillian Krah to promote their ideology, for example in case of genocide denial in Xinjiang or in Ukraine.
14/20
One example of relatively successful information laundering was the "secret bioweapons labs in Ukraine" hoax. The bioweapons lab theory (re-)surfaced on the same day when Russia started its invasion in Ukraine, and it was initially suggested by QAnon podcast host Jacob...
15/20
..."RedPill78" Creech from the US. He claimed that the Russians only bombed the locations of these bioweapons labs, and that this was the reason for the invasion. Three days later the story was tweeted out by the Russian Embassy in Bosnia and Herzegovina and from there...
16/20
...it quickly spread to fake news blogs and eventually to the mainstream media. It was even discussed on Tucker Carlson on many occasions and Russia's UN Ambassador Vassily Nebenzia repeated the allegations.
17/20
But information laundering isn't limited just to news blogs - it is often used in academia, too. The information laundered through these "scientific journals" is often outright propaganda camouflaged as research. Especially Alexander Dugin has been active in establishing...
18/20
...credibility for pro-Kremlin narratives in academia. Once published, these propaganda pieces can be used to counter publications of Western academia (which to be fair, can also be straight up propaganda).
19/20
To conclude: information laundering is relatively effective method of injecting disinformation and propaganda into the mainstream, and it's yet another form of hybrid warfare in which Russia and China have been forerunners for a good decade.
In today’s Vatnik Soup, I’ll explain the Alaska Fiasco and how it marks the peak of Trump’s two-year betrayal of Ukraine. What was sold as “peace talks” turned into a spectacle of weakness, humiliation, empty promises, and photo-ops that handed Putin exactly what he wanted.
1/24
Let’s start with the obvious: Trump desperately wants the gold medal of the Nobel Peace Prize, mainly because Obama got one. That’s why he’s now LARPing as a “peace maker” in every conflict: Israel-Gaza, Azerbaijan-Armenia, India-Pakistan, and of course Ukraine-Russia.
2/24
Another theory is that Putin holds kompromat — compromising material such as videos or documents — that would put Trump in an extremely bad light. Some have suggested it could be tied to the Epstein files or Russia’s interference in the 2016 US presidential election.
In today’s Vatnik Soup, I’ll talk about engagement farming: a cynical social media tactic to rack up likes, shares, and comments. From rage farming to AI-powered outrage factories, engagement farming is reshaping online discourse and turning division into profit.
1/23
Engagement farming is a social media tactic aimed at getting maximum likes, shares, and comments, with truth being optional. It thrives on provocative texts, images, or videos designed to spark strong reactions, boost reach, and turn online outrage into clicks and cash.
2/23
One subset of engagement farming is rage farming: a tactic built to provoke strong negative emotions through outrageous or inflammatory claims. By triggering anger or moral outrage, these posts often generate 100s or even 1,000s of heated comments, amplifying their reach.
In today’s Vatnik Soup, I’ll cover the autocratic concept of “Good Tsar, Bad Boyars”: the idea that the leader is wise and just, but constantly sabotaged by corrupt advisors. This narrative shields the ruler from blame, and it’s used by both Putin and Trump today.
1/20
The phrase “Good Tsar, Bad Boyars” (Царь хороший, бояре плохие), also known as Naïve Monarchism, refers to a long-standing idea in Russian political culture: the ruler is good and benevolent, but his advisors are corrupt, incompetent and responsible for all failures.
2/20
From this perception, any positive action taken by the government is viewed as being an accomplishment of the benevolent leader, whereas any negative one is viewed as being caused by lower-level bureaucrats or “boyars”, without the approval of the leader.
In today’s Vatnik Soup, I’ll introduce a Russian politician and First Deputy Chief of Staff of the Presidential Administration of Russia, Sergey Kiriyenko. He’s best known for running both domestic and foreign disinformation and propaganda operations for the Kremlin.
1/20
On paper, and in photos, Kiriyenko is just as boring as most of the Kremlin’s “political technologists”: between 2005-2016 he headed the Rosatom nuclear energy company, but later played a leading role in the governance of Russia-occupied territories in Ukraine.
2/20
What is a political technologist? In Russia, they’re spin doctors & propaganda architects who shape opinion, control narratives, and manage elections — often by faking opposition, staging events, and spreading disinfo to maintain Putin’s power and the illusion of democracy.
Let me show you how a Pakistani (or Indian, they're usually the same) AI slop farm/scam operates. The account @designbonsay is a prime example: a relatively attractive, AI-generated profile picture and a ChatGPT-style profile description are the first red flags.
1/5
The profile's posts are just generic engagement farming, usually using AI-generated photos of celebrities or relatively attractive women.
These posts are often emotionally loaded and ask the user to interact with them ("like and share if you agree!").
2/5
Then there's the monetization part. This particular account sells "pencil art", which again are just AI-generated slop.
In today’s Vatnik Soup, I’ll introduce an American lawyer and politician, Mike Lee (@BasedMikeLee). He’s best-known for opposing the aid to Ukraine, undermining NATO by calling the US to withdraw from the alliance, and for fighting with a bunch of braindead dogs online.
1/21
Like many of the most vile vatniks out there, “Based Mike” is a lawyer by profession. He hails from the holy land of Mormons, Utah, where he faces little political competition, allowing him to make the most outrageous claims online without risking his Senate seat.
2/21
Before becoming a senator, Mike fought to let a nuclear waste company dump Italian radioactive waste in Utah, arguing it was fine if they just diluted it. The state said no, the public revolted, and the courts told poor Mikey to sit down.