Health Nerd Profile picture
23 Nov, 33 tweets, 11 min read
The anonymous people behind ivm meta dot com have put together a response to this excellent piece. Their main argument is that it doesn't look at all the evidence

So, following Scott Alexander's fine example, let's briefly review the prophylaxis literature 1/n Image
2/n Here are all 15 prophylaxis studies, in their wonderful glory

I'm going to try and be brief, but we'll see how that goes Image
3/n First up, Shouman. In this 'randomized' study, the authors ceased allocating people into the control group at some point, no allocation concealment, and massive differences between groups at baseline

Incredibly low-quality. Exclude
4/n Then we've got Carvallo, twice. This is the study with numeric mistakes in the reporting of results, and where the hospitals and researchers involved have said they have never heard of the study

Likely fraud. Exclude
5/n Next, Behera #1. A case-control study that matched on gender, age, and date of diagnosis. Obviously, this is not sufficient to control for confounding from any number of reasons ImageImage
6/n However, worse still there appear to be serious concerns about both this and the other Behera study. At least one of the data files has demonstrable concerns
7/n On top of that, at an absolute minimum, both studies were conducted on the same patient population at the same time with almost exactly the same number of positive tests

Many red flags. Exclude
8/n Moving on, we've got Hellwig. This study and Tanioka ran fundamentally the same analysis, so I'd group them together

The analysis is, in a word, shit
9/n Both studies are ecological in nature, and simply compare places where ivermectin was used in mass drug administration programs in Africa against reported cases/deaths
10/n Obviously, this is a very weak methodology that you'd usually exclude from any meta-analysis. However, in addition (for the reasons outlined in the thread above), it's also gibberish

Exclude
11/n Benigaud. Short case-series on 69 people who received ivermectin while living in a nursing home, compared to a synthetic control drawn from other nursing homes

Pretty useless as evidence. Exclude ImageImage
12/n Next, Alam et al. Non-randomized trial conducted in Bangladesh, which found vast benefits for ivermectin as prophylaxis Image
13/n This is obviously very low-quality, given that the methodology is scant, there's virtually no information on key metrics of interest (COVID-19 exposure, for example), and as an observational trial we'd usually exclude it simply because of those issues
14/n But on top of that, there's some incorrect/impossible numbers in the document, there are passages that appear to have been copied from other work (h/t @JackMLawrence), and the attack rate in the control group is wildly implausible

Exclude ImageImage
15/n Next, we've got IVERCOR PREP, which it turns out is a single slide tweeted out once by an anonymous account

Obviously exclude, but I thought I'd investigate further anyway Image
16/n I emailed the senior author on the IVERCOR paper, and asked about this slide. It turns out, this was merely an illustrative slide from a presentation on the paper given by one of the authors, and he was shocked that anyone would be using it as evidence of anything
17/n That brings us to Chahla et al. The second randomized trial on our list!

It's really, deeply shit medrxiv.org/content/10.110…
18/n The randomization process makes allocation concealment impossible. They report 100% follow-up of healthcare personnel over several weeks. There are numeric errors in the tables. They have issues with the sample size

So many red flags. Exclude ImageImage
19/n Moving past the second Behera study and Tanioka, which I've discussed already, brings us to the first half-decent study so far, Seet. A cluster-RCT from Singapore that divided dorms into different treatments, and looked at the effect on risk of COVID

ijidonline.com/article/S1201-… Image
20/n There are definitely some quibbles here. Clustering at the dorm level leads to issues with sharing of meds (that the authors acknowledge). Also, there were some issues with part of the primary outcome measure of seropositivity
21/n However, the study is reasonably solid as such things go. For the primary, pre-registered outcome measure there was a small, non-significant benefit for ivermectin compared to vitamin C (65% vs 70% infected) Image
22/n There was a lower proportion of people reporting symptoms in the ivermectin group when compared to placebo, but since this was open-label and that wasn't part of the main cluster-controlled analysis it's hard to interpret that result

Include
23/n Here we're getting to the last few studies. Morgenstern is an observational study from the Dominican Republic that used a propensity scoring model to compare healthcare workers who did and didn't take ivermectin cureus.com/articles/63131…
24/n From an initial sample of 943, the authors here whittled their numbers down to 542 in an odd series of exclusions. Why you'd discard half your sample in an observational trial is confusing to me Image
25/n The propensity model controlled for gender (m/f), exposure (l/m/h), and role (doctor/nurse/assistant/other)

That's just...not sufficient. Huge potential for residual confounding, especially as they had to discard ~1/4 of their sample to use the propensity model Image
26/n This is the sort of study that might've been useful if the authors had had 20,000 patients matched across dozens of variables, but this methodology doesn't eliminate key biases on a sample this small

Exclude
27/n Then we've got Mondal. This is a slightly odd serosurvey of healthcare workers, where the authors ran a logistic regression looking at factors that predict a positive IgG result for SARS-CoV-2

onlinejima.com/read_journals.…
28/n As far as uncontrolled studies go, it's a bit weak. Some of the percentages reported in the paper are off, the numbers don't quite match, and the methodology is pretty mediocre. Massive potential for residual confounding, and also no real control group

Exclude
29/n Finally, the newest trial, Samajdar, where a group of Indian researchers ran an online survey about which medications healthcare workers liked

Just...what? This is not evidence about ivermectin in any sense. Exclude

japi.org/x2a464b4/iverm… Image
30/n Overall, that leaves us with this picture. One study among 15 that isn't potentially fake, extremely bad, or just useless as evidence Image
31/n There's not a whole lot you can take home from Seet, to be honest. The potential benefits are mostly in reported symptoms which may be biased due to the lack of blinding. The more objective lab tests found very little, if any, benefit for ivermectin
32/n So what do you say to the question of whether ivermectin works as a prophylactic?

NO ONE KNOWS

The evidence is shit, and the one decent-ish study inconclusive. A big old shrug 🤷‍♂️🤷‍♂️
33/n Anyway, the simple fact is that the prophylaxis literature is, if anything, WORSE than the treatment studies. This is evidence so woeful that you could not possibly draw a reasonable conclusion about whether ivermectin does or doesn't work

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Health Nerd

Health Nerd Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @GidMK

22 Nov
What's really interesting is that this has >4,000 retweets and no one bothered to read the source website which explicitly says not to do this Image
The thing about reporting systems such as this is that they rely entirely on self-reports, which are themselves influenced by media attention. This makes comparisons problematic
The HPV vaccine is probably the most talked about product of all time, and even then we were lucky to get a single news story in a global publication once every few months since it was rolled out
Read 5 tweets
13 Nov
It's always fun being a debunker of sorts because whether people love it hate me depends on the popular myth du jour
I've had right-wing people super happy with me when I pointed out that fearmongering about glyphosate didn't really make sense, and very upset when I pointed out that fearmongering about lockdowns was also overblown
The low-carb people got really angry when I pointed out that the evidence isn't strong for any particular diet for long-term weightloss, but love me every time I note that meat is not as bad as headlines often paint it
Read 4 tweets
10 Nov
One interesting thing that I think this underscores is how terrible people are at judging the benefits of treatments based on their own experience. The true figures show that people who had MATH+ may have been *more* likely to die
Now, I don't think that's likely to be causal - the treating clinicians were probably just giving the treatment to the sickest people - but the basic narrative supporting the FLCCC protocol has been clinical expertise
The central idea driving FLCCC is that these experts in clinical care - which make no mistake, they are - could tell based on their own recollections which treatments worked best
Read 5 tweets
9 Nov
Wow. The key FLCCC paper on their main protocol has been retracted after understating the true number of deaths in the treatment group by ~78%
retractionwatch.com/2021/11/09/bad…
From the hospital where this work was conducted:

"...a careful review of our data for patients with COVID-19...only 25 of 191 patients (13.1%) received all 4 MATH+ therapies, and their mortality rate was 28%" rather than 6.1% cited in the article

journals.sagepub.com/doi/pdf/10.117…
"This would be an incorrect calculation of a hospital mortality rate, but might explain the incorrect number of 6.1% in Table 2. Using this incorrect mortality
rate to compare with the published reports and claim a “75% absolute risk reduction” is thus an incorrect conclusion"
Read 5 tweets
8 Nov
One of the weirdest things about the ivermectin discussion has been watching people go from "the evidence is good quality" to "it's just one fraud" to "lots of fraud isn't an issue" as the story has progressed
Most of the time, having the single biggest and most positive study looking at a topic withdrawn due to research misconduct would be enough for any researcher to reassess their position, but not with ivermectin!
Instead, people argue that it's fine because there's a "baseline" of fraud in other research, or because there are still (mostly null) trials which are not fraudulent
Read 4 tweets
7 Nov
Agreeing to do free work (reviewing papers) should not require more free work (creating 100s of logins on editorial management systems)
Oh, I forgot the specific, unique username you assigned to me 5 years ago? WHAT A SURPRISE GUESS I'LL RESET THAT
"Use your ORCID to login!"

*tries ORCID*

"Looks like you don't have an account. Create one today!"

*Creates account*

"Looks like you DO have an account. Forgotten your username?" 🙃🙃🙃
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Thank you for your support!

Follow Us on Twitter!

:(