Kert Viele Profile picture
11 Sep, 14 tweets, 3 min read
(1/n) I’ve been following discussions of #Bayesian sequential analyses, type I error, alpha spending etc. At the risk of offending everyone (please be kind!), I see reasons Bayesian sponsors and regulators can still find value in type I error rates and so forth.
(2/n) I’m focused on the design phase. After the trial, the data is the data. Lots of good stuff has been written on the invariance of Bayesian rules to stopping decisions. But in prospectively evaluating a trial design, even for a Bayesian there is a cost to interims, etc.
(3/n) Example…ultra simple trial. Normally distributed data. Known sigma=2. Mean is either 0 (null) or 1 (good). Simple decision space at end of trial…approve or not. A Bayesian utility would place values over the 4 combinations of truth/decision.
(4/n) To keep to 240 characters in what follows, those 4 combinations of truth/decision are approve good drug (AG), approve null drug (AN), do not approve good drug (DG) or do not approve null drug (DN).
(5/n) Approving a good drug is good. Approving a null drug is bad. Definitely can argue about the utilities of non approval. It won’t affect my main point so I’m going to keep it simple and say U(approve good=AG)=1, U(AN)=(-a), U(DG)=0, and U(DN)=0.
(6/n) After you collect data, you get a posterior probability the drug is good (Pi). You should approve if the expected utility of approval is greater than the expected utility of non-approval. With a little math, approve if Pi > a/(1+a).
(7/n) Suppose I run a fixed (non-adaptive) trial with 50 patients and use a=1. My prior probability of a good drug is 10%. I approve if my sample mean exceeds .675. In the design phase, what is the expected utility of my entire trial?
(8/n) The expected utility is the sum, over the 4 decision/truth outcomes, of Pr(outcome)*U(outcome). One of these terms is
Pr(approve null drug)*U(approve null drug) = Pr(app | null) Pr(null) U(AN). Pr(app|null) is the type I error.
(9/n) Another term is Pr(approve good drug)*U(approve good drug) = Pr(app | good) Pr(good) U(AG). Pr(app | good) is the power. Both the type I error and power are key components of the Bayesian expected trial utility.
(10/n) With a fixed n=50 trial, my expected utility is 0.080 (an infinite sized trial would have expected utility 0.100). I’d need to have meaningfully scaled utilities for this to be interpretable (lives saved, etc.), but here I just want to do some comparisons.
(11/n) Now consider a Bayesian sequential trial with max n=50. I look after every patient, and stop/approve whenever my posterior probability meets the Pi > a/(1+a) condition. What is the expected utility of this sequential trial? It’s MUCH lower….0.023 (compared to 0.080).
(12/n) If I’m a regulator, this matters to me. I want trials that meet some minimum standard so the utility of the drug supply I approve stays above a certain level. It’s reasonable for me to ask trials to have sufficient expected utility.
(13/n) I can certainty create an adaptive Bayesian trial that has the same expected utility as the N=50 trial. I have to have a bigger maximum than n=50 (expected N might still be lower), and I may want to think about my interim schedule.
(14/14) None of this argues for type I error etc, after the trial is done. Nor is it anti-adaptive. Just simply that many of the things frequentist think about (have a low chance of approving a bad drug) still have relevance somewhere in Bayesian trial design.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Kert Viele

Kert Viele Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @KertViele

1 Aug
(1/12) I receive emails warning about hackers targeting COVID data. Huge issues..it’s a crime and a danger to trial integrity. But also begs another discussion…if your loved one could be randomized but you could just select the current best arm, how much is gained?
(2/12) Your timing matters. The information you could use to select an arm accumulates over time.
The design matters. Some designs treat patients within the trial better.
(3/12) What is “the cost of randomization”? In a mortality trial, we might define it as the difference in survival probability between being randomized versus just being allowed to choose the current observed best arm.
Read 12 tweets
2 May
(1/n) I saw this recently arguing against response adaptive randomization (RAR) in platform trials. There are good counterarguments to the objections…here is a tweetorial on the RAR debate (refs at end) and why RAR is used in many modern platform trials.
ncbi.nlm.nih.gov/pubmed/32222766
(2/n) RAR is an adaptive experimental method that employs multiple interim analyses. At each interim allocation for poorly performing arms is reduced (perhaps to 0) and allocation for better performing arms is increased.
(3/n) RAR was initially proposed to benefit patients within a trial (“internal patients”). The intent is that as the trial evolves, more patients are placed on better performing arms, resulting in better outcomes.
Read 28 tweets
7 Apr
(1/20) Tweetorial on this statement…

”…red flag…The investigators….already taken an interim look…did not stop the study early….dampens optimism for an overall positive outcome”

Is this a red flag? Maybe…maybe not…depends on how optimistic you were to start.
(2/20) I don’t have the details of the interims. If someone has them, I’ll followup with this specific trial. Without them, I’m going to discuss the general principles. The principles are easier to see in a single arm trial. They generalize completely to two arms.
(3/20) Example…dichotomous outcome, want to beat a 40% control rate. A 200 patient single arm trial has 80% power for a true 50% treatment rate. Being adaptive, we will add interim looks at N=100 and N=150 that allow us to stop early for success or futility.
Read 20 tweets
18 Feb
1) Utilizing any external data in a clinical trial, including real world evidence, presents risks and benefits. These should be quantified and balanced by choosing an appropriate design (a tweetorial explaining this graph…)
2) Simple example…dichotomous outcome, you have a novel therapy and resources for 60 patients. You also have a database of untreated patients showing a 40% response rate. How can you incorporate the database into the trial?
3) Statistically, the risks/benefits of external information depend on the difference between the 40% (database rate) and the true control rate for the clinical trial. This is often called “drift” in the literature, but has other names.
Read 17 tweets
8 Dec 19
1) If you achieve an observed effect in a (standard) study equal to the effect you powered for, the one-sided p-value is 0.0006. This has two implications, the first on avoiding “significant but meaningless” results, and the second on resource allocation (tweetorial)
2) Standard clinical trials, testing means and proportions with alpha=0.025, protect against bad luck. If you have 90% power for an effect X, you reject the null for anything above 0.604 X. This protects you against missing an effective therapy that was unlucky in your trial.
3) You should always ask your statistician “what is the smallest observed effect where I reject the null?”. If this value (0.604 X for standard trials, might be different in others) is clinically meaningless, you should rethink your experiment.
Read 7 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!