Exciting update to our open source ranking algorithm that improves note quality, making it less likely that mediocre or unhelpful notes appear on Tweets. In the spirit of #AlgorithmicTransparency, here’s how and why it works…
Some notes appear on Tweets, then later disappear. As more ratings come in, they lose their Helpful status. This kind of self-correction is a good thing, but we consider these “false positives” — not necessarily bad notes, but ultimately not broadly helpful enough to show.
Analyzing data over the past few months, we’ve found that rating tags (for example, "sources not included") can be powerful signals to identify false positive notes early in the rating process.
Going forward, notes that receive an atypically high number of such tags, relative to other Helpful notes, need to reach a higher level of helpfulness to show on Twitter.
The intuition behind this: if a note doesn’t cite a source, or might contain language that people see as biased or argumentative, it’s reasonable to hold it to a higher bar of quality before showing it.
To prevent this from being gamed, tags are inversely weighted by the statistical likelihood a rater would find the note helpful. If someone tries to game tags, their ratings will lose their intended effect.
In practice, we find that this significantly reduces the number of notes that appear on Tweets and then later disappear as they get noticed by more raters. The updates are live in GitHub at twitter.github.io/communitynotes…
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Birdwatch is now Community Notes. We’ve seen our contributors trying to explain what Birdwatch is, and we think a simpler, more intuitive name will help understanding. Hence Community Notes. It’s… what it is.
This isn’t just a random new name. It’s actually Birdwatch’s very first name. Before there was a Birdwatch, before there was anyone building this thing, there was a design mockup envisioning the idea, and you know what it was called?
In the spirit of transparency, Birdwatch makes everything and the kitchen sink public and open source. Today, we’re sharing a paper that details the research, analysis and outcomes that informed its development and helped us understand its performance. 📊 github.com/twitter/birdwa…
Birdwatch seeks to provide context that will be informative and helpful to a wide range of people. It does this using a bridging-based ranking system — different from typical engagement-based approaches — which we have made open source.
This paper dives deep into the Birdwatch algorithm, the research methods used to evaluate its effect on core metrics — informing understanding, informing sharing behavior, and perceived helpfulness — and what we learned from the analysis.
Buckle up! We have a handful of exciting new updates to share with y’all:
- Research on the impact Birdwatch is having
- A new system to keep contribution quality high
- Enrolling more contributors
- Expanding the visibility of notes on Twitter blog.twitter.com/en_us/topics/p…
🧵
Our research shows that people who see a Tweet with a Helpful Birdwatch note are, on average, 15-35% less likely to Like/Retweet it. And, on average, people from across the political spectrum are 20-40% less likely to agree with a misleading Tweet after reading a note on it.
These results indicate that Birdwatch notes help people make more informed decisions about what they share on Twitter and we can't wait to bring this value to everyone! But before that can happen, we need to ensure that the quality of notes remains high, at scale.