Brian Guay Profile picture
Aug 24, 2021 11 tweets 5 min read Read on X
🚨New Paper Alert🚨 in Public Opinion Quarterly w @Jesse_L_Lopez

Partisan Bias in Bipartisan Places? A Field Experiment Measuring Attitudes Toward the Presidential Alert in Real Time

tinyurl.com/POQ21
ungated: tinyurl.com/POQ21ungated
WaPo: tinyurl.com/POQ21wapo

🧵👇
Americans report liking bipartisanship, but attitudes toward bipartisan issues quickly become polarized when associated w/ partisan identities. Lots of work showing this in lab setting, but hard to examine experimentally in real world. How do you randomize party cues outside lab?
We leverage a bipartisan issue that became
associated with a partisan identities suddenly in 2018: the presidential alert. Remember this thing? The Trump admin sent it to all Americans' phones in 2018.

If you don't, SNL did a great sketch on it:
Image
The alert had bipartisan roots (Bush and Obama created it after Hurricane Katrina), but quickly became polarized before its inaugural test under Trump admin.
We rapidly recruited a sample of U.S. adults immediately
before the alert was sent, so that participants received the alert during the survey. We exploited the timing of the alert to randomize whether they answered questions about the alert moments before or after receiving it. Image
While prior research suggests that associating bipartisan issues with partisan identities polarizes attitudes, we find little evidence that receiving the
alert from the Trump admin elicited a partisan reaction. We looked at attitudes toward both the alert and privacy. Image
But maybe receiving the presidential alert from the Trump admin wasn't a strong enough cue. So, in the same study we ran a 2nd experiment--respondents were randomly assigned to receive
info explicitly associating the alert with either the Trump or Obama administration
But similar results here. Image
takeaway #1: Online surveys--especially those that draw respondents from many panels at once--provide opportunity to exploit timing of political events with large samples.
takeaway #2: We find little evidence that associating
the alert w/ the Trump administration had any polarizing effect on attitudes, even when using explicit partisan cues, suggesting that at least some bipartisan attitudes are not as easily polarized as prior work implies.
and here are the paper links (w/ lots of robustness checks) and Washington Post writeup:

tinyurl.com/POQ21 @AAPOR #publicopinion
ungated: tinyurl.com/POQ21ungated
WaPo: tinyurl.com/POQ21wapo

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Brian Guay

Brian Guay Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @BrianMGuay

Sep 28, 2022
🚨New WP🚨

How should researchers determine whether misinformation interventions work?

We argue that researchers should 1) measure whether people believe or share both false *and* true content and 2) assess efficacy using a measure of discernment 🧵

psyarxiv.com/gv8qx Image
Who believes and shares misinfo? Why? What can we do about it?

Answering these questions requires measuring whether ppl believe and share misinfo

But studies purporting to answer the same question often use different research designs, inhibiting progress on combating misinfo
We provide a framework for differentiating between research designs on the basis of the normative claims they make about how people should interact with information

Then we show different designs lead to different conclusions about whether misinfo interventions work
Read 15 tweets
Apr 14, 2022
🚨New WP🚨
Are Republicans really more inclined to share fake news? Or just exposed to more of it? And are they resistant to accuracy nudges?

To find out, we presented a national sample from YouGov with a large set of politically balanced headlines

psyarxiv.com/y762k

1/
The problem of fake news sharing appears to be largely concentrated among the political right. Republicans share *a lot* (up to 8X) more fake news than Democrats on FB and Twitter
science.org/doi/10.1126/sc…
science.org/doi/full/10.11…
h/t @andyguess @grinbergnir
But WHY do Reps share more fake news than Dems? Reps could be more *susceptible* to sharing fake news (more close-minded, less attentive to accuracy...)

Or, Reps could simply be more *exposed* to more fake news.

From social media data, we can't easily tell which it is...
Read 13 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(