E.g., say you are creating a new community for discussing a StarWar game with a few rules. Given this description, our tool generated a simulacrum like this: (2/10)
Why are these useful? In social computing design, understanding our design decisions’ impact is hard since many challenges do not arise until a system is populated by *many*. Think about: newcomers with unintentional norm-breaking, trolling, or other antisocial behaviors (3/10)
What if we could generate an unbounded number of synthetic users and the social interactions between them that can realistically reflect how actual users might behave in our system designs?
Social simulacra lets you do that. (4/10)
What power social simulacra are LLMs (e.g., GPT-3). We observe that their training data has a wide range of social behavior & they can generate compelling simulacra of possible interactions w/ proper prompting. This lets us ask “what if” questions to iterate on our design. (5/10)
*So what does this all mean?*
For social computing designers, this means they can design “proactively.” Participating designers in our study said that they are in the practice of reactive design, implementing interventions only after a dumpster fire damages the community. (6/10)
Social simulacra could change this equation by helping the designers understand possible modes of failures in their design before they arise and cause harm. (7/10)
Also important: social simulacra (intentionally) generates not only the good, but also the bad behaviors so it can help the community leaders to understand what could go well *and* wrong in their community. But this puts in focus… (8/10)
... the need for a close collaboration with social computing stakeholders and accountability measures/evaluation techniques such as auditing to ensure that our prototyping approach is used for its intended purpose—empowering online communities—and not for auto-trolling. (9/10)
Simulating human behavior with AI agents promises a testbed for policy and the social sciences. We interviewed 1,000 people for two hours each to create generative agents of them. These agents replicate their source individuals’ attitudes and behaviors. 🧵 arxiv.org/abs/2411.10109
When we presented generative agents last year, we pointed to a future where we can simulate life to understand ourselves better in situations where direct engagement or observation is impossible (e.g., health policies, product launches, or external shocks). (2/14)
But we felt our story was incomplete: to trust these simulations, they ought to avoid flattening agents to demographic stereotypes, and measurement of their accuracy needs to advance beyond replication success or failure on average treatment effects. (3/14)
Our new research estimates that *one in twenty* comments on Reddit are violations of its norms: anti-social behaviors that most subreddits try to moderate. But almost none are moderated.
First, what does this mean? It means if you are scrolling through a post on Reddit, in a single scroll, you will likely see at least one comment that exemplifies bad behaviors such as personal attacks or bigotry that most communities would choose not to see. (2/13)
So let’s get into the details. What did we measure exactly? We measured the proportion of unmoderated comments in the 97 most popular subreddits that are violations of one of its platform norms that most subreddits try to moderate (e.g., personal attacks, bigotry). (3/13)