MMitchell Profile picture
Feb 25 16 tweets 4 min read Read on X
I really love the active discussion abt the role of ethics in AI, spurred by Google Gemini's text-to-image launch & its relative lack of white representation. As one of the most experienced AI ethics people in the world (>4 years! ha), let me help explain what's going on a bit. This screen grab shows CNN asking Google Gemini to create an AI-generated image of a pope, as well as the tool's response. Clare Duffy/CNN via Google Gemini. Source: https://www.cnn.com/2024/02/22/tech/google-gemini-ai-image-generator/index.html
One of the critical pieces of operationalizing ethics in AI dev is to articulate *foreseeable use* (including misuse): Once the model we're thinking of building is deployed, how will people use it? And how can we design it to be as beneficial as possible in these contexts? 2/
This question overlaps with "Responsible AI" as well. Where ethics begs the question "should we build this at all?", once the decision at your company is YES, Responsible AI and Ethical AI share a lot of similarities. 3/
What you first do is figure out ways it's likely to be used, and by whom. This is harder for some people than others -- ppl with expertise in social & cog sci really shine at this (interdisciplinarity in tech hiring ftw!)
To help, I made this chart: Fill out the cells. 4/ Title: "Chart to help guide Foresight in AI"  It's a matrix of People x Use Contexts: People: 1. Users  -- intended and unintended (Both malicious actors &  people un-accounted for in development) 2. Those affected -- intended and unintended (Both people in training data &  people the technology is used on)  Use Contexts:  1.  Intended 2. Unintended (Both harmful contexts & those unmodeled in development) 3. Out of scope  The intended use contexts x intended users cell is green. The unintended use contexts x unintended users is red. Etc.  Notes below: Unintended: Results unpredict...
Now here's the same chart filled out for text-to-image. The "green" cells are those *where beneficial AI is most likely possible*, not where it always will be beneficial. Within intended uses are "historic" depictions of reality -- such as white popes 5/ Matrix filled out: Intended Use Context: Historic bias depiction, dream world depiction Unintended Use Context: Mis/disinformation (e.g., Deep Fakes) Out of scope Use Context: Medical images, world as-it-is  Intended Users: People who'd like to create art Unintended Users: Malicious ex-partner Those Affected, Intended: Art appreciators Those Affected, Unintended: Unconsented artists whose art is in training data
-- as well as "Dream world" depictions, where popes come in all shapes, colors, sizes. 6/
When designing a system in light of these foreseeable uses, you see that there are many use cases that should be accounted for:
- Historic depictions (what do popes tend to look like?)
- Diverse depictions (what could the world look like with less white supremacy?) 7/
Things go wrong when you treat all use cases as ONE use case, or don't model the use cases at all. 8/
That can mean, without an ethics/responsible AI-focused analysis of use cases in different context, you don't develop models "under the hood" that help to identify what the user is asking for (and whether that should be generated). 9/
We saw this same error in the generation of TSwift pornography: They forgot to have models "under the hood" that identify user requests that *should not* be generated. 10/zdnet.com/article/micros…
In Gemini, they erred towards the "dream world" approach, understanding that defaulting to the historic biases that the model learned would (minimally) result in massive public pushback. I explained how this could work technically here (gift link):
11/wapo.st/3OSmXo4
With an ethics or responsible AI approach to deployment -- I mean, the expert kind, not the PR kind -- you would leverage the fact that Gemini is a system, not just a single model, & build multiple classifiers given a user request. These can determine: 12/
1. Intent
2. Whether intent is ambiguous
3. Multiple potential responses given (1) & (2). E.g., Generate a few sets of images when the intent is ambiguous, telling user you're generating both the world *as the model learned it* and the world *as it could be* (Wording TBD). 13/
And further -- as is outlined in AI Safety, Responsible AI, AI ethics, etc., we're all in agreement on this AFAIK -- give the user a way to provide feedback as to their preferences (within bounds defined by the company's explicitly defined values). 14/
I think I've covered the basics. The high-level point is that it is possible to have technology that benefits users & minimizes harm to those most likely to be negatively affected. But you have to have experts that are good at doing this! 15/
And these people are often disempowered (or worse) in tech.
It doesn't have to be this way: We can have different paths for AI that empower the right people for what they're most qualified to help with. Where diverse perspectives are *sought out*, not shut down. 🩵 /end

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with MMitchell

MMitchell Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @mmitchell_ai

Feb 19
OpenAI's Sora is out! Creating video from text. Similar models from other tech companies will likely follow. There are COOL technical things and NOT COOL social things to know about. A super quick 🧵. 1/
COOL TECHNICAL:
Why is this a particularly notable launch? In part bc we see realistic images with *multiple frame coherence*. A few yrs ago, we started being able to produce *single realistic images* from text prompts. Now, we have *hundreds* logically following one another. 2/
A key ML technique here is "DiT" -- Diffusion transformers -- which operates on latent patches .
This is also *open* code: We continue to see AI innovation from open science work. 2/huggingface.co/docs/diffusers…
Read 29 tweets
Feb 12
With the rise of AI-generated "fake" human content--"deepfake" imagery, voice cloning scams & chatbot babble plagiarism--those of us working on social impact @huggingface put together a collection of some of the state-of-the-art technology that can help:
huggingface.co/collections/so…
@huggingface 1. Audio watermarking. This embeds an imperceptible signal that can be used to identify synthetic voices as fake. Work from Guangyu Chen, Yu Wu, Shujie Liu, Tao Liu, Xiaoyong Du, Furu Wei ()
Demo by @ezi_ozoani github.com/wavmark/wavmark
huggingface.co/spaces/Ezi/Aud…
@huggingface @ezi_ozoani 2. LLM Watermarking. From @tomgoldsteincs's awesome lab, implemented by @jwkirchenbauer. Also won the ICML 2023 Outstanding Paper award.
huggingface.co/spaces/tomg-gr…
Read 10 tweets
Jan 4
Reflecting on Claudine Gay, I'm reminded that a fundamental of racism--that we should all be aware of--is the disparate application of rules: People from one race* are disproportionately punished for "breaking a rule" that ppl from another are virtually never punished for.🧵
This has a few parts:
1. Being flagged as breaking rule where ppl from other races wouldn't be flagged
2. Having the system *determine* that you've broken a rule when the system wouldn't determine that for others
3. Being subjected to more extreme punishment for the rule-break
2/
On 1: one of the most obvious/clear/agreed-upon areas this happens is with traffic policing.
E.g., Black drivers are ~20% more likely to be stopped than white drivers.
On 3: Once stopped, Black drivers are searched ~2 times more than white drivers.

3/openpolicing.stanford.edu
Read 29 tweets
Nov 5, 2023
AI regulation: As someone who has worked for years in both "open" and "closed" AI companies, operationalising ethical AI, I'm dismayed by battle lines being drawn between “open” and “closed”. That's not where the battle should be--it's a distraction from what we all agree on. 🧵
Within tech, across the spectrum from fully closed to fully open, everyone generally agrees that peoples’ safety and security must be protected. That can mean everything from stopping identity theft or scamming, to mitigating psychological trauma from abusive bots.(2/n)
To protecting national secrets of how the AI sausage is made. (3/n)
Read 16 tweets
May 30, 2023
Another good from @jjvincent.
2 important points: (1/2ish)
- These systems are being used as search, whether or not it's what OpenAI intended. By recognizing the *use* (both intended & unintended but foreseeable), companies can do much more to situate their products responsibly.
*cough cough* model cards. Here's the most recent annotated version from me @huggingface:
huggingface.co/docs/hub/model…
(....more...)
And here is my How To Foresight chart. We can do this!
(Boarding plane, can't alt text, will do shortly) Image
Read 15 tweets
May 1, 2023
Reporting from @CadeMetz on Geoff Hinton's Google departure. A few things stand out to me; 🧵time (promise it'll be short).
nytimes.com/2023/05/01/tec…
@CadeMetz One of the most personally pressing issues is that this would have been a moment for Dr. Hinton to denormalize the firing of @timnitGebru (not to mention many that have recently followed). To say it was the wrong choice. Especially given his statements supporting ethical work.
He did not.
This is how systemic discrimination works. The people in positions of power normalize. They do the discrimination, they watch their peers do it, they say nothing and carry on: This makes discrimination normal.
Read 16 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(