I really love the active discussion abt the role of ethics in AI, spurred by Google Gemini's text-to-image launch & its relative lack of white representation. As one of the most experienced AI ethics people in the world (>4 years! ha), let me help explain what's going on a bit.
One of the critical pieces of operationalizing ethics in AI dev is to articulate *foreseeable use* (including misuse): Once the model we're thinking of building is deployed, how will people use it? And how can we design it to be as beneficial as possible in these contexts? 2/
This question overlaps with "Responsible AI" as well. Where ethics begs the question "should we build this at all?", once the decision at your company is YES, Responsible AI and Ethical AI share a lot of similarities. 3/
What you first do is figure out ways it's likely to be used, and by whom. This is harder for some people than others -- ppl with expertise in social & cog sci really shine at this (interdisciplinarity in tech hiring ftw!)
To help, I made this chart: Fill out the cells. 4/
Now here's the same chart filled out for text-to-image. The "green" cells are those *where beneficial AI is most likely possible*, not where it always will be beneficial. Within intended uses are "historic" depictions of reality -- such as white popes 5/
-- as well as "Dream world" depictions, where popes come in all shapes, colors, sizes. 6/
When designing a system in light of these foreseeable uses, you see that there are many use cases that should be accounted for:
- Historic depictions (what do popes tend to look like?)
- Diverse depictions (what could the world look like with less white supremacy?) 7/
Things go wrong when you treat all use cases as ONE use case, or don't model the use cases at all. 8/
That can mean, without an ethics/responsible AI-focused analysis of use cases in different context, you don't develop models "under the hood" that help to identify what the user is asking for (and whether that should be generated). 9/
We saw this same error in the generation of TSwift pornography: They forgot to have models "under the hood" that identify user requests that *should not* be generated. 10/zdnet.com/article/micros…
In Gemini, they erred towards the "dream world" approach, understanding that defaulting to the historic biases that the model learned would (minimally) result in massive public pushback. I explained how this could work technically here (gift link):
11/wapo.st/3OSmXo4
With an ethics or responsible AI approach to deployment -- I mean, the expert kind, not the PR kind -- you would leverage the fact that Gemini is a system, not just a single model, & build multiple classifiers given a user request. These can determine: 12/
1. Intent 2. Whether intent is ambiguous 3. Multiple potential responses given (1) & (2). E.g., Generate a few sets of images when the intent is ambiguous, telling user you're generating both the world *as the model learned it* and the world *as it could be* (Wording TBD). 13/
And further -- as is outlined in AI Safety, Responsible AI, AI ethics, etc., we're all in agreement on this AFAIK -- give the user a way to provide feedback as to their preferences (within bounds defined by the company's explicitly defined values). 14/
I think I've covered the basics. The high-level point is that it is possible to have technology that benefits users & minimizes harm to those most likely to be negatively affected. But you have to have experts that are good at doing this! 15/
And these people are often disempowered (or worse) in tech.
It doesn't have to be this way: We can have different paths for AI that empower the right people for what they're most qualified to help with. Where diverse perspectives are *sought out*, not shut down. 🩵 /end
• • •
Missing some Tweet in this thread? You can try to
force a refresh
OpenAI's Sora is out! Creating video from text. Similar models from other tech companies will likely follow. There are COOL technical things and NOT COOL social things to know about. A super quick 🧵. 1/
COOL TECHNICAL:
Why is this a particularly notable launch? In part bc we see realistic images with *multiple frame coherence*. A few yrs ago, we started being able to produce *single realistic images* from text prompts. Now, we have *hundreds* logically following one another. 2/
A key ML technique here is "DiT" -- Diffusion transformers -- which operates on latent patches .
This is also *open* code: We continue to see AI innovation from open science work. 2/huggingface.co/docs/diffusers…
With the rise of AI-generated "fake" human content--"deepfake" imagery, voice cloning scams & chatbot babble plagiarism--those of us working on social impact @huggingface put together a collection of some of the state-of-the-art technology that can help: huggingface.co/collections/so…
@huggingface 1. Audio watermarking. This embeds an imperceptible signal that can be used to identify synthetic voices as fake. Work from Guangyu Chen, Yu Wu, Shujie Liu, Tao Liu, Xiaoyong Du, Furu Wei ()
Demo by @ezi_ozoani github.com/wavmark/wavmark huggingface.co/spaces/Ezi/Aud…
@huggingface @ezi_ozoani 2. LLM Watermarking. From @tomgoldsteincs's awesome lab, implemented by @jwkirchenbauer. Also won the ICML 2023 Outstanding Paper award. huggingface.co/spaces/tomg-gr…
Reflecting on Claudine Gay, I'm reminded that a fundamental of racism--that we should all be aware of--is the disparate application of rules: People from one race* are disproportionately punished for "breaking a rule" that ppl from another are virtually never punished for.🧵
This has a few parts: 1. Being flagged as breaking rule where ppl from other races wouldn't be flagged 2. Having the system *determine* that you've broken a rule when the system wouldn't determine that for others 3. Being subjected to more extreme punishment for the rule-break
2/
On 1: one of the most obvious/clear/agreed-upon areas this happens is with traffic policing.
E.g., Black drivers are ~20% more likely to be stopped than white drivers.
On 3: Once stopped, Black drivers are searched ~2 times more than white drivers.
AI regulation: As someone who has worked for years in both "open" and "closed" AI companies, operationalising ethical AI, I'm dismayed by battle lines being drawn between “open” and “closed”. That's not where the battle should be--it's a distraction from what we all agree on. 🧵
Within tech, across the spectrum from fully closed to fully open, everyone generally agrees that peoples’ safety and security must be protected. That can mean everything from stopping identity theft or scamming, to mitigating psychological trauma from abusive bots.(2/n)
To protecting national secrets of how the AI sausage is made. (3/n)
Another good from @jjvincent.
2 important points: (1/2ish)
- These systems are being used as search, whether or not it's what OpenAI intended. By recognizing the *use* (both intended & unintended but foreseeable), companies can do much more to situate their products responsibly.
Reporting from @CadeMetz on Geoff Hinton's Google departure. A few things stand out to me; 🧵time (promise it'll be short). nytimes.com/2023/05/01/tec…
@CadeMetz One of the most personally pressing issues is that this would have been a moment for Dr. Hinton to denormalize the firing of @timnitGebru (not to mention many that have recently followed). To say it was the wrong choice. Especially given his statements supporting ethical work.
He did not.
This is how systemic discrimination works. The people in positions of power normalize. They do the discrimination, they watch their peers do it, they say nothing and carry on: This makes discrimination normal.