For the next few days, our timelines are gonna be full of cutesy images made by a new Google AI called #Imagen.

What you won't see are any pictures of Imagen's ugly side. Images that would reveal its astonishing toxicity. And yet these are the real images we need to see. 🧵
How do we know about these images? Because the team behind Imagen has acknowledged this dark side in a technical report, which you can read for yourself here. Their findings and admissions are troubling, to say the least.
gweb-research-imagen.appspot.com/paper.pdf
First, the researchers did not conduct a systematic study of the system's potential for harm. But even in their limited evaluations they found that it "encodes several social biases and stereotypes."
They don't list all these observed biases and stereotypes, but they admit that they include "an overall bias towards generating images of people with lighter skin tones and a tendency for images portraying different professions to align with Western gender stereotypes." 💔
The team didn't publish any examples of such images. But the pictures made by a similar system called DALL-E 2 a few weeks ago can give us an idea of what they probably look like. (Warning: they're upsetting)
That's not all. Even when Imagen isn't depicting people, it has a tendency to "encode[] a range of social and cultural biases when generating images of activities, events, and objects." Again, they don't give examples, but you can imagine that these depictions are dismaying.
Again, no concrete examples are provided, but in some of the other images in the paper it is possible to see these biases rearing their ugly heads.
For example, when prompted to make a picture of "detailed oil painting[s]" of a "royal raccoon king" and "royal raccoon queen," the system went for decidedly European symbols of royalty. An AI-generated image of oi...
We can also safely assume that Imagen has a high capacity to generate overtly toxic content. Why? Because one of the datasets the team used to train it contains "a wide range of inappropriate content including pornographic imagery, racist slurs, and harmful social stereotypes."
Let's pause on that. They *knowingly* trained Imagen on porn, violence, racist content, and other toxic material. It wasn't an accident.
They also knowingly trained the system on "uncurated web-scale data." AKA large volumes of stuff from the internet that *they admit* "risks reproducing" things like "oppressive viewpoints, and derogatory, or otherwise harmful, associations to marginalized identity groups."
To be sure, because of all this, the team decided not to release Imagen to the public. It's also why they won't prompt it to make images of people. But the fact remains that they have knowingly created a machine with serious dangers for which they do not have any clear fixes.
Again, this transparency and honesty is commendable. But as the authors admit, mitigating these risks is going to require serious new research. Research that they knowingly chose not to address before revealing the system.
And while they do say that they intend to address these issues in future work, it's going to take as many minds coming together as possible to fix them. A whole community of effort...
But if most folks within that community are only focused on the cute stuff—or on philosophical debates over whether this system is truly "intelligent"—it's going to be hard to generate a groundswell of effort to do anything about the nasty stuff.
Bottom line: if we want these types of AI to benefit all of humanity, we should probably look beyond all the pictures of wizard corgi dogs or cows in tuxedos. Because all that cute stuff will mean nothing if the system ends up hurting people.
Ok that’s it. That's the thread.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Arthur Holland Michel

Arthur Holland Michel Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @WriteArthur

May 4
Meta has released a huge new AI language model called OPT-175B and made it available to a broad array of researchers. It also released a technical report with some truly extraordinary findings about just how dangerous this machine can be. 🧵

#AI #OPT175B
Here's the report. Everyone should read it.
arxiv.org/pdf/2205.01068…
Bottom line is this: across tests, they found that "OPT-175B has a high propensity to generate toxic language and reinforce harmful stereotypes.”
Read 24 tweets
Apr 8
With all the cute, quirky #dalle2 AI images that have been circulating these last few days, I wanted to share some other images* that DALL-E 2 also made that you may not have seen.

*Warning: these are quite distressing

1/ 🧵
2/ I hope OpenAI is cool with me reposting them. They are all available here in OpenAI’s report on the system's “Risks and Limitations.” github.com/openai/dalle-2…
3/ Here’s what it does when told to make an image of “nurse.” Notice any patterns? Anything missing? Image
Read 39 tweets
Mar 18
With reports that kamikaze drones are entering the fray in Ukraine, I'd urge people not to spend too much time debating whether or not they are "autonomous weapons."
I was really hoping to avoid adding another thread to your TL, but let me explain.
Here's the rub. These systems probably have some capacity to be used in ways that *would fit most definitions of "lethal autonomous weapon." BUT they also can be used in ways that would *not qualify them as autonomous weapons by these same definitions.
Eg. A weapon with some target recognition capability (which many of these loitering munitions seem to have) could probably "select and engage targets" without human intervention, which would qualify it as a LAWS.
Read 21 tweets
May 18, 2021
🧵Yesterday @UNIDIR published my new report about how autonomous military systems will have failures that are both inevitable and impossible to anticipate. Here's a mega-thread on how such "known unknown" accidents arise, and why they're such a big deal.
unidir.org/press-release/…
*Deep breath*
Ok. For our purposes here today, think of autonomous systems as data processing machines.
i.e. when they operate in the real-world, they take in data from the environment and use that data to "decide" on the appropriate course of action.
These data must be complete, of good enough quality, accurate, and true. MOST IMPORTANTLY, these data must align with the data that the system was designed and tested for. This is true of all autonomous systems.

(Image credit: Eric Shulz/UNIDIR)
Read 28 tweets
May 13, 2021
A cornerstone of the ICRC's proposed rules for #LAWS is a ban on "unpredictable weapon systems." i.e. systems whose "effects cannot be sufficiently understood, predicted and explained."

So here's a quick thread🧵 on predictability and understandability.
icrc.org/en/document/ic…
First, what is "Predictability"? Well as it happens, there are different types of predictability.
1. Technical (un)predictability
2. Operational (un)predictability
3. The (un)predictability of effects/outcomes.
They're all different in crucial ways.
Read 22 tweets
Sep 18, 2020
Lots to unpack from this major test of a previously very quiet system to automate the "kill chain" leading up to a strike using...yep, you guessed it, Artificial Intelligence. (1/5)
breakingdefense.com/2020/09/kill-c…
2/5 Basically, this technology enables drones to autonomously feed intel directly to algorithms that identify threats and suggest how to destroy those targets. All the humans have to do is review the recommendations and pull the trigger.
3/5 The implications of this automated "kill chain" technology are massive. In another recent test, the Army used a similar system to shorten an artillery kill chain from 10 minutes to just 20 SECONDS. breakingdefense.com/2020/09/target…
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(