For the next few days, our timelines are gonna be full of cutesy images made by a new Google AI called #Imagen.
What you won't see are any pictures of Imagen's ugly side. Images that would reveal its astonishing toxicity. And yet these are the real images we need to see. 🧵
How do we know about these images? Because the team behind Imagen has acknowledged this dark side in a technical report, which you can read for yourself here. Their findings and admissions are troubling, to say the least. gweb-research-imagen.appspot.com/paper.pdf
First, the researchers did not conduct a systematic study of the system's potential for harm. But even in their limited evaluations they found that it "encodes several social biases and stereotypes."
They don't list all these observed biases and stereotypes, but they admit that they include "an overall bias towards generating images of people with lighter skin tones and a tendency for images portraying different professions to align with Western gender stereotypes." 💔
The team didn't publish any examples of such images. But the pictures made by a similar system called DALL-E 2 a few weeks ago can give us an idea of what they probably look like. (Warning: they're upsetting)
That's not all. Even when Imagen isn't depicting people, it has a tendency to "encode[] a range of social and cultural biases when generating images of activities, events, and objects." Again, they don't give examples, but you can imagine that these depictions are dismaying.
Again, no concrete examples are provided, but in some of the other images in the paper it is possible to see these biases rearing their ugly heads.
For example, when prompted to make a picture of "detailed oil painting[s]" of a "royal raccoon king" and "royal raccoon queen," the system went for decidedly European symbols of royalty.
We can also safely assume that Imagen has a high capacity to generate overtly toxic content. Why? Because one of the datasets the team used to train it contains "a wide range of inappropriate content including pornographic imagery, racist slurs, and harmful social stereotypes."
Let's pause on that. They *knowingly* trained Imagen on porn, violence, racist content, and other toxic material. It wasn't an accident.
They also knowingly trained the system on "uncurated web-scale data." AKA large volumes of stuff from the internet that *they admit* "risks reproducing" things like "oppressive viewpoints, and derogatory, or otherwise harmful, associations to marginalized identity groups."
To be sure, because of all this, the team decided not to release Imagen to the public. It's also why they won't prompt it to make images of people. But the fact remains that they have knowingly created a machine with serious dangers for which they do not have any clear fixes.
Again, this transparency and honesty is commendable. But as the authors admit, mitigating these risks is going to require serious new research. Research that they knowingly chose not to address before revealing the system.
And while they do say that they intend to address these issues in future work, it's going to take as many minds coming together as possible to fix them. A whole community of effort...
But if most folks within that community are only focused on the cute stuff—or on philosophical debates over whether this system is truly "intelligent"—it's going to be hard to generate a groundswell of effort to do anything about the nasty stuff.
Bottom line: if we want these types of AI to benefit all of humanity, we should probably look beyond all the pictures of wizard corgi dogs or cows in tuxedos. Because all that cute stuff will mean nothing if the system ends up hurting people.
Ok that’s it. That's the thread.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Meta has released a huge new AI language model called OPT-175B and made it available to a broad array of researchers. It also released a technical report with some truly extraordinary findings about just how dangerous this machine can be. 🧵
With all the cute, quirky #dalle2 AI images that have been circulating these last few days, I wanted to share some other images* that DALL-E 2 also made that you may not have seen.
*Warning: these are quite distressing
1/ 🧵
2/ I hope OpenAI is cool with me reposting them. They are all available here in OpenAI’s report on the system's “Risks and Limitations.” github.com/openai/dalle-2…
3/ Here’s what it does when told to make an image of “nurse.” Notice any patterns? Anything missing?
With reports that kamikaze drones are entering the fray in Ukraine, I'd urge people not to spend too much time debating whether or not they are "autonomous weapons."
I was really hoping to avoid adding another thread to your TL, but let me explain.
Here's the rub. These systems probably have some capacity to be used in ways that *would fit most definitions of "lethal autonomous weapon." BUT they also can be used in ways that would *not qualify them as autonomous weapons by these same definitions.
Eg. A weapon with some target recognition capability (which many of these loitering munitions seem to have) could probably "select and engage targets" without human intervention, which would qualify it as a LAWS.
🧵Yesterday @UNIDIR published my new report about how autonomous military systems will have failures that are both inevitable and impossible to anticipate. Here's a mega-thread on how such "known unknown" accidents arise, and why they're such a big deal. unidir.org/press-release/…
*Deep breath*
Ok. For our purposes here today, think of autonomous systems as data processing machines.
i.e. when they operate in the real-world, they take in data from the environment and use that data to "decide" on the appropriate course of action.
These data must be complete, of good enough quality, accurate, and true. MOST IMPORTANTLY, these data must align with the data that the system was designed and tested for. This is true of all autonomous systems.
A cornerstone of the ICRC's proposed rules for #LAWS is a ban on "unpredictable weapon systems." i.e. systems whose "effects cannot be sufficiently understood, predicted and explained."
First, what is "Predictability"? Well as it happens, there are different types of predictability. 1. Technical (un)predictability 2. Operational (un)predictability 3. The (un)predictability of effects/outcomes.
Lots to unpack from this major test of a previously very quiet system to automate the "kill chain" leading up to a strike using...yep, you guessed it, Artificial Intelligence. (1/5) breakingdefense.com/2020/09/kill-c…
2/5 Basically, this technology enables drones to autonomously feed intel directly to algorithms that identify threats and suggest how to destroy those targets. All the humans have to do is review the recommendations and pull the trigger.
3/5 The implications of this automated "kill chain" technology are massive. In another recent test, the Army used a similar system to shorten an artillery kill chain from 10 minutes to just 20 SECONDS. breakingdefense.com/2020/09/target…