While a lot of the feedback we received was constructive, some of the comments need to be addressed.
A thread, with some new gibberish text and some discussion 🧵 (1/N)
@benjamin_hilton said that we got lucky with the whales example.
We found another similar example.
"Two men talking about soccer, with subtitles" gives the word "tiboer". This seems to give sports in ~4/10 images. (2/N)
A few people, including @realmeatyhuman, asked whether our method works beyond natural images (of birds, etc).
Yes, we found some examples that seem statistically significant.
E.g. "doitcdces" seems related (~4/10 images) to students (or learning). (3/N)
Similarly, "comafuruder" seems correlated (~4/10) to sickness/hospitals/patients. (4/N)
@BarneyFlames, @mattgroh pointed out that "Apoploe", our gibberish word for birds, has similar BPE encoding to "Apodidae".
Interestingly, "Apodidae" produces ~1/10 birds (but many flying insects), while our gibberish "Apoploe" gives 10/10.
(5/N)
However, "Apodidae Ploceidae" (two names of real bird families) indeed gives 10/10 birds.
Therefore, one possible explanation is that our gibberish tokens are mashups of parts of real words. This seems reasonable.
It is interesting that DALLE-2 generates those mashups.
(6/N)
Our gibberish tokens might have many meanings.
@benjamin_hilton run "Contarra ccetnxniams luryca tanniounons" and pointed out that not all are bugs. Indeed, our gibberish text produces a statistically significant fraction, but rarely a 100% match to the target concept. (7/N)
Our gibberish tokens have varying degrees of robustness in combinations with contexts.
E.g. if xx produces birds, ‘xx flying’ is an easy prompt
‘xx on a table’ is a neutral prompt, and ‘xx in space’ is a hard prompt.
(8/N)
Our hidden vocabulary seems robust in easy and sometimes neutral prompts but not in hard ones.
These tokens may produce low confidence in the generator and small perturbations move it in random directions.
"vicootes" means vegetables in some contexts and not in others. (9/N)
We want to emphasize that this is an adversarial attack and hence does not need to work all the time.
If a system behaves in an unpredictable way, even if that happens 1/10 times, that is still a massive security and interpretability issue, worth understanding. (10/N, N=10).
Given a few images, we learn pseudo-words that represent a concept at different resolutions.
"A painting of a dog in the style of <jane(number)>" gives different levels of artistic freedom to match the <jane> style based on the number index.
The key idea of our method is to condition the embedding of the learned concept on the diffusion time.
Instead of learning one embedding to represent the concept, we learn a set of embeddings: each element of the set represents the object at different resolutions.
During inference, we can use the embeddings in many creative ways to access the learned object at different resolutions.
For example, given a painting made of buttons, we can isolate the buttons and create new objects with that texture.
Announcing Soft Diffusion: A framework to correctly schedule, learn and sample from general diffusion processes.
State-of-the-art results on CelebA, outperforms DDPMs and vanilla score-based models.
A 🧵to learn about Soft Score Matching, Momentum Sampling and the role of noise
Typically, diffusion models generate images by reversing a known corruption process that gradually adds noise.
We show how to learn to reverse diffusions that involve a linear deterministic degradation and a stochastic part (additive noise).
Ingredient 1: Soft Score Matching.
Soft Score Matching incorporates the filtering process in the network. It trains the model to predict an image that after corruption matches the diffused observation.
DALLE-2 has a secret language.
"Apoploe vesrreaitais" means birds.
"Contarra ccetnxniams luryca tanniounons" means bugs or pests.
The prompt: "Apoploe vesrreaitais eating Contarra ccetnxniams luryca tanniounons" gives images of birds eating bugs.
A thread (1/n)🧵
A known limitation of DALLE-2 is that it struggles with text. For example, the prompt: "Two farmers talking about vegetables, with subtitles" gives an image that appears to have gibberish text on it.
However, the text is not as random as it initially appears... (2/n)
We feed the text "Vicootes" from the previous image to DALLE-2. Surprisingly, we get (dishes with) vegetables! We then feed the words: "Apoploe vesrreaitars" and we get birds. It seems that the farmers are talking about birds, messing with their vegetables! (3/n)