So here's the gist of "promptgen". Without any code.
1. Copy-paste these good "prompt-English" examples, from this file in the repo: github.com/sharonzhou/lon… 2. On a newline, write your prompt in plain English 3. Generate prompt-English suffixes and use it in your fave model!
Can you guess which is pre-promptgen and which are using promptgen? Literally just using the prompts in the mini-video above.
Yeah, the two images from promptgen are better
It’s out!! Illustrations from this script, accompanying this short story (written by GPT3)
The OpenAI Minecraft paper is a great push to getting AI to work in Photoshop, Figma, or any software product — using just the keyboard & mouse, like a person would.
1. First, hire people to play Minecraft, who are OK at it. Record their screen and keyboard & mouse strokes. This costs $2k for 2k hrs of video in total.
This is your small dataset.
2/
2. Train a model on this small dataset. Let the model to look a little bit in the past and a little bit in the future in the videos. Let it predict the key & mouse strokes the person used, aligned to the video.
Huge appreciation for all reviewers esp R5 in making our work better.
My goal in 🧵: Explain our work in my simplest terms to you. Don't worry if you get lost, it's admittedly dense :)
Disentanglement in your generative model means dimensions in its latent space can change a corresponding feature in its data space, e.g. adapting just 1️⃣ dim can make the output "sunnier" ☁️→🌥→⛅️→🌤→☀️ Contrast w/ this entangled mess ☁️→🌥→🌩→🌪→☀️
Lasting collaborations can come from transient places.
One of my collaborations came out of a train ride 🚊 where I traded notes with a mathematician, en route to Stockholm. Followed by making these corgis happen.
Story 🧵👇🏿
On the 4 hr 🚊ride, I talked about generative models and neural networks. He talked about fractals and Mobius transformations — and even how all this ties into making better compression socks 🧦.
Hours 1-2: Just a pen and a few loose pages.
Mobius transformations generalize affine transformations
(see cute dogs). They are found naturally in biology. Maybe... we could use these for data augmentation, without much tuning across a ton of different augmentations.