Xander Steenbrugge Profile picture
Apr 23, 2022 β€’ 7 tweets β€’ 3 min read β€’ Read on X
I discovered a bug in my own Diffusion + CLIP pipeline and suddenly the samples are unreal.. 🀯
Here's
"Just a liquid reality..."
#AIart #notdalle2 #Diffusion #clip Image
"The magnificent portal of mother Gaia" Image
"Framing reality" Image
"Gathering at the great elder sphere" Image
"Why such a rush? It's all twisting and bending anyway" Image
"My hair is a living creature" Image
Caveat: all these pieces are the result of a tremendous amount (months) of code and parameter tuning, careful selection of initialization images, prompt engineering and cherry picking.

#dalle2 is incredible at compositionality and realism, but I haven't seen it do this yetπŸ‘¨β€πŸŽ¨πŸ§™β€β™‚οΈπŸ˜‹

β€’ β€’ β€’

Missing some Tweet in this thread? You can try to force a refresh
γ€€

Keep Current with Xander Steenbrugge

Xander Steenbrugge Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @xsteenbrugge

Sep 6, 2022
I continued exploring #stablediffusion's latent space over the weekend and oh my; there's still a LOT of treasure to be discovered inside this magnificent neural universe!

Here's a quick thread with some of my personal favorites and how I found them.. ImageImageImageImage
The fact that all this visual splendor is compressed in just 4Gb of neural network weights totally blows my mind. Call it compression, call it emergence, it's just 🀯🀯

Getting bored by a StyleGAN model after looking at samples for 20 minutes seems like a very distant past now.. ImageImageImageImage
Reminiscent of cut-up poetry, one cool trick I implemented is to:
1. Start with a list of great, proven prompts
2. Chunk the prompts into word groups of ~2-5 words
3. Randomly recombine multiple word groups into new 'pseudo-prompts'

Turns out, some of those work really well πŸŽ‰πŸ˜‹ Image
Read 7 tweets
Aug 17, 2022
Ok, so first of all, #stablediffusion did not come with code to make videos, so I came up with a way to interpolate between encoded prompt vectors (no worries if you don't know what that means) and thereby create video sequences from prompt sequences (1/n)
Next, I had to come up with a visual narrative that would work well with the style of the Diffusion interpolations. You can't just tell any story here: like with any medium, you have to work within the constraints of the technology. (2/n)
Once I settled on the "evolution" narrative, I wrote about a thousand different prompts, containing many variations on the narrative sequence I wanted. I then rendered all the corresponding stills with multiple seeds over roughly two nights of GPU time. (3/n)
Read 13 tweets
Aug 13, 2022
"Voyage through Time"
is my first artpiece using #stablediffusion and I am blown away with the possibilities...

We're crossing a threshold where generative AI is no longer just about novel aesthetics, but evolving into an amazing tool to build powerful, human-centered narratives
This video was created using 36 consecutive phrases that define the visual narrative.

To find the best possible sequence, I tried over a thousand different prompts and seeds and applied many "prompt engineering" tricks in my code, to figure out what works and what doesn't
The way this model "interpolates" between the meaning of two sentences (in semantic rather than visual latent space) is a huge gamechanger for storytelling, and this is only just the beginning of a MASSIVE revolution in digital content creation powered by generative AI..
Read 5 tweets
Feb 24, 2022
This is a "3D-diffusion" video created using a combination of four different AI models🀯

Welcome to the metaverse! 🌌😎

There's such incredible potential here that I want to explain how I made this, so here's a thread! (1/n)
The two main models that draw the pixels are a diffusion model guided by a language prompt through @OpenAI's CLIP model.
This idea was introduced by @advadnoun and later refined by many other creatives. My talk at @Kikk_Festival further explains this:
The diffusion model (I integrated code from @RiversHaveWings and @Somnai_dreams for this) generates images by iteratively denoising noisy-pixel images, every time you run this from different noise, you get a different image, guided by the language prompt:
Read 10 tweets
Jan 21, 2022
Finally playing around with CLIP + diffusion models.

12 GPU hours in I gotta say I'm pretty impressed with the difference in esthetics compared to VQGANπŸ‘Œ
Big thanks to @RiversHaveWings & @Somnai_dreams for providing great starting code!

"a dystopian city"
"The real problem of humanity is that we have Paleolithic emotions, medieval institutions and godlike technology"
"The being"
Read 5 tweets
Dec 19, 2021
Just felt like sharing some beautiful images, these are still hot from the GPU...

"The elder sphere"
"The Engine"
"We live in golden bubbles of the mind"
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(