Then I took a single still frame from that video and ran it through #img2img in a local instance of #stablediffusion WebUI. After generating a few options I ended up with this image. 3/9 #aiartprocess
Here's a split screen highlighting the consistency and coherence between the driving video and the results. I upscaled the video a bit using #TopazVideoEnhance. Notice the quality of the blinking and the coherence of the "birthmark" under her left eye. 6/9 #aiartprocess
Good as this was, there were some artifacts in the eyes I wanted to reduce. Made the video an img sequence. Used a local instance of GFPGAN to batch process all the frames. This increased the quality of the eyes and sharpened the image by enlarging 4x.(1st video)7/9 #aiartprocess
GFPGAN changed the alignment of some things slightly for some reason and shifted the color. It also faded the birthmark a bit. Not sure how to control the strength of the effect locally just yet. GFPGAN link below. 8/9 #aiartprocess
There's still some artifacts here and there, but pretty good considering the look is from a single keyframe. I'll continue experimenting and see how to improve on this initial workflow. Hope this thread gives you some ideas! 9/9 #aiartprocess
• • •
Missing some Tweet in this thread? You can try to
force a refresh
If you have a style or consistent aesthetic you’re going for with AI art, it might be a good idea to figure out how to get to that look with different tools, starting points, and process pathways. 1/8
We already see how fast these tools are changing. And as we iterate, as the tool space evolves, we can’t know how backwards compatible changes will be. Maybe we have more control with local SD, but maybe 2/8
parts of your future process will take on powerful pieces that are app based or otherwise in a blackbox. You might ask yourself, how much of my style is locked in a specific process/tool combination? How fragile is that to change? To parent companies pivoting? 3/8
I imagine if you're coming to AI art from a different set of creative fields and client expectations, you'd approach the latent space from a different collection of frameworks and processes. Very curious how different a UI/UX catering to painters could look. 2/7 #aiartprocess
After an exhaustive process of prompting, I use different combos of X/Y Plot to make variations. The image I posted looked at sampler method+step count. Still trying to figure out what flow makes sense for me. I'm sure I'll throw it out as soon as new stuff gets released lol 3/7