Made this video with iPhone photos I took of my friend Stephanie that I used as keyframes in @LumaLabsAI! With the camera controls I can gen transitions between shots. I also built a custom web app in Next.js to help me speedramp and edit all the clips! Breakdown 🧵(1/18)
Basically if you have some photos, throw in a start and end frame, and start your prompt with the camera move, and then stuff like “smooth camera, steadicam” I find minimal prompts work best. And don’t enhance the prompt (that tends to add hand held). (2/18)
Jul 20 • 7 tweets • 2 min read
Testing how LivePortrait works lip syncing 24fps lyrics on top of slow motion footage. Was curious to see if it might help with music videos. Quick explanation below! 🧵
Started with a clip from an Eminem song and passed it through Adobe Podcast to get the acapella. Passed that through @hedra_labs with a Midjourney portrait for the face animation. Used that as input into LivePortrait using ComfyUI and a slowmo clip from Die Hard.
Dec 24, 2023 • 12 tweets • 5 min read
Made this video (🎶) with a Midjourney v6 image! Started by upscaling/refining with @Magnific_AI, pulled a Marigold Depth Map from that in ComfyUI, then used as a displacement map in Blender where I animated this camera pass with some relighting and narrow depth of field.🧵1/12
Here's the base image and the before/after in @Magnific_AI. Even though MJv6 has an upscaler, Magnific gave me better eyelid and skin details for this case. (Fun fact, this image was from a v4 prompt from summer last year, when MJ had just released a new beta upscaler.) 2/12
Nov 15, 2023 • 11 tweets • 3 min read
Testing LCM LORAs in an AnimateDiff & multi-controlnet workflow in ComfyUI. I was able to process this entire Black Pink music video as a single .mp4 input. The LCM lets me render at 6 steps (vs 20+) on my 4090 and uses up only 10.5 GB of VRAM. Here's a breakdown 🧵[1/11]
Entire thing took 81 minutes to render 2,467 frames, so about 2 seconds per frame. This isn't including the time to extract the img sequence from video and gen the ControlNet maps. Used Zoe Depth and Canny ControlNets in SD 1.5 at 910 x 512. [2/11]
May 25, 2023 • 4 tweets • 3 min read
Timelapse of using #photoshop’s new generative fill feature to connect two images and build a scene around them using blank prompts. Was inspired by @MatthieuGB’s post doing something similar! Notice how I’m not adding any descriptions, but letting gen fill present options for… twitter.com/i/web/status/1…
Here’s the final image! 2/4
Feb 28, 2023 • 15 tweets • 7 min read
Testing Multi-ControlNet on a scene with extended dialog, a range of facial expressions, and head turning. No EBSynth used. #AImusic from aiva.ai. Breakdown🧵1/15 #aicinema#controlnet#stablediffusion#aiia#aiartcommunity#aiva#machinelearning#deeplearning#ai
Overall flow: pre-process video > img seq > play with prompts > initial controlnet settings > control net batch render > upscale > clean up in post > key out background > deflicker > video post-production 2/15
I wonder what the future of UX design (maybe apps in general) may be like if AI really allows us to customize our experience. Not to mention blend programs together through a 3rd party/custom UI if an AI can understand onscreen what's being displayed by the app's GUI. 1/5
Combined with no code platforms of the future and advanced templates, you could probably do weird stuff like Premiere Pro x Unreal Engine x a fighting game template x an anime you like and custom gameify your interface. 2/5
Jan 6, 2023 • 11 tweets • 2 min read
Lately I've been thinking about how much of "reality" is a negotiation with useful illusions and the material world. I think it's safe to say that a portion of how we view things is through shortcuts and narratives. 1/11
To what degree we engage in fictions probably differs from person to person. Some believe the entire thing is a fiction passed on to us from evolution to navigate the food chain. Others think they concretely engage in reality the whole time. 2/11
A thought on resistance to change. I recently had a convo with a friend of mine who went thru a serious breakup that left her rattled. She talked about how hard it was to let go of the future she had envisioned for herself; that she felt so sure was going to come. 1/7
I feel like part of the resistance to change isn’t just rooted in the past and present, but also your perception of how you thought the world was going to look like and your place in it. Expectations are set and not met. 2/7
Dec 15, 2022 • 7 tweets • 5 min read
I’m so fascinated by how much of understanding a concept can sometimes just be a language issue. Being able to ask #chatgpt to summarize, expand, rephrase and format explanations in different ways is so refreshing. 1/7
Like here’s #ChatGPT explaining how to cook a steak in pseudo code format. 2/7
One thing I learned from working in fashion—what’s considered “good” changes from region to region, country to country. From working in video and photo, “good” changed depending on if the person worked mostly with photographers or filmmakers. 1/10
Then there’s differences in genre, branding strategy, or individual ppl. There’s overlap of course, what we might consider the transcendant or timeless qualities of certain aesthetics, but the differences in taste always had relevant impact on budgets, rates, and networking. 2/10
If you have a style or consistent aesthetic you’re going for with AI art, it might be a good idea to figure out how to get to that look with different tools, starting points, and process pathways. 1/8
We already see how fast these tools are changing. And as we iterate, as the tool space evolves, we can’t know how backwards compatible changes will be. Maybe we have more control with local SD, but maybe 2/8
Sep 7, 2022 • 7 tweets • 4 min read
Using X/Y Plot in a local #stablediffusion WebUI to create contact sheets exploring the latent space for my previous post. As a photo/video person I'm trying to bring the #aiartprocess closer to the workflows I use with clients.1/7 #aiphotography#aifashion#aiart#aiartcommunity
I imagine if you're coming to AI art from a different set of creative fields and client expectations, you'd approach the latent space from a different collection of frameworks and processes. Very curious how different a UI/UX catering to painters could look. 2/7 #aiartprocess