As soon as the model was trained, the first step was randomly generating a large set of images w. a simple prompt ("golem, detailed, realistic, 3D rendering")
Each golem can be extracted by removing the images' backgrounds (reco: @photoroom_app).
Some of the designs are amazing. However, the golems all look very similar to each other. Let's separate them into categories.
The following step is to prompt some more precise description (e.g. "golem with stone, lava, steel, fire")
Here's a first set of 16 golems that look consistent.
Because that initial set of lava golems was a bit simplistic, I used #img2img and an improved prompt to slightly increase the level of details (e.g. "3D rendering, trending on Artstation, etc.)
Here's the result:
And now, we can keep the same prompt and introduce many other variations while keeping the overall shape, posture, and level of detail.
Here are the ice golems (like the ones in Clash of Clans)
The forest golems! My favorites 🌲🌳
The sand golems 🏜️
And... the golden golem (directly inspired by the golden golem in Minecraft, via #img2img)
Now it's also possible to use the same finetune, and generate pictures that could be used elsewhere than in the game (ads, splash screens, etc).
Just add "Greg Rutkowski" in the prompt and start seing some cinematic composition, dramatic lighting, etc.
Some other tests I made...
As always, feel free to add any questions, feeback, remarks (and don't forget to like/follow/RT if you find this content interesting. I really appreciate it!)
Tomorrow, I'll keep sharing other explorations with #StableDiffusion 🚀🚀
(11/06 update - here's another run to create more characters, this time with "Space Marines" heavy infantry)
Here are the key steps to creating stunning turnaround, using #Scenario ()
1/ Train or pick a character model (A).
2/ Optionaly>, pick a style model (B). Use it to create training images for (A), or you can merge both (A + B = C) for example.
3/ Utilize the custom model (A or C) to generate consistent characters. Then select a reference image to produce initial character turnarounds in your desired poses.
4/ Refine these initial outputs using Sketching and image2image.
5/ Select the best result and refine details in the Canvas for maximum consistency.
6/ Finally, upscale your final image (up to 8K resolution.)
@ClaireSilver12 - (I hope you don't mind me RT this for a broader reach and to share it with more users.)
Here's an advanced use case for the IP Adapter. You can adjust or remove the steps depending on the desired output/goal. Bear with me; it's actually quite straightforward.
1 - Train a LoRA on a specific subject (e.g., character).
2 - Blend the LoRA to perfectly capture the style (e.g., comic, cartoon, oil painting, 3D...).
3 - Run inference on that "blended" model.
4 - Select an image that stands out and use it as a reference with the IP Adapter.
5 - Modify the prompt to create variations of the subject.
Let's get started 👇👇
1/ The first step is to train one (or more LoRA) models on a specific subject (e.g. character or object), or also a style.
The process is straightforward. I'll use the example of the "girl with pink hair" (😊🫠) that I shared before (12 training images)
Simply select "New Model - Train" on . I use 9 images of the model, showcasing various angles and zoom levels, accompanied by concise captions (details below).
This could be the best model I've ever created for generating isometric buildings, on Scenario.
Output consistently match the style I wanted, and the model responds perfectly to (short) prompts, without any reference images needed.
It's a LoRA composition. More below.
Process: it's pretty simple.
I created a LoRA composition from 4⃣ distinct LoRA.
(i) - My own "Fantasy Buildings" LoRA
(ii) - Three LoRAs available on #Scenario: "Isometric Storybook", "Stylized Fantasy Iconic Imagery" and "Belgian School Comics".
The influence of each LoRA is below.
My prompt structure was dead simple... less than 10 words!
(type of building/scene), solid color background, highly detailed, centered.