>>8964
Thanks!
Unfortunately, I'm not a videographer or writer by trade, so I can't assure you that this tutorial will be "good." It is, however, how I achieve my results.
If you're going to use my methods, you'll need a computer with decent-ish specs. Mine has an RTX 3060, but I don't think you need something high-end if you're fine with waiting a bit longer.
You will no doubt run into problems trying to get this working. It's not as easy as something like NovelAI, but it is free, and it is powerful. I ran into many hurdles where things just didn't work for reasons I didn't fully understand. Keep trying! I promise, it's possible. A lot of other people have good tutorials on how to fix some issues. Others, you'll have to figure out yourself.
You'll want to use lykos.ai's "stability matrix" application. It's a handy wrapper for stable diffusion and ComfyUI. This part is pretty basic, so you should be able to find a youtube tutorial for running ComfyUI via stability matrix (that's what I did, after all).
Once you've got the application set up and running, you'll want to install the ComfyUI manager. This will help a lot in later steps. This is another place where you'd be better off looking up "how to install ComfyUI manager" than listening to me.
Now you can launch ComfyUI! It'll spray out a bunch of console stuff (i'm not a computer scientist either), and eventually "To see the GUI go to: [local server link]." Click the link, and this will open up the ComfyUI instance in your browser. Don't worry, it's a purely local instance - no-one else can connect (although if you wanted to, you could set that up).
ComfyUI is a sort of backend tool. It's got a plug-and-play system where YOU need to place the nodes to load the model, encode the prompt, and sample the image. This has upsides and downsides - a major upside, for example, is the fact that you can install custom nodes (which is why we installed ComfyUI manager earlier) to do things like refine faces or hands automatically.
I prefer the EasyFluff V11.2 model. You can find this on Huggingface, posted by Zatochu. Find the "EasyFluffV11.2.yaml" (config) and "EasyFluffV11.2.safetensors" (model) files in the file tab, and download them. You'll want to put these both in your "stablediffusion" folder, which you can find by going to the "Checkpoints" tab in StabilityMatrix and clicking "models folder" in the top right. Pay attention to the model card - you'll need to install the "sampler_rescaleCFG.py" node to use EasyFluff properly.
Here is where things get a little difficult. You'll want to install "ComfyUI Impact Pack", "ComfyUI Inspire Pack", and "ComfyUI Prompt Control" via the ComfyUI manager. If you can't figure out how to do this, this is another step you can probably find on youtube.
Once you've got EasyFluff and the custom nodes installed, you'll want to pick a blank spot on the ComfyUI canvas (or make one), right click, and click "add node > advanced > Load Checkpoint with Config (DEPRECATED)" This is necessary to run EasyFluff properly. Select EasyFluffV11.2 for both the config_name and ckpt_name fields.
Okay, so now if we click "queue prompt" we'll successfully load the model - just not use it yet. You'll notice that the checkpoint loader node comes with three outputs - "MODEL", "CLIP", and "VAE". To use the model, you'll want to click and drag these outputs to the appropriate inputs on other nodes. Create a new "RescaleCFG" node through "add node>advanced>model>RescaleCFG". Drag the "model" output from the loader to the "model" input on the "RescaleCFG" node. Set the multiplier to something between 0.70 and 0.80.
Next, you'll want to create the prompting nodes. "add node>conditioning>CLIP Text Encode". Make two. One will serve as the positive prompt, and one as the negative. These nodes take "CLIP" input from the loader, and output "Conditioning".
One last step before we can sample (generate the image!!!). "add node>latent>empty latent image". This is where you can set the size of the image to be generated, as well as set up multiple batches to be run if you so choose. "add node>sampling>KSampler". You'll notice "model" "positive" "negative" and "latent_image" inputs. Link these up with the appropriate outputs from your other nodes. The sampler will show you a preview of your image, but this isn't final. I find that 25 steps with 7.5 cfg on euler_ancestral sampler with a normal scheduler and a denoising of 1.00 works well for EasyFluff.
Now, when you click "queue prompt", it will take the input from your model and text encoding and generate an image with the sampler. You still need to decode the latent image though. "add node>latent>VAE decode". This will take the latent output from your sampler, and the VAE output from your model loader, and output an image!. I like to upscale the image by 2x before displaying it with "add node>image>preview image".
At this stage, you should have a functional image generator where you can input a prompt and get out an image. I prefer EasyFluff specifically for its ability to adhere to artist styles, avoiding the usual "hi-res fixed AI image" look that a lot of AI products have. This gives you a lot more control over how you want the final product to look. Like a painting? OK! Like your favorite artist on E621? Sure! (works best if the artist has 100+ tags). A wholly unique style formed from a composite of one or more wildly different artists? Absolutely!
I can't really explain well how to piece together the automatic detailing functions. It took me a lot of time, and at this point I'm just using my past work - not really "understanding" it. Luckily for you, you can use this past work too! If you've figured out ComfyUI well enough, and have installed all the custom nodes (including bbox FacesV1 (the furry face identification model) and the other default bbox models) then you can simply download my ComfyUI workspace and use it! It's messy, complex, and hard to figure out, but it is functional, and it includes a lot of separated/grouped nodes with brief annotations to (try) to explain things. It also includes a lot of artists that I've found to work well in prompts, as well as "default" negative and positive prompts to base your work on. If you make it this far and would like to try using my workspace, let me know.
Otherwise, if you have any questions or feel as though you need assistance, I may (or may not) be able to help. This took a lot of trial and error on my end, and got frustrating on many separate occasions. Keep at it!