Comfy UI, simple Txt To Image setup

 

The default Comfy UI graph looks like this.

Load Checkpoint: Here's where you specify the model, eg Stable Diffusion 1.5, XL, ToonYou whatever

Clip Text Encode Prompt: your text prompts (the red and green are the same type of node, just coloured), both of them are connected to the Load Checkpoint's Clip output

Empty Latent Image: Here you specify the resolution of the image & batch size, that we'll be working with

These 3 all feed into the KSampler, as shown above. The KSampler lets you specify Seed number, sampling steps, type of sampler (Euler, etc), how much denoise to use.

KSampler & the VAE output of the initial Load Checkpoint feed into the VAE Decode node. Finally the Image output from this gets fed into a Save Image.

VAE - it is worth noting that most checkpoints based on SD 1.5 has a VAE built in, so you can feed that straight to the VAE Decoder, from the Load Checkpoint. I had rather drab desaturated results when I used a separate VAE that was recommended vs the more vibrant expected results from the built in one. So this last sentence is not entirely true. I think it might be AnimateDiff's loop feature causing this!



Comments

Popular posts from this blog

Comfy UI AnimateDiff BASIC setup

face swap "deep fake" with stable diffusion and roop, in Automatic1111