Comfy UI, simple Txt To Image setup

 

The default Comfy UI graph looks like this.

Load Checkpoint: Here's where you specify the model, eg Stable Diffusion 1.5, XL, ToonYou whatever

Clip Text Encode Prompt: your text prompts (the red and green are the same type of node, just coloured), both of them are connected to the Load Checkpoint's Clip output

Empty Latent Image: Here you specify the resolution of the image & batch size, that we'll be working with

These 3 all feed into the KSampler, as shown above. The KSampler lets you specify Seed number, sampling steps, type of sampler (Euler, etc), how much denoise to use.

KSampler & the VAE output of the initial Load Checkpoint feed into the VAE Decode node. Finally the Image output from this gets fed into a Save Image.

VAE - it is worth noting that most checkpoints based on SD 1.5 has a VAE built in, so you can feed that straight to the VAE Decoder, from the Load Checkpoint. I had rather drab desaturated results when I used a separate VAE that was recommended vs the more vibrant expected results from the built in one. So this last sentence is not entirely true. I think it might be AnimateDiff's loop feature causing this!



Comments

Popular posts from this blog

Comfy UI AnimateDiff BASIC setup

ComfyUI ControlNet + AnimateDiff