Comfy Tool Notes
Summary from Link
Key Notes
- Model files - civitai, hugging face
- CLIP, Main Model, VAE
- CheckpointLoader - Outputs Model, Clip, VAE
- Clip Model - Encode the text to main model, Positive and Negative prompt
- Encoded positive and Negative prompts sent to MODEL at each step and used to guide denoising
- VAE transalate image in latent space to pixel space
Samplername - uni_pc_bh2
- AutocodePro
- Finetuned Stable Diffusion for Anime
- AlphaCTR
- Low Rank Optimization LoRA models are essentially compact versions of Stable Diffusion that introduce minor, yet impactful modifications to the standard models.
- ControlNet/T2I adapter needs the image that is passed to it to be in a specific format like depthmaps
- Stable Zero123 is a diffusion model that given an image with an object and a simple background can generate images of that object from different angles.
- SDXL Turbo is a SDXL model that can generate consistent images in a single step.
- CLIP model: to convert text into a format the Unet can understand
- Unet: to perform the "diffusion" process, the step-by-step processing of images that we call generation
- VAE: to decode the image from latent space into pixel space (also used to encode a regular image from pixel space to latent space when we are doing img2img)
- KSampler node. This is the actual "generation" part, so you'll notice the KSampler takes the most time to run when you queue a prompt.
Checkpoints
- Place checkpoints in the folder ComfyUI/models/checkpoints:
- SDXL 1.0 base checkpoint, SDXL 1.0 refiner checkpoint
- VAE - Place VAEs in the folder ComfyUI/models/vae
- Fixed SDXL 0.9 VAE
- LoRAs - Place LoRAs in the folder ComfyUI/models/loras
- Stable Diffusion Hub
Keep Exploring!!!
No comments:
Post a Comment