ComfyUI is a free, open-source, node-based interface for Stable Diffusion that gives you complete visual control over every step of AI image generation.Instead of filling in text fields and clicking "Generate," you connect processing nodes together like building blocks, creating reusable workflows that produce consistent, high-quality images. It is the tool behind most AI influencer content, product photography, and concept art created in 2026.
This guide covers everything from installation to advanced workflows, with step-by-step instructions for complete beginners.
What Is ComfyUI and Why Use It?
ComfyUI is a graphical interface for running Stable Diffusion and other AI image models. It uses a node-based workflow system, similar to Blender's shader editor or Unreal Engine's Blueprints. Each node performs one operation: loading a model, encoding text, sampling noise, decoding an image. You connect nodes with wires to create a pipeline from text prompt to final image.
The advantage over simpler tools like Automatic1111 is control. You can see exactly what happens at each step, swap components in and out, and build complex workflows that combine multiple models, LoRAs, ControlNets, and post-processing steps. Once you build a workflow, you can save it, share it, and rerun it with different prompts indefinitely. This is why professionals and AI influencer creators use ComfyUI as their primary tool.
ComfyUI vs Other Tools
ComfyUI
Node-based visual workflows. Full control over every generation step. Handles all model types (SDXL, Flux, SD3). Steepest learning curve but most powerful. Free and open source.
Automatic1111
Traditional form-based UI. Fill in settings and click Generate. Easier to start with but limited for complex workflows. Slower to adopt new model architectures. Free and open source.
Midjourney / DALL-E
Cloud-based, prompt-only. No workflow control, no custom models, no LoRA support. Easiest to use but least flexible. Monthly subscription required. No local processing.
Installing ComfyUI (Windows, Mac, and Cloud)
ComfyUI runs locally on your computer or on a cloud GPU. Choose the method that matches your hardware.
Option 1: Windows (NVIDIA GPU)
Download the Portable Package
Go to the ComfyUI GitHub releases page and download the latest Windows portable package (ComfyUI_windows_portable.7z). Extract it to a folder with a short path like C:\ComfyUI. Avoid paths with spaces or special characters.
Install NVIDIA Drivers
Make sure you have the latest NVIDIA drivers installed from nvidia.com/drivers. The portable package includes CUDA and PyTorch, so no separate Python installation is needed.
Launch ComfyUI
Double-click run_nvidia_gpu.bat. A terminal window opens showing the loading progress. Once you see "Starting server", open your browser to http://127.0.0.1:8188. ComfyUI loads with a default text-to-image workflow.
Option 2: Mac (Apple Silicon)
Install Python 3.11+ from python.org. Open Terminal and clone the repository: git clone https://github.com/comfyanonymous/ComfyUI. Navigate into the folder, install dependencies with pip install -r requirements.txt, and run python main.py --force-fp16. The --force-fp16 flag is required for Apple Silicon. Performance is roughly 2-4x slower than a comparable NVIDIA GPU, but M2 Pro and M3 chips handle SDXL reasonably well.
Option 3: Cloud (RunPod)
If you do not have a capable GPU, RunPod is the best cloud option. Create an account at runpod.io, deploy a GPU pod with the ComfyUI template (search "ComfyUI" in the template marketplace), and select an A40 or RTX 4090 GPU. The template comes pre-installed with ComfyUI and common models. Access it through the pod's web interface. Costs range from $0.39 to $0.69 per hour depending on GPU selection. Stop the pod when not in use to avoid charges.
Understanding ComfyUI Nodes
Every workflow in ComfyUI is built from nodes. Each node takes inputs, processes them, and produces outputs. Understanding the core nodes is the key to using ComfyUI effectively. Here are the essential ones you will use in every workflow.
Load Checkpoint
Loads the main AI model (like SDXL or Flux). This is always the starting point of any workflow. Place your model files (.safetensors) in the ComfyUI/models/checkpoints/ folder and they appear in the dropdown.
CLIP Text Encode
Converts your text prompt into a format the AI model understands. You need two of these: one for the positive prompt (what you want) and one for the negative prompt (what you want to avoid). Connect both to the sampler.
KSampler
The core generation engine. It takes noise and iteratively refines it into an image guided by your prompts. Key settings include steps (20-30 for quality), CFG scale (7-8 for balanced results), and sampler name (euler_ancestral or dpmpp_2m are reliable defaults).
VAE Decode
Converts the sampler's output from latent space into a visible image. The VAE (Variational Autoencoder) is loaded from the checkpoint by default. Connect the KSampler output to VAE Decode, then connect that to a Save Image or Preview Image node.
Empty Latent Image
Creates a blank canvas of noise at your desired resolution. For SDXL use 1024x1024, for SD 1.5 use 512x512. This connects to the KSampler's latent_image input. Batch size controls how many images generate at once.
Save Image / Preview Image
Save Image writes the final output to your ComfyUI/output/ folder. Preview Image shows it in the UI without saving. Always end your workflow with one of these nodes so you can see the results.
Your First Workflow: Text-to-Image
When you first open ComfyUI, it loads a default text-to-image workflow. Here is how to use it to generate your first AI image step by step.
Step-by-Step First Generation
Download a Model
Download SDXL 1.0 Base from Hugging Face (stabilityai/stable-diffusion-xl-base-1.0). Place the .safetensors file in ComfyUI/models/checkpoints/. The file is about 6.5 GB. Refresh ComfyUI or restart it to see the new model.
Select the Model
Click on the Load Checkpoint node in the default workflow. Select your SDXL model from the dropdown. This model outputs three connections: MODEL (the neural network weights), CLIP (the text encoder), and VAE (the image decoder).
Write Your Prompt
Click the positive CLIP Text Encode node and enter your prompt. Start simple: "professional portrait photo of a young woman, studio lighting, sharp focus, 8k resolution". In the negative prompt node, enter: "blurry, low quality, distorted, deformed hands, extra fingers".
Set Resolution and Generate
In the Empty Latent Image node, set width and height to 1024x1024 for SDXL. In the KSampler, set steps to 25, CFG to 7.5, and sampler to euler_ancestral. Click "Queue Prompt" (or press Ctrl+Enter). Your first image generates in 10-30 seconds depending on your GPU.
Essential Models to Download
The model you use determines the style and quality of your output. Here are the essential models every ComfyUI user should have.
| Model | Best For | VRAM | Size |
|---|---|---|---|
| SDXL 1.0 Base | General purpose, largest LoRA ecosystem | 8 GB | 6.5 GB |
| Flux Dev | Photorealism, text rendering in images | 12 GB | 23 GB |
| Flux Schnell | Fast generation (4 steps), drafts and iteration | 12 GB | 23 GB |
| RealVisXL V4 | Photorealistic portraits and scenes (SDXL fine-tune) | 8 GB | 6.5 GB |
| Juggernaut XL | AI influencer content, cinematic photos | 8 GB | 6.5 GB |
Download models from CivitAI (civitai.com) or Hugging Face (huggingface.co). Always download the .safetensors format, not .ckpt, for security. Place all checkpoint files in ComfyUI/models/checkpoints/.
LoRA Integration for Character Consistency
LoRAs (Low-Rank Adaptations) are small model add-ons that teach the base model new concepts: a specific face, an art style, a clothing type, or a pose. They are essential for creating consistent AI characters across multiple images, which is the foundation of any AI influencer brand.
How to Use LoRAs in ComfyUI
Download and Place the LoRA
Download a LoRA .safetensors file from CivitAI. Place it in ComfyUI/models/loras/. LoRA files are typically 50-200 MB, much smaller than full checkpoints.
Add the Load LoRA Node
Right-click the canvas, go to Add Node > Loaders > Load LoRA. Connect the MODEL and CLIP outputs from your checkpoint loader to the LoRA node's inputs. Then connect the LoRA node's MODEL and CLIP outputs to your KSampler and CLIP Text Encode nodes.
Adjust LoRA Strength
Set the strength_model and strength_clip values between 0.7 and 0.9 for most LoRAs. Lower values (0.5-0.7) give a subtle effect. Higher values (0.9-1.0) give a stronger effect but can introduce artifacts. Include the LoRA trigger words in your positive prompt as described on the model's download page.
Training Your Own LoRA
To create a LoRA for your own AI character, collect 15-30 high-quality reference images (varied angles, lighting, and expressions). Use Kohya_ss (free, local) or CivitAI on-site training (paid, simpler). Train for 1,500-3,000 steps with a learning rate of 1e-4. The result is a .safetensors file under 200 MB that captures your character's likeness across any scene or outfit.
Advanced Workflows: IPAdapter and ControlNet
Once you are comfortable with basic text-to-image, these two tools unlock professional-grade control over your generations.
IPAdapter (Face Transfer)
IPAdapter takes a reference image and transfers the face, style, or composition into your generation. Upload a photo of your AI character, connect it through the IPAdapter node, and every generated image will have that face. This is more flexible than LoRAs because you do not need to train anything: just provide a reference photo.
Setup: Install the ComfyUI IPAdapter Plus custom nodes. Download the ip-adapter-plus-face_sdxl model. Connect: Load Image > IPAdapter > KSampler. Set weight to 0.7-0.85 for face transfer.
ControlNet (Pose and Composition)
ControlNet lets you guide the pose, depth, edges, or composition of your generation using a reference image. Upload a photo with the pose you want, run it through a ControlNet preprocessor (like OpenPose for body pose or Canny for edges), and the AI generates a new image matching that structure.
Setup: Install the ComfyUI ControlNet Auxiliary Preprocessors nodes. Download the appropriate ControlNet model (OpenPose for poses, Depth for spatial layout). Set strength to 0.6-0.8 for natural results.
Combining Everything: The AI Influencer Workflow
The professional AI influencer workflow combines all three: a base checkpoint (Juggernaut XL or RealVisXL) provides the foundation, a character LoRA ensures facial consistency, IPAdapter reinforces face similarity from a reference photo, and ControlNet controls the pose from a stock photo reference. This stack produces images that look like the same person in different outfits, locations, and poses. It is the same pipeline used by AI influencers with 100K+ followers.
ComfyUI Manager and Community Resources
The ComfyUI ecosystem includes thousands of custom nodes, workflows, and models. Here are the essential resources to level up your workflow.
ComfyUI Manager
The essential extension. Install it first. It lets you browse, install, and update custom nodes directly from the ComfyUI interface. It also detects missing nodes when you load a shared workflow and offers to install them automatically. Install by cloning the repo into ComfyUI/custom_nodes/.
CivitAI
The largest repository for Stable Diffusion models, LoRAs, and embeddings. Filter by base model (SDXL, Flux), sort by popularity, and read reviews before downloading. Most models include example images and recommended settings. Free to download with optional membership.
OpenArt Workflow Library
A collection of pre-built ComfyUI workflows for common tasks: face swap, upscaling, inpainting, video generation, and more. Download a workflow JSON, drag it into ComfyUI, install any missing nodes via ComfyUI Manager, and run it. A massive time saver when learning.
r/comfyui and Discord
The ComfyUI subreddit and Discord server are the best places to troubleshoot issues, share workflows, and learn new techniques. When you hit an error, search the subreddit first because someone has almost certainly encountered the same issue.
Frequently Asked Questions
Do I need a powerful GPU to run ComfyUI?
For local use, you need an NVIDIA GPU with at least 6GB VRAM (GTX 1660 or better). 8-12GB VRAM is recommended for SDXL and Flux models. If you do not have a compatible GPU, use a cloud option like RunPod where you can rent an A40 or 4090 GPU for $0.39-$0.69 per hour. Mac users with M1/M2/M3 chips can run ComfyUI natively, though generation is slower than NVIDIA cards.
What is the difference between ComfyUI and Automatic1111?
Automatic1111 (also called A1111 or SD WebUI) uses a traditional settings-panel interface where you fill in fields and click Generate. ComfyUI uses a node-based visual workflow where you connect processing blocks together. ComfyUI gives you more control over every step of image generation, makes complex workflows reusable and shareable, and handles new model architectures (like Flux) faster. A1111 is simpler for basic text-to-image but ComfyUI is the standard for serious AI image work in 2026.
Which model should I start with in ComfyUI?
Start with SDXL 1.0 Base. It produces high-quality 1024x1024 images, has the largest ecosystem of LoRAs and fine-tunes, and runs on GPUs with 8GB VRAM. Once comfortable, try Flux Dev for photorealistic images or Flux Schnell for fast generation. Avoid SD 1.5 models for new projects as the quality gap with SDXL is significant.
How do I keep my AI character looking consistent across images?
Use a LoRA (Low-Rank Adaptation) trained on your character. You can train a custom LoRA on 15-30 reference images using Kohya_ss or CivitAI training. In ComfyUI, add a Load LoRA node between your checkpoint loader and CLIP encoder, set the strength to 0.7-0.9, and your character will appear consistently. For additional control, combine the LoRA with IPAdapter (face reference) and ControlNet (pose control).
Is ComfyUI free?
Yes, ComfyUI is completely free and open source. The software itself costs nothing. You only pay if you choose to run it on a cloud GPU service like RunPod. Models are also free to download from CivitAI and Hugging Face. The only potential costs are cloud GPU rental (if you do not have a local GPU) and optional premium LoRAs from some creators.
Want to master AI Influencers Academy? Get it + 3 more complete courses
Complete Creator Academy - All Courses
Master Instagram growth, AI influencers, n8n automation, and digital products for just $99/month. Cancel anytime.
All 4 premium courses (Instagram, AI Influencers, Automation, Digital Products)
100+ hours of training content
Exclusive templates and workflows
Weekly live Q&A sessions
Private community access
New courses and updates included
Cancel anytime - no long-term commitment
✨ Includes: Instagram Ignited • AI Influencers Academy • AI Automations • Digital Products Empire