flux-2-klein-9b.safetensors
FLUX.2 Klein 9Bβ
Look, I've seen a lot of model files waddle through ComfyUI in my time, but flux-2-klein-9b.safetensors is something special. This isn't your grandmother's image generation modelβunless your grandmother has a thing for sub-second inference times and 9 billion parameters, in which case, respect.
π§ File Formatβ
Type: SafeTensors (.safetensors)
Size: Approximately 17-20GB (because excellence takes space, folks)
Architecture: Rectified flow transformer with integrated Qwen3 text embedder
SafeTensors format means you're not loading some sketchy pickle file that could potentially order pizza to your address at 3 AM. It's a safe, efficient tensor storage format that loads faster than you can say "why is my VRAM full again?"
π Function in ComfyUI Workflowsβ
FLUX.2 Klein 9B is your Swiss Army knife for image generation. This bad boy handles:
- Text-to-Image Generation: Type words, get pixels. Revolutionary, I know.
- Image-to-Image Multi-Reference Editing: Feed it reference images and watch it work its magic
- Unified Architecture: One model to rule them allβgeneration AND editing in the same package
In ComfyUI, this model slots into your checkpoint loader nodes and becomes the beating heart of your workflow. It's the difference between "I made an image" and "I made an image that doesn't look like it was rendered on a potato."
π§ Technical Detailsβ
Alright, time to get nerdy (as if we haven't been already):
- Parameters: 9 billion (9B flow model + 8B Qwen3 text embedder)
- Inference Steps: 4 (step-distilled for speed demons)
- VRAM Requirements: ~24GB (hope you have a GPU that didn't come from a garage sale)
- Architecture Type: Rectified flow transformer
- Training Methodology: Step-distilled from base model
- Speed: Sub-second generation (yes, really)
- Resolution Support: Variable, optimized for standard aspect ratios
This model sits at the Pareto frontier for quality vs. latency, which is a fancy way of saying it punches way above its weight class. Models 5x its size are side-eyeing this thing with concern.
β Benefitsβ
Speed That'll Make You Question Reality
Sub-second generation times. I've seen cold starts take longer than this model takes to generate a masterpiece.
Quality Without Compromise
Matches or exceeds models with 45B+ parameters. It's like bringing a tactical nuke to a pillow fight.
Unified Model Architecture
One model for text-to-image AND multi-reference editing. No more juggling seventeen different checkpoints like some kind of ML circus performer.
Excellent Prompt Adherence
Actually listens to what you tell it. Unlike my last three neural networks, which had the listening comprehension of a goldfish.
Output Diversity
Great for creative exploration. Generate variations without everything looking like the same image with a different hat.
Real-Time Application Integration
Built for production use, not just showing off on Reddit.
βοΈ Usage Tipsβ
Sampler Settings:
Since this model is step-distilled to 4 steps, don't go crazy with 50+ steps. You're not making it better, you're just making your GPU cry. Stick to 4-8 steps for optimal results.
CFG Scale:
Start around 3.5-7.0. This model doesn't need aggressive guidance to understand what you want. It's not a rebellious teenager.
Prompt Engineering:
Be specific but don't write a novel. This model has an 8B parameter text embedderβit understands nuance better than your autocorrect understands what you're trying to type.
Batch Processing:
With sub-second inference, you can actually do batch generation without aging noticeably. Live your best life.
Reference Images for Editing:
When using image-to-image workflows, quality reference images = quality outputs. Garbage in, garbage outβa tale as old as computing itself.
π ComfyUI Setup Instructionsβ
- Download the Model
Acquireflux-2-klein-9b.safetensorsfrom your preferred source (see Additional Resources below). - Installation Location
Place the model file in your ComfyUI models directory:
ComfyUI/models/checkpoints/Don't put it in/Downloads/random_stuff/maybe_models/idk/like some kind of digital hoarder. - Verify VRAM
Check that you have at least 24GB VRAM available. If you don't, this is a great opportunity to explain to your significant other why you need a new GPU. - Load in ComfyUI
- Add a "Load Checkpoint" node to your workflow
- Select
flux-2-klein-9b.safetensorsfrom the dropdown - Connect to your KSampler or other sampling nodes
- Set steps to 4-8 (seriously, don't overdo it)
- Configure Sampling
- Use appropriate samplers (Euler, DPM++ recommended)
- Keep CFG scale reasonable (3.5-7.0 range)
- Set resolution to your target output size
- Test Generation
Run a simple prompt first. Something straightforward to make sure everything works properly before you start generating your magnum opus.
π₯ What-Not-To-Do-Unless-You-Want-a-Fireβ
Don't Use 50+ Sampling Steps
This model was step-distilled to 4 steps. Using 50 steps is like taking a Ferrari through a school zoneβpointless and you're just wasting resources.
Don't Crank CFG to 15
High CFG values will make your outputs look like they went through a deep-frying process. Twice.
Don't Load This on 8GB VRAM
Physics says no. Your GPU will say no. Your computer's fans will achieve liftoff velocity trying to say no.
Don't Mix with Incompatible VAEs
Use the recommended VAE or auto-select. Mixing random VAEs is like putting diesel in a sports carβtechnically possible, hilariously inadvisable.
Don't Skip the Text Encoder
This model uses an 8B Qwen3 text embedder. It's not optional. It's PART of the architecture. Skipping it is like trying to make a sandwich without bread.
Don't Use on Ancient Hardware
If your GPU still thinks HDMI is "newfangled technology," this model is not for you.
π Additional Resourcesβ
Model Download: FLUX.2 Klein 9B download
License Information:
FLUX Non-Commercial License (read it before your lawyer has to)
Community Resources:
- ComfyUI Official Documentation
- FLUX Model Family Documentation
- ComfyUI Community Forums
π Example Node Configurationβ
[Load Checkpoint]
βββ checkpoint_name: flux-2-klein-9b.safetensors
βββ output: MODEL, CLIP, VAE
[CLIP Text Encode (Prompt)]
βββ text: "a majestic mountain landscape at sunset, detailed, professional photography"
βββ clip: [from Load Checkpoint]
βββ output: CONDITIONING
[KSampler]
βββ model: [from Load Checkpoint]
βββ seed: 42 (or random, live dangerously)
βββ steps: 4
βββ cfg: 5.0
βββ sampler_name: euler
βββ scheduler: normal
βββ positive: [from CLIP Text Encode]
βββ negative: [from Negative Prompt]
βββ latent_image: [from Empty Latent Image]
βββ output: LATENT
[VAE Decode]
βββ samples: [from KSampler]
βββ vae: [from Load Checkpoint]
βββ output: IMAGE
Alternative: Multi-Reference Editing Configuration
[Load Checkpoint]
βββ flux-2-klein-9b.safetensors
[Load Image] (Reference Image 1)
[Load Image] (Reference Image 2)
[VAE Encode] β Connect reference images
[KSampler]
βββ steps: 4-6
βββ cfg: 4.5
βββ denoise: 0.7-0.85 (for editing)
βββ [Connect everything appropriately]
π Notesβ
Performance Optimization:
If you're experiencing slower-than-advertised inference times, check your:
- CUDA/ROCm versions (update if you're still running drivers from 2019)
- System RAM (yes, it matters)
- Background processes (close those 47 Chrome tabs)
- Cooling (thermal throttling is real)
Model Variants:
This documentation covers the 9B model specifically. There's also a 4B variant with Apache 2.0 licensing if you need commercial use, but that's a story for another documentation file.
Fine-Tuning Potential:
While this is the step-distilled version, you can technically fine-tune it. However, if you want maximum fine-tuning potential, the Base 9B variant (undistilled) is your friend.
Compatibility:
Works with ComfyUI's standard node ecosystem. Most custom nodes should play nicely, but if something breaks, check for FLUX-specific compatibility notes.
Future Updates:
The FLUX model family is actively developed. Keep an eye on official channels for updates, improvements, and new variants that might make this documentation obsolete (the circle of tech life).
Remember: With great parameters comes great VRAM requirements. Use responsibly, generate beautifully, and may your inference times be ever in your favor.
P.S. If this model doesn't work for you, check your setup before assuming the model is broken. 99% of the time it's user error. The other 1% is cosmic rays flipping bits in your RAM.