Skip to content

Flux.2 Klein

A lightweight 4B parameter FLUX variant. Fast 4-step generation with low VRAM requirements.

Variants

ModelStepsSizeNotes
flux2-klein:q844.3 GBGood quality
flux2-klein:q643.4 GBBetter quality
flux2-klein:q442.6 GBSmallest FLUX
flux2-klein:bf1647.8 GBFull precision 4B

Defaults

  • Resolution: 1024x1024
  • Guidance: 0.0
  • Steps: 4

Flux.2 Klein-9B (alpha)

Note: Klein-9B is in alpha. The Qwen3 encoder (hidden_size=4096) differs from Klein-4B (hidden_size=2560) and is not yet fully supported.

A larger 9B parameter FLUX variant. Distilled for fast 4-step generation with higher quality than the 4B Klein.

Variants

ModelStepsSizeNotes
flux2-klein-9b:q8410 GBGood quality
flux2-klein-9b:q647.9 GBBetter quality
flux2-klein-9b:q445.9 GBSmallest 9B
flux2-klein-9b:bf16418 GBFull precision, gated, 2 shards

Defaults

  • Resolution: 1024x1024
  • Guidance: 1.0
  • Steps: 4

Note: The 9B model requires ~29GB VRAM for BF16. The BF16 variant is gated on HuggingFace and requires license acceptance before download.

WidthHeightAspect Ratio
102410241:1 (native)
10247684:3
76810243:4
102457616:9
57610249:16
7687681:1

Using non-recommended dimensions will trigger a warning. All values must be multiples of 16.

Example

Flux.2 Klein Q8 — 4 steps, seed 100:

bash
mold run flux2-klein:q8 "A minimalist zen garden with raked sand patterns, a single cherry blossom tree, morning mist" --seed 100

Zen garden — Flux.2 Klein

Flux.2 Klein BF16 — 4 steps:

bash
mold run flux2-klein:bf16 "a majestic owl perched on a mossy branch in a moonlit forest"

Owl — Flux.2 Klein BF16

Architecture

Flux.2 Klein uses a Qwen3 text encoder (BF16 or GGUF, layers 9/18/27), a shared modulation transformer (BF16 or GGUF), and a BN-VAE decoder.