Edit model card

GGUF quantized version of Stable Diffusion 3.5 Medium

screenshot

Setup (once)

  • drag sd3.5_medium-q5_0.gguf (2.02GB) to > ./ComfyUI/models/unet
  • drag clip_g.safetensors (1.39GB) to > ./ComfyUI/models/clip
  • drag clip_l.safetensors (246MB) to > ./ComfyUI/models/clip
  • drag t5xxl_fp8_e4m3fn.safetensors (4.89GB) to > ./ComfyUI/models/clip
  • drag diffusion_pytorch_model.safetensors (168MB) to > ./ComfyUI/models/vae

Run it straight (no installation needed way)

  • run the .bat file in the main directory (assuming you are using the gguf-comfy pack below)
  • drag the workflow json file (see below) to > your browser
  • generate your first picture with sd3, awesome!

Workflows

  • example workflow for gguf (if it doesn't work, upgrade your pack: ggc y) 👻
  • example workflow for the original safetensors 🎃

Bug reports (or brief review)

  • t/q1_0 and t/q2_0; not working recently (invalid GGMLQ type error)
  • q2_k is super fast; not usable; but might be good for medical research or abstract painter
  • q3 family is fast; finger issue can be easily detected but picture quality is interestingly good
  • btw, q3 family is usable; just need some effort on +/- prompt(s); good for old machine user
  • notice the same file size in some quantized models; but they are different (according to its SHA256 hash) even they are exactly the same in size; still keep them all here (in full set); see who can deal with it
  • q4 and above should be no problem for general-to-high quality production (demo picture shown above generated by just Q4_0 - 1.74GB); and sd team is pretty considerable, though you might not think this model is useful if you have good hardware; imagine you can run it with merely an ancient CPU, should probably appreciate their great effort; good job folks, thumbs up 👍

Upper tier options

References

Downloads last month
3,562
GGUF
Model size
2.47B params
Architecture
sd3

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for calcuis/sd3.5-medium-gguf

Quantized
(2)
this model