Description
Stable Diffusion 3.5 Large and Large Turbo just got released publicly.
https://huggingface.co/stabilityai/stable-diffusion-3.5-large
https://huggingface.co/stabilityai/stable-diffusion-3.5-large-turbo
Inference code here (warning: weird licence): https://github.com/Stability-AI/sd3.5
It's a model that should perform fairly well (SD3-Large is ranked slightly above Flux Schnell on artificialanalysis areana leaderboard, and this is an upgraded version of SD3-Large), while being smaller than Flux (it's has 8B parameters).
Right now, these two models are not supported by sdcpp (I tried).
What's required:
- Differentiate SD3 2B and SD3.5 Large/Large Turbo
- Add
--clip_g
argument: Add--clip_g
argument and support split SD3 2B models (for SD3.5 support) #444 - Fix k quantization (q3_k I generated is 13.5 GB, while the full fp16 model is 16GB)
- Maybe more?
Sidenote: SD3.5 Medium (2B) is also going to be released soon, hopefully it will work as a drop-in replacement for SD3 2B
Edit: About quantization, the majority of tensors in sd3.5 large do not fit nicely in a whole number of blocks of size 256, so they are skipped when trying to quantize to q3_k, q4_k and so on.