Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Does AutoGGUF support quantization for SD3 or Flux and other DIT drawing models? #79

Open
Amazon90 opened this issue Jan 20, 2025 · 1 comment
Assignees

Comments

@Amazon90
Copy link

Amazon90 commented Jan 20, 2025

Your GUI tool is very beginner-friendly for those new to model quantization, but I’m unsure if your tool supports quantizing drawing models. I hope you can provide an answer.

https://github.com/city96/ComfyUI-GGUF/blob/main/tools/README.md

@leafspark
Copy link
Owner

It should support quantizing DIT models like SD3 and Flux. Here's how:

  1. Convert your safetensors model to GGUF format:
python convert.py --src model.safetensors
  1. Build and install the patched llama.cpp version and copy its binaries into a new folder in the llama_bin folder in your AutoGGUF directory.

  2. Launch AutoGGUF - the new folder should appear in the backend selection.

  3. Quantize normally, but:

    • Skip text model parameters
    • Use the Extra Arguments box for vision model settings

If there's prebuilt binaries in a repo you'd like to use, you could also set AUTOGGUF_BACKEND_REPO to owner/repo to pull new releases from that repository instead (although the binaries have to be in one zip file, same as the llama.cpp repo).

@leafspark leafspark self-assigned this Jan 27, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants