-
Notifications
You must be signed in to change notification settings - Fork 2
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Flux1 nf4 #1
Comments
Currently, it only supports the official models from BlackForestLabs. The main reason being that this is my first project running inference, and I have no idea how to load models manually. I'll try to look into it, because I'm interested in getting the quantized models to run myself, but can't promise that I'll get it to work. |
@freecode-ai I've pushed a few commits with significant changes to model loading, theoretically allowing for loading of quantized models. Experimental quantized model supportAn fp8 version of dev and schnell is now available for selection, but I can't test it myself, because fp8 isn't supported on MPS, which is currently the only device I have access to. If you or anyone else reading this could test it, please let me know if it works correctly! Issues regarding NF4 supportHonestly, I just don't know how to load them properly.. After the recent refactoring, there's support for loading a I've tried finding out how I can load these models and so far, I'm looking into the If anyone knows how to do this, please let me know, or even better, submit a PR :) Future GGUF supportNow that the codebase is a bit cleaner and there's more support for doing more manual stuff regarding model loading, maybe we could support GGUF too to get access to all the nice integer GGUF quants? My issues regarding this are basically the same as my nf4 issues: How can I load these models, and is it even compatible with the |
Does this support the nf4 model?
https://huggingface.co/lllyasviel/flux1-dev-bnb-nf4
The text was updated successfully, but these errors were encountered: