Replies: 1 comment
-
The convert functionality is already working. It's still missing weight quantization on model loading, but I guess the file conversion would be more useful anyway. I'll send a PR later today. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
@wbruna This was your message on #696 issue, so i wanted to know if you were able to finish your work on this feature. I recently experienced the power of using --override-tensor in llama.cpp which helped me load bigger model easily and i was able to improve performance of inference, so i was interested to see if it was possible for image generation models like flux, so i searched for it and got to know about this another great project.
I tried to find info regarding similar feature in this project but was not able to find it and then i tried to search in this project issue and got to know that someone was working on it, so was just interested to know if you were able achieve similar thing like llama.cpp with --override-tensor in this project :-))
Beta Was this translation helpful? Give feedback.
All reactions