I need help to quantize video-llava model to quant size 4 or 8 #6029
dnabanita7
started this conversation in
General
Replies: 1 comment
-
you can use bitsandbytes to convert the model, i assume it is an easiest way load the model from transformer |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I am trying to convert video-llava according to instructions at #2948 but I am having issues because of
convert.py
file. Since, this is llava + vicuna should I add anything extra in the convert.py file or let it remain the same?Beta Was this translation helpful? Give feedback.
All reactions