-
-
Notifications
You must be signed in to change notification settings - Fork 81
Vicuna problem #160
Comments
I don't think basaran supports GPTQ pre-quantized models. Feature request? 👀 |
Basaran should work with Vicuna models. The model repo seems to contain outdated configs that point to non-existing weight files: https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g/discussions/15 Also, you may want to install |
Do you know how to make it work? I get this error.
|
Oops, I forgot to install the extra dependencies since I was inside a venv. Works good! |
Hey I'd appreciate it if you could help me out running this model with basaran. Maybe point me to the right huggingface repository? Thanks. |
Thank you, you rock. |
@zhound420 did you manage to run GPTQ model? |
@karfly no I did not yet. I'll have to come back to it in a couple days. |
@zhound420 looking forward to hear from you! |
Did you manage to do it? Kinda stuck on the same. Edit: I'm trying to use it with docker image(1st option) |
Has anyone got this model to work yet? Running into this:
OSError: anon8231489123/vicuna-13b-GPTQ-4bit-128g does not appear to have a file named pytorch_model-00001-of-00003.bin. Checkout 'https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g/main' for available files.
The text was updated successfully, but these errors were encountered: