Skip to content
This repository was archived by the owner on Jan 24, 2024. It is now read-only.

Vicuna problem #160

Closed
zhound420 opened this issue Apr 20, 2023 · 13 comments
Closed

Vicuna problem #160

zhound420 opened this issue Apr 20, 2023 · 13 comments
Labels
question Further information is requested

Comments

@zhound420
Copy link

Has anyone got this model to work yet? Running into this:

OSError: anon8231489123/vicuna-13b-GPTQ-4bit-128g does not appear to have a file named pytorch_model-00001-of-00003.bin. Checkout 'https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g/main' for available files.

@zhound420 zhound420 changed the title Vacuna problem Vicuna problem Apr 20, 2023
@jota2rz
Copy link

jota2rz commented Apr 20, 2023

I don't think basaran supports GPTQ pre-quantized models.
https://github.com/oobabooga/text-generation-webui supports this model.
Documentation at https://github.com/oobabooga/text-generation-webui/wiki/GPTQ-models-(4-bit-mode)

Feature request? 👀

@peakji
Copy link
Member

peakji commented Apr 21, 2023

Basaran should work with Vicuna models. The model repo seems to contain outdated configs that point to non-existing weight files: https://huggingface.co/anon8231489123/vicuna-13b-GPTQ-4bit-128g/discussions/15

Also, you may want to install safetensors, as the repo only provides weights in safetensors format.

@jota2rz
Copy link

jota2rz commented Apr 21, 2023

Basaran should work with Vicuna models.

Do you know how to make it work?

I get this error.

ValueError: Couldn't instantiate the backend tokenizer from one of:
(1) a `tokenizers` library serialization file,
(2) a slow tokenizer instance to convert or
(3) an equivalent slow tokenizer class to instantiate and convert.
You need to have sentencepiece installed to convert a slow tokenizer to a fast one.

@jota2rz
Copy link

jota2rz commented Apr 21, 2023

Oops, I forgot to install the extra dependencies since I was inside a venv.
I needed transformers, sentenpiece and safetensors.
pip install safetensors transformers[sentencepiece]

Works good!

@fardeon
Copy link
Member

fardeon commented Apr 21, 2023

We will add safetensors support in the next release: #174 #175

@zhound420
Copy link
Author

Oops, I forgot to install the extra dependencies since I was inside a venv. I needed transformers, sentenpiece and safetensors. pip install safetensors transformers[sentencepiece]

Works good!

Hey I'd appreciate it if you could help me out running this model with basaran. Maybe point me to the right huggingface repository? Thanks.

@jota2rz
Copy link

jota2rz commented Apr 21, 2023

@zhound420
Copy link
Author

@zhound420 https://rentry.org/nur779

Thank you, you rock.

@fardeon fardeon added the question Further information is requested label Apr 24, 2023
@karfly
Copy link

karfly commented Apr 25, 2023

@fardeon @peakji

Hi, guys!
In the end, I did not understand whether GPTQ 4bit models are supported or not?

@karfly
Copy link

karfly commented Apr 26, 2023

@zhound420 did you manage to run GPTQ model?

@zhound420
Copy link
Author

@karfly no I did not yet. I'll have to come back to it in a couple days.

@karfly
Copy link

karfly commented Apr 26, 2023

@zhound420 looking forward to hear from you!

@advaitdeshmukh
Copy link

advaitdeshmukh commented May 30, 2023

@karfly no I did not yet. I'll have to come back to it in a couple days.

Did you manage to do it? Kinda stuck on the same.

Edit: I'm trying to use it with docker image(1st option)

Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
question Further information is requested
Projects
None yet
Development

No branches or pull requests

6 participants