-
Notifications
You must be signed in to change notification settings - Fork 75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dolly example no longer works ... #156
Comments
Which version of Transformers.jl and Pickle.jl? |
(@v1.9) pkg> status Transformers |
Ok, so you would need to update Pickle.jl to 0.3.3 which adds support for bfloat16. |
pkg>update Pickle |
on linux / ubuntu ..is this platform specific ? |
Not really. What happens if you explicitly |
(@v1.9) pkg> add [email protected] |
I will create a local project and activate and see if it works .. |
Similar warning as before ... will load the model and see if it works. |
This time ...no error , but model hangs .. |
The warning can usually be ignored
That is a big model, which takes time to be moved to GPU. And there is an extra |
I have a 3080 ti ..will wait to see and let you know . Thank you. I was able to run Ollama with other 13b models fairly quickly. |
finally it put out an error: |
That is because the extra |
Let me see where it is happening ... it is this line: |
The other thing , I noticed ..nvidia-smi was constant ...almost telling it was not copying there .. |
I will copy into a vim editor and try it out ..that way , double quotes sometimes comes up with special chars...that can be eliminated as a cause. |
Still no luck. It just kills my shell after a while . Will come back and try later . |
It sounds like the process might be killed due to OOM. So currently you would need about 70GB CPU memory to load the |
This example doesn't work : https://info.juliahub.com/large-language-model-llm-tutorial-with-julias-transformers.jl
Looks like this package ..is no longer current and doesn't work with LLMs in HuggingFace
textenc = hgf"databricks/dolly-v2-12b:tokenizer"
Warning: fuse_unk is unsupported, the tokenization result might be slightly different in some cases
julia> model = todevice(hgf"databricks/dolly-v2-12b:ForCausalLM")
ERROR: unknown type: torch.BFloat16Storage
The text was updated successfully, but these errors were encountered: