-
Notifications
You must be signed in to change notification settings - Fork 45
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Export of Llama2 fails #76
Comments
I have the exact same issue |
I get the same error when trying to convert: https://huggingface.co/HuggingFaceTB/SmolLM-1.7B I'm wondering if this is one of the situations where the torch operation -> CoreML Operation mapping does not automatically work (i.e. requires us to write our own operator: https://apple.github.io/coremltools/docs-guides/source/custom-operators.html |
@rradjabi try installing coremltools 8 and a newer version of transformers! I was able to run this conversion just fine 👏 (with my own memory fixing patch ofcourse). |
@Proryanator Could you please provide more details about your fix? Thanks |
Yeah sure! Let me collect the specific details (it was a bit complicated in the end). In a nutshell though: Out of Memory Issue Unsupported 'full' op |
I'm unable to use
exporters
formeta-llama/Llama-2-7b-chat-hf
model.Here is my command
And here is the output
I was able to generate a
mlpackage
fordistilbert-base-uncased-finetuned-sst-2-english
, with this command:python -m exporters.coreml --model=distilbert-base-uncased-finetuned-sst-2-english --feature=sequence-classification models/defaults.mlpackage
, so I have some confidence that the environment is correct and working.The text was updated successfully, but these errors were encountered: