-
Notifications
You must be signed in to change notification settings - Fork 28.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Running LlamaForCausalLM with MPS provokes "RuntimeError: MPS does not support cumsum op with int64 input" #22502
Comments
Relevant stack trace (can provide more if needed):
This seems to happen during greedy search and subsequently precisely at:
|
Actually, this could be PyTorch/MPS issue, that the int64 version of cumsum is not implemented. Found the issue there: I wonder if long is necessary for attention_mask? should int32 be good enough? |
According to the issue it should be fixed with a nightly install of PyTorch and MacOS 13.3 |
@sgugger thanks for responding. I just updated to 13.3 and the torch nightly, and indeed, no more problem. Closing issue. |
just for fun, increase length to 256 my prompt is "Is facebook a bad company?" " Is facebook a bad company?\nI'm not sure if this is the right place to post this, but I'm not sure where else to post it.\nI'm not a facebook user, but I've heard a lot of bad things about it. I've heard that it's a bad company, that it's a bad product, that it's a bad service, that it's a bad website, that it's a bad social network, that it's a bad company, that it's a bad product, that it's a bad service, that it's a bad website, that it's a bad social network, that it's a bad company, that it's a bad product, that it's a bad service, that it's a bad website, that it's a bad social network, that it's a bad company, that it's a bad product, that it's a bad service, that it's a bad website, that it's a bad social network, that it's a bad company, that it's a bad product, that it's a bad service, that it's a bad website" it started repeating things. Maybe this is 7B, and it would behave better for larger one? This must have not been an encouraging sign for earlier pioneers. So it is amazing openAi stuck at it and arrived all the way to chatGPT level of great. |
This is a problem for me now - running 13.5.2 MacOS, python 3.10.9. Cannot find a solution to this other than workarounds that I can't understand. Any advice on how to get past this? Must be a problem for a lot of people? Thanks in advance. |
I have the same issue (RuntimeError: MPS does not support cumsum op with int64 input) with my MacOS Version 14.0 and nightly torch. Any idea how I can solve this issue? |
I have same issue, anyone can help me ? |
m1 macOS 14.1.1 (23B81), also has this problem |
Running against ChatMusician, which has been trained from Llama 2 7b I see the same thing. Solved with:
Relevant from the installation:
|
System Info
transformers
version: 4.28.0.dev0Who can help?
No response
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
To reproduce, just run this on a M1/M2 Mac with Apple silicon
Expected behavior
No error. Will post stack trace.
The text was updated successfully, but these errors were encountered: