-
-
Notifications
You must be signed in to change notification settings - Fork 5.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Bug]: Qwen2-VL AssertionError: assert "factor" in rope_scaling. #8281
Comments
Qwen2-VL is not supported yet. Please note that Qwen-VL in the list of supported models refers to version 1. |
Thanks a lot, and looking forward to the support! |
Can we support Qwen2-VL now? |
Please check the status of PR #7905 |
目前官网上看是支持qwen2-vl模型但是为什么还是会有相同的报错 |
It is caused by a bug in transformers v4.45. For now, you'll have to either downgrade to a lower version of vLLM that doesn't use transformers v4.45, or build vLLM from source to use our patched Qwen2VL config which doesn't have this problem. |
我采用以下方式进行: |
Can you run |
You should install vLLM using |
correction: that is python-only dev, and is correct. see https://docs.vllm.ai/en/latest/getting_started/installation.html#python-only-build-without-compilation |
Just for those who don't wanna upgrade vllm and still encounter this problem, the following code works for me: model = LLM(
model=model_path,
rope_scaling={
"mrope_section": [
16,
24,
24
],
"rope_type": "mrope",
"type": "mrope"
}
) and I'm using Qwen2VL 7B |
python -m vllm.entrypoints.openai.api_server How to add these parameters if started in this way? |
You can pass |
Traceback (most recent call last): |
Please show the command you used. |
python -m vllm.entrypoints.openai.api_server --served-model-name vlmodel --model /model --dtype=half --gpu-memory-utilization 0.9 --max_model_len 1 --rope-scaling '{"mrope_section": [16,24, 24], "rope_type": "mrope", "type": "mrope"}' |
Can you manually update the |
Otherwise it would be best to just upgrade vLLM. |
hello, I met the same problem and for some reason. I have to use 0.6.1(or is there any other version support for cuda 11.8?). I want to know what do you mean update config.json, where should I change the code? |
I mean that you go to the location where the model is downloaded on your machine and edit the |
yeah, I change the code. You can see I add the { |
I think you just have to update vLLM then. I have no time to debug this in an older version. |
Would I like to ask how can i install the higher version vllm with cuda 11.8 ? I try to install vllm-0.6.4 by pip install -e. and it will install torch 2.5 with cuda 12 for me automatically. (By the way , python=3.10) |
You can try installing torch manually: https://pytorch.org/get-started/previous-versions/ and follow this section to keep your installed version. @youkaichao might have more context on this. |
OK, I will try out . Thank you very much ! You are so kind and friendly! |
Your current environment
The output of `python collect_env.py`
🐛 Describe the bug
Obtaining a bug as follows:
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: