-
Notifications
You must be signed in to change notification settings - Fork 67
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
请问是否支持Qwen1.5系列模型(不同量化方式 / 非量化) #2
Comments
这个应该看 vllm是支持,vllm 支持,这个仓库就支持 |
好的 您这边有尝试嘛 |
官方文档提到vllm 0.3.0以上是支持的 |
热插拔的lora加载如何实现,我看VLLM在加载lora时会有词表大小限制? |
ok的 |
我切换到qwen1.5之后,提示报错了,自己改了下,发现推理能力变差了,大神是否可以做一个适配的版本谢谢 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
No description provided.
The text was updated successfully, but these errors were encountered: