-
Notifications
You must be signed in to change notification settings - Fork 5.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Integrate mistral.rs LLM #13105
Integrate mistral.rs LLM #13105
Conversation
Check out this pull request on See visual diffs & provide feedback on Jupyter Notebooks. Powered by ReviewNB |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Awesome to see rust-based LLM inference libraries!
Took a look, and it seems there is an excessive amount of lint/fmt type of changes in this PR. Not sure they're entirely needed as in that our own static checks should pass with out these.
Thank you! I've removed formatting the whole codebase and just formatted my new changes. |
thanks! |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Left a few comments in my first pass :)
llama-index-integrations/llms/llama-index-llms-mistral-rs/pyproject.toml
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/pyproject.toml
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/llama_index/llms/mistral_rs/base.py
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/llama_index/llms/mistral_rs/base.py
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/llama_index/llms/mistral_rs/base.py
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/llama_index/llms/mistral_rs/base.py
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/llama_index/llms/mistral_rs/base.py
Outdated
Show resolved
Hide resolved
@EricLBuehler do you mind adding me to your fork? Looks like we need to do some pants related stuff (i.e. run |
@nerdai, I addressed your comments and have added you to the repo. |
llama-index-integrations/llms/llama-index-llms-mistral-rs/README.md
Outdated
Show resolved
Hide resolved
llama-index-integrations/llms/llama-index-llms-mistral-rs/tests/test_llms_mistral-rs.py
Outdated
Show resolved
Hide resolved
@EricLBuehler looks like we're running into error still:
|
@nerdai, sorry for that mistake. It should be fixed now. |
All good -- thanks for the quick fix! |
It seems like the CI tests are failing because this integration depends on the |
yes, please ensure all the required deps are listed in the pyproject.toml. Best to run:
have you published |
@nerdai, I just released |
Ah okay. Can it work with just he standard rust installation? |
@nerdai, yes, it can. It depends on |
Sorry @EricLBuehler not sure if I'm following. To my knowledge, Rust Toolchain is installed in our github runners by default (source). Can we not just do a |
@nerdai, thanks for clarifying! I have added it as a dependency now. |
Hi @nerdai! I have updated this PR to use our latest PyPi release. Additionally, I made sure the tests pass by running the following commands:
I think that the CI tests should pass now. |
@EricLBuehler we're getting close! Looks like tests are still failing. From the traceback captured in the logs, i see this:
Maybe we need to do the following:
where |
@nerdai, that should be fixed now! Not sure why the tests I ran locally didn't catch that though. |
Happens to me too sometimes. I think there's a mismatch between versions perhaps on python or other testing/formatting dependencies. 🤔 |
@EricLBuehler time to 🛳️! Thanks for this :) |
@nerdai thank you! |
Description
In this PR, I have added support for the
mistral.rs
LLM inference platform via a new integration.mistral.rs
is a new LLM inference platform with key features such as prefix caching, optimized X-LoRA support, LoRA support via weight merging and grammar support.New Package?
Did I fill in the
tool.llamahub
section in thepyproject.toml
and provide a detailed README.md for my new integration or package?Version Bump?
Did I bump the version in the
pyproject.toml
file of the package I am updating? (Except for thellama-index-core
package)Type of Change
Please delete options that are not relevant.
How Has This Been Tested?
Please describe the tests that you ran to verify your changes. Provide instructions so we can reproduce. Please also list any relevant details for your test configuration
Suggested Checklist:
make format; make lint
to appease the lint gods