-
Notifications
You must be signed in to change notification settings - Fork 1.4k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: pull model list for openai-compatible endpoints #630
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Could you also call the same function for the case where vllm
is the endpoint type? e.g. with our endpoint you can do http://api.memgpt.ai/v1/models
(since we require the OpenAI compatible vLLM endpoint).
Also might be nice to add Azure too, but not that important.
Added vLLM support + tested against our vLLM API (see picture in OP) |
…ai api key again in case you made a mitake / want to swap it out
@sarahwooders added a new method to resolve the "too many models" problem: Start by showing a curated list: Allow the user to expand the list (if they expand the list, they get a warning): Finally, they can still enter a manual model if the internet / GET request is broken and they know what they're doing: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
lgtm!
* allow entering custom model name when using openai/azure * pull models from endpoint * added/tested vllm and azure * no print * make red * make the endpoint question give you an opportunity to enter your openai api key again in case you made a mitake / want to swap it out * add cascading workflow for openai+azure model listings * patched bug w/ azure listing
* allow entering custom model name when using openai/azure * pull models from endpoint * added/tested vllm and azure * no print * make red * make the endpoint question give you an opportunity to enter your openai api key again in case you made a mitake / want to swap it out * add cascading workflow for openai+azure model listings * patched bug w/ azure listing
* allow entering custom model name when using openai/azure * pull models from endpoint * added/tested vllm and azure * no print * make red * make the endpoint question give you an opportunity to enter your openai api key again in case you made a mitake / want to swap it out * add cascading workflow for openai+azure model listings * patched bug w/ azure listing
Closes #592
Please describe the purpose of this pull request.
memgpt configure
if using an openai endpoint, pull the list of modelsHow to test
Run
memgpt configure
OpenAI
Azure
vLLM
Have you tested this PR?
Other comments
@sarahwooders
gpt-
gpt-4-vision-preview
shows up, but we don't support vision inputs at the moment)