-
-
Notifications
You must be signed in to change notification settings - Fork 324
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Command showing available options for installed models #82
Comments
I'll do this: llm models list --options And introspect the |
I got this working:
I don't like it outputting the same help multiple times though. I'm going to have it only output the detailed descriptions once per model of each class. |
Tests are failing now - |
Could be related to plugin order. It shouldn't though, I would expect this order to be the same each time with only the Lines 49 to 56 in 18f34b5
|
Could be the order of this bit: Line 381 in 18f34b5
|
Actually the problem was something else. I copied the generated text to my local environment and did a diff --git a/docs/usage.md b/docs/usage.md
index 8e76010..dfc4c16 100644
--- a/docs/usage.md
+++ b/docs/usage.md
@@ -98,53 +98,53 @@ cog.out("```\n{}\n```".format(result.output))
OpenAI Chat: gpt-3.5-turbo (aliases: 3.5, chatgpt)
- temperature: float
+ temperature: Union[float, NoneType]
What sampling temperature to use, between 0 and 2. Higher values like
0.8 will make the output more random, while lower values like 0.2 will
make it more focused and deterministic.
- max_tokens: int
+ max_tokens: Union[int, NoneType]
Maximum number of tokens to generate
- top_p: float
+ top_p: Union[float, NoneType]
An alternative to sampling with temperature, called nucleus sampling,
where the model considers the results of the tokens with top_p
probability mass. So 0.1 means only the tokens comprising the top 10%
probability mass are considered. Recommended to use top_p or
temperature but not both.
- frequency_penalty: float
+ frequency_penalty: Union[float, NoneType]
Number between -2.0 and 2.0. Positive values penalize new tokens based
on their existing frequency in the text so far, decreasing the model's
likelihood to repeat the same line verbatim.
- presence_penalty: float
+ presence_penalty: Union[float, NoneType]
Number between -2.0 and 2.0. Positive values penalize new tokens based
on whether they appear in the text so far, increasing the model's
likelihood to talk about new topics.
- stop: str
+ stop: Union[str, NoneType]
A string where the API will stop generating further tokens.
logit_bias: Union[dict, str, NoneType]
Modify the likelihood of specified tokens appearing in the completion.
OpenAI Chat: gpt-3.5-turbo-16k (aliases: chatgpt-16k, 3.5-16k)
- temperature: float
- max_tokens: int
- top_p: float
- frequency_penalty: float
- presence_penalty: float
- stop: str
+ temperature: Union[float, NoneType]
+ max_tokens: Union[int, NoneType]
+ top_p: Union[float, NoneType]
+ frequency_penalty: Union[float, NoneType]
+ presence_penalty: Union[float, NoneType]
+ stop: Union[str, NoneType]
logit_bias: Union[dict, str, NoneType]
OpenAI Chat: gpt-4 (aliases: 4, gpt4)
- temperature: float
- max_tokens: int
- top_p: float
- frequency_penalty: float
- presence_penalty: float
- stop: str
+ temperature: Union[float, NoneType]
+ max_tokens: Union[int, NoneType]
+ top_p: Union[float, NoneType]
+ frequency_penalty: Union[float, NoneType]
+ presence_penalty: Union[float, NoneType]
+ stop: Union[str, NoneType]
logit_bias: Union[dict, str, NoneType]
OpenAI Chat: gpt-4-32k (aliases: 4-32k)
- temperature: float
- max_tokens: int
- top_p: float
- frequency_penalty: float
- presence_penalty: float
- stop: str
+ temperature: Union[float, NoneType]
+ max_tokens: Union[int, NoneType]
+ top_p: Union[float, NoneType]
+ frequency_penalty: Union[float, NoneType]
+ presence_penalty: Union[float, NoneType]
+ stop: Union[str, NoneType]
logit_bias: Union[dict, str, NoneType] So clearly on Python 3.8 this bit of code has a different output: Lines 382 to 384 in 18f34b5
|
To test locally I ran: pyenv install 3.8.17 Then waited for that to compile. Then: ~/.pyenv/versions/3.8.17/bin/python -m venv /tmp/pvenv
source /tmp/pvenv/bin/activate
pip install -e '.[test]'
/tmp/pvenv/bin/cog --check docs/usage.md And to rewrite it:
|
Extracted a TIL: https://til.simonwillison.net/python/quick-testing-pyenv |
Now documented here: https://llm.datasette.io/en/latest/usage.html#listing-available-models - including a |
This might be part of
llm models list
or may be something else.Follows:
register_models
#53The text was updated successfully, but these errors were encountered: