-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: [support return multiple choices] #264
Comments
according to this comment, llamacpp hasn't supported it yet. this issue need to be transferred to handle at the cortex.cpp layer |
This was referenced Oct 30, 2024
Now we can get multiple choices from 1 request by adding
|
✅ QA API - thank you @nguyenhoangthuan99! |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Problem
n
integer
or nullHow many chat completion choices to generate for each input message. Note that you will be charged based on the number of generated tokens across all of the choices. Keep n as 1 to minimize costs.
-> need to check if llama.cpp support this option.
reference: https://platform.openai.com/docs/api-reference/chat/create#chat-create-n
related issue: https://github.com/janhq/internal/issues/160
The text was updated successfully, but these errors were encountered: