-
Notifications
You must be signed in to change notification settings - Fork 494
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Handling OpenAI 429's gracefully #4153
Labels
Comments
vga91
added a commit
that referenced
this issue
Dec 10, 2024
vga91
added a commit
that referenced
this issue
Dec 10, 2024
RobertoSannino
pushed a commit
that referenced
this issue
Dec 11, 2024
vga91
added a commit
that referenced
this issue
Dec 11, 2024
github-project-automation
bot
moved this from In Progress
to Done (check if cherry-pick)
in APOC Extended Larus
Dec 11, 2024
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Expected Behavior (Mandatory)
Ability to control OpenAI backoff strategy for large volume of embeddings calls. This is standard practice in almost any library I've used because we cannot assume we have infinite capacity from our API providers.
Actual Behavior (Mandatory)
How to Reproduce the Problem
Try embedding 5M nodes at 2000 nodes batched per API request (to maximise throughput) so you end up hitting the 429 for too many tokens per minute
Specifications (Mandatory)
Currently used versions
Versions
The text was updated successfully, but these errors were encountered: