This repository was archived by the owner on Jul 11, 2023. It is now read-only.
Adds token and request-based rate limiting with an example#29
Open
Adds token and request-based rate limiting with an example#29
Conversation
|
The latest updates on your projects. Learn more about Vercel for Git βοΈ
|
Contributor
|
I think this is useful but limited. The issue is that in a serverless environment (like nextjs api functions) every api call is made in a different lambda function that has no knowledge of the other lambdas so this will rate limit individual lambda functions but we can still have like 100 lambda functions doing api calls at the same time and getting rate limited. To account for that the only solution is some kind of retry with exponential back off. |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
The rate limits on OpenAI seem a little wonky. Not sure if that's only true for Codex or for all the models but I was definitely getting rate limited even when using less than half the requests per minute.
But it's better than nothing (maybe?)