Multi-resource rate limiting for LLM APIs. Reserve tokens before you call, refund what you don't use, stay under the limit across workers.
ai rate-limiting tokens openai rate-limit rate-limiter throttler ai-agents throttle-requests rate-limit-redis ai-engineering openai-api llm llms llm-token
-
Updated
Mar 7, 2026 - Python