-
Notifications
You must be signed in to change notification settings - Fork 4.6k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Rate limited #2397
Comments
You will typically always see a few of those messages, but it should eventually succeed in computing the embeddings after it pauses. One way to avoid seeing as many messages is to increase the TPM capacity on your embedding model - we default to requesting 30K TPM but you may have more capacity available. Or did none of your embedding calls succeed at all? Please share full log if so. |
I'm getting [13:55:26] INFO Rate limited on the OpenAI embeddings API, sleeping before retrying... embeddings.py:63 The above exception was the direct cause of the following exception: Traceback (most recent call last): The above exception was the direct cause of the following exception: Traceback (most recent call last): Deploying services (azd deploy) (x) Failed: Deploying service backend ERROR: error executing step command 'deploy --all': failed deploying service 'backend': POST https://management.azure.com/subscriptions/c645d3fb-d531-43b4-b061-945f80bf073c/resourceGroups/rg-openai-demo-dev/providers/Microsoft.ContainerRegistry/registries/openaidemodevacrm67nlh27l373u/scheduleRunRESPONSE 400: 400 Bad Request
|
I'm getting "Rate limited on the OpenAI embeddings API, sleeping before retrying..." when trying to deploy to AZURE, is there any solution?
The text was updated successfully, but these errors were encountered: