We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Hi, we are using batch inference with gemini-1.5-pro, while we found:
when we submit > 1 tasks, at most 1 task can succeed, the other will failed (is this because the quota?)
when we run batch inference job (even though we only submit 1 task), it sometimes occurs the following issue & the whole task will fail:
RESOURCE_EXHAUSTED error occurred: {"error": {"code": 429, "message": "Online prediction request quota exceeded for gemini-1.5-pro. Please try again later with backoff.", "status": "RESOURCE_EXHAUSTED"}}
Can we have any methods that:
if the quota exceed, the job can automatically pending & retry
if some of the samples fails to be inferred, keep the results of other samples, do not let the complete job failed.
The text was updated successfully, but these errors were encountered:
weichungw
No branches or pull requests
Hi, we are using batch inference with gemini-1.5-pro, while we found:
when we submit > 1 tasks, at most 1 task can succeed, the other will failed (is this because the quota?)
when we run batch inference job (even though we only submit 1 task), it sometimes occurs the following issue & the whole task will fail:
RESOURCE_EXHAUSTED error occurred: {"error": {"code": 429, "message": "Online prediction request quota exceeded for gemini-1.5-pro. Please try again later with backoff.", "status": "RESOURCE_EXHAUSTED"}}
Can we have any methods that:
if the quota exceed, the job can automatically pending & retry
if some of the samples fails to be inferred, keep the results of other samples, do not let the complete job failed.
The text was updated successfully, but these errors were encountered: