Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Issue when sending parallel requests #3361

Open
lschaupp opened this issue Nov 8, 2024 · 0 comments
Open

Issue when sending parallel requests #3361

lschaupp opened this issue Nov 8, 2024 · 0 comments

Comments

@lschaupp
Copy link

lschaupp commented Nov 8, 2024

Hello,

I am getting the following error when I send multiple requests in parallel to the inference endpoint:

ERROR: 503
{
"code": 503,
"type": "ServiceUnavailableException",
"message": "Model "restorer" has no worker to serve inference request. Please use scale workers API to add workers. If this is a sequence inference, please check if it is closed, or expired; or exceeds maxSequenceJobQueueSize"
}

I have two separate processes that can access the inference API.
Any ideas?

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant