-
Notifications
You must be signed in to change notification settings - Fork 249
Error: could not create backend -> jinaai/jina-reranker-v1-turbo-en #579
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
Comments
The error persists under the following scenarios involving model artifacts stored in S3 that were:
For each of these cases, AutoMap from config.json was altered such that it points to (1) artifacts from jinaai or (2) artifacts stored in the same model archive (after adding them: e.g. modeling_bert.py, configuration_bert.py) The above scenarios incorporate the modifications listed here: https://huggingface.co/jinaai/jina-reranker-v1-turbo-en/discussions/13 |
I am not familiar with Rust, but I looked a bit into
It seems the conversion from Classifier to Reranker needs to happen in get_backend_model_type() before the backend is initialized, rather than after. The current code structure means that even though we correctly determine it should be a reranker in the router code, we're still passing it to the backend as a classifier. I hope this is useful and provides a starting point towards a solution |
Facing the same issue when attempting to deploy jina reranker models on GPUs devices |
Hello, |
Thank you all for reporting, I'll have a look into it in the coming days hoping to push a patch soon! 🤗 |
Thank you @alvarobartt for looking into this. Btw, are there any TEI environment variables for tuning the performance of the containers (e.g. batch size). |
Hey @CoolFish88, I just got back, and was about to look into this; before going on, just to confirm that the fix in https://huggingface.co/jinaai/jina-reranker-v1-turbo-en/discussions/13 seems to be working fine with ONNX (on CPU), which is the issue I intentionally fixed; I'll investigate the issue further on both MPS and CUDA devices as we may need to refactor the current |
@CoolFish88 I've created a draft for it at #582, in case you want to give it an early look! |
@alvarobartt, |
@alvarobartt Awaiting for the reviewers to approve the merge |
System Info
Hello,
When deploying jinaai/jina-reranker-v1-turbo-en to a Sagemaker endpoint using model artifacts stored in S3, the following error was raised:
The issue is different from the one reported in #556 for
jinaai/jina-embeddings-v2-small-en
in terms of the backend error having a different source. If "_name_or_path": "jinaai/jina-bert-implementation" is missing from config.json, then the error #556 will emerge instead.Cloudwatch logs:
Information
Tasks
Reproduction
Model artifacts
Config.json
serde
deserializer for JinaBERT models #559 (without this a Bert engine is erroneously initialized)TEI version:
Expected behavior
Model deployed successfully with TEI
The text was updated successfully, but these errors were encountered: