Skip to content
New issue

Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? # to your account

Onnx quantized backend for Clip-ViT-B-16 #3006

Open
PraNavKumAr01 opened this issue Oct 20, 2024 · 1 comment
Open

Onnx quantized backend for Clip-ViT-B-16 #3006

PraNavKumAr01 opened this issue Oct 20, 2024 · 1 comment

Comments

@PraNavKumAr01
Copy link

@tomaarsen
Just wanted to know if clip (text + image) embedding models will have an onnx quantized model? i tried finding it everywhere but had no luck. If it is there can you please point me to it? And if not, is it possible to create a model_qint8_avx512_vnni.onnx for it.
Can we expect it to be there with future updates or would i have to run some experiments and convert it on my own?

@tomaarsen
Copy link
Collaborator

Hello!

I'm afraid that CLIP models don't have ONNX support in Sentence Transformers right now. In short, CLIP models are loaded with the CLIPModel module, whereas the ONNX support is implemented in the Transformer module. This latter module is used for text-based embedding models.

  • Tom Aarsen

# for free to join this conversation on GitHub. Already have an account? # to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants