-
-
Notifications
You must be signed in to change notification settings - Fork 6.2k
New issue
Have a question about this project? # for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “#”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? # to your account
[Usage]: Dose vLLM support embedding api of multimodal llm? #8483
Comments
No, this is not supported yet. |
In fact, this isn't even available for most language-only models. The only one supported right now is Mistral. See also #7915 |
I am working on it. #8453 #8452 @DarkLight1337 |
According to my understanding, MiniCPM-V 2.6 is a generative model, not a retrieval model specifically used to generate embeddings. (Maybe you need multimodal retrieval models such as BAAI/bge-visualized-m3 https://huggingface.co/BAAI/bge-visualized) Can you send some sample code and tell me how you want to use MiniCPM-V 2.6 to generate embedding |
@noooop sample code
|
|
|
Simple but inefficient method:Output last hidden states #853, A hacker’s method is mentioned below. https://github.com/WuNein/vllm4mteb/tree/main More efficient implementationImplement a model yourself. adding_model You must think of a cool name to avoid routing to the previous model. |
You can now modify any existing model to support embeddings, please see #9314 (comment). |
does vllm support BAAI/bge-visualized now? |
It doesn't look like the HF repo is compatible with |
Your current environment
How would you like to use vllm
eg: get embedding of minicpmv 2.6
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: