-
-
Notifications
You must be signed in to change notification settings - Fork 4.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Usage]: Dose vLLM support embedding api of multimodal llm? #8483
Comments
No, this is not supported yet. |
In fact, this isn't even available for most language-only models. The only one supported right now is Mistral. See also #7915 |
I am working on it. #8453 #8452 @DarkLight1337 |
According to my understanding, MiniCPM-V 2.6 is a generative model, not a retrieval model specifically used to generate embeddings. (Maybe you need multimodal retrieval models such as BAAI/bge-visualized-m3 https://huggingface.co/BAAI/bge-visualized) Can you send some sample code and tell me how you want to use MiniCPM-V 2.6 to generate embedding |
@noooop sample code
|
|
|
Simple but inefficient method:Output last hidden states #853, A hacker’s method is mentioned below. https://github.com/WuNein/vllm4mteb/tree/main More efficient implementationImplement a model yourself. adding_model You must think of a cool name to avoid routing to the previous model. |
You can now modify any existing model to support embeddings, please see #9314 (comment). |
Your current environment
How would you like to use vllm
eg: get embedding of minicpmv 2.6
Before submitting a new issue...
The text was updated successfully, but these errors were encountered: