Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Issue Using Qwen Model with Lotus (vLLM Deployment) #94

Open
KH-Khadija opened this issue Jan 30, 2025 · 0 comments
Open

Issue Using Qwen Model with Lotus (vLLM Deployment) #94

KH-Khadija opened this issue Jan 30, 2025 · 0 comments
Labels
bug Something isn't working

Comments

@KH-Khadija
Copy link

I am trying to use the Qwen2.5-14B-Instruct-GPTQ-Int8 model with Lotus and have it deployed via vLLM on a GPU. Based on the Lotus documentation, models deployed with vLLM should be usable like this:

from lotus.models import LM
lm = LM(
    model='hosted_vllm/Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8',
    api_base='http://localhost:8010/v1',
)

However, when I try this, I get the following error:

Exception: This model isn't mapped yet. model=hosted_vllm/Qwen/Qwen2.5-14B-Instruct-GPTQ-Int8, custom_llm_provider=hosted_vllm. Add it here - https://github.com/BerriAI/litellm/blob/main/model_prices_and_context_window.json.

From my understanding, Lotus only supports models available in LiteLLM, so does this mean LiteLLM does not yet support Qwen models?

If that is the case and I am constrained to using Qwen, what are my options to work around this? Would I need to manually extend LiteLLM's supported models, or is there another way to integrate Qwen with Lotus?

@KH-Khadija KH-Khadija added the bug Something isn't working label Jan 30, 2025
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

1 participant