Does vllm support deploy models on vuda? #9691
Unanswered
Gaohang0804
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I'm trying to deploy internVL2 model with vllm on k8s,I have succeed on 1 A800 80G GPU,but because the low memory required of my model,many gpu memory is wasted.
I just wondering could deploy my model on virtualized cuda,and required lower memory on one k8s pod.
Beta Was this translation helpful? Give feedback.
All reactions