-
-
Notifications
You must be signed in to change notification settings - Fork 290
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Using Venv or Dockerisation #80
Comments
I would advise you to use miniconda3 on your local environment. After installation, create a python 3.10 or 3.11 env: If you are on NVIDIA and not AMD, it totally suffices to just run the turboderp/exllama#192 (comment) If you still run into issues please let me know. I hope this helps |
Also, are you running on WSL? Sometimes you need to link libraries manually there (although the above setup works fine for me) |
Hey thanks for the comment. I'm on a Ubuntu server 22. I did exactly what you suggested: I tried miniconda, and venv, but it failed to compile. I suspect that the cuda toolkit is most definitely needed for compilation, because after I installed the latest cuda using apt(i followed nvidia docs) it worked. Probably if I get the precompiled package, it runs with my environments. |
You could also try to install the prebuilt wheels published, look in the README for more information |
When I got the same error message, doing this on ubuntu helped. Here |
Hi, terrific work as always. I was wondering if we could use venvs (or maybe even Docker) for running exllamav2.
A lot of tools in the generative AI space have some kind of option for self containment(eg: oobabooga, ComfyUI).
I had a lot of troubles when I cleaned up my system from nvidia bloat. I tried running it without having nvcc installed system wide, but even when using a venv that contains all the required dependencies leads to missing files during compile.
nvcc is available inside my env.
The error:
The text was updated successfully, but these errors were encountered: