-
Notifications
You must be signed in to change notification settings - Fork 486
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
onnxruntime-gpu but GPU dependencies are not loaded #1402
Comments
Thank you @SuperSecureHuman, I can reproduce the issue. It appears |
I am not very familiar with onnx internals. But from my understanding, if something is there in the available providers, it should work right? Why is the check needed? |
The check is "needed" because ORT falls back to For the record, here is how Fixed in #1403, will publish a patch today! |
Got it! Thanks! |
System Info
Optimum 1.13.1 Python 3.9 pip list | grep onnx onnx 1.14.1 onnxruntime-gpu 1.16.0
Who can help?
@JingyaHuang @ech
Reproduction (minimal, reproducible, runnable)
Trying to use ORT with CUDA raises error -
onnxruntime-gpu
is installed, but GPU dependencies are not loaded. It is likely there is a conflicting install betweenonnxruntime
andonnxruntime-gpu
. Please install onlyonnxruntime-gpu
in order to use CUDAExecutionProvider."While
Clearly, only ort-gpu is installed
Same issue can be reproduced in colab
%pip install optimum[onnxruntime-gpu]
ORTModelForSeq2SeqLM.from_pretrained(.... providers=["CUDAExecutionProvider"], use_io_binding=True)
Expected behavior
No error, expecting to run ort pipeline on GPU.
Workaround
Currently I am commenting out
optimum/optimum/onnxruntime/utils.py
Line 214 in 7fc27f6
Now the pipeline runs on GPU without any issues
The text was updated successfully, but these errors were encountered: