-
Notifications
You must be signed in to change notification settings - Fork 1.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Segmentation Fault error when crafting pb_utils.Tensor object Triton BLS model #7953
Comments
After debugging with GBD, I've obtained a backtrace that shows the following logs indicating why the crash occurs:
From my understanding the segmentation fault occurs post-inference. I also noticed via ps aux, that a triton backend stub process is created when starting the tritonserver then another one is created when the server is being queried. Does this extra context assist with debugging the issue? |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
We are using Triton to serve a BLS model. Inside the model.py file for this BLS model, there is a function that uses the triton gRPC client to query another model hosted on the server. While this process works correctly, the issue arises in the execute function. When the final output tensor is extracted, I attempt to cast it as a pb_utils.Tensor object and append it to the InferenceResponse class as documented. However, during the pb_utils.Tensor casting, a Segmentation fault error occurs.
My triton inference server docker image is 24.07-py3
Cuda is 12.5
Error Stack
I've confirmed that the dtype, tensor shape that the config.pbtxt for the BLS model is in alignment with what is being sent in the execute function.
Here is the execute function that is throwing the error in model.py:
My model repository is as shown:
I reference my custom execution environment in config.pbtxt (the tarball), and I have a custom triton_python_backend_stub. If possible could you assist me in finding where the source of the error is coming from.
The text was updated successfully, but these errors were encountered: