model transform to int8 problem #536
Unanswered
steven12356789
asked this question in
Q&A
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
HI, thanks for your work.
I use ResNeSt model to train my own dataset. Now I can transform the model to ONNX.
I faced this warning when I wanted to use tensorRT to speed up inference and use C++ to convert float32 to int8.
How to solve this problem?
onnx 1.7.0
onnxruntime 1.5.2
tensorrt 7.2.1.4
torch 1.7.0
WARN TRT: No implementation of layer (Unnamed Layer* 69) [Shuffle] + Transpose_52 obeys the requested constraints in strict mode. No conforming implementation was found i.e. requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_68 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_103 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer (Unnamed Layer* 158) [Shuffle] obeys the requested constraints in strict mode. No conforming implementa
WARN TRT: No implementation of layer Softmax_116 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer (Unnamed Layer* 160) [Shuffle] + Transpose_117 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_133 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_166 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer (Unnamed Layer* 247) [Shuffle] obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer Softmax_179 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer (Unnamed Layer* 249) [Shuffle] + Transpose_180 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_196 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
WARN TRT: No implementation of layer ReduceSum_229 obeys the requested constraints in strict mode. No conforming implementation was found i.e requested layer computation precision and output precision types are ignored, using the fastest implementation. trt_utils.cpp:253
Beta Was this translation helpful? Give feedback.
All reactions