-
Notifications
You must be signed in to change notification settings - Fork 2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Inconsistent results between TensorRT and ONNX #3850
Labels
triaged
Issue has been triaged by maintainers
Comments
Sukeysun
changed the title
XXX failure of TensorRT X.Y when running XXX on GPU XXX
Inconsistent results between TensorRT and ONNX
May 9, 2024
max=0.0016785 polygraphy run pose_detection.onnx --onnxrt \
--save-inputs inputs.json \
--onnx-outputs mark all --save-outputs layerwise_golden.json
polygraphy run pose_detection.onnx --trt \
--validate --trt-outputs mark all --save-outputs trt_out.json compare each layer |
Please also try the latest TRT release. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Description
When attempting to convert the ONNX model 'pose_detect' of mediapipe to TensorRT, I observed a significant loss in model accuracy. The model fails to maintain its original precision after the conversion process.
Environment
TensorRT Version: 8.6.1
CUDA Version: 11.8
Relevant Files
Model link:
https://storage.googleapis.com/ailia-models/blazepose-fullbody/pose_detection.onnx
Steps To Reproduce
The full traceback of errors encountered is as follows: [Error details here]
The text was updated successfully, but these errors were encountered: