-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
GroupNormalization plugin failure of TensorRT 10.0.1.6 when running trtexec on GPU A4000 #3950
Comments
Can you upload full log with |
Make sure libcudnn.so load successed. Add path to |
The problem is that the NVIDIA container contains cudnn 9.1.0, but the plugin is trying to load libcudnn.so.8. There is a version mismatch, not that cudnn is not available. |
You should make sure your env has one cudnn, and why your nvinfer plugin will load cudnn.8.0 ? |
This is not my plugin - this is the plugin provided in this repo - https://github.com/NVIDIA/TensorRT/tree/release/10.1/plugin/groupNormalizationPlugin And it loads cudnn 8, not 9 because it has the wrong macro defined here: https://github.com/NVIDIA/TensorRT/blob/release/10.1/plugin/common/cudnnWrapper.cpp#L26 |
From https://github.com/NVIDIA/TensorRT/tree/release/10.0, trt version 10.0.1.6,
map to https://github.com/NVIDIA/TensorRT/blob/release/10.1/plugin/common/cudnnWrapper.cpp#L26-L42 You can try to creat a soft link |
Why does the container include cudnn 9 then? If TensorRT doesn't work in an NVIDIA container with cudnn 9, why does it ship with it? |
@appearancefnp , we now use native could you replace the |
@appearancefnp , I will be closing this ticket due to our policy to close tickets with no activity for more than 21 days after a reply had been posted. Please reopen a new ticket if you still need help. |
I was able to run the native |
Description
Hey guys!
I wanted to upgrade from TensorRT 8.6 to 10.0. I have a ONNX model that contains GroupNormalization plugin. It creates a serialized version, but it fails when deserializing the model while trying to load cudnn 8 instead of cudnn 9.
Environment
Using docker: nvcr.io/nvidia/tensorrt:24.05-py3
TensorRT Version: 10.0.1
NVIDIA GPU: A4000
NVIDIA Driver Version: 550.67
CUDA Version: 12.4
CUDNN Version: 9.1 (per container documentation)
Operating System:
Python Version (if applicable): -
Tensorflow Version (if applicable): -
PyTorch Version (if applicable): -
Baremetal or Container (if so, version): nvcr.io/nvidia/tensorrt:24.05-py3
Relevant Files
Model link: https://drive.google.com/file/d/1vmGZpWJ_1sfz2ejbZoO3fFaR5udxOLTi/view?usp=sharing
Steps To Reproduce
Commands or scripts:
trtexec --onnx=model.onnx
Have you tried the latest release?: yes
The text was updated successfully, but these errors were encountered: