-
Notifications
You must be signed in to change notification settings - Fork 2.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
UINT8-to-FLOAT cast after transpose breaks the graph. #3985
Comments
In my env it runs normal.
|
trt version 8510 |
It also fails on my 8.6.1 env. |
You can upload the two onnx here. |
|
Also run passed. |
Can you try the latest trt ? |
Same thing happens with TRT 10.2.0 and 10.3.0 |
also has the same problem: tensorrt 10.1.0 |
This is expected behavior. All UINT8 inputs must directly feed into a cast layer in TensorRT. No other layers support UINT8 inputs, even just data-movement ops like reshape / transpose. For current users of TensorRT, the recommendation is to change your network structure for UINT8 inputs to feed directly into a cast. Marking this a feature request. |
Description
Casting an UINT8 to FLOAT16/32 after a
transpose
operation breaks the graph (network.num_outputs = 0
).Casting before the transpose works fine.
Environment
TensorRT Version: 10.1.0
NVIDIA GPU: A100-SXM4-80GB
NVIDIA Driver Version: 550.90.07
CUDA Version: 12.5
CUDNN Version: 9.1.0
Operating System: Ubuntu 22.04.4 LTS
Python Version: Python 3.10.12
PyTorch Version: 2.3.1+cu121
ONNX version: 1.16.1
Container: nvcr.io/nvidia/tensorrt:24.06-py3
Steps To Reproduce
Minimal example
Output
Both should have one output.
Logs:
TransposeCast
CastTranspose
The text was updated successfully, but these errors were encountered: