|
2 | 2 |
|
3 | 3 | # Supported ONNX Operators |
4 | 4 |
|
5 | | -TensorRT 10.7 supports operators in the inclusive range of opset 9 to opset 22. Latest information of ONNX operators can be found [here](https://github.com/onnx/onnx/blob/main/docs/Operators.md). More details and limitations are documented in the chart below. |
| 5 | +TensorRT 10.8 supports operators in the inclusive range of opset 9 to opset 22. Latest information of ONNX operators can be found [here](https://github.com/onnx/onnx/blob/main/docs/Operators.md). More details and limitations are documented in the chart below. |
6 | 6 |
|
7 | | -TensorRT supports the following ONNX data types: DOUBLE, FLOAT32, FLOAT16, BFLOAT16, INT32, INT64, FP8, INT8, INT4, UINT8, and BOOL |
| 7 | +TensorRT supports the following ONNX data types: DOUBLE, FLOAT32, FLOAT16, BFLOAT16, FP8, FP4, INT32, INT64, INT8, INT4, UINT8, and BOOL |
8 | 8 |
|
9 | 9 | > Note: There is limited support for DOUBLE type. TensorRT will attempt to cast DOUBLE down to FLOAT, clamping values to `+-FLT_MAX` if necessary. |
10 | 10 |
|
11 | | -> Note: INT8, INT4, and FP8 are treated as `Quantized Types` in TensorRT, where support is available only through quantization from a floating-point type with higher precision. See [section 7.4.2](https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#qat-models-work) of the developer guide for more information. |
| 11 | +> Note: INT8, INT4, FP8 and FP4 are treated as `Quantized Types` in TensorRT, where support is available only through quantization from a floating-point type with higher precision. See [section 7.4.2](https://docs.nvidia.com/deeplearning/tensorrt/developer-guide/index.html#qat-models-work) of the developer guide for more information. |
12 | 12 |
|
13 | 13 | > Note: UINT8 is only supported as network input or output tensor types. |
14 | 14 |
|
@@ -47,22 +47,22 @@ TensorRT supports the following ONNX data types: DOUBLE, FLOAT32, FLOAT16, BFLOA |
47 | 47 | | Compress | N | |
48 | 48 | | Concat | Y | FP32, FP16, BF16, INT32, INT64, BOOL | |
49 | 49 | | ConcatFromSequence | N | |
50 | | -| Constant | Y | FP32, FP16, BF16, INT32, INT64, BOOL | `sparse_value`, `value_string`, and `value_strings` attributes are unsupported. |
51 | | -| ConstantOfShape | Y | FP32, FP16, BF16, INT32, INF64, BOOL | |
| 50 | +| Constant | Y | FP32, FP16, BF16, FP8, FP4, INT4, INT32, INT64, BOOL | `sparse_value`, `value_string`, and `value_strings` attributes are unsupported. |
| 51 | +| ConstantOfShape | Y | FP32, FP16, BF16, FP8, FP4, INT4, INT32, INF64, BOOL | |
52 | 52 | | Conv | Y | FP32, FP16, BF16 | |
53 | 53 | | ConvInteger | N | |
54 | 54 | | ConvTranspose | Y | FP32, FP16, BF16 | |
55 | 55 | | Cos | Y | FP32, FP16, BF16 | |
56 | 56 | | Cosh | Y | FP32, FP16, BF16 | |
57 | | -| CumSum | Y | FP32, FP16, BF16 | `axis` must be an initializer | |
| 57 | +| CumSum | Y | FP32, FP16, BF16 | `axis` must be a build-time constant | |
58 | 58 | | DFT | N | |
59 | 59 | | DeformConv | Y | FP32, FP16 | `input` must have 1D or 2D spatial dimensions. `pads` for the beginning and end along each spatial axis must be the same |
60 | 60 | | DepthToSpace | Y | FP32, FP16, BF16, INT32, INT64 | |
61 | | -| DequantizeLinear | Y | INT8, FP8, INT4 | `x_zero_point` must be zero | |
| 61 | +| DequantizeLinear | Y | INT8, FP8, FP4, INT4 | `x_zero_point` must be zero | |
62 | 62 | | Det | N | |
63 | 63 | | Div | Y | FP32, FP16, BF16, INT32, INT64 | |
64 | 64 | | Dropout | Y | FP32, FP16, BF16 | `is_traning` must be an initializer and evaluate to False. |
65 | | -| DynamicQuantizeLinear | N | |
| 65 | +| DynamicQuantizeLinear | N | Not supported. TensorRT's IDynamicQuantize can be composed from ONNX operators in the form of a model local function. |
66 | 66 | | Einsum | Y | FP32, FP16, BF16 | |
67 | 67 | | Elu | Y | FP32, FP16, BF16 | |
68 | 68 | | Equal | Y | FP32, FP16, BF16, INT32, INT64 | |
|
0 commit comments