-
Notifications
You must be signed in to change notification settings - Fork 2.3k
Description
I have applied post-training quantization to my pretty complex neural network architecture. Now I am using a TRT engine inspector to retrieve information about each individual layer to see if it was quantized or not, as not all layer types and operations are supported for conversion to INT8 by TensorRT.
After collecting layer information and discarding some columns which I do not need I am left with something that I cannot really parse or understand honestly.
I know for a fact the very first layers of my architecture are the very first 3 layers of a ResNet34, so a Convolution, ReLU and MaxPool, followed by another custom convolutional bloc. Why do the first 4 layers in my quantized model show up as having LayerType Constant, Reformat, Constant and Reformat respectively? With names like onnx::Conv_1063, Identity_76, onnx::Conv_1062 and Identity_77?
I have been so far unable to find a guide on how to interpret these LayerTypes unfortunately.