Following the sample code from https://colab.research.google.com/drive/1AmcITfN2ELQe07WKQ9szaxq-WSu4hdQb#scrollTo=nQabIE9Guyrh
if in the QuantizedDeployable part, the model is not loaded again (i.e., state_dict = torch.load('checkpoint/mnist_fq_mixed.pth')['state_dict']), the accuracy is just 17.87% in the deployable version, which should work.