Hi! I’m using the api docs tutorial listed here to save a quantized model:
https://quic.github.io/aimet-pages/AimetDocs/api_docs/torch_quantsim.html#api-torch-quantsim
Especifically, the command used to export the quantized model using is:
sim.export(path=‘./’, filename_prefix=‘quantized_mnist’, input_shape=input_shape)
This command export three files: quantized_mnist.pth
, quantized_mnist.onnx
and quantized_mnist.encodings
My question is: How to load again the quantized model? Is it possible to load the quantized model into a machine without aimet tool installed?
Thanks!