Hi! I’m using the api docs tutorial listed here to save a quantized model:
Especifically, the command used to export the quantized model using is:
sim.export(path=’./’, filename_prefix=‘quantized_mnist’, input_shape=input_shape)
This command export three files:
My question is: How to load again the quantized model? Is it possible to load the quantized model into a machine without aimet tool installed?