Even with my model that is less than 250KB in size, I get the onnx_data file after quantization. https://github.com/onnx/neural-compressor/blob/aabbf967cf7ea91c078c28c7b4dab043add5257b/onnx_neural_compressor/onnx_model.py#L245