-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Model size check doesn't seem to work #39
Comments
Hi, is your original model format .onnx file + weight file?
If the original format is .onnx file + weight file, saving operation will follow it. |
My model is just the .onnx file, no external weight file. |
Do you quantize the model with this API
If so, during optimization it will save external data if model size > min_size:
|
Yes, I do quantize using that API. Why is the size limit so low? |
Since onnxruntime can load external data automatically and we find model with external data format can be processed faster. But according to you case, the size limitation may need optimization. |
Even with my model that is less than 250KB in size, I get the onnx_data file after quantization.
neural-compressor/onnx_neural_compressor/onnx_model.py
Line 245 in aabbf96
The text was updated successfully, but these errors were encountered: