-
Notifications
You must be signed in to change notification settings - Fork 3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
onnx.onnx_cpp2py_export.checker.ValidationError when call quantize_static() in onnxruntime==1.20.1 #23268
Comments
The model is referring to a weight that is expected to be stored in a file, but the file is not found. Whatever was used to create the model, externalized the weight and placed it in a file that is usually expected to be found next to the model. |
@yuslepukhin I think you are right. But somehow the algorithm cannot properly load the model with external data and give errors in this simple cases, maybe the root cause is related to the save and reload the model part? |
Please look at the test/repro code attached by @dzk9528. The ModelProto that is passed to The weights are externalized by the function
If you look at what is happening within the ONNXRuntime quantization tool:
So yes, the weights are externalized and then deleted. But this not happen in the model creation in user code. The broken external weight path is introduced inside of |
Cc: @xiaoyu-work |
Describe the issue
When I try to quantize model with larger weight size in onnxruntime 1.20.1, following error appeared:
To reproduce
Urgency
This is a urgent request and it is very close to our overall model quantization software product development.
Platform
Linux
OS Version
Ubuntu 22.04
ONNX Runtime Installation
Released Package
ONNX Runtime Version or Commit ID
1.20.1
ONNX Runtime API
Python
Architecture
X64
Execution Provider
Default CPU
Execution Provider Library Version
No response
The text was updated successfully, but these errors were encountered: