-
Notifications
You must be signed in to change notification settings - Fork 355
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
How to get a Qoperator format ONNX model after quantization? #2825
Comments
Hi @JiliangNi I'm not aware of Qoperator support in AIMET. However, you can obtain QDQ format in your ONNX using use_embedded_encodings=true with AIMET's ONNX export feature. If you're unfamiliar with QDQ format, you can find more information in this link PS: Please note that AIMET QDQ format is supported only for int8 quantization (W8A8) due to limitation in the ONNX opset version related to the Torch version (1.13) of AIMET |
So AIMET does not support Qoperator format in ONNX, right? Is there any method which could convert QDQ to Qoperator in ONNX? |
@JiliangNi, we don't have the support to convert currently. |
How to get a Qoperator format ONNX model after quantization?
The text was updated successfully, but these errors were encountered: