Replies: 2 comments 1 reply
-
@yufenglee Can you answer it? |
Beta Was this translation helpful? Give feedback.
0 replies
-
Yes, the QAT model is fine-tuned with either PyTorch or Tensorflow. And then use the PyTorch exporter and tf2onnx to convert the QAT model to onnx model. We don't have plan to finetune ONNX model directly with QAT for now. @mrry could provide more on the timeline. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello,
The input to the
quantize_qat
functionhttps://github.com/microsoft/onnxruntime/blob/0b9f7bb1b0d56d7fffccd0557525d6c03229d137/onnxruntime/python/tools/quantization/quantize.py#L248
is supposed to be
quantize-aware traning onnx model
, based on the function descriptionGiven a quantize-aware traning onnx model, create a quantized onnx model and save it into a file
.quantize-aware traning onnx model
a QAT model trained and fine-tuned in other platforms (e.g. pytorch), and converted to onnx?Beta Was this translation helpful? Give feedback.
All reactions