Gliner quantization #145
Unanswered
carloronsi
asked this question in
Q&A
Replies: 2 comments 1 reply
-
Beta Was this translation helpful? Give feedback.
0 replies
-
I'm also having a similar problem. When quantizing to 8bit I'm measuring a quite strong drop in performance. More than I usually see for this type of models. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Hi everyone! I've tried the quantization code present in this report:
https://github.com/urchade/GLiNER/blob/main/examples/convert_to_onnx.ipynb
Followed step-by-step, but the quantized model (last cell) is not returning any entity (if the threshold is reduced, texts are labeled incorrectly).
Am I missing something? Is there an additional step required to make the quantized model work?
Thank you all!
Beta Was this translation helpful? Give feedback.
All reactions