Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

INT8 accuracy drops to zero with explicit quantization #126

Open
sriram487 opened this issue Feb 6, 2025 · 0 comments
Open

INT8 accuracy drops to zero with explicit quantization #126

sriram487 opened this issue Feb 6, 2025 · 0 comments

Comments

@sriram487
Copy link

I am currently using the ONNX_PTQ pipeline to generate an int8 quantized ONNX model, and the int8 quantized ONNX model performs well. However, when I convert the quantized ONNX model into a .engine file, the accuracy drops to zero. I've tried both the --best and --int8 flags, but the results are the same in both cases.

Originally posted by @sriram487 in #5

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant