Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

INT8 Quantization #298

Open
sriram487 opened this issue Feb 4, 2025 · 2 comments
Open

INT8 Quantization #298

sriram487 opened this issue Feb 4, 2025 · 2 comments

Comments

@sriram487
Copy link

I'm working on quantizing the model, and while the accuracy remains stable when using FP16 precision, the model's performance degrades significantly when quantizing to INT8 or FP8. The issue could be related to the calibration data. For INT8 extrinsic quantization, it's typically recommended to use at least 500 images for CNNs. However, I'm not sure about the minimum number of samples required and the appropriate batch size for the calibration data, especially the batch size changes during registration and tracking, which could affect the calibration parameters.

What data should be used for calibration? Since the model is trained on synthetic data, is it ok to use a portion of that synthetic training data for calibration?

@sriram487 sriram487 changed the title INT8 Calibration. INT8 Quantization Feb 5, 2025
@wenbowen123
Copy link
Collaborator

I haven't tried quantizing. But I'm also curious. You are welcome to report back your findings : )

@A7eNg
Copy link

A7eNg commented Feb 5, 2025

+1 waiting for the quantization result

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants