-
Notifications
You must be signed in to change notification settings - Fork 370
Breaking Change: Remove the deprecated int8 calibrator related #3759
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
No ciflow labels are configured for this repo. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Mostly looks good, if the fx changes still work I would leave those since Meta might still be using them.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. just a minor comment
from torch_tensorrt.fx.converters.converter_utils import ( | ||
mark_as_int8_layer, | ||
set_layer_name, | ||
) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Any reasons for partially replacing import lib?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Its looking good, only thing is if we are switching to dynamo.types.TRTTensor we should just go to trt.ITensor and be closer to removing that types file. Dont think I need to review again after those changes
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. I think we can directly use trt.ITensor instead of alias, but it doesn't matter. that's fine for now
An error pops up.
you can import from torch_tensorrt.dynamo.utils |
Description
Remove all the int8 calibrator, int8 implicit quantization and dynamic range related.
All the quantization should be explicit.
Fixes # (issue)
Type of change
Please delete options that are not relevant and/or add your own.
Checklist: