Skip to content

Commit

Permalink
Add WOQ int8 test with Inductor Freeze
Browse files Browse the repository at this point in the history
  • Loading branch information
leslie-fang-intel committed Jun 14, 2024
1 parent ff42c85 commit fac758a
Showing 1 changed file with 7 additions and 0 deletions.
7 changes: 7 additions & 0 deletions test/integration/test_integration.py
Original file line number Diff line number Diff line change
Expand Up @@ -1046,6 +1046,13 @@ def test_int8_weight_only_quant_subclass_api(self, device, dtype):
change_linear_weights_to_int8_woqtensors, device, 40, test_dtype=dtype
)

@parameterized.expand(COMMON_DEVICE_DTYPE)
@torch._inductor.config.patch({"freezing": True})
def test_int8_weight_only_quant_with_freeze(self, device, dtype):
self._test_lin_weight_subclass_api_impl(
change_linear_weights_to_int8_woqtensors, device, 40, test_dtype=dtype
)

@parameterized.expand(COMMON_DEVICE_DTYPE)
@unittest.skipIf(not TORCH_VERSION_AFTER_2_3, "int4 requires torch nightly.")
def test_int4_weight_only_quant_subclass_api(self, device, dtype):
Expand Down

0 comments on commit fac758a

Please sign in to comment.