From 5395ae6964037da4004b3a0c58b76927bd42c99f Mon Sep 17 00:00:00 2001 From: Max Ren <40742183+mcr229@users.noreply.github.com> Date: Tue, 27 Aug 2024 18:32:22 -0700 Subject: [PATCH] fold quantize in convert Differential Revision: D61814397 Pull Request resolved: https://github.com/pytorch/executorch/pull/4889 --- examples/models/phi-3-mini/export_phi-3-mini.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/examples/models/phi-3-mini/export_phi-3-mini.py b/examples/models/phi-3-mini/export_phi-3-mini.py index 553fded67f..c2e97a21b1 100644 --- a/examples/models/phi-3-mini/export_phi-3-mini.py +++ b/examples/models/phi-3-mini/export_phi-3-mini.py @@ -69,7 +69,7 @@ def export(args) -> None: ) model = prepare_pt2e(model, xnnpack_quantizer) # pyre-fixme[6] model(*example_inputs) - model = convert_pt2e(model, fold_quantize=False) + model = convert_pt2e(model) DuplicateDynamicQuantChainPass()(model) # TODO(lunwenh): update it to use export once # https://github.com/pytorch/pytorch/issues/128394 is resolved.