From 83ec9d7f042b013dd67595fdac51f588c59d6744 Mon Sep 17 00:00:00 2001 From: pytorchbot Date: Thu, 10 Oct 2024 21:35:24 -0700 Subject: [PATCH] Move mediatek.passes to mediatek._passes (#6089) Move mediatek.passes to mediatek._passes (#5922) Summary: Changing mediatek.passes to mediatek._passes to indicate that these passes are not covered under the API stability guarantee. Pull Request resolved: https://github.com/pytorch/executorch/pull/5922 Reviewed By: helunwencser Differential Revision: D63926846 fbshipit-source-id: fcea1a560892200f9ca0e110104908de0be25867 (cherry picked from commit e540bcb5511fde9799254881657975560323568a) Co-authored-by: Tarun Karuturi --- backends/mediatek/{passes => _passes}/__init__.py | 0 .../decompose_scaled_dot_product_attention.py | 0 backends/mediatek/quantizer/quantizer.py | 2 +- 3 files changed, 1 insertion(+), 1 deletion(-) rename backends/mediatek/{passes => _passes}/__init__.py (100%) rename backends/mediatek/{passes => _passes}/decompose_scaled_dot_product_attention.py (100%) diff --git a/backends/mediatek/passes/__init__.py b/backends/mediatek/_passes/__init__.py similarity index 100% rename from backends/mediatek/passes/__init__.py rename to backends/mediatek/_passes/__init__.py diff --git a/backends/mediatek/passes/decompose_scaled_dot_product_attention.py b/backends/mediatek/_passes/decompose_scaled_dot_product_attention.py similarity index 100% rename from backends/mediatek/passes/decompose_scaled_dot_product_attention.py rename to backends/mediatek/_passes/decompose_scaled_dot_product_attention.py diff --git a/backends/mediatek/quantizer/quantizer.py b/backends/mediatek/quantizer/quantizer.py index 44e35ef54e..4e78d6dff1 100644 --- a/backends/mediatek/quantizer/quantizer.py +++ b/backends/mediatek/quantizer/quantizer.py @@ -7,7 +7,7 @@ from torch.ao.quantization.quantizer import Quantizer from torch.fx import GraphModule -from ..passes.decompose_scaled_dot_product_attention import ( +from .._passes.decompose_scaled_dot_product_attention import ( DecomposeScaledDotProductAttention, ) from .annotator import annotate