diff --git a/pyproject.toml b/pyproject.toml index 1ad3595ab29..4cf3562886f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -208,6 +208,7 @@ select = [ "PLC1802", # len({expression}) used as condition without comparison "PLC0205", # string as __slots__ "PLC3002", # unnecessary-direct-lambda-call + "PLC0414", # Import alias does not rename original package "PLE", "PLR0133", # constant comparison "PLR0206", # property with params diff --git a/test/jit/test_save_load_for_op_version.py b/test/jit/test_save_load_for_op_version.py index 37df00ab8ef..fdb0b085044 100644 --- a/test/jit/test_save_load_for_op_version.py +++ b/test/jit/test_save_load_for_op_version.py @@ -4,7 +4,7 @@ import io import os import sys -from itertools import product as product +from itertools import product from typing import Union import hypothesis.strategies as st diff --git a/test/profiler/test_execution_trace.py b/test/profiler/test_execution_trace.py index 7da2898ffbe..2cd51136ab4 100644 --- a/test/profiler/test_execution_trace.py +++ b/test/profiler/test_execution_trace.py @@ -426,7 +426,7 @@ class TestExecutionTrace(TestCase): @skipCPUIf(True, "skip CPU device for testing profiling triton") def test_execution_trace_env_enabled_with_pt2(self, device): # clean up the local cache for triton kernel - from torch._inductor.codecache import PyCodeCache as PyCodeCache + from torch._inductor.codecache import PyCodeCache PyCodeCache.cache_clear(purge=True) @@ -488,7 +488,7 @@ class TestExecutionTrace(TestCase): @skipCPUIf(True, "skip CPU device for testing profiling triton") def test_triton_fx_graph_with_et(self, device): # clean up the local cache for triton kernel - from torch._inductor.codecache import PyCodeCache as PyCodeCache + from torch._inductor.codecache import PyCodeCache PyCodeCache.cache_clear(purge=True) diff --git a/test/quantization/jit/test_ondevice_quantization.py b/test/quantization/jit/test_ondevice_quantization.py index ce23b155810..7fac580af2e 100644 --- a/test/quantization/jit/test_ondevice_quantization.py +++ b/test/quantization/jit/test_ondevice_quantization.py @@ -18,7 +18,7 @@ from torch.testing._internal.common_quantization import ( LinearAddModel, ) from torch.testing._internal.common_utils import TestCase -from torch.utils import bundled_inputs as bundled_inputs +from torch.utils import bundled_inputs class myMod(torch.nn.Module): diff --git a/tools/test/heuristics/test_heuristics.py b/tools/test/heuristics/test_heuristics.py index 575d1b5732b..9c52ea55e31 100644 --- a/tools/test/heuristics/test_heuristics.py +++ b/tools/test/heuristics/test_heuristics.py @@ -14,7 +14,6 @@ REPO_ROOT = Path(__file__).resolve().parents[3] sys.path.append(str(REPO_ROOT)) from tools.test.heuristics.test_interface import TestTD -from tools.testing.target_determination.determinator import TestPrioritizations from tools.testing.target_determination.heuristics.filepath import ( file_matches_keyword, get_keywords, @@ -22,6 +21,7 @@ from tools.testing.target_determination.heuristics.filepath import ( from tools.testing.target_determination.heuristics.historical_class_failure_correlation import ( HistoricalClassFailurCorrelation, ) +from tools.testing.target_determination.heuristics.interface import TestPrioritizations from tools.testing.target_determination.heuristics.previously_failed_in_pr import ( get_previous_failures, ) diff --git a/tools/testing/do_target_determination_for_s3.py b/tools/testing/do_target_determination_for_s3.py index 27a0fbb5b42..1272181a91f 100644 --- a/tools/testing/do_target_determination_for_s3.py +++ b/tools/testing/do_target_determination_for_s3.py @@ -19,9 +19,9 @@ from tools.stats.import_test_stats import ( ) from tools.stats.upload_metrics import emit_metric from tools.testing.discover_tests import TESTS -from tools.testing.target_determination.determinator import ( +from tools.testing.target_determination.determinator import get_test_prioritizations +from tools.testing.target_determination.heuristics.interface import ( AggregatedHeuristics, - get_test_prioritizations, TestPrioritizations, ) diff --git a/tools/testing/target_determination/determinator.py b/tools/testing/target_determination/determinator.py index 9207e62c28b..0d00b85d073 100644 --- a/tools/testing/target_determination/determinator.py +++ b/tools/testing/target_determination/determinator.py @@ -4,9 +4,9 @@ import sys from typing import Any from tools.testing.target_determination.heuristics import ( - AggregatedHeuristics as AggregatedHeuristics, + AggregatedHeuristics, HEURISTICS, - TestPrioritizations as TestPrioritizations, + TestPrioritizations, ) diff --git a/torch/_inductor/compile_fx.py b/torch/_inductor/compile_fx.py index 6153daac47c..5787cfdfc51 100644 --- a/torch/_inductor/compile_fx.py +++ b/torch/_inductor/compile_fx.py @@ -114,7 +114,7 @@ from .fx_passes.post_grad import post_grad_passes, view_to_reshape from .fx_passes.pre_grad import pre_grad_passes from .graph import GraphLowering from .ir import get_device_type, IRNode -from .output_code import complex_memory_overlap as complex_memory_overlap # noqa: F401 +from .output_code import complex_memory_overlap # noqa: F401 from .triton_bundler import TritonBundler from .utils import ( align_inputs_from_check_idxs, diff --git a/torch/_inductor/compile_fx_async.py b/torch/_inductor/compile_fx_async.py index 05c896ae864..4faa48a0044 100644 --- a/torch/_inductor/compile_fx_async.py +++ b/torch/_inductor/compile_fx_async.py @@ -9,7 +9,7 @@ import torch._inductor.async_compile # noqa: F401 required to warm up AsyncComp from torch._inductor.output_code import CompiledFxGraphConstants, OutputCode from .compile_fx import _CompileFxKwargs, _InProcessFxCompile, FxCompile -from .output_code import complex_memory_overlap as complex_memory_overlap # noqa: F401 +from .output_code import complex_memory_overlap # noqa: F401 # When async compile works with cache, remove the disabling below diff --git a/torch/_inductor/compile_fx_ext.py b/torch/_inductor/compile_fx_ext.py index 113a7c92606..7e3c5bb7cfa 100644 --- a/torch/_inductor/compile_fx_ext.py +++ b/torch/_inductor/compile_fx_ext.py @@ -30,7 +30,7 @@ from . import config from .compile_fx import _CompileFxKwargs, _InProcessFxCompile, FxCompile, log from .debug import DebugContext from .graph import GraphLowering -from .output_code import complex_memory_overlap as complex_memory_overlap # noqa: F401 +from .output_code import complex_memory_overlap # noqa: F401 from .virtualized import V diff --git a/torch/_inductor/compile_fx_subproc.py b/torch/_inductor/compile_fx_subproc.py index 3a1535ec1e2..58d5195046f 100644 --- a/torch/_inductor/compile_fx_subproc.py +++ b/torch/_inductor/compile_fx_subproc.py @@ -20,7 +20,7 @@ from .compile_fx_ext import ( _WireProtocolPickledInput, _WireProtocolPickledOutput, ) -from .output_code import complex_memory_overlap as complex_memory_overlap # noqa: F401 +from .output_code import complex_memory_overlap # noqa: F401 if TYPE_CHECKING: diff --git a/torch/_inductor/cudagraph_trees.py b/torch/_inductor/cudagraph_trees.py index 566db12e492..25b60c3e732 100644 --- a/torch/_inductor/cudagraph_trees.py +++ b/torch/_inductor/cudagraph_trees.py @@ -102,7 +102,7 @@ S = TypeVar("S", bound="StorageWeakRefWrapper") if torch.backends.cuda.is_built(): from torch._C import ( _cuda_CUDAAllocator_AllocatorState as AllocatorState, - _set_cached_tensors_enabled as _set_cached_tensors_enabled, + _set_cached_tensors_enabled, ) else: diff --git a/torch/_inductor/debug.py b/torch/_inductor/debug.py index 5dbe849af09..18b644b8539 100644 --- a/torch/_inductor/debug.py +++ b/torch/_inductor/debug.py @@ -19,7 +19,7 @@ from unittest.mock import patch import torch from functorch.compile import draw_graph, get_aot_graph_name, get_graph_being_compiled -from torch import fx as fx +from torch import fx from torch._dynamo.repro.after_aot import save_graph_repro from torch._dynamo.utils import get_debug_dir from torch._inductor import utils diff --git a/torch/_subclasses/fake_tensor.py b/torch/_subclasses/fake_tensor.py index 3c2d609b736..a4dadc558f4 100644 --- a/torch/_subclasses/fake_tensor.py +++ b/torch/_subclasses/fake_tensor.py @@ -2674,7 +2674,7 @@ class FakeTensorMode(TorchDispatchMode): return maybe_propagate_real_tensors(fast_impl(self, *args, **kwargs)) # If there's a Python meta, prefer that over the decomposition - from torch._decomp import meta_table as meta_table + from torch._decomp import meta_table if ( func not in meta_table diff --git a/torch/profiler/profiler.py b/torch/profiler/profiler.py index 1e8e6591d49..e92aa3fafb7 100644 --- a/torch/profiler/profiler.py +++ b/torch/profiler/profiler.py @@ -1038,7 +1038,7 @@ class ExecutionTraceObserver(_ITraceObserver): return # Save the kernel paths for the generated kernels - from torch._inductor.codecache import PyCodeCache as PyCodeCache + from torch._inductor.codecache import PyCodeCache kernel_files = [ v.__file__ diff --git a/torch/types.py b/torch/types.py index 01a62ff4d01..0388c9c66ae 100644 --- a/torch/types.py +++ b/torch/types.py @@ -18,15 +18,15 @@ from typing_extensions import Self # `as` imports have better static analysis support than assignment `ExposedType: TypeAlias = HiddenType` from torch import ( # noqa: F401 device as _device, - DispatchKey as DispatchKey, + DispatchKey, dtype as _dtype, layout as _layout, qscheme as _qscheme, - Size as Size, - SymBool as SymBool, - SymFloat as SymFloat, - SymInt as SymInt, - Tensor as Tensor, + Size, + SymBool, + SymFloat, + SymInt, + Tensor, ) diff --git a/torch/utils/_cxx_pytree.py b/torch/utils/_cxx_pytree.py index d2ba9791371..9ffd8190a5e 100644 --- a/torch/utils/_cxx_pytree.py +++ b/torch/utils/_cxx_pytree.py @@ -21,13 +21,13 @@ from typing_extensions import deprecated, TypeIs import torch.utils._pytree as python_pytree from torch.torch_version import TorchVersion as _TorchVersion from torch.utils._pytree import ( - is_namedtuple as is_namedtuple, - is_namedtuple_class as is_namedtuple_class, - is_namedtuple_instance as is_namedtuple_instance, - is_structseq as is_structseq, - is_structseq_class as is_structseq_class, - is_structseq_instance as is_structseq_instance, - KeyEntry as KeyEntry, + is_namedtuple, + is_namedtuple_class, + is_namedtuple_instance, + is_structseq, + is_structseq_class, + is_structseq_instance, + KeyEntry, ) diff --git a/torch/utils/data/datapipes/gen_pyi.py b/torch/utils/data/datapipes/gen_pyi.py index bce38547986..9f16f6f4552 100644 --- a/torch/utils/data/datapipes/gen_pyi.py +++ b/torch/utils/data/datapipes/gen_pyi.py @@ -8,7 +8,7 @@ from typing_extensions import deprecated try: from torchgen.api.python import format_function_signature - from torchgen.utils import FileManager as FileManager + from torchgen.utils import FileManager except ImportError: import sys