pytorch/torch/_inductor/codegen
Davide Italiano 35b46a75f1 [mps/inductor] Add support for round() (#144731)
With this change, inductor/test_view_on_aliased passes.

Pull Request resolved: https://github.com/pytorch/pytorch/pull/144731
Approved by: https://github.com/malfet
2025-01-14 05:56:13 +00:00
..
aoti_runtime [AOTI][refactor] Separate header codegen (#138882) 2024-10-27 14:14:27 +00:00
cuda Add instantiation level to CutlassArgs (#144506) 2025-01-10 02:01:40 +00:00
rocm [ROCm][Inductor][CK] hackfix for segfault in addmm op (#144519) 2025-01-10 19:29:14 +00:00
xpu [AOTI XPU] Enable Cpp wraper for Intel GPU. (#135318) 2024-11-26 11:51:32 +00:00
__init__.py
aoti_hipify_utils.py remove allow-untyped-defs from _inductor/codegen/aoti_hipify_utils.py (#143916) 2024-12-27 23:25:37 +00:00
block_analysis.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
common.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpp_bmm_template.py [inductor][cpu] Fix bmm b_index for dynamic expressions in inductor autotuner (#143141) 2025-01-05 18:02:37 +00:00
cpp_flex_attention_template.py Remove is_reduced_floating_point from namespace std (#144502) 2025-01-10 03:24:10 +00:00
cpp_gemm_template.py [inductor][cpu] Fix bmm b_index for dynamic expressions in inductor autotuner (#143141) 2025-01-05 18:02:37 +00:00
cpp_micro_gemm.py Simplify & rectify dequantized B buffer loading for AMX GEMM micro-kernel for WoQ int8 case (#140258) 2024-11-22 01:34:06 +00:00
cpp_prefix.h Remove is_reduced_floating_point from namespace std (#144502) 2025-01-10 03:24:10 +00:00
cpp_template_kernel.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpp_template.py [AOTI] Remove the non-ABI-compatible mode (part 1) (#138009) 2024-10-17 02:48:26 +00:00
cpp_utils.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpp_wrapper_cpu_array_ref.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpp_wrapper_cpu.py Support nanj in inductor (#144064) 2025-01-13 14:29:38 +00:00
cpp_wrapper_gpu.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpp.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
cpu_device_op_overrides.py remove allow-untyped-defs from _inductor/codegen/cpu_device_op_overrides.py (#143881) 2024-12-27 04:10:47 +00:00
cuda_combined_scheduling.py Prologue Fusion (#134532) 2024-12-13 04:18:25 +00:00
debug_utils.py Rename convert_arrayref_tensor_to_tensor to copy_arrayref_tensor_to_tensor (#142182) 2024-12-09 22:23:21 +00:00
halide.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
memory_planning.py [inductor] Replace set by OrderedSet (#138466) 2024-12-13 16:08:45 +00:00
mps_device_op_overrides.py [Inductor] Add MPS device op overrides (#143892) 2024-12-28 02:11:45 +00:00
mps.py [mps/inductor] Add support for round() (#144731) 2025-01-14 05:56:13 +00:00
multi_kernel.py Revert "Use absolute path path.resolve() -> path.absolute() (#129409)" 2025-01-04 14:17:20 +00:00
simd_kernel_features.py Skip L1 cache for single-use buffers (#143115) 2025-01-07 19:35:40 +00:00
simd.py [Inductor] Restrict ND tiling analysis to MemoryDeps (#144497) 2025-01-11 05:16:47 +00:00
triton_combo_kernel.py Migrate from Tuple -> tuple in torch/_inductor (#144264) 2025-01-07 03:27:27 +00:00
triton_split_scan.py [inductor] Replace set by OrderedSet (#138466) 2024-12-13 16:08:45 +00:00
triton_utils.py [inductor] Move V.graph.scheduler.current_device to V.graph.current_device (#138252) 2024-10-18 23:05:54 +00:00
triton.py Skip L1 cache for single-use buffers (#143115) 2025-01-07 19:35:40 +00:00
wrapper.py Support nanj in inductor (#144064) 2025-01-13 14:29:38 +00:00