..
aoti_runtime
[AOTI][refactor] Separate header codegen ( #138882 )
2024-10-27 14:14:27 +00:00
cuda
Add instantiation level to CutlassArgs ( #144506 )
2025-01-10 02:01:40 +00:00
rocm
[ROCm][Inductor][CK] hackfix for segfault in addmm op ( #144519 )
2025-01-10 19:29:14 +00:00
xpu
[AOTI XPU] Enable Cpp wraper for Intel GPU. ( #135318 )
2024-11-26 11:51:32 +00:00
__init__.py
aoti_hipify_utils.py
remove allow-untyped-defs from _inductor/codegen/aoti_hipify_utils.py ( #143916 )
2024-12-27 23:25:37 +00:00
block_analysis.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
common.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp_bmm_template.py
[inductor][cpu] Fix bmm b_index for dynamic expressions in inductor autotuner ( #143141 )
2025-01-05 18:02:37 +00:00
cpp_flex_attention_template.py
Remove is_reduced_floating_point from namespace std ( #144502 )
2025-01-10 03:24:10 +00:00
cpp_gemm_template.py
[inductor][cpu] Fix bmm b_index for dynamic expressions in inductor autotuner ( #143141 )
2025-01-05 18:02:37 +00:00
cpp_micro_gemm.py
Simplify & rectify dequantized B buffer loading for AMX GEMM micro-kernel for WoQ int8 case ( #140258 )
2024-11-22 01:34:06 +00:00
cpp_prefix.h
Remove is_reduced_floating_point from namespace std ( #144502 )
2025-01-10 03:24:10 +00:00
cpp_template_kernel.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp_template.py
[AOTI] Remove the non-ABI-compatible mode (part 1) ( #138009 )
2024-10-17 02:48:26 +00:00
cpp_utils.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp_wrapper_cpu_array_ref.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp_wrapper_cpu.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp_wrapper_gpu.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpp.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
cpu_device_op_overrides.py
remove allow-untyped-defs from _inductor/codegen/cpu_device_op_overrides.py ( #143881 )
2024-12-27 04:10:47 +00:00
cuda_combined_scheduling.py
Prologue Fusion ( #134532 )
2024-12-13 04:18:25 +00:00
debug_utils.py
Rename convert_arrayref_tensor_to_tensor to copy_arrayref_tensor_to_tensor ( #142182 )
2024-12-09 22:23:21 +00:00
halide.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
memory_planning.py
[inductor] Replace set by OrderedSet ( #138466 )
2024-12-13 16:08:45 +00:00
mps_device_op_overrides.py
[Inductor] Add MPS device op overrides ( #143892 )
2024-12-28 02:11:45 +00:00
mps.py
[MPSInductor] Add support for sizevars ( #144662 )
2025-01-13 06:22:38 +00:00
multi_kernel.py
Revert "Use absolute path path.resolve() -> path.absolute() ( #129409 )"
2025-01-04 14:17:20 +00:00
simd_kernel_features.py
Skip L1 cache for single-use buffers ( #143115 )
2025-01-07 19:35:40 +00:00
simd.py
[Inductor] Restrict ND tiling analysis to MemoryDeps ( #144497 )
2025-01-11 05:16:47 +00:00
triton_combo_kernel.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00
triton_split_scan.py
[inductor] Replace set by OrderedSet ( #138466 )
2024-12-13 16:08:45 +00:00
triton_utils.py
[inductor] Move V.graph.scheduler.current_device to V.graph.current_device ( #138252 )
2024-10-18 23:05:54 +00:00
triton.py
Skip L1 cache for single-use buffers ( #143115 )
2025-01-07 19:35:40 +00:00
wrapper.py
Migrate from Tuple -> tuple in torch/_inductor ( #144264 )
2025-01-07 03:27:27 +00:00