pytorch/test/cpp/aoti_inference
Mu-Chu Lee 2291199e9b [AOTInductor] Use CudaCachingAllocator for memory allocation (#162893)
Summary:
Use c10::CudaCachingAllocator for AOTInductor's initial constant buffer
allocation.

Test Plan:
Activate test under test/cpp/aoti_inference/test.cpp

Reviewers:

Subscribers:

Tasks:

Tags:

Pull Request resolved: https://github.com/pytorch/pytorch/pull/162893
Approved by: https://github.com/desertfire
2025-09-17 17:08:20 +00:00
..
aoti_custom_class.cpp [AOTI] Fix #140546 and support AOTI package load for Intel GPU. (#140664) 2024-12-10 05:05:08 +00:00
aoti_custom_class.h
CMakeLists.txt [AOTI] Fix AOT inductor CMake build dependency order (#157557) 2025-07-04 14:33:36 +00:00
compile_model.py [AOTI] Fix test_aoti_inference CPU build issue (#134675) 2024-08-28 17:42:19 +00:00
generate_lowered_cpu.py [AOTInductor] Add standalone test for compilation from ExportedProgram (#142327) 2024-12-10 06:50:09 +00:00
standalone_compile.sh [AOTInductor] Add standalone test for compilation from ExportedProgram (#142327) 2024-12-10 06:50:09 +00:00
standalone_test.cpp [AOTInductor] Add standalone test for compilation from ExportedProgram (#142327) 2024-12-10 06:50:09 +00:00
test.cpp [AOTInductor] Use CudaCachingAllocator for memory allocation (#162893) 2025-09-17 17:08:20 +00:00
test.py [AOTInductor] Add test for enabling CUDACachingAllocator for AOTInductor's Weight (#159279) 2025-07-29 02:52:10 +00:00