mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-07 12:21:27 +01:00
Summary: This PR supports _scaled_dot_product_flash_attention fallback kernel. Note that in the abi_compatible mode, we retrieve outputs by passing output argument pointers rather than relying on std::get. It also fixes an issue related to dynamic shapes, where we wrongfully query undefined dynamic symbols. Test Plan: ci Reviewed By: frank-wei Differential Revision: D49620191 Pull Request resolved: https://github.com/pytorch/pytorch/pull/110085 Approved by: https://github.com/desertfire |
||
|---|---|---|
| .. | ||
| aoti_runtime | ||
| cuda | ||
| __init__.py | ||
| common.py | ||
| cpp_prefix.h | ||
| cpp.py | ||
| triton_foreach.py | ||
| triton_utils.py | ||
| triton.py | ||
| wrapper.py | ||