mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
[Inductor UT][XPU] Skip fft_c2c case since it's not implemented on XPU. (#147351)
Pull Request resolved: https://github.com/pytorch/pytorch/pull/147351 Approved by: https://github.com/jansel
This commit is contained in:
parent
ed83b0b70b
commit
8cbf7d0d6e
|
|
@ -4526,7 +4526,7 @@ GPU_TEST_FAILURES = {
|
|||
# No scaled_dot_product_efficient_attention implementation for XPU yet.
|
||||
"test_scaled_dot_product_efficient_attention": fail_gpu(("xpu",)),
|
||||
# No fft implementation for XPU yet.
|
||||
"test_fft_c2c": fail_gpu(("xpu",)),
|
||||
"test_fft_c2c": fail_gpu(("xpu",), is_skip=True),
|
||||
"test_stft": fail_gpu(("xpu",)),
|
||||
}
|
||||
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user