mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 00:20:18 +01:00
[Fix XPU CI][Inductor UT] Fix test cases broken by community. (#160403)
Fixes #160243, Fixes #160244, Fixes #160245 Pull Request resolved: https://github.com/pytorch/pytorch/pull/160403 Approved by: https://github.com/janeyx99
This commit is contained in:
parent
bbc7c03e93
commit
1853f71b4f
|
|
@ -348,7 +348,7 @@ test_failures = {
|
||||||
"test_rand_like_deterministic_dynamic_shapes": TestFailure(
|
"test_rand_like_deterministic_dynamic_shapes": TestFailure(
|
||||||
("cpu", "cuda", "xpu"), is_skip=True
|
("cpu", "cuda", "xpu"), is_skip=True
|
||||||
),
|
),
|
||||||
"test_repeat_interleave_2_dynamic_shapes": TestFailure(("cpu", "xpu")),
|
"test_repeat_interleave_2_dynamic_shapes": TestFailure(("cpu",)),
|
||||||
"test_slice_mutation2_dynamic_shapes": TestFailure(
|
"test_slice_mutation2_dynamic_shapes": TestFailure(
|
||||||
("cpu", "cuda", "xpu"), is_skip=True
|
("cpu", "cuda", "xpu"), is_skip=True
|
||||||
),
|
),
|
||||||
|
|
|
||||||
|
|
@ -682,6 +682,14 @@ inductor_override_kwargs["xpu"] = {
|
||||||
("nn.functional.unfold", f16): {
|
("nn.functional.unfold", f16): {
|
||||||
"reference_in_float": True,
|
"reference_in_float": True,
|
||||||
},
|
},
|
||||||
|
# Reference crash on Intel LTS2 driver.
|
||||||
|
("nn.functional.interpolate.trilinear", f32): {
|
||||||
|
"check_gradient": False,
|
||||||
|
},
|
||||||
|
# Reference crash on Intel LTS2 driver.
|
||||||
|
("nn.functional.interpolate.trilinear", f64): {
|
||||||
|
"check_gradient": False,
|
||||||
|
},
|
||||||
}
|
}
|
||||||
if TEST_WITH_ROCM:
|
if TEST_WITH_ROCM:
|
||||||
inductor_override_kwargs["cuda"].update(
|
inductor_override_kwargs["cuda"].update(
|
||||||
|
|
|
||||||
|
|
@ -2351,7 +2351,7 @@ class TestImports(TestCase):
|
||||||
# fail, so just set CWD to this script's directory
|
# fail, so just set CWD to this script's directory
|
||||||
cwd=os.path.dirname(os.path.realpath(__file__)),).decode("utf-8")
|
cwd=os.path.dirname(os.path.realpath(__file__)),).decode("utf-8")
|
||||||
|
|
||||||
# The test is flaky on ROCm and has been open and close multiple times
|
# The test is flaky on ROCm/XPU and has been open and close multiple times
|
||||||
# https://github.com/pytorch/pytorch/issues/110040
|
# https://github.com/pytorch/pytorch/issues/110040
|
||||||
@skipIfRocm
|
@skipIfRocm
|
||||||
def test_circular_dependencies(self) -> None:
|
def test_circular_dependencies(self) -> None:
|
||||||
|
|
|
||||||
|
|
@ -7160,10 +7160,10 @@ class DeviceCopy(ExternKernelOut):
|
||||||
# x.get_stride() may be unimplemented if x's size is empty
|
# x.get_stride() may be unimplemented if x's size is empty
|
||||||
stride = x.get_stride()
|
stride = x.get_stride()
|
||||||
is_destination_pinned = (
|
is_destination_pinned = (
|
||||||
x_device.type == "cuda" and device.type == "cpu" and non_blocking
|
is_gpu(x_device.type) and device.type == "cpu" and non_blocking
|
||||||
)
|
)
|
||||||
is_source_pinned = (
|
is_source_pinned = (
|
||||||
x_device.type == "cpu" and device.type == "cuda" and non_blocking
|
x_device.type == "cpu" and is_gpu(device.type) and non_blocking
|
||||||
)
|
)
|
||||||
if is_source_pinned and is_storage_and_layout(x):
|
if is_source_pinned and is_storage_and_layout(x):
|
||||||
x.get_layout().is_pinned = True
|
x.get_layout().is_pinned = True
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue
Block a user