mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
Remove outdated skipCUDAIfCudnnVersionLessThan decoration (#148940)
Test conditions for CUDNN 7 and 8 were removed because we have moved to CUDNN 9. Pull Request resolved: https://github.com/pytorch/pytorch/pull/148940 Approved by: https://github.com/mikaylagawarecki
This commit is contained in:
parent
c73c72b1e1
commit
970fefcc53
|
|
@ -26,7 +26,6 @@ from torch.testing._internal.common_device_type import (
|
|||
onlyNativeDeviceTypes,
|
||||
precisionOverride,
|
||||
skipCPUIfNoMkldnn,
|
||||
skipCUDAIfCudnnVersionLessThan,
|
||||
skipCUDAIfMiopen,
|
||||
skipCUDAIfNoCudnn,
|
||||
skipCUDAIfNoMiopen,
|
||||
|
|
@ -3639,7 +3638,6 @@ class TestConvolutionNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfNotMiopenSuggestNHWC
|
||||
@skipCUDAIfCudnnVersionLessThan(7603)
|
||||
@dtypes(torch.half, torch.float, torch.cfloat)
|
||||
def test_conv_cudnn_nhwc(self, device, dtype):
|
||||
def helper(n, c, h, w, out_channels, kernel_size, groups):
|
||||
|
|
@ -3693,7 +3691,6 @@ class TestConvolutionNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfRocm
|
||||
@skipCUDAIfCudnnVersionLessThan(8005)
|
||||
@dtypes(torch.half, torch.float)
|
||||
def test_conv_cudnn_ndhwc(self, device, dtype):
|
||||
def helper(n, c, d, h, w, out_channels, kernel_size, groups):
|
||||
|
|
@ -3824,7 +3821,6 @@ class TestConvolutionNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfNotMiopenSuggestNHWC
|
||||
@skipCUDAIfCudnnVersionLessThan(7603)
|
||||
@tf32_on_and_off(0.05)
|
||||
def test_conv_cudnn_mismatch_memory_format(self, device):
|
||||
configs = [
|
||||
|
|
@ -3958,7 +3954,6 @@ class TestConvolutionNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfRocm
|
||||
@skipCUDAIfCudnnVersionLessThan(7603)
|
||||
def test_convert_conv2d_weight_memory_format(self, device):
|
||||
input = torch.randint(1, 10, (2, 8, 4, 4), dtype=torch.float32, device=device)
|
||||
model = nn.Sequential(nn.Conv2d(8, 4, 3), nn.BatchNorm2d(4)).to(device).float()
|
||||
|
|
@ -3979,7 +3974,6 @@ class TestConvolutionNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfRocm
|
||||
@skipCUDAIfCudnnVersionLessThan(7603)
|
||||
def test_convert_conv3d_weight_memory_format(self, device):
|
||||
input = torch.randint(
|
||||
1, 10, (2, 8, 4, 4, 4), dtype=torch.float32, device=device
|
||||
|
|
|
|||
|
|
@ -41,7 +41,7 @@ from torch.testing._internal.common_nn import NNTestCase, NewModuleTest, Criteri
|
|||
module_tests, criterion_tests, loss_reference_fns, _create_basic_net, \
|
||||
ctcloss_reference, get_new_module_tests, single_batch_reference_fn, _test_bfloat16_ops, _test_module_empty_input
|
||||
from torch.testing._internal.common_device_type import dtypesIfMPS, instantiate_device_type_tests, dtypes, \
|
||||
dtypesIfCUDA, precisionOverride, skipCUDAIfCudnnVersionLessThan, onlyCUDA, onlyCPU, \
|
||||
dtypesIfCUDA, precisionOverride, onlyCUDA, onlyCPU, \
|
||||
skipCUDAIfRocm, skipCUDAIf, skipCUDAIfNotRocm, \
|
||||
onlyNativeDeviceTypes, deviceCountAtLeast, largeTensorTest, expectedFailureMeta, expectedFailureMPS, \
|
||||
skipMeta, get_all_device_types
|
||||
|
|
@ -10913,7 +10913,6 @@ class TestNNDeviceType(NNTestCase):
|
|||
out = m(inp)
|
||||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfCudnnVersionLessThan(7600)
|
||||
def test_CTCLoss_cudnn(self, device):
|
||||
def _helper(zero_infinity):
|
||||
target_lengths = [30, 25, 20]
|
||||
|
|
@ -11456,7 +11455,6 @@ class TestNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfRocm(msg="skipped Cudnn test on ROCm")
|
||||
@skipCUDAIfCudnnVersionLessThan(7600)
|
||||
def test_ctc_loss_cudnn(self, device):
|
||||
batch_size = 16
|
||||
input_length = 30
|
||||
|
|
@ -11481,7 +11479,6 @@ class TestNNDeviceType(NNTestCase):
|
|||
|
||||
@onlyCUDA
|
||||
@skipCUDAIfRocm(msg="skipped Cudnn test on ROCm")
|
||||
@skipCUDAIfCudnnVersionLessThan(8000)
|
||||
def test_ctc_loss_cudnn_tensor(self, device):
|
||||
batch_size = 16
|
||||
input_length = 30
|
||||
|
|
|
|||
|
|
@ -16,7 +16,7 @@ from torch.testing._internal.common_dtype import (
|
|||
floating_types, floating_and_complex_types_and, get_all_fp_dtypes)
|
||||
from torch.testing._internal.common_device_type import (
|
||||
_TestParametrizer, _update_param_kwargs, expectedFailureMPS, toleranceOverride, tol,
|
||||
skipCUDAIfCudnnVersionLessThan, skipCUDAIfRocm, precisionOverride, skipMeta, skipMPS,
|
||||
skipCUDAIfRocm, precisionOverride, skipMeta, skipMPS,
|
||||
skipCUDAVersionIn)
|
||||
from torch.testing._internal.common_methods_invocations import DecorateInfo
|
||||
from torch.testing._internal.common_nn import (
|
||||
|
|
@ -3506,8 +3506,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# See #119108: MPSNDArrayConvolutionA14.mm:3976: failed assertion `destination datatype must be fp32'
|
||||
|
|
@ -3525,8 +3523,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# This was wrongly being skipped before and needs investigation.
|
||||
|
|
@ -3551,8 +3547,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 8005
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=8005), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Conv3d is not supported on MPS backend
|
||||
|
|
@ -3570,8 +3564,6 @@ module_db: list[ModuleInfo] = [
|
|||
module_memformat_affects_out=True,
|
||||
dtypes=floating_and_complex_types_and(torch.chalf),
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Not implemented for chalf on CPU
|
||||
|
|
@ -3593,8 +3585,6 @@ module_db: list[ModuleInfo] = [
|
|||
module_memformat_affects_out=True,
|
||||
dtypes=floating_and_complex_types_and(torch.chalf),
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Fails on backward check because ViewAsRealBackward apply contiguous for grad
|
||||
|
|
@ -3627,8 +3617,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 8005
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=8005), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# ConvTranspose3d is not supported on MPS backend
|
||||
|
|
@ -3698,8 +3686,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
@ -3720,8 +3706,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
@ -3749,8 +3733,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 8005
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=8005), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
@ -3770,8 +3752,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
@ -3792,8 +3772,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 7603
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=7603), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
@ -3821,8 +3799,6 @@ module_db: list[ModuleInfo] = [
|
|||
gradcheck_nondet_tol=GRADCHECK_NONDET_TOL,
|
||||
module_memformat_affects_out=True,
|
||||
skips=(
|
||||
# channels_last support on cuda requires cudnn >= 8005
|
||||
DecorateInfo(skipCUDAIfCudnnVersionLessThan(version=8005), 'TestModule', 'test_memory_format'),
|
||||
# Failure on ROCM for float32 issue #70125
|
||||
DecorateInfo(skipCUDAIfRocm, 'TestModule', 'test_memory_format', dtypes=[torch.float32]),
|
||||
# Lazy modules don't currently play well with ModuleInfo tests on the meta device.
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user