mirror of
https://github.com/zebrajr/pytorch.git
synced 2025-12-06 12:20:52 +01:00
Revert "[CI] Move CUDA-11.6 to Python-3.10 configuration (#81233)"
This reverts commit 7ccf693cf6.
Reverted https://github.com/pytorch/pytorch/pull/81233 on behalf of https://github.com/janeyx99 due to this should have been reverted along with 81372 for breaking internal builds
This commit is contained in:
parent
1233c3c256
commit
c96485804f
|
|
@ -141,7 +141,7 @@ case "$image" in
|
||||||
pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7)
|
pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7)
|
||||||
CUDA_VERSION=11.6.2
|
CUDA_VERSION=11.6.2
|
||||||
CUDNN_VERSION=8
|
CUDNN_VERSION=8
|
||||||
ANACONDA_PYTHON_VERSION=3.10
|
ANACONDA_PYTHON_VERSION=3.7
|
||||||
GCC_VERSION=7
|
GCC_VERSION=7
|
||||||
PROTOBUF=yes
|
PROTOBUF=yes
|
||||||
DB=yes
|
DB=yes
|
||||||
|
|
|
||||||
|
|
@ -76,10 +76,7 @@ if [ -n "$ANACONDA_PYTHON_VERSION" ]; then
|
||||||
# DO NOT install cmake here as it would install a version newer than 3.13, but
|
# DO NOT install cmake here as it would install a version newer than 3.13, but
|
||||||
# we want to pin to version 3.13.
|
# we want to pin to version 3.13.
|
||||||
CONDA_COMMON_DEPS="astunparse pyyaml mkl=2022.0.1 mkl-include=2022.0.1 setuptools cffi future six"
|
CONDA_COMMON_DEPS="astunparse pyyaml mkl=2022.0.1 mkl-include=2022.0.1 setuptools cffi future six"
|
||||||
if [ "$ANACONDA_PYTHON_VERSION" = "3.10" ]; then
|
if [ "$ANACONDA_PYTHON_VERSION" = "3.9" ]; then
|
||||||
# Install llvm-8 as it is required to compile llvmlite-0.30.0 from source
|
|
||||||
conda_install numpy=1.21.2 ${CONDA_COMMON_DEPS} llvmdev=8.0.0
|
|
||||||
elif [ "$ANACONDA_PYTHON_VERSION" = "3.9" ]; then
|
|
||||||
# Install llvm-8 as it is required to compile llvmlite-0.30.0 from source
|
# Install llvm-8 as it is required to compile llvmlite-0.30.0 from source
|
||||||
conda_install numpy=1.19.2 ${CONDA_COMMON_DEPS} llvmdev=8.0.0
|
conda_install numpy=1.19.2 ${CONDA_COMMON_DEPS} llvmdev=8.0.0
|
||||||
elif [ "$ANACONDA_PYTHON_VERSION" = "3.8" ]; then
|
elif [ "$ANACONDA_PYTHON_VERSION" = "3.8" ]; then
|
||||||
|
|
|
||||||
|
|
@ -100,7 +100,6 @@ networkx==2.6.3
|
||||||
|
|
||||||
numba==0.49.0 ; python_version < "3.9"
|
numba==0.49.0 ; python_version < "3.9"
|
||||||
numba==0.54.1 ; python_version == "3.9"
|
numba==0.54.1 ; python_version == "3.9"
|
||||||
numba==0.55.2 ; python_version == "3.10"
|
|
||||||
#Description: Just-In-Time Compiler for Numerical Functions
|
#Description: Just-In-Time Compiler for Numerical Functions
|
||||||
#Pinned versions: 0.54.1, 0.49.0, <=0.49.1
|
#Pinned versions: 0.54.1, 0.49.0, <=0.49.1
|
||||||
#test that import: test_numba_integration.py
|
#test that import: test_numba_integration.py
|
||||||
|
|
@ -179,8 +178,7 @@ scikit-image
|
||||||
#Pinned versions: 0.20.3
|
#Pinned versions: 0.20.3
|
||||||
#test that import:
|
#test that import:
|
||||||
|
|
||||||
scipy==1.6.3 ; python_version < "3.10"
|
scipy==1.6.3
|
||||||
scipy==1.8.1 ; python_version == "3.10"
|
|
||||||
# Pin SciPy because of failing distribution tests (see #60347)
|
# Pin SciPy because of failing distribution tests (see #60347)
|
||||||
#Description: scientific python
|
#Description: scientific python
|
||||||
#Pinned versions: 1.6.3
|
#Pinned versions: 1.6.3
|
||||||
|
|
|
||||||
16
.github/workflows/pull.yml
vendored
16
.github/workflows/pull.yml
vendored
|
|
@ -148,20 +148,20 @@ jobs:
|
||||||
{ config: "default", shard: 1, num_shards: 1, runner: "linux.2xlarge" },
|
{ config: "default", shard: 1, num_shards: 1, runner: "linux.2xlarge" },
|
||||||
]}
|
]}
|
||||||
|
|
||||||
linux-bionic-cuda11_6-py3_10-gcc7-build:
|
linux-bionic-cuda11_6-py3_7-gcc7-build:
|
||||||
name: linux-bionic-cuda11.6-py3.10-gcc7
|
name: linux-bionic-cuda11.6-py3.7-gcc7
|
||||||
uses: ./.github/workflows/_linux-build.yml
|
uses: ./.github/workflows/_linux-build.yml
|
||||||
with:
|
with:
|
||||||
build-environment: linux-bionic-cuda11.6-py3.10-gcc7
|
build-environment: linux-bionic-cuda11.6-py3.7-gcc7
|
||||||
docker-image-name: pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7
|
docker-image-name: pytorch-linux-bionic-cuda11.6-cudnn8-py3-gcc7
|
||||||
|
|
||||||
linux-bionic-cuda11_6-py3_10-gcc7-test:
|
linux-bionic-cuda11_6-py3_7-gcc7-test:
|
||||||
name: linux-bionic-cuda11.6-py3.10-gcc7
|
name: linux-bionic-cuda11.6-py3.7-gcc7
|
||||||
uses: ./.github/workflows/_linux-test.yml
|
uses: ./.github/workflows/_linux-test.yml
|
||||||
needs: linux-bionic-cuda11_6-py3_10-gcc7-build
|
needs: linux-bionic-cuda11_6-py3_7-gcc7-build
|
||||||
with:
|
with:
|
||||||
build-environment: linux-bionic-cuda11.6-py3.10-gcc7
|
build-environment: linux-bionic-cuda11.6-py3.7-gcc7
|
||||||
docker-image: ${{ needs.linux-bionic-cuda11_6-py3_10-gcc7-build.outputs.docker-image }}
|
docker-image: ${{ needs.linux-bionic-cuda11_6-py3_7-gcc7-build.outputs.docker-image }}
|
||||||
test-matrix: |
|
test-matrix: |
|
||||||
{ include: [
|
{ include: [
|
||||||
{ config: "default", shard: 1, num_shards: 4, runner: "linux.4xlarge.nvidia.gpu" },
|
{ config: "default", shard: 1, num_shards: 4, runner: "linux.4xlarge.nvidia.gpu" },
|
||||||
|
|
|
||||||
|
|
@ -75,10 +75,7 @@ class TestHash(JitTestCase):
|
||||||
self.checkScript(fn, (1.2345, float("inf")))
|
self.checkScript(fn, (1.2345, float("inf")))
|
||||||
self.checkScript(fn, (float("inf"), float("inf")))
|
self.checkScript(fn, (float("inf"), float("inf")))
|
||||||
self.checkScript(fn, (1.2345, float('nan')))
|
self.checkScript(fn, (1.2345, float('nan')))
|
||||||
if sys.version_info < (3, 10):
|
self.checkScript(fn, (float("nan"), float("nan")))
|
||||||
# Hash of two nans are not guaranteed to be equal. From https://docs.python.org/3/whatsnew/3.10.html :
|
|
||||||
# Hashes of NaN values of both float type and decimal.Decimal type now depend on object identity.
|
|
||||||
self.checkScript(fn, (float("nan"), float("nan")))
|
|
||||||
self.checkScript(fn, (float("nan"), float("inf")))
|
self.checkScript(fn, (float("nan"), float("inf")))
|
||||||
|
|
||||||
def test_hash_int(self):
|
def test_hash_int(self):
|
||||||
|
|
|
||||||
|
|
@ -3280,7 +3280,6 @@ class TestFX(JitTestCase):
|
||||||
.run(scripted.code)
|
.run(scripted.code)
|
||||||
|
|
||||||
@unittest.skipIf(IS_WINDOWS, "Python Windows bug? https://bugs.python.org/issue45108")
|
@unittest.skipIf(IS_WINDOWS, "Python Windows bug? https://bugs.python.org/issue45108")
|
||||||
@unittest.skipIf(sys.version_info >= (3, 10), "Does not work on Python-3.10")
|
|
||||||
def test_assert(self):
|
def test_assert(self):
|
||||||
def f(x):
|
def f(x):
|
||||||
assert x > 1
|
assert x > 1
|
||||||
|
|
@ -4056,7 +4055,7 @@ class TestFunctionalTracing(JitTestCase):
|
||||||
|
|
||||||
def functional_test(self):
|
def functional_test(self):
|
||||||
if func_name in self.UNTRACEABLE_FUNCTIONALS_PY38 and \
|
if func_name in self.UNTRACEABLE_FUNCTIONALS_PY38 and \
|
||||||
sys.version_info >= (3, 8) and sys.version_info < (3, 11):
|
sys.version_info >= (3, 8) and sys.version_info < (3, 10):
|
||||||
exc, err = self.UNTRACEABLE_FUNCTIONALS_PY38[func_name]
|
exc, err = self.UNTRACEABLE_FUNCTIONALS_PY38[func_name]
|
||||||
with self.assertRaisesRegex(exc, err):
|
with self.assertRaisesRegex(exc, err):
|
||||||
symbolic_trace(fn)
|
symbolic_trace(fn)
|
||||||
|
|
|
||||||
|
|
@ -7442,9 +7442,6 @@ a")
|
||||||
# tensor from empty list is type float in python and annotated type in torchscript
|
# tensor from empty list is type float in python and annotated type in torchscript
|
||||||
if "annotate" in li and "dtype" not in option:
|
if "annotate" in li and "dtype" not in option:
|
||||||
continue
|
continue
|
||||||
# Skip unsigned tensor initializaton for signed values on 3.10
|
|
||||||
if sys.version_info[:2] >= (3, 10) and "torch.uint8" in option and "-" in li:
|
|
||||||
continue
|
|
||||||
code = tensor_template.format(list_create=li, tensor_op=op, options=option)
|
code = tensor_template.format(list_create=li, tensor_op=op, options=option)
|
||||||
scope = {}
|
scope = {}
|
||||||
exec(code, globals(), scope)
|
exec(code, globals(), scope)
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue
Block a user