diff --git a/SETUP.md b/SETUP.md index 3c5f84240f..e9f58560b4 100644 --- a/SETUP.md +++ b/SETUP.md @@ -42,7 +42,7 @@ Another alternative is to run all the recommender utilities directly from a loca ### Requirements * A machine running Linux, MacOS or Windows -* An optional requirement is Anaconda with Python version >= 3.6 +* An optional requirement is Anaconda with Python version >= 3.6, <= 3.9 * This is pre-installed on Azure DSVM such that one can run the following steps directly. To setup on your local machine, [Miniconda](https://docs.conda.io/en/latest/miniconda.html) is a quick way to get started. Alternatively a [virtual environment](#using-a-virtual-environment) can be used instead of Anaconda. diff --git a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_cpu.yml b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_cpu.yml index c1182a5d7e..8c7a17d69b 100644 --- a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_cpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_cpu.yml @@ -32,7 +32,7 @@ extends: task_name: "Test - Nightly Linux CPU" timeout: 180 conda_env: "nightly_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[examples,dev,experimental] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "not spark and not gpu" pytest_params: "-x" diff --git a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_gpu.yml b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_gpu.yml index 02c3270b5a..0da58a8176 100644 --- a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_gpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_gpu.yml @@ -31,7 +31,7 @@ extends: task_name: "Test - Nightly Linux GPU" timeout: 420 conda_env: "nightly_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,examples,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "not spark and gpu" pytest_params: "-x" diff --git a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_pyspark.yml b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_pyspark.yml index f542f059ff..5087d3caa4 100644 --- a/tests/ci/azure_pipeline_test/dsvm_nightly_linux_pyspark.yml +++ b/tests/ci/azure_pipeline_test/dsvm_nightly_linux_pyspark.yml @@ -32,7 +32,7 @@ extends: task_name: "Test - Nightly Linux Spark" timeout: 180 conda_env: "nightly_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,examples,dev]" pytest_markers: "spark and not gpu" pytest_params: "-x" diff --git a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_cpu.yml b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_cpu.yml index d1efb71754..7118cb34b1 100644 --- a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_cpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_cpu.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Notebook Linux CPU" conda_env: "unit_notebook_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[examples,dev,experimental] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "notebooks and not spark and not gpu" diff --git a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_gpu.yml b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_gpu.yml index c6c07525ce..98e9a6ab83 100644 --- a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_gpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_gpu.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Notebook Linux GPU" conda_env: "unit_notebook_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,examples,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "notebooks and not spark and gpu" diff --git a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_pyspark.yml b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_pyspark.yml index 31d699588d..3abe815aa1 100644 --- a/tests/ci/azure_pipeline_test/dsvm_notebook_linux_pyspark.yml +++ b/tests/ci/azure_pipeline_test/dsvm_notebook_linux_pyspark.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Notebook Linux Spark" conda_env: "unit_notebook_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,examples,dev]" pytest_markers: "notebooks and spark and not gpu" diff --git a/tests/ci/azure_pipeline_test/dsvm_unit_linux_cpu.yml b/tests/ci/azure_pipeline_test/dsvm_unit_linux_cpu.yml index f5c4057f29..eac67ec5ef 100644 --- a/tests/ci/azure_pipeline_test/dsvm_unit_linux_cpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_unit_linux_cpu.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Linux CPU" conda_env: "unit_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[dev,experimental] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "not notebooks and not spark and not gpu" diff --git a/tests/ci/azure_pipeline_test/dsvm_unit_linux_gpu.yml b/tests/ci/azure_pipeline_test/dsvm_unit_linux_gpu.yml index 5f166b87fd..7331eae4e6 100644 --- a/tests/ci/azure_pipeline_test/dsvm_unit_linux_gpu.yml +++ b/tests/ci/azure_pipeline_test/dsvm_unit_linux_gpu.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Linux GPU" conda_env: "unit_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "not notebooks and not spark and gpu" diff --git a/tests/ci/azure_pipeline_test/dsvm_unit_linux_pyspark.yml b/tests/ci/azure_pipeline_test/dsvm_unit_linux_pyspark.yml index 1f3006a05e..d32a1930fb 100644 --- a/tests/ci/azure_pipeline_test/dsvm_unit_linux_pyspark.yml +++ b/tests/ci/azure_pipeline_test/dsvm_unit_linux_pyspark.yml @@ -59,6 +59,6 @@ extends: - unit task_name: "Test - Unit Linux Spark" conda_env: "unit_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,dev]" pytest_markers: "not notebooks and spark and not gpu" diff --git a/tests/ci/azure_pipeline_test/release_pipeline.yml b/tests/ci/azure_pipeline_test/release_pipeline.yml index a37d9a549c..cad472cd3c 100644 --- a/tests/ci/azure_pipeline_test/release_pipeline.yml +++ b/tests/ci/azure_pipeline_test/release_pipeline.yml @@ -22,7 +22,7 @@ jobs: - unit task_name: "Test - Unit Linux CPU" conda_env: "release_unit_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[experimental,dev] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "not notebooks and not spark and not gpu" install: "release" @@ -34,7 +34,7 @@ jobs: - unit task_name: "Test - Unit Notebook Linux CPU" conda_env: "release_unit_notebook_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[experimental,examples,dev] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "notebooks and not spark and not gpu" install: "release" @@ -45,7 +45,7 @@ jobs: - unit task_name: "Test - Unit Linux GPU" conda_env: "release_unit_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "not notebooks and not spark and gpu" install: "release" @@ -56,7 +56,7 @@ jobs: - unit task_name: "Test - Unit Notebook Linux GPU" conda_env: "release_unit_notebook_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,examples,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "notebooks and not spark and gpu" install: "release" @@ -67,7 +67,7 @@ jobs: - unit task_name: "Test - Unit Linux Spark" conda_env: "release_unit_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,dev]" pytest_markers: "not notebooks and spark and not gpu" install: "release" @@ -78,7 +78,7 @@ jobs: - unit task_name: "Test - Unit Notebook Linux Spark" conda_env: "release_unit_notebook_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,examples,dev]" pytest_markers: "notebooks and spark and not gpu" install: "release" @@ -92,7 +92,7 @@ jobs: task_name: "Test - Nightly Linux CPU" timeout: 180 conda_env: "release_nightly_linux_cpu" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[experimental,examples,dev] 'scikit-surprise@https://github.com/NicolasHug/Surprise/archive/refs/tags/v1.1.1.tar.gz' 'pymanopt@https://github.com/pymanopt/pymanopt/archive/fb36a272cdeecb21992cfd9271eb82baafeb316d.zip' --no-cache --no-binary scikit-surprise" pytest_markers: "not spark and not gpu" install: "release" @@ -105,7 +105,7 @@ jobs: task_name: "Test - Nightly Linux GPU" timeout: 360 conda_env: "release_nightly_linux_gpu" - conda_opts: "-c conda-forge python=3.6 cudatoolkit=11.2 cudnn=8.1" + conda_opts: "-c conda-forge python=3.7 cudatoolkit=11.2 cudnn=8.1" pip_opts: "[gpu,examples,dev] -f https://download.pytorch.org/whl/cu111/torch_stable.html" pytest_markers: "not spark and gpu" install: "release" @@ -118,7 +118,7 @@ jobs: task_name: "Test - Nightly Linux Spark" timeout: 180 conda_env: "release_nightly_linux_spark" - conda_opts: "python=3.6" + conda_opts: "python=3.7" pip_opts: "[spark,examples,dev]" pytest_markers: "spark and not gpu" install: "release" diff --git a/tests/ci/cpu_unit_tests.yml b/tests/ci/cpu_unit_tests.yml index 5432593e65..14aec98ea6 100644 --- a/tests/ci/cpu_unit_tests.yml +++ b/tests/ci/cpu_unit_tests.yml @@ -39,9 +39,9 @@ steps: - task: UsePythonVersion@0 inputs: - versionSpec: '3.6' + versionSpec: '3.7' architecture: 'x64' - displayName: 'Use Python 3.6' + displayName: 'Use Python 3.7' - script: | az login --service-principal -u $(ClientID) -p $(ClientSecret) --tenant $(TenantID) diff --git a/tests/ci/env-setup.yml b/tests/ci/env-setup.yml index 210833f59a..17d807fd63 100644 --- a/tests/ci/env-setup.yml +++ b/tests/ci/env-setup.yml @@ -7,9 +7,9 @@ steps: - task: UsePythonVersion@0 inputs: - versionSpec: '3.6' + versionSpec: '3.7' architecture: 'x64' - displayName: 'Use Python 3.6' + displayName: 'Use Python 3.7' - bash: | echo "##vso[task.prependpath]/data/anaconda/bin" diff --git a/tests/ci/gpu_unit_test.yml b/tests/ci/gpu_unit_test.yml index e59bad1ccf..c44c3bd633 100644 --- a/tests/ci/gpu_unit_test.yml +++ b/tests/ci/gpu_unit_test.yml @@ -40,9 +40,9 @@ steps: - task: UsePythonVersion@0 inputs: - versionSpec: '3.6' + versionSpec: '3.7' architecture: 'x64' - displayName: 'Use Python 3.6' + displayName: 'Use Python 3.7' - script: | az login --service-principal -u $(ClientID) -p $(ClientSecret) --tenant $(TenantID) diff --git a/tools/generate_conda_file.py b/tools/generate_conda_file.py index 3dfd5df288..23fd2a3fec 100644 --- a/tools/generate_conda_file.py +++ b/tools/generate_conda_file.py @@ -35,7 +35,7 @@ CHANNELS = ["defaults", "conda-forge", "pytorch", "fastai"] CONDA_BASE = { - "python": "python>=3.6,<3.8", + "python": "python>=3.6,<3.10", "bottleneck": "bottleneck==1.2.1", "cornac": "cornac>=1.11.0", "ipykernel": "ipykernel>=4.6.1",