Skip to content

Fix UnboundLocalError when using the predict method with return_predictions=False. #12163

Fix UnboundLocalError when using the predict method with return_predictions=False.

Fix UnboundLocalError when using the predict method with return_predictions=False. #12163

Workflow file for this run

name: Docs builds
# https://github.com/marketplace/actions/sphinx-build
on:
push:
branches: ["master", "release/*"]
tags: ["*"]
pull_request:
branches: ["master", "release/*"]
types: [opened, reopened, ready_for_review, synchronize] # added `ready_for_review` since draft is skipped
paths:
- ".actions/*"
- ".github/workflows/docs-build.yml"
- "requirements/ci.txt"
- "docs/**"
- "_notebooks"
- "requirements/**"
- "src/lightning/fabric/**"
- "src/lightning_fabric/*"
- "src/lightning/pytorch/**"
- "src/pytorch_lightning/*"
- "setup.py"
- "pyproject.toml" # includes metadata used in the package creation
- "!*.md"
- "!**/*.md"
workflow_dispatch:
inputs:
checkout:
description: "checkout specific git state"
type: string
required: false
default: ""
version:
description: "specify a version / tag to be uploaded"
type: string
required: true
concurrency:
group: ${{ github.workflow }}-${{ github.ref }}-${{ github.head_ref }}
cancel-in-progress: ${{ github.event_name == 'pull_request' }}
defaults:
run:
shell: bash
env:
FREEZE_REQUIREMENTS: "1"
TORCH_URL: "https://download.pytorch.org/whl/cpu/torch_stable.html"
PYPI_CACHE_DIR: "_pip-wheels"
PYPI_LOCAL_DIR: "pypi_pkgs/"
jobs:
docs-make:
if: github.event.pull_request.draft == false
runs-on: ubuntu-20.04
strategy:
fail-fast: false
matrix:
pkg-name: ["fabric", "pytorch"]
target: ["html", "doctest", "linkcheck"]
env:
DOCS_COPY_NOTEBOOKS: 1
PIN_RELEASE_VERSIONS: 1
ARTIFACT_DAYS: 0
steps:
- uses: actions/checkout@v4
with:
ref: ${{ inputs.checkout }}
# only Pytorch has/uses notebooks
submodules: ${{ matrix.pkg-name == 'pytorch' }}
lfs: ${{ matrix.pkg-name == 'pytorch' }}
- uses: actions/setup-python@v5
with:
python-version: "3.9"
- name: List notebooks
if: ${{ matrix.pkg-name == 'pytorch' }}
working-directory: _notebooks/
run: |
pip install -q py-tree
py-tree .notebooks/
ls -lhR .notebooks/
- name: Pull sphinx template
run: |
pip install -q -r requirements/ci.txt
aws s3 sync --no-sign-request s3://sphinx-packages/ ${PYPI_LOCAL_DIR}
pip install lai-sphinx-theme -U -f ${PYPI_LOCAL_DIR}
- name: pip wheels cache
uses: actions/cache/restore@v4
with:
path: ${{ env.PYPI_CACHE_DIR }}
key: pypi_wheels
- name: Install pandoc & texlive
if: ${{ matrix.pkg-name == 'pytorch' }}
timeout-minutes: 5
run: |
sudo apt-get update --fix-missing
sudo apt-get install -y pandoc
- name: Install package & dependencies
timeout-minutes: 20
run: |
mkdir -p ${PYPI_CACHE_DIR} # in case cache was not hit
ls -lh ${PYPI_CACHE_DIR}
pip install .[all] -U -r requirements/${{ matrix.pkg-name }}/docs.txt \
-f ${PYPI_LOCAL_DIR} -f ${PYPI_CACHE_DIR} -f ${TORCH_URL}
pip list
- name: Install req. for Notebooks/tutorials
if: matrix.pkg-name == 'pytorch'
timeout-minutes: 10
run: pip install -q -r _notebooks/.actions/requires.txt
- name: Full build for deployment
if: github.event_name != 'pull_request'
run: echo "DOCS_FETCH_ASSETS=1" >> $GITHUB_ENV
- name: Build without warnings
if: github.event_name != 'workflow_dispatch'
run: echo "BUILD_SPHINX_OPTS=-W --keep-going" >> $GITHUB_ENV
- name: Make ${{ matrix.target }}
working-directory: ./docs/source-${{ matrix.pkg-name }}
# allow failing link check and doctest if you run with dispatch
continue-on-error: ${{ (matrix.target == 'doctest' || matrix.target == 'linkcheck') && github.event_name == 'workflow_dispatch' }}
run: make ${{ matrix.target }} --debug --jobs $(nproc) SPHINXOPTS="$BUILD_SPHINX_OPTS"
- name: Keep artifact
if: github.event_name == 'pull_request'
run: echo "ARTIFACT_DAYS=7" >> $GITHUB_ENV
- name: Upload built docs
if: ${{ matrix.target == 'html' }}
uses: actions/upload-artifact@v3
with:
name: docs-${{ matrix.pkg-name }}-${{ github.sha }}
path: docs/build/html/
retention-days: ${{ env.ARTIFACT_DAYS }}
include-hidden-files: true
#- name: Dump handy wheels
# if: github.event_name == 'push' && github.ref == 'refs/heads/master'
# continue-on-error: true
# uses: ./.github/actions/pip-wheels
# with:
# wheel-dir: ${{ env.PYPI_CACHE_DIR }}
# torch-url: ${{ env.TORCH_URL }}
# cache-key: "pypi_wheels"
deploy-docs:
needs: docs-make
if: github.repository_owner == 'Lightning-AI' && github.event_name != 'pull_request'
runs-on: ubuntu-20.04
strategy:
fail-fast: false
matrix:
pkg-name: ["fabric", "pytorch"]
env:
GCP_TARGET: "gs://lightning-docs-${{ matrix.pkg-name }}"
# use input if dispatch or git tag
VERSION: ${{ inputs.version || github.ref_name }}
steps:
- uses: actions/download-artifact@v3
with:
name: docs-${{ matrix.pkg-name }}-${{ github.sha }}
path: docs/build/html/
- name: Authenticate to Google Cloud
uses: google-github-actions/auth@v2
with:
credentials_json: ${{ secrets.GCS_SA_KEY }}
- name: Setup gcloud
uses: google-github-actions/setup-gcloud@v2
with:
project_id: ${{ secrets.GCS_PROJECT }}
# Uploading docs to GCS, so they can be served on lightning.ai
- name: Upload docs/${{ matrix.pkg-name }}/stable to GCS 🪣
if: startsWith(github.ref, 'refs/heads/release/') && github.event_name == 'push'
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/stable
# Uploading docs to GCS, so they can be served on lightning.ai
- name: Upload docs/${{ matrix.pkg-name }}/latest to GCS 🪣
if: github.ref == 'refs/heads/master' && github.event_name == 'push'
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/latest
# Uploading docs to GCS, so they can be served on lightning.ai
- name: Upload docs/${{ matrix.pkg-name }}/release to GCS 🪣
if: startsWith(github.ref, 'refs/tags/') || github.event_name == 'workflow_dispatch'
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/${{ env.VERSION }}
# Uploading docs as archive to GCS, so they can be as backup
- name: Upload docs as archive to GCS 🪣
if: startsWith(github.ref, 'refs/tags/') || github.event_name == 'workflow_dispatch'
working-directory: docs/build
run: |
zip ${{ env.VERSION }}.zip -r html/
gsutil cp ${{ env.VERSION }}.zip ${GCP_TARGET}