StreamingDataloader: Add profiling support #9533
Workflow file for this run
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Docs builds | |
# https://github.com/marketplace/actions/sphinx-build | |
on: | |
push: | |
branches: ["master", "release/*"] | |
tags: ["*"] | |
pull_request: | |
branches: ["master", "release/*"] | |
types: [opened, reopened, ready_for_review, synchronize] # added `ready_for_review` since draft is skipped | |
paths: | |
- ".actions/*" | |
- ".github/workflows/docs-build.yml" | |
- "requirements/ci.txt" | |
- "docs/**" | |
- "_notebooks" | |
- "requirements/**" | |
- "src/lightning/app/**" | |
- "src/lightning_app/*" | |
- "src/lightning/fabric/**" | |
- "src/lightning_fabric/*" | |
- "src/lightning/pytorch/**" | |
- "src/pytorch_lightning/*" | |
- "setup.py" | |
- "pyproject.toml" # includes metadata used in the package creation | |
- "!*.md" | |
- "!**/*.md" | |
workflow_dispatch: | |
inputs: | |
checkout: | |
description: "checkout specific git state" | |
type: string | |
required: false | |
default: "" | |
version: | |
description: "specify a version / tag to be uploaded" | |
type: string | |
required: true | |
concurrency: | |
group: ${{ github.workflow }}-${{ github.ref }}-${{ github.head_ref }} | |
cancel-in-progress: ${{ github.event_name == 'pull_request' }} | |
defaults: | |
run: | |
shell: bash | |
env: | |
FREEZE_REQUIREMENTS: "1" | |
TORCH_URL: "https://download.pytorch.org/whl/cpu/torch_stable.html" | |
PYPI_CACHE_DIR: "_pip-wheels" | |
PYPI_LOCAL_DIR: "pypi_pkgs/" | |
jobs: | |
docs-make: | |
if: github.event.pull_request.draft == false | |
runs-on: ubuntu-20.04 | |
strategy: | |
fail-fast: false | |
matrix: | |
pkg-name: ["app", "fabric", "pytorch"] | |
target: ["html", "doctest", "linkcheck"] | |
env: | |
DOCS_COPY_NOTEBOOKS: 1 | |
ARTIFACT_DAYS: 0 | |
steps: | |
- uses: actions/checkout@v4 | |
with: | |
ref: ${{ inputs.checkout }} | |
# only Pytorch has/uses notebooks | |
submodules: ${{ matrix.pkg-name == 'pytorch' }} | |
- uses: actions/setup-python@v5 | |
with: | |
python-version: "3.9" | |
- name: Pull sphinx template | |
run: | | |
pip install -q -r requirements/ci.txt | |
aws s3 sync --no-sign-request s3://sphinx-packages/ ${PYPI_LOCAL_DIR} | |
pip install lai-sphinx-theme -U -f ${PYPI_LOCAL_DIR} | |
- name: pip wheels cache | |
uses: actions/cache/restore@v3 | |
with: | |
path: ${{ env.PYPI_CACHE_DIR }} | |
key: pypi_wheels | |
- name: Install pandoc & texlive | |
if: ${{ matrix.pkg-name == 'pytorch' }} | |
timeout-minutes: 5 | |
run: | | |
sudo apt-get update --fix-missing | |
sudo apt-get install -y pandoc texlive-latex-extra dvipng texlive-pictures | |
- name: Install package & dependencies | |
timeout-minutes: 20 | |
run: | | |
mkdir -p ${PYPI_CACHE_DIR} # in case cache was not hit | |
ls -lh ${PYPI_CACHE_DIR} | |
pip install .[all] -U -r requirements/${{ matrix.pkg-name }}/docs.txt \ | |
-f ${PYPI_LOCAL_DIR} -f ${PYPI_CACHE_DIR} -f ${TORCH_URL} | |
pip list | |
- name: Install req. for Notebooks/tutorials | |
if: matrix.pkg-name == 'pytorch' | |
timeout-minutes: 10 | |
run: pip install -q -r _notebooks/.actions/requires.txt | |
- name: Full build for deployment | |
if: github.event_name != 'pull_request' | |
run: echo "DOCS_FETCH_ASSETS=1" >> $GITHUB_ENV | |
- name: Make ${{ matrix.target }} | |
working-directory: ./docs/source-${{ matrix.pkg-name }} | |
run: make ${{ matrix.target }} --debug --jobs $(nproc) SPHINXOPTS="-W --keep-going" | |
- name: Keep artifact | |
if: github.event_name == 'pull_request' | |
run: echo "ARTIFACT_DAYS=7" >> $GITHUB_ENV | |
- name: Upload built docs | |
if: ${{ matrix.target == 'html' }} | |
uses: actions/upload-artifact@v3 | |
with: | |
name: docs-${{ matrix.pkg-name }}-${{ github.sha }} | |
path: docs/build/html/ | |
retention-days: ${{ env.ARTIFACT_DAYS }} | |
- name: Dump handy wheels | |
if: github.event_name == 'push' && github.ref == 'refs/heads/master' | |
continue-on-error: true | |
uses: ./.github/actions/pip-wheels | |
with: | |
wheel-dir: ${{ env.PYPI_CACHE_DIR }} | |
torch-url: ${{ env.TORCH_URL }} | |
cache-key: "pypi_wheels" | |
deploy-docs: | |
needs: docs-make | |
if: github.repository_owner == 'Lightning-AI' && github.event_name != 'pull_request' | |
runs-on: ubuntu-20.04 | |
strategy: | |
fail-fast: false | |
matrix: | |
pkg-name: ["app", "fabric", "pytorch"] | |
env: | |
GCP_TARGET: "gs://lightning-docs-${{ matrix.pkg-name }}" | |
# use input if dispatch or git tag | |
VERSION: ${{ inputs.version || github.ref_name }} | |
steps: | |
- uses: actions/download-artifact@v3 | |
with: | |
name: docs-${{ matrix.pkg-name }}-${{ github.sha }} | |
path: docs/build/html/ | |
- name: Authenticate to Google Cloud | |
uses: google-github-actions/auth@v2 | |
with: | |
credentials_json: ${{ secrets.GCS_SA_KEY }} | |
- name: Setup gcloud | |
uses: google-github-actions/setup-gcloud@v2 | |
with: | |
project_id: ${{ secrets.GCS_PROJECT }} | |
# Uploading docs to GCS, so they can be served on lightning.ai | |
- name: Upload docs/${{ matrix.pkg-name }}/stable to GCS 🪣 | |
if: startsWith(github.ref, 'refs/heads/release/') && github.event_name == 'push' | |
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/stable | |
# Uploading docs to GCS, so they can be served on lightning.ai | |
- name: Upload docs/${{ matrix.pkg-name }}/latest to GCS 🪣 | |
if: github.ref == 'refs/heads/master' && github.event_name == 'push' | |
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/latest | |
# Uploading docs to GCS, so they can be served on lightning.ai | |
- name: Upload docs/${{ matrix.pkg-name }}/release to GCS 🪣 | |
if: startsWith(github.ref, 'refs/tags/') || github.event_name == 'workflow_dispatch' | |
run: gsutil -m rsync -d -R docs/build/html/ ${GCP_TARGET}/${{ env.VERSION }} | |
# Uploading docs as archive to GCS, so they can be as backup | |
- name: Upload docs as archive to GCS 🪣 | |
if: startsWith(github.ref, 'refs/tags/') || github.event_name == 'workflow_dispatch' | |
working-directory: docs/build | |
run: | | |
zip ${{ env.VERSION }}.zip -r html/ | |
gsutil cp ${{ env.VERSION }}.zip ${GCP_TARGET} |