From 897d541dba332b87e16f3ee9cd93149342784955 Mon Sep 17 00:00:00 2001 From: John Freeman Date: Mon, 8 Jul 2024 13:54:16 -0500 Subject: [PATCH] JCF: with removal of SL7 support, drop the SL7 builds, etc. --- .../workflows/build-candidate-release-sl7.yml | 193 ------------ .../workflows/build-frozen-release-sl7.yml | 195 ------------ .../workflows/build-nightly-release-sl7.yml | 273 ----------------- .github/workflows/build-v4-release-sl7.yml | 277 ------------------ .github/workflows/slim_externals.yaml | 156 ---------- .github/workflows/test-single-pkg.yml | 1 + docs/README.md | 4 - 7 files changed, 1 insertion(+), 1098 deletions(-) delete mode 100644 .github/workflows/build-candidate-release-sl7.yml delete mode 100644 .github/workflows/build-frozen-release-sl7.yml delete mode 100644 .github/workflows/build-nightly-release-sl7.yml delete mode 100644 .github/workflows/build-v4-release-sl7.yml diff --git a/.github/workflows/build-candidate-release-sl7.yml b/.github/workflows/build-candidate-release-sl7.yml deleted file mode 100644 index 6ab5d868f..000000000 --- a/.github/workflows/build-candidate-release-sl7.yml +++ /dev/null @@ -1,193 +0,0 @@ -name: SL7 build v4 or v5 candidate release - -on: - workflow_dispatch: - inputs: - base-release: - description: 'targeted base release (e.g. v4.4.0)' - required: true - det-release: - description: 'targeted detector release (e.g. v4.3.0)' - required: true - det: - description: 'detector type (nd or fd)' - required: true - build-number: - description: 'candidate release number, e.g. 1' - required: true - -jobs: - build_the_candidate_release_spack: - name: build_dev_release_spack - runs-on: daq - container: - image: ghcr.io/dune-daq/sl7-slim-externals:v1.1 - defaults: - run: - shell: bash - - steps: - - name: Checkout daq-release - uses: actions/checkout@v3 - with: - path: daq-release - - - name: setup directories and install spack for the base release - run: | - export DET=${{ github.event.inputs.det }} - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/candidates/coredaq-${{ github.event.inputs.base-release }}-rc${{ github.event.inputs.build-number }}-c8 - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/candidates/${DET}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/coredaq/coredaq-${{ github.event.inputs.base-release }}/release.yaml -a -c -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR core $OS - - cd $BASE_RELEASE_DIR/.. - tar_and_stage_release $BASE_RELEASE_TAG - - - name: upload spack candidate release tarball for base release - uses: actions/upload-artifact@v3 - with: - name: candidates_coredaq - path: ${{ github.workspace }}/tarballs_for_upload/coredaq-${{ github.event.inputs.base-release }}-rc${{ github.event.inputs.build-number }}-c8.tar.gz - - - name: setup directories and install spack for the detector release - run: | - export DET=${{ github.event.inputs.det }} - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/candidates/coredaq-${{ github.event.inputs.base-release }}-rc${{ github.event.inputs.build-number }}-c8 - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/candidates/${DET}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/fddaq/fddaq-${{ github.event.inputs.det-release }}/release.yaml -a -c -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR $DET $OS - - cd $DET_RELEASE_DIR/.. - tar_and_stage_release $DET_RELEASE_TAG - - - name: upload spack candidate release tarball for detector release - uses: actions/upload-artifact@v3 - with: - name: candidates_${{ github.event.inputs.det }}daq - path: ${{ github.workspace }}/tarballs_for_upload/${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8.tar.gz - - update_image: - name: update_spack_image_rc - runs-on: ubuntu-latest - strategy: - matrix: - include: - - input_image: "ghcr.io/dune-daq/sl7-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/candidate-release-sl7" - tag: ${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - - input_image: "ghcr.io/dune-daq/c8-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/candidate-release-c8" - tag: ${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - needs: build_the_candidate_release_spack - environment: dockerhub - permissions: - packages: write - contents: read - - steps: - - name: clean docker-build - run: | - mkdir -p ${GITHUB_WORKSPACE}/docker-build - rm -rf ${GITHUB_WORKSPACE}/docker-build/* - - - name: Download spack candidate release tarball artifact for base release - uses: actions/download-artifact@v3 - with: - name: candidates_coredaq - path: ${{ github.workspace }}/docker-build - - - name: Download spack candidate release tarball artifact for detector release - uses: actions/download-artifact@v3 - with: - name: candidates_${{ github.event.inputs.det }}daq - path: ${{ github.workspace }}/docker-build - - - name: prepare cvmfs mirror spack-candidates - run: | - cd ${{ github.workspace }}/docker-build - mkdir -p candidates - - cd candidates - base_tag=coredaq-${{ github.event.inputs.base-release }}-rc${{ github.event.inputs.build-number }}-c8 - ls .. - tar xf ../${base_tag}.tar.gz - rm -rf ../${base_tag}.tar.gz - - tar xf ../${{ matrix.tag }}.tar.gz - rm -rf ../${{ matrix.tag }}.tar.gz - - - cd .. - echo "FROM "${{ matrix.input_image }} > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_FOR "${{ matrix.tag }} >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/pypi-repo:latest /cvmfs/dunedaq.opensciencegrid.org/pypi-repo /cvmfs/dunedaq.opensciencegrid.org/pypi-repo" >> Dockerfile - echo "ADD candidates /cvmfs/dunedaq-development.opensciencegrid.org/candidates" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Log in to the Container registry - uses: docker/login-action@v3 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v5 - with: - images: | - ${{ matrix.output_image }} - tags: | - type=raw,value=${{ matrix.tag }} - - - name: Build and push Docker images - uses: docker/build-push-action@v5 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - - generate_dbt_setup_release_env: - name: generate_dbt_setup_release_env - runs-on: ubuntu-latest - needs: update_image - container: - image: ghcr.io/dune-daq/candidate-release-sl7:${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - defaults: - run: - shell: bash - - steps: - - - name: create dbt-setup-release-env.sh and daq_app_rte.sh for detector release - run: | - DET=${{ github.event.inputs.det }} - - source /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh - setup_dbt latest || true - dbt-setup-release -b candidate ${DET}daq-${{ github.event.inputs.det-release }}-rc${{ github.event.inputs.build-number }}-c8 - declare -x > ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - declare -f >> ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - egrep "declare -x (PATH|.*_SHARE|CET_PLUGIN_PATH|DUNEDAQ_SHARE_PATH|LD_LIBRARY_PATH|LIBRARY_PATH|PYTHONPATH)=" ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh > ${GITHUB_WORKSPACE}/${DET}daq_app_rte.sh - - - - name: upload ${{ github.event.inputs.det }}daq-dbt-setup-release-env.sh - uses: actions/upload-artifact@v3 - with: - name: ${{ github.event.inputs.det }}daq-dbt_setup_release_env - path: ${{ github.workspace }}/${{ github.event.inputs.det }}daq-dbt-setup-release-env.sh - - - name: upload ${{ github.event.inputs.det }}daq_app_rte.sh - uses: actions/upload-artifact@v3 - with: - name: ${{ github.event.inputs.det }}daq_app_rte - path: ${{ github.workspace }}/${{ github.event.inputs.det }}daq_app_rte.sh diff --git a/.github/workflows/build-frozen-release-sl7.yml b/.github/workflows/build-frozen-release-sl7.yml deleted file mode 100644 index 048fa44b0..000000000 --- a/.github/workflows/build-frozen-release-sl7.yml +++ /dev/null @@ -1,195 +0,0 @@ -name: SL7 build v4 or v5 frozen release - - -on: - workflow_dispatch: - inputs: - base-release: - description: 'targeted base release (e.g. v4.4.0)' - required: true - det-release: - description: 'targeted detector release (e.g. v4.3.0)' - required: true - det: - description: 'detector type (nd or fd)' - required: true - build-number: - description: 'frozen release number, e.g. 1' - required: true - -jobs: - build_the_frozen_release_spack: - name: build_frozen_release_spack - runs-on: daq - container: - image: ghcr.io/dune-daq/sl7-slim-externals:v1.1 - defaults: - run: - shell: bash - - steps: - - name: Checkout daq-release - uses: actions/checkout@v3 - with: - path: daq-release - - - name: setup directories and install spack for the base release - run: | - export DET=${{ github.event.inputs.det }} - export BASE_RELEASE_DIR=/cvmfs/dunedaq.opensciencegrid.org/spack/releases/coredaq-${{ github.event.inputs.base-release }}-c8-${{ github.event.inputs.build-number }} - export DET_RELEASE_DIR=/cvmfs/dunedaq.opensciencegrid.org/spack/releases/${DET}daq-${{ github.event.inputs.det-release }}-c8-${{ github.event.inputs.build-number }} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/coredaq/coredaq-${{ github.event.inputs.base-release }}/release.yaml -a -c -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR core $OS - - cd $BASE_RELEASE_DIR/.. - tar_and_stage_release $BASE_RELEASE_TAG - - - name: upload spack frozen release tarball for base release - uses: actions/upload-artifact@v3 - with: - name: frozens_coredaq - path: ${{ github.workspace }}/tarballs_for_upload/coredaq-${{ github.event.inputs.base-release }}-c8-${{ github.event.inputs.build-number }}.tar.gz - - - name: setup directories and install spack for the detector release - run: | - export DET=${{ github.event.inputs.det }} - export BASE_RELEASE_DIR=/cvmfs/dunedaq.opensciencegrid.org/spack/releases/coredaq-${{ github.event.inputs.base-release }}-c8-${{ github.event.inputs.build-number }} - export DET_RELEASE_DIR=/cvmfs/dunedaq.opensciencegrid.org/spack/releases/${DET}daq-${{ github.event.inputs.det-release }}-c8-${{ github.event.inputs.build-number }} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/fddaq/fddaq-${{ github.event.inputs.det-release }}/release.yaml -a -c -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR $DET $OS - - cd $DET_RELEASE_DIR/.. - tar_and_stage_release $DET_RELEASE_TAG - - - - name: upload spack frozen release tarball for detector release - uses: actions/upload-artifact@v3 - with: - name: frozens_${{ github.event.inputs.det }}daq - path: ${{ github.workspace }}/tarballs_for_upload/${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-c8-${{ github.event.inputs.build-number }}.tar.gz - - update_image: - name: update_spack_image_rc - runs-on: ubuntu-latest - strategy: - matrix: - include: - - input_image: "ghcr.io/dune-daq/sl7-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/frozen-release-sl7" - tag: ${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-c8 - - input_image: "ghcr.io/dune-daq/c8-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/frozen-release-c8" - tag: ${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-c8 - needs: build_the_frozen_release_spack - environment: dockerhub - permissions: - packages: write - contents: read - - steps: - - name: clean docker-build - run: | - mkdir -p ${GITHUB_WORKSPACE}/docker-build - rm -rf ${GITHUB_WORKSPACE}/docker-build/* - - - name: Download spack frozen release tarball artifact for base release - uses: actions/download-artifact@v3 - with: - name: frozens_coredaq - path: ${{ github.workspace }}/docker-build - - - name: Download spack frozen release tarball artifact for detector release - uses: actions/download-artifact@v3 - with: - name: frozens_${{ github.event.inputs.det }}daq - path: ${{ github.workspace }}/docker-build - - - name: prepare cvmfs mirror spack-frozens - run: | - cd ${{ github.workspace }}/docker-build - mkdir -p releases - - cd releases - base_tag=coredaq-${{ github.event.inputs.base-release }}-c8 - tar xf ../${base_tag}-${{ github.event.inputs.build-number }}.tar.gz - rm -rf ../${base_tag}-${{ github.event.inputs.build-number }}.tar.gz - - tar xf ../${{ matrix.tag }}-${{ github.event.inputs.build-number }}.tar.gz - rm -rf ../${{ matrix.tag }}-${{ github.event.inputs.build-number }}.tar.gz - ln -s ${{ matrix.tag }}-${{ github.event.inputs.build-number }} ${{ matrix.tag }} - - - cd .. - echo "FROM "${{ matrix.input_image }} > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_FOR "${{ matrix.tag }} >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/pypi-repo:latest /cvmfs/dunedaq.opensciencegrid.org/pypi-repo /cvmfs/dunedaq.opensciencegrid.org/pypi-repo" >> Dockerfile - echo "ADD releases /cvmfs/dunedaq.opensciencegrid.org/spack/releases" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Log in to the Container registry - uses: docker/login-action@v3 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v5 - with: - images: | - ${{ matrix.output_image }} - tags: | - type=raw,value=${{ matrix.tag }} - - - name: Build and push Docker images - uses: docker/build-push-action@v5 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - - generate_dbt_setup_release_env: - name: generate_dbt_setup_release_env - runs-on: ubuntu-latest - needs: update_image - container: - image: ghcr.io/dune-daq/frozen-release-sl7:${{ github.event.inputs.det }}daq-${{ github.event.inputs.det-release }}-c8 - defaults: - run: - shell: bash - - steps: - - - name: create dbt-setup-release-env.sh and daq_app_rte.sh for detector release - run: | - DET=${{ github.event.inputs.det }} - - source /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh - setup_dbt latest || true - dbt-setup-release ${DET}daq-${{ github.event.inputs.det-release }}-c8 - declare -x > ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - declare -f >> ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - egrep "declare -x (PATH|.*_SHARE|CET_PLUGIN_PATH|DUNEDAQ_SHARE_PATH|LD_LIBRARY_PATH|LIBRARY_PATH|PYTHONPATH)=" ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh > ${GITHUB_WORKSPACE}/${DET}daq_app_rte.sh - - - - name: upload ${{ github.event.inputs.det }}daq-dbt-setup-release-env.sh - uses: actions/upload-artifact@v3 - with: - name: ${{ github.event.inputs.det }}daq-dbt_setup_release_env - path: ${{ github.workspace }}/${{ github.event.inputs.det }}daq-dbt-setup-release-env.sh - - - name: upload ${{ github.event.inputs.det }}daq_app_rte.sh - uses: actions/upload-artifact@v3 - with: - name: ${{ github.event.inputs.det }}daq_app_rte - path: ${{ github.workspace }}/${{ github.event.inputs.det }}daq_app_rte.sh diff --git a/.github/workflows/build-nightly-release-sl7.yml b/.github/workflows/build-nightly-release-sl7.yml deleted file mode 100644 index 7d58e3030..000000000 --- a/.github/workflows/build-nightly-release-sl7.yml +++ /dev/null @@ -1,273 +0,0 @@ -name: SL7 build v5 nightly release - -on: - schedule: - - cron: "0 3 * * 0" - - workflow_dispatch: - inputs: - tag-prefix: - description: 'nightly tag prefix' - default: '' - feature-branch: - description: 'feature branch to be used across all DAQ repos wherever possible.' - default: develop - cvmfs-deployment: - description: 'whether to deploy the release to cvmfs' - default: 'no' - -jobs: - make_nightly_tag: - name: create nightly tag - runs-on: ubuntu-latest - outputs: - tag: ${{ steps.create_nightly_tag.outputs.nightly_tag }} - defaults: - run: - shell: bash - steps: - - id: create_nightly_tag - run: | - date_tag=$(date +%y%m%d) - echo "nightly_tag=${{ github.event.inputs.tag-prefix }}_DEV_${date_tag}_C8" >> "$GITHUB_OUTPUT" - cat "$GITHUB_OUTPUT" - - build_the_develop_release_spack: - name: build_dev_release_spack - runs-on: ubuntu-latest - needs: make_nightly_tag - container: - image: ghcr.io/dune-daq/sl7-slim-externals:v1.1 - defaults: - run: - shell: bash - - steps: - - - name: Checkout daq-release - uses: actions/checkout@v3 - with: - path: daq-release - - - name: setup directories and install spack for the coredaq release - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NB${NIGHTLY_TAG} - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NFD${NIGHTLY_TAG} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/coredaq/coredaq-develop/release.yaml -a -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR core $OS ${{ github.event.inputs.feature-branch }} - - cd $BASE_RELEASE_DIR/../ - tar_and_stage_release ${BASE_RELEASE_TAG} - - - name: upload spack nightly tarball for base release - uses: actions/upload-artifact@v3 - with: - name: nightly_coredaq - path: ${{ github.workspace }}/tarballs_for_upload/NB${{needs.make_nightly_tag.outputs.tag}}.tar.gz - - - name: setup directories and install spack for the fddaq release - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NB${NIGHTLY_TAG} - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NFD${NIGHTLY_TAG} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/fddaq/fddaq-develop/release.yaml -a -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR fd $OS ${{ github.event.inputs.feature-branch }} - - cd $DET_RELEASE_DIR/../ - tar_and_stage_release ${DET_RELEASE_TAG} - - - name: upload spack nightly tarball for far detector release - uses: actions/upload-artifact@v3 - with: - name: nightly_fddaq - path: ${{ github.workspace }}/tarballs_for_upload/NFD${{needs.make_nightly_tag.outputs.tag}}.tar.gz - - update_image: - name: update_spack_image_nightly - runs-on: ubuntu-latest - strategy: - matrix: - include: - - input_image: "ghcr.io/dune-daq/sl7-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/nightly-release-sl7" - tag: "development_v5" - needs: [build_the_develop_release_spack, make_nightly_tag] - environment: dockerhub - permissions: - packages: write - contents: read - - steps: - - name: clean docker-build - run: | - mkdir -p ${GITHUB_WORKSPACE}/docker-build - rm -rf ${GITHUB_WORKSPACE}/docker-build/* - - - name: Download spack nightly release tarball artifact for base release - uses: actions/download-artifact@v3 - with: - name: nightly_coredaq - path: ${{ github.workspace }}/docker-build - - - name: Download spack nightly release tarball artifact for far detector release - uses: actions/download-artifact@v3 - with: - name: nightly_fddaq - path: ${{ github.workspace }}/docker-build - - - name: prepare cvmfs mirror spack-nightly - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - cd ${{ github.workspace }}/docker-build - mkdir -p nightly - - cd nightly - base_tag="NB${NIGHTLY_TAG}" - tar xf ../${base_tag}.tar.gz - rm -rf ../${base_tag}.tar.gz - - fddaq_tag="NFD${NIGHTLY_TAG}" - tar xf ../${fddaq_tag}.tar.gz - rm -rf ../${fddaq_tag}.tar.gz - ln -s ${fddaq_tag} last_fddaq_c8 - - cd .. - echo "FROM "${{ matrix.input_image }} > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_AT ${NIGHTLY_TAG}" >> Dockerfile - echo "ADD nightly /cvmfs/dunedaq-development.opensciencegrid.org/nightly" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Log in to the Container registry - uses: docker/login-action@v3 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v5 - with: - images: | - ${{ matrix.output_image }} - tags: | - type=raw,value=${{ matrix.tag }} - - - name: Build and push Docker images - uses: docker/build-push-action@v5 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - - generate_dbt_setup_release_env: - name: generate_dbt_setup_release_env - runs-on: ubuntu-latest - needs: update_image - container: - image: ghcr.io/dune-daq/nightly-release-sl7:development_v5 - defaults: - run: - shell: bash - - steps: - - - name: create dbt-setup-release-env.sh and daq_app_rte.sh for fddaq - run: | - export DET=fd - source /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh - setup_dbt latest|| true - dbt-setup-release -n last_${DET}daq_c8 - declare -x > ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - declare -f >> ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - egrep "declare -x (PATH|.*_SHARE|CET_PLUGIN_PATH|DUNEDAQ_SHARE_PATH|LD_LIBRARY_PATH|LIBRARY_PATH|PYTHONPATH)=" ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh > ${GITHUB_WORKSPACE}/${DET}daq_app_rte.sh - - - name: upload fddaq-dbt-setup-release-env.sh - uses: actions/upload-artifact@v3 - with: - name: fddaq-dbt_setup_release_env - path: ${{ github.workspace }}/fddaq-dbt-setup-release-env.sh - - - name: upload fddaq_app_rte.sh - uses: actions/upload-artifact@v3 - with: - name: fddaq_app_rte - path: ${{ github.workspace }}/fddaq_app_rte.sh - - publish_to_cvmfs: - name: publish to cvmfs - needs: generate_dbt_setup_release_env - runs-on: daq - - steps: - - name: check_cvmfs_deployment - id: do_deployment - run: | - cvmfs_deployment_input=${{ github.event.inputs.cvmfs-deployment }} - cvmfs_deployment_flag=${cvmfs_deployment_input:-"yes"} - echo "Do cvmfs deployment: ${cvmfs_deployment_flag}" - [[ ${cvmfs_deployment_flag} == 'yes' ]] && exit 0 || exit 1 - continue-on-error: true - - - name: download fddaq-dbt-setup-release-env.sh - if: steps.do_deployment.outcome == 'success' - uses: actions/download-artifact@v3 - with: - name: fddaq-dbt_setup_release_env - path: ${{ github.workspace }}/fddaq-release - - - name: download fddaq_app_rte.sh - if: steps.do_deployment.outcome == 'success' - uses: actions/download-artifact@v3 - with: - name: fddaq_app_rte - path: ${{ github.workspace }}/fddaq-rte - - - name: start docker container, and then rsync... - if: steps.do_deployment.outcome == 'success' - run: | - [[ ${DO_DEPLOYMENT} == 'no' ]] && exit 0 - - cd $GITHUB_WORKSPACE - - IMAGE="ghcr.io/dune-daq/nightly-release-sl7:development_v5" - unique_name=$( date | tr " :" _ ) - - docker pull $IMAGE - files_to_copy=$( docker run --rm --entrypoint ls $IMAGE /cvmfs/dunedaq-development.opensciencegrid.org/nightly ) - - docker run --name $unique_name $IMAGE - mkdir $unique_name - - for file in $files_to_copy; do - docker cp $unique_name:/cvmfs/dunedaq-development.opensciencegrid.org/nightly/$file $unique_name - done - docker rm $unique_name - docker image rm $IMAGE || true - - dir_for_env_scripts=$( find $unique_name -mindepth 1 -maxdepth 1 -type d -name "NFD*" ) - cp ./fddaq-release/fddaq-dbt-setup-release-env.sh $dir_for_env_scripts/dbt-setup-release-env.sh - cp ./fddaq-rte/fddaq_app_rte.sh $dir_for_env_scripts/daq_app_rte.sh - - kinit -k -t $HOME/daq-nightly.keytab nightly-build/dune/daq.dunescience.org@FNAL.GOV - - ssh -o StrictHostKeyChecking=no -l cvmfsdunedaqdev oasiscfs05.fnal.gov "cvmfs_server transaction dunedaq-development.opensciencegrid.org" - - rsync -e "ssh -o StrictHostKeyChecking=no" -rlpt --stats $unique_name/* cvmfsdunedaqdev@oasiscfs05.fnal.gov:/cvmfs/dunedaq-development.opensciencegrid.org/nightly - - ssh -o StrictHostKeyChecking=no -l cvmfsdunedaqdev oasiscfs05.fnal.gov "cvmfs_server publish dunedaq-development.opensciencegrid.org" - - rm -rf $unique_name diff --git a/.github/workflows/build-v4-release-sl7.yml b/.github/workflows/build-v4-release-sl7.yml deleted file mode 100644 index aae27bcde..000000000 --- a/.github/workflows/build-v4-release-sl7.yml +++ /dev/null @@ -1,277 +0,0 @@ -name: SL7 build v4 production nightly release - -on: - schedule: - - cron: "0 7 * * 0" - - workflow_dispatch: - inputs: - tag-prefix: - description: 'nightly tag prefix' - default: '' - cvmfs-deployment: - description: 'whether to deploy the release to cvmfs' - default: 'no' - -jobs: - make_nightly_tag: - name: create nightly tag and decide if deploying to cvmfs - runs-on: ubuntu-latest - outputs: - tag: ${{ steps.create_nightly_tag.outputs.nightly_tag }} - defaults: - run: - shell: bash - steps: - - id: create_nightly_tag - run: | - date_tag=$(date +%y%m%d) - echo "nightly_tag=${{ github.event.inputs.tag-prefix }}_PROD4_${date_tag}_C8" >> "$GITHUB_OUTPUT" - cat "$GITHUB_OUTPUT" - - build_the_develop_release_spack: - name: build_dev_release_spack - runs-on: ubuntu-latest - needs: make_nightly_tag - container: - image: ghcr.io/dune-daq/sl7-slim-externals:v1.1 - defaults: - run: - shell: bash - - steps: - - - name: Checkout daq-release - uses: actions/checkout@v3 - with: - path: daq-release - - - name: setup directories and install spack for the coredaq release - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NB${NIGHTLY_TAG} - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NFD${NIGHTLY_TAG} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/coredaq/coredaq-production_v4/release.yaml -a -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR core $OS production/v4 - - cd $BASE_RELEASE_DIR/../ - tar_and_stage_release ${BASE_RELEASE_TAG} - - - name: upload spack nightly tarball for base release - uses: actions/upload-artifact@v3 - with: - name: nightly_coredaq - path: ${{ github.workspace }}/tarballs_for_upload/NB${{needs.make_nightly_tag.outputs.tag}}.tar.gz - - - name: setup directories and install spack for the fddaq release - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - export BASE_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NB${NIGHTLY_TAG} - export DET_RELEASE_DIR=/cvmfs/dunedaq-development.opensciencegrid.org/nightly/NFD${NIGHTLY_TAG} - export OS=scientific7 - source daq-release/.github/workflows/wf-setup-tools.sh - - daq-release/scripts/checkout-daq-package.py -i daq-release/configs/fddaq/fddaq-production_v4/release.yaml -a -o $DET_RELEASE_DIR/sourcecode - daq-release/scripts/spack/build-release.sh $BASE_RELEASE_DIR $DET_RELEASE_DIR fd $OS production/v4 - - cd $DET_RELEASE_DIR/../ - tar_and_stage_release ${DET_RELEASE_TAG} - - - name: upload spack nightly tarball for far detector release - uses: actions/upload-artifact@v3 - with: - name: nightly_fddaq - path: ${{ github.workspace }}/tarballs_for_upload/NFD${{needs.make_nightly_tag.outputs.tag}}.tar.gz - - update_image: - name: update_spack_image_nightly - runs-on: daq - strategy: - matrix: - include: - - input_image: "ghcr.io/dune-daq/sl7-slim-externals:v1.1" - output_image: "ghcr.io/dune-daq/nightly-release-sl7" - tag: "production_v4" - needs: [build_the_develop_release_spack, make_nightly_tag] - environment: dockerhub - permissions: - packages: write - contents: read - - steps: - - name: clean docker-build - run: | - mkdir -p ${GITHUB_WORKSPACE}/docker-build - rm -rf ${GITHUB_WORKSPACE}/docker-build/* - - - name: Download spack nightly release tarball artifact for base release - uses: actions/download-artifact@v3 - with: - name: nightly_coredaq - path: ${{ github.workspace }}/docker-build - - - name: Download spack nightly release tarball artifact for far detector release - uses: actions/download-artifact@v3 - with: - name: nightly_fddaq - path: ${{ github.workspace }}/docker-build - - - name: prepare cvmfs mirror spack-nightly - env: - NIGHTLY_TAG: ${{needs.make_nightly_tag.outputs.tag}} - run: | - cd ${{ github.workspace }}/docker-build - mkdir -p nightly - - cd nightly - base_tag="NB${NIGHTLY_TAG}" - tar xf ../${base_tag}.tar.gz - rm -rf ../${base_tag}.tar.gz - - fddaq_tag="NFD${NIGHTLY_TAG}" - tar xf ../${fddaq_tag}.tar.gz - rm -rf ../${fddaq_tag}.tar.gz - ln -s ${fddaq_tag} last_fddaq_c8 - - cd .. - echo "FROM "${{ matrix.input_image }} > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_AT ${NIGHTLY_TAG}" >> Dockerfile - echo "ADD nightly /cvmfs/dunedaq-development.opensciencegrid.org/nightly" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Log in to the Container registry - uses: docker/login-action@v3 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v5 - with: - images: | - ${{ matrix.output_image }} - tags: | - type=raw,value=${{ matrix.tag }} - - - name: Build and push Docker images - uses: docker/build-push-action@v5 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - - generate_dbt_setup_release_env: - name: generate_dbt_setup_release_env - runs-on: ubuntu-latest - needs: update_image - container: - image: ghcr.io/dune-daq/nightly-release-sl7:production_v4 - defaults: - run: - shell: bash - - steps: - - - name: create dbt-setup-release-env.sh and daq_app_rte.sh for fddaq - run: | - export DET=fd - source /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh - setup_dbt latest|| true - dbt-setup-release -n last_${DET}daq_c8 - declare -x > ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - declare -f >> ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh - egrep "declare -x (PATH|.*_SHARE|CET_PLUGIN_PATH|DUNEDAQ_SHARE_PATH|LD_LIBRARY_PATH|LIBRARY_PATH|PYTHONPATH)=" ${GITHUB_WORKSPACE}/${DET}daq-dbt-setup-release-env.sh > ${GITHUB_WORKSPACE}/${DET}daq_app_rte.sh - - - name: upload fddaq-dbt-setup-release-env.sh - uses: actions/upload-artifact@v3 - with: - name: fddaq-dbt_setup_release_env - path: ${{ github.workspace }}/fddaq-dbt-setup-release-env.sh - - - name: upload fddaq_app_rte.sh - uses: actions/upload-artifact@v3 - with: - name: fddaq_app_rte - path: ${{ github.workspace }}/fddaq_app_rte.sh - - publish_to_cvmfs: - name: publish to cvmfs - needs: generate_dbt_setup_release_env - runs-on: daq - - steps: - - name: check_cvmfs_deployment - id: do_deployment - run: | - cvmfs_deployment_input=${{ github.event.inputs.cvmfs-deployment }} - cvmfs_deployment_flag=${cvmfs_deployment_input:-"yes"} - echo "Do cvmfs deployment: ${cvmfs_deployment_flag}" - [[ ${cvmfs_deployment_flag} == 'yes' ]] && exit 0 || exit 1 - continue-on-error: true - - - name: download fddaq-dbt-setup-release-env.sh - if: steps.do_deployment.outcome == 'success' - uses: actions/download-artifact@v3 - with: - name: fddaq-dbt_setup_release_env - path: ${{ github.workspace }}/fddaq-release - - - name: download fddaq_app_rte.sh - if: steps.do_deployment.outcome == 'success' - uses: actions/download-artifact@v3 - with: - name: fddaq_app_rte - path: ${{ github.workspace }}/fddaq-rte - - - name: start docker container, and then rsync... - if: steps.do_deployment.outcome == 'success' - run: | - [[ ${DO_DEPLOYMENT} == 'no' ]] && exit 0 - - cd $GITHUB_WORKSPACE - - IMAGE="ghcr.io/dune-daq/nightly-release-sl7:production_v4" - unique_name=$( date | tr " :" _ ) - - docker system df - docker pull $IMAGE - files_to_copy=$( docker run --rm --entrypoint ls $IMAGE /cvmfs/dunedaq-development.opensciencegrid.org/nightly ) - docker system df - - docker run --name $unique_name $IMAGE - docker system df - mkdir $unique_name - - for file in $files_to_copy; do - docker cp $unique_name:/cvmfs/dunedaq-development.opensciencegrid.org/nightly/$file $unique_name - done - - docker system df - docker rm $unique_name - docker system df - docker image rm $IMAGE || true - docker system df - - dir_for_env_scripts=$( find $unique_name -mindepth 1 -maxdepth 1 -type d -name "NFD*" ) - cp ./fddaq-release/fddaq-dbt-setup-release-env.sh $dir_for_env_scripts/dbt-setup-release-env.sh - cp ./fddaq-rte/fddaq_app_rte.sh $dir_for_env_scripts/daq_app_rte.sh - - kinit -k -t $HOME/daq-nightly.keytab nightly-build/dune/daq.dunescience.org@FNAL.GOV - - ssh -o StrictHostKeyChecking=no -l cvmfsdunedaqdev oasiscfs05.fnal.gov "cvmfs_server transaction dunedaq-development.opensciencegrid.org" - - rsync -e "ssh -o StrictHostKeyChecking=no" -rlpt --stats $unique_name/* cvmfsdunedaqdev@oasiscfs05.fnal.gov:/cvmfs/dunedaq-development.opensciencegrid.org/nightly - - ssh -o StrictHostKeyChecking=no -l cvmfsdunedaqdev oasiscfs05.fnal.gov "cvmfs_server publish dunedaq-development.opensciencegrid.org" - - rm -rf $unique_name diff --git a/.github/workflows/slim_externals.yaml b/.github/workflows/slim_externals.yaml index 3cacbd5c1..1473e4b24 100644 --- a/.github/workflows/slim_externals.yaml +++ b/.github/workflows/slim_externals.yaml @@ -12,162 +12,6 @@ on: jobs: - build_sl7_slim_externals: - name: build_sl7 - runs-on: daq - permissions: - packages: write - contents: read - steps: - - name: Intro - run: | - echo "Will create a ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }} image" - echo "Image contents will include $HOSTNAME:/home/nfs/${{ github.event.inputs.account-to-copy }}/docker-scratch/cvmfs_dunedaq/spack/externals/ext-v1.1" - - - name: Log in to the Container registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v4 - with: - images: | - ghcr.io/dune-daq/sl7-slim-externals - tags: | - type=raw,value=v1.1${{ github.event.inputs.tag-suffix }}-initial - - - name: prepare build dir - id: prep_build_dir - run: | - build_dir="/home/nfs/${{ github.event.inputs.account-to-copy }}/docker-scratch/cvmfs_dunedaq" - cd $build_dir - refresh_date=$(date +%y-%m-%d) - echo "FROM ghcr.io/dune-daq/sl7-spack:latest" > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_AT $refresh_date" >> Dockerfile - echo "ADD spack/externals/ext-v1.1 /cvmfs/dunedaq.opensciencegrid.org/spack/externals/ext-v1.1" >> Dockerfile - echo "RUN source /cvmfs/dunedaq.opensciencegrid.org/spack/externals/ext-v1.1/spack-0.20.0/share/spack/setup-env.sh && spack uninstall -y llvm" >> Dockerfile - echo "ADD tools /cvmfs/dunedaq.opensciencegrid.org/tools" >> Dockerfile - echo "ADD setup_dunedaq.sh /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh" >> Dockerfile - echo "ADD pypi-repo /cvmfs/dunedaq.opensciencegrid.org/pypi-repo" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Build and push Docker images - uses: docker/build-push-action@v3 - with: - context: /home/nfs/${{ github.event.inputs.account-to-copy }}/docker-scratch/cvmfs_dunedaq - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - - name: Clean up - if: always() - run: | - - rm -f /home/nfs/${{ github.event.inputs.account-to-copy }}/docker-scratch/cvmfs_dunedaq/Dockerfile - docker image prune -f - docker container prune -f - docker network prune -f - docker volume prune -f - - build_c7_slim_externals: - name: build_c7 - runs-on: daq - needs: build_sl7_slim_externals - permissions: - packages: write - contents: read - steps: - - name: Log in to the Container registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v4 - with: - images: | - ghcr.io/dune-daq/sl7-slim-externals - tags: | - type=raw,value=v1.1${{ github.event.inputs.tag-suffix }} - - - name: prepare build dir - run: | - build_dir="${{ github.workspace }}/docker-build" - mkdir -p $build_dir - cd $build_dir - refresh_date=$(date +%y-%m-%d) - echo "FROM ghcr.io/dune-daq/sl7-spack:latest" > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_AT $refresh_date" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/spack /cvmfs/dunedaq.opensciencegrid.org/spack" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/tools /cvmfs/dunedaq.opensciencegrid.org/tools" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/pypi-repo /cvmfs/dunedaq.opensciencegrid.org/pypi-repo" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Build and push Docker images - uses: docker/build-push-action@v3 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - - build_c8_slim_externals: - name: build_c8 - runs-on: daq - needs: build_sl7_slim_externals - permissions: - packages: write - contents: read - steps: - - name: Log in to the Container registry - uses: docker/login-action@v2 - with: - registry: ghcr.io - username: ${{ github.actor }} - password: ${{ secrets.GITHUB_TOKEN }} - - - name: Extract metadata (tags, labels) for Docker - id: meta - uses: docker/metadata-action@v4 - with: - images: | - ghcr.io/dune-daq/c8-slim-externals - tags: | - type=raw,value=v1.1${{ github.event.inputs.tag-suffix }} - - - name: prepare build dir - run: | - build_dir="${{ github.workspace }}/docker-build" - mkdir -p $build_dir - cd $build_dir - refresh_date=$(date +%y-%m-%d) - echo "FROM ghcr.io/dune-daq/c8-spack:latest" > Dockerfile - echo 'MAINTAINER John Freeman "jcfree@fnal.gov"' >> Dockerfile - echo "ENV REFRESHED_AT $refresh_date" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/spack /cvmfs/dunedaq.opensciencegrid.org/spack" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/tools /cvmfs/dunedaq.opensciencegrid.org/tools" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh /cvmfs/dunedaq.opensciencegrid.org/setup_dunedaq.sh" >> Dockerfile - echo "COPY --from=ghcr.io/dune-daq/sl7-slim-externals:v1.1${{ github.event.inputs.tag-suffix }}-initial /cvmfs/dunedaq.opensciencegrid.org/pypi-repo /cvmfs/dunedaq.opensciencegrid.org/pypi-repo" >> Dockerfile - echo 'ENTRYPOINT ["/bin/bash"]' >> Dockerfile - - - name: Build and push Docker images - uses: docker/build-push-action@v3 - with: - context: ${{ github.workspace }}/docker-build - push: true - tags: ${{ steps.meta.outputs.tags }} - labels: ${{ steps.meta.outputs.labels }} - build_alma9_externals: name: build_al9_initial runs-on: daq diff --git a/.github/workflows/test-single-pkg.yml b/.github/workflows/test-single-pkg.yml index a7e43a00f..ab708a649 100644 --- a/.github/workflows/test-single-pkg.yml +++ b/.github/workflows/test-single-pkg.yml @@ -34,6 +34,7 @@ jobs: - name: restore dir permission run: | + echo "JCF, Jul-8-2024: for the time being this Workflow is deprecated" && false sudo chown -R cvmfs.cvmfs /var/lib/cvmfs sudo chmod -R go-rwx /var/lib/cvmfs diff --git a/docs/README.md b/docs/README.md index 6bd51d2b4..48b8aaa18 100644 --- a/docs/README.md +++ b/docs/README.md @@ -3,12 +3,8 @@ [![AL9 Spack Nightly Workflow (prod, v4) ](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-v4-release-alma9.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-v4-release-alma9.yml) -[![Spack Nightly Workflow (prod, v4)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-v4-release-sl7.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-v4-release-sl7.yml) - [![AL9 Spack Nightly Workflow (dev, v5) ](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-nightly-release-alma9.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-nightly-release-alma9.yml) -[![Spack Nightly Workflow (dev, v5)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-nightly-release-sl7.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/build-nightly-release-sl7.yml) - [![Nightly v5 Integration Test Workflow](https://github.com/DUNE-DAQ/daq-release/actions/workflows/nightly-v5-integtest.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/nightly-v5-integtest.yml) [![Nightly v4 Integration Test Workflow](https://github.com/DUNE-DAQ/daq-release/actions/workflows/nightly-v4-integtest.yml/badge.svg)](https://github.com/DUNE-DAQ/daq-release/actions/workflows/nightly-v4-integtest.yml)