Build All Packages CI #744
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
name: Build All Packages CI | |
on: | |
push: | |
branches: [ 'dev-**', 'pr-**', staging, master ] | |
tags: [ '**' ] | |
pull_request: | |
branches: [ staging, master ] | |
schedule: | |
# At 02:30 on Saturday | |
- cron: '30 2 * * 6' | |
jobs: | |
integration_tests: | |
name: Run integration tests | |
runs-on: ubuntu-latest | |
steps: | |
# Fetch shallow git repository | |
- uses: actions/checkout@v4 | |
# Restore test video's cache | |
- name: Restore test videos cache | |
uses: actions/cache@v4 | |
id: test_videos_cache | |
with: | |
path: tests/support_/videos | |
key: ${{ runner.os }}-test_videos-${{ hashFiles('tests/scripts_/download_test_files.sh') }} | |
restore-keys: | | |
${{ runner.os }}-test_videos- | |
# Download any missing test videos | |
- name: Download any missing test videos | |
if: success() && steps.test_videos_cache.outputs.cache-hit != 'true' | |
run: | | |
bash -c "tests/scripts_/download_test_files.sh" | |
# Setup python environment | |
- name: Set up Python 3.7 | |
uses: actions/setup-python@v5 | |
with: | |
python-version: 3.7 | |
# Restore the python cache if it exists | |
- name: Restore python cache | |
uses: actions/cache@v4 | |
with: | |
path: ~/.cache/pip | |
key: ${{ runner.os }}-pip-${{ hashFiles('**/requirements.txt') }} | |
restore-keys: | | |
${{ runner.os }}-pip- | |
# Restore the apt cache if it exists | |
# TODO: Fix this cache. Its not really working at all. It would be nice to have tho. | |
- name: Restore python cache | |
id: apt-cache | |
uses: actions/cache@v4 | |
with: | |
path: /var/cache/apt/archives | |
key: ${{ runner.os }}-apt-cache | |
# Download FFMPEG | |
- name: Download FFMPEG | |
if: success() | |
run: | | |
[[ ! -z `sudo find /var/cache/apt/archives -type f -name "ffmpeg*.deb"` ]] && echo "no update" || sudo apt-get update | |
sudo apt-get install -yq --no-install-recommends --download-only ffmpeg | |
# Install FFMPEG | |
- name: Install FFMPEG | |
if: success() | |
run: | | |
sudo apt-get install -yq --no-install-recommends ffmpeg | |
# Install python dependencies for testing unmanic | |
- name: Install python dependencies | |
if: success() | |
run: | | |
python -m pip install --upgrade pip | |
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi | |
if [ -f requirements-dev.txt ]; then pip install -r requirements-dev.txt; fi | |
# Run pytest unit tests | |
- name: Test with pytest | |
if: success() | |
run: | | |
echo "Using test videos:" | |
ls -l ./tests/support_/videos/*/ | |
pytest -m integrationtest | |
py_build: | |
name: Build Python package | |
needs: integration_tests | |
runs-on: ubuntu-latest | |
steps: | |
# Fetch full git repository and all submodules | |
- name: Checkout project | |
uses: actions/checkout@v4 | |
with: | |
submodules: 'recursive' | |
# Checkout tags | |
- name: Fetch tags | |
run: | | |
git fetch --prune --unshallow --tags | |
echo exit code $? | |
git tag --list | |
# Setup python environment | |
- name: Set up Python 3.11 | |
uses: actions/setup-python@v5 | |
with: | |
python-version: 3.11 | |
# Restore the python cache if it exists | |
- name: Restore python cache | |
uses: actions/cache@v4 | |
with: | |
path: ~/.cache/pip | |
key: ${{ runner.os }}-pip-${{ hashFiles('**/requirements.txt') }} | |
restore-keys: | | |
${{ runner.os }}-pip- | |
# Install python dependencies for building unmanic | |
- name: Install dependencies | |
run: | | |
python -m pip install --upgrade pip | |
if [ -f requirements.txt ]; then pip install -r requirements.txt; fi | |
if [ -f requirements-dev.txt ]; then pip install -r requirements-dev.txt; fi | |
# Setup node environment for building unmanic's webui | |
- uses: actions/setup-node@v4 | |
with: | |
node-version: '20' | |
# Build python dist package | |
- name: Build python dist package | |
id: build_py | |
run: | | |
echo "Short version:" | |
python ./setup.py --quiet --version | |
echo "Long version:" | |
python ./setup.py --quiet fullversion | |
python -m build --no-isolation --skip-dependency-check --wheel | |
python -m build --no-isolation --skip-dependency-check --sdist | |
# Read the python package distribution data (save version to file) | |
- name: Read python package distribution data | |
id: py_build_data | |
run: | | |
PY_VERSION=$(python ./setup.py --quiet --version) | |
PY_BDIST_PATH=$(ls dist/*.whl | head -n1) | |
PY_BDIST_FILE=${PY_BDIST_PATH#*/} | |
echo "py_version=${PY_VERSION}" >> $GITHUB_OUTPUT | |
echo "py_bdist_file=${PY_BDIST_FILE}" >> $GITHUB_OUTPUT | |
echo "py_bdist_path=${PY_BDIST_PATH}" >> $GITHUB_OUTPUT | |
echo ${PY_VERSION} > dist/VERSION.txt | |
# Upload python package distribution data artifact | |
- uses: actions/upload-artifact@v4 | |
with: | |
name: unmanic-py-dist-data-${{ steps.py_build_data.outputs.py_version }} | |
path: dist/ | |
build_docker: | |
name: Build Docker Image | |
needs: py_build | |
runs-on: ubuntu-latest | |
steps: | |
# Fetch shallow git repository | |
- name: Checkout | |
uses: actions/checkout@v4 | |
# Fetch all artifacts | |
- name: Download Artifact | |
uses: actions/download-artifact@v4 | |
with: | |
path: ./artifacts/ | |
# Restore python package distribution data | |
- name: Restore python package distribution data | |
id: py_build_data | |
run: | | |
mkdir -p ./dist | |
find ./artifacts/ -type f -name "*.whl" -exec cp -n {} ./dist/ \; | |
find ./artifacts/ -type f -name "VERSION.txt" -exec cp -n {} ./dist/ \; | |
ls -l ./dist/ | |
PY_VERSION=$(cat ./dist/VERSION.txt) | |
echo "py_version=${PY_VERSION}" >> $GITHUB_OUTPUT | |
# Use QEMU to build for other arch | |
- name: Set up QEMU | |
if: success() | |
id: qemu | |
uses: docker/setup-qemu-action@v3 | |
with: | |
image: tonistiigi/binfmt:latest | |
platforms: all | |
# Use configure docker to use buildx to build the image | |
- name: Set up Docker Buildx | |
uses: docker/setup-buildx-action@v3 | |
if: success() | |
id: buildx | |
with: | |
version: latest | |
# List available build platforms | |
- name: Available platforms | |
if: success() | |
run: echo ${{ steps.buildx.outputs.platforms }} | |
# Generate 'prepare' build arguments to be retrieved later on | |
- name: Prepare | |
if: success() | |
id: prepare | |
run: | | |
echo "GITHUB_REF:${GITHUB_REF}" | |
echo "GITHUB_REPOSITORY:${GITHUB_REPOSITORY}" | |
DOCKER_IMAGE=docker.io/josh5/unmanic | |
VERSION_TAG=${GITHUB_REF#refs/*/} | |
DOCKER_TAGS="" | |
if [[ ${VERSION_TAG%/merge} == 'master' ]]; then | |
DOCKER_TAGS="${DOCKER_TAGS}${DOCKER_IMAGE}:latest," | |
elif [[ ${VERSION_TAG%/merge} == 'staging' ]]; then | |
DOCKER_TAGS="${DOCKER_TAGS}${DOCKER_IMAGE}:staging," | |
elif [[ ${VERSION_TAG%/merge} =~ "dev-"* ]]; then | |
DOCKER_TAGS="${DOCKER_TAGS}${DOCKER_IMAGE}:${VERSION_TAG%/merge}," | |
fi | |
if [[ ${GITHUB_REF} == refs/tags/* ]]; then | |
VERSION=${GITHUB_REF#refs/tags/} | |
if [[ ${VERSION} =~ ^[0-9]{1,3}\.[0-9]{1,3}\.[0-9]{1,3}[-\w]*$ ]]; then | |
DOCKER_TAGS="${DOCKER_TAGS}${DOCKER_IMAGE}:${VERSION}," | |
DOCKER_TAGS="${DOCKER_TAGS}${DOCKER_IMAGE}:latest," | |
fi | |
fi | |
DOCKER_PUSH="true" | |
if [[ ${DOCKER_IMAGE} != 'docker.io/josh5/unmanic' ]]; then | |
DOCKER_PUSH="false" | |
fi | |
if [[ ${VERSION_TAG%/merge} =~ "pr-"* ]]; then | |
DOCKER_PUSH="false" | |
fi | |
if [[ ${VERSION_TAG%/merge} =~ ^[0-9]+$ ]]; then | |
DOCKER_PUSH="false" | |
fi | |
echo "docker_image=${DOCKER_IMAGE}" >> $GITHUB_OUTPUT | |
echo "docker_tags=$(echo ${DOCKER_TAGS} | sed 's/,$//')" >> $GITHUB_OUTPUT | |
echo "docker_platforms=linux/amd64,linux/arm64" >> $GITHUB_OUTPUT | |
echo "docker_push=${DOCKER_PUSH}" >> $GITHUB_OUTPUT | |
# Cache the build | |
- name: Cache Docker layers | |
uses: actions/cache@v4 | |
id: cache | |
with: | |
path: /tmp/.buildx-cache | |
key: ${{ runner.os }}-buildx-${{ github.sha }} | |
restore-keys: | | |
${{ runner.os }}-buildx- | |
# Login to Docker Hub | |
- name: Login to Docker Hub | |
if: success() && (startsWith(github.ref, 'refs/heads/') || startsWith(github.ref, 'refs/tags/')) | |
uses: docker/login-action@v3 | |
with: | |
username: ${{ secrets.DOCKER_USERNAME }} | |
password: ${{ secrets.DOCKER_PASSWORD }} | |
# Run docker build and push | |
- name: Docker Build and Push | |
if: success() | |
uses: docker/build-push-action@v3 | |
with: | |
context: . | |
file: docker/Dockerfile | |
platforms: ${{ steps.prepare.outputs.docker_platforms }} | |
pull: true | |
push: ${{ steps.prepare.outputs.docker_push }} | |
tags: | | |
${{ steps.prepare.outputs.docker_tags }} | |
cache-from: type=local,src=/tmp/.buildx-cache | |
cache-to: type=local,mode=max,dest=/tmp/.buildx-cache-new | |
# Keep only latest cache | |
# https://github.com/docker/build-push-action/issues/252 | |
# https://github.com/moby/buildkit/issues/1896 | |
- name: Move cache | |
if: always() | |
run: | | |
if [[ -e /tmp/.buildx-cache-new ]]; then | |
echo "Cleaning up old cache..." | |
rm -rf /tmp/.buildx-cache | |
mv -v /tmp/.buildx-cache-new /tmp/.buildx-cache | |
fi | |
build_pypi: | |
name: Publish package to PyPI | |
needs: py_build | |
runs-on: ubuntu-latest | |
steps: | |
# Fetch shallow git repository | |
- name: Checkout | |
uses: actions/checkout@v4 | |
# Fetch all artifacts | |
- name: Download Artifact | |
uses: actions/download-artifact@v4 | |
with: | |
path: ./artifacts/ | |
# Restore python package distribution data | |
- name: Restore python package distribution data | |
id: py_build_data | |
run: | | |
mkdir -p ./dist | |
find ./artifacts/ -type f -name "*.whl" -exec cp -n {} ./dist/ \; | |
find ./artifacts/ -type f -name "*.tar.gz" -exec cp -n {} ./dist/ \; | |
ls -l ./dist/ | |
# Push the artifacts to PyPI repo | |
- name: Publish distribution package to PyPI | |
if: github.ref == 'refs/heads/master' || startsWith(github.ref, 'refs/tags/') | |
uses: pypa/gh-action-pypi-publish@master | |
with: | |
password: ${{ secrets.PYPI_TOKEN }} |