From ae9cc77063d5b30a3475882f2d5f312eda7caaad Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 19:42:37 +0100 Subject: [PATCH 1/7] update the docker file --- Dockerfile-cpu | 27 ++++++++++----------------- Dockerfile-gpu | 43 +++++++++++++++++++++++++++++-------------- README.md | 45 ++++++++++++++++++++++++++++++++++++++++----- 3 files changed, 79 insertions(+), 36 deletions(-) diff --git a/Dockerfile-cpu b/Dockerfile-cpu index ccb2a58..5f10e2f 100644 --- a/Dockerfile-cpu +++ b/Dockerfile-cpu @@ -1,27 +1,20 @@ -# This Dockerfile configures a Docker environment that -# contains all the required packages for the tool FROM ubuntu:22.04 - USER root -# Install apt packages ADD install-pkgs.sh install-pkgs.sh RUN bash install-pkgs.sh -CMD ["bash"] - -# Install PyTorch and Torch-MLIR RUN pip3 install --upgrade pip -RUN pip3 install --pre torch-mlir torchvision \ - -f https://github.com/llvm/torch-mlir-release/releases/expanded_assets/dev-wheels \ - --extra-index-url https://download.pytorch.org/whl/nightly/cpu -# Install pip packages +# Pin CPU-only PyTorch before MASE install to prevent pip from pulling in the CUDA variant +RUN pip3 install --no-cache-dir \ + torch==2.6 torchvision==0.21.0 --index-url https://download.pytorch.org/whl/cpu + +# Install MASE dependencies, remove any NVIDIA packages pulled in, then re-pin CPU torch ADD install-pips.sh install-pips.sh -RUN bash install-pips.sh +RUN bash install-pips.sh && \ + pip3 list | grep -E '^nvidia-' | awk '{print $1}' | xargs pip3 uninstall -y 2>/dev/null || true && \ + pip3 install --force-reinstall --no-deps --no-cache-dir \ + torch==2.6 torchvision==0.21.0 --index-url https://download.pytorch.org/whl/cpu -# Add environment variables -ARG VHLS_PATH -ARG VHLS_VERSION -ADD install-env.sh install-env.sh -RUN bash install-env.sh $VHLS_PATH $VHLS_VERSION +CMD ["bash"] \ No newline at end of file diff --git a/Dockerfile-gpu b/Dockerfile-gpu index 8759223..a5e6ce9 100644 --- a/Dockerfile-gpu +++ b/Dockerfile-gpu @@ -1,25 +1,40 @@ -# This Dockerfile configures a Docker environment that -# contains all the required packages for the tool -FROM ubuntu:22.04 +# GPU Dockerfile for MASE +# +# Build arguments: +# CUDA_VERSION - CUDA version matching your host driver (default: 12.1.1) +# Run `nvidia-smi` on the host to check the max supported CUDA version. +# Common options: 11.8.0 / 12.1.1 / 12.4.0 +# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag matching CUDA_VERSION (default: cu121) +# Common options: cu118 / cu121 / cu124 +# +# Example build command: +# docker buildx build --platform linux/amd64 \ +# --build-arg CUDA_VERSION=12.1.1 \ +# --build-arg TORCH_CUDA_ARCH=cu121 \ +# -f Dockerfile-gpu \ +# --tag aaronyirenzhao/mase-gpu:latest \ +# --push . +ARG CUDA_VERSION=12.1.1 +FROM nvidia/cuda:${CUDA_VERSION}-cudnn8-devel-ubuntu22.04 USER root -# Install apt packages +# Install system packages ADD install-pkgs.sh install-pkgs.sh RUN bash install-pkgs.sh -CMD ["bash"] - -# Install PyTorch and Torch-MLIR RUN pip3 install --upgrade pip -RUN pip3 install torch torchvision torchaudio -# Install pip packages +# Install CUDA-enabled PyTorch. +# The wheel index must match CUDA_VERSION (e.g. cu121 for CUDA 12.1). +# If your host driver is older, override TORCH_CUDA_ARCH at build time. +ARG TORCH_CUDA_ARCH=cu121 +RUN pip3 install --no-cache-dir \ + torch==2.6 torchvision==0.21.0 torchaudio==2.6.0 \ + --index-url https://download.pytorch.org/whl/${TORCH_CUDA_ARCH} + +# Install MASE and its dependencies ADD install-pips.sh install-pips.sh RUN bash install-pips.sh -# Add environment variables -ARG VHLS_PATH -ARG VHLS_VERSION -ADD install-env.sh install-env.sh -RUN bash install-env.sh $VHLS_PATH $VHLS_VERSION +CMD ["bash"] diff --git a/README.md b/README.md index 2c9ef3e..cc74bd3 100644 --- a/README.md +++ b/README.md @@ -1,11 +1,46 @@ # MASE Docker -The docker container used to run MASE. +Docker images for running MASE. Two variants are provided: -The deps are in `setup.py` of deepwok/mase +- `Dockerfile-cpu` — CPU-only, used for CI and general development +- `Dockerfile-gpu` — CUDA-enabled, for GPU-accelerated training and inference -## How to update dependencies? +## Prerequisites -1. Create a PR to update deps in mase-tools +- [Docker Desktop](https://docs.docker.com/get-docker/) with `buildx` support +- A [Docker Hub](https://hub.docker.com/) account +- For GPU image: an NVIDIA GPU with drivers installed on the host -2. Manually trigger the build in mase-docker +## Build and Push + +Replace `` with your Docker Hub username before running. + +**CPU image:** + +```bash +DOCKER_USER= +TAG=$(date -u +%Y%m%d%H%M%S) && \ +docker buildx build --no-cache \ + --platform linux/amd64 \ + -f Dockerfile-cpu \ + --tag docker.io/$DOCKER_USER/mase-cpu:$TAG \ + --tag docker.io/$DOCKER_USER/mase-cpu:latest \ + --push \ + . +``` + +**GPU image** (see `Dockerfile-gpu` comments for CUDA version options): + +```bash +DOCKER_USER= +TAG=$(date -u +%Y%m%d%H%M%S) && \ +docker buildx build --no-cache \ + --platform linux/amd64 \ + --build-arg CUDA_VERSION=12.1.1 \ + --build-arg TORCH_CUDA_ARCH=cu121 \ + -f Dockerfile-gpu \ + --tag docker.io/$DOCKER_USER/mase-gpu:$TAG \ + --tag docker.io/$DOCKER_USER/mase-gpu:latest \ + --push \ + . +``` \ No newline at end of file From f139175ea6b97082474a114bdeaaf6491cff5bc4 Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:00:27 +0100 Subject: [PATCH 2/7] Update Dockerfiles, workflows and README - Dockerfile-cpu: remove VHLS/MLIR args and install-env.sh, fix CMD placement - Dockerfile-gpu: switch to NVIDIA CUDA base image, parameterize CUDA_VERSION and TORCH_CUDA_ARCH, remove VHLS/MLIR deps - Workflows: remove VHLS build args; add CUDA_VERSION/TORCH_CUDA_ARCH to GPU workflow - README: document official DeepWok images and custom build instructions Made-with: Cursor --- .github/workflows/docker-image-cpu.yml | 17 +++++-------- .github/workflows/docker-image-gpu.yml | 17 ++++++------- README.md | 34 ++++++++++++++++++++++---- 3 files changed, 42 insertions(+), 26 deletions(-) diff --git a/.github/workflows/docker-image-cpu.yml b/.github/workflows/docker-image-cpu.yml index b0435d9..19bd882 100644 --- a/.github/workflows/docker-image-cpu.yml +++ b/.github/workflows/docker-image-cpu.yml @@ -6,7 +6,7 @@ on: pull_request: branches: [ "main" ] schedule: - - cron: '0 1 * * *' + - cron: '0 1 * * *' jobs: @@ -16,25 +16,20 @@ jobs: steps: - uses: actions/checkout@v3 - - name: Build the CPU Docker image + - name: Build and push the CPU Docker image run: | TAG=$(date -u +%Y%m%d%H%M%S) - docker build --no-cache --build-arg VHLS_PATH=/mnt/applications/Xilinx/23.1 \ - --build-arg VHLS_VERSION=2023.1 \ + docker build --no-cache \ -f Dockerfile-cpu \ --tag docker.io/deepwok/mase-docker-cpu:$TAG . || exit 1 - # Push to dockerhub if [ "${{ secrets.DOCKER_HUB_PASSWORD }}" != "" ]; then - echo ${{secrets.DOCKER_HUB_PASSWORD}} | docker login docker.io -u deepwok --password-stdin + echo ${{ secrets.DOCKER_HUB_PASSWORD }} | docker login docker.io -u deepwok --password-stdin docker push docker.io/deepwok/mase-docker-cpu:$TAG docker tag docker.io/deepwok/mase-docker-cpu:$TAG docker.io/deepwok/mase-docker-cpu:latest docker push docker.io/deepwok/mase-docker-cpu:latest - echo "MASE Docker image for CPU pushed." + echo "MASE CPU Docker image pushed." else - echo "Skipped pushing docker image for CPU." + echo "Skipped pushing: DOCKER_HUB_PASSWORD secret not configured." fi - - - diff --git a/.github/workflows/docker-image-gpu.yml b/.github/workflows/docker-image-gpu.yml index c9ae749..928bd7a 100644 --- a/.github/workflows/docker-image-gpu.yml +++ b/.github/workflows/docker-image-gpu.yml @@ -14,25 +14,22 @@ jobs: steps: - uses: actions/checkout@v3 - - name: Build the GPU Docker image + - name: Build and push the GPU Docker image run: | TAG=$(date -u +%Y%m%d%H%M%S) - docker build --no-cache --build-arg VHLS_PATH=/mnt/applications/Xilinx/23.1 \ - --build-arg VHLS_VERSION=2023.1 \ + docker build --no-cache \ + --build-arg CUDA_VERSION=12.1.1 \ + --build-arg TORCH_CUDA_ARCH=cu121 \ -f Dockerfile-gpu \ --tag docker.io/deepwok/mase-docker-gpu:$TAG . || exit 1 - # Push to dockerhub if [ "${{ secrets.DOCKER_HUB_PASSWORD }}" != "" ]; then - echo ${{secrets.DOCKER_HUB_PASSWORD}} | docker login docker.io -u deepwok --password-stdin + echo ${{ secrets.DOCKER_HUB_PASSWORD }} | docker login docker.io -u deepwok --password-stdin docker push docker.io/deepwok/mase-docker-gpu:$TAG docker tag docker.io/deepwok/mase-docker-gpu:$TAG docker.io/deepwok/mase-docker-gpu:latest docker push docker.io/deepwok/mase-docker-gpu:latest - echo "MASE Docker image for GPU pushed." + echo "MASE GPU Docker image pushed." else - echo "Skipped pushing docker image for GPU." + echo "Skipped pushing: DOCKER_HUB_PASSWORD secret not configured." fi - - - diff --git a/README.md b/README.md index cc74bd3..67eb83e 100644 --- a/README.md +++ b/README.md @@ -5,13 +5,37 @@ Docker images for running MASE. Two variants are provided: - `Dockerfile-cpu` — CPU-only, used for CI and general development - `Dockerfile-gpu` — CUDA-enabled, for GPU-accelerated training and inference -## Prerequisites +## Official Images (DeepWok) +The official MASE Docker images are automatically built and pushed to Docker Hub +by GitHub Actions whenever changes are merged into `main`: + +| Image | Docker Hub | +|-------|-----------| +| CPU | `deepwok/mase-docker-cpu:latest` | +| GPU | `deepwok/mase-docker-gpu:latest` | + +To use the official images: + +```bash +# CPU +docker pull deepwok/mase-docker-cpu:latest + +# GPU +docker pull deepwok/mase-docker-gpu:latest +``` + +--- + +## Build Your Own Image + +If you need to customise the image (e.g. different CUDA version, private fork), +you can build and push it yourself. + +**Prerequisites:** - [Docker Desktop](https://docs.docker.com/get-docker/) with `buildx` support - A [Docker Hub](https://hub.docker.com/) account -- For GPU image: an NVIDIA GPU with drivers installed on the host - -## Build and Push +- For GPU image: check your host driver version with `nvidia-smi` Replace `` with your Docker Hub username before running. @@ -29,7 +53,7 @@ docker buildx build --no-cache \ . ``` -**GPU image** (see `Dockerfile-gpu` comments for CUDA version options): +**GPU image** (see `Dockerfile-gpu` for available CUDA version options): ```bash DOCKER_USER= From 90e0fa482bed62a7511ba073888d76a5a35481a3 Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:07:24 +0100 Subject: [PATCH 3/7] Fix GPU Dockerfile: upgrade to CUDA 12.4 to support torch==2.6 Made-with: Cursor --- .github/workflows/docker-image-gpu.yml | 4 ++-- Dockerfile-gpu | 10 +++++----- README.md | 4 ++-- 3 files changed, 9 insertions(+), 9 deletions(-) diff --git a/.github/workflows/docker-image-gpu.yml b/.github/workflows/docker-image-gpu.yml index 928bd7a..664711b 100644 --- a/.github/workflows/docker-image-gpu.yml +++ b/.github/workflows/docker-image-gpu.yml @@ -19,8 +19,8 @@ jobs: TAG=$(date -u +%Y%m%d%H%M%S) docker build --no-cache \ - --build-arg CUDA_VERSION=12.1.1 \ - --build-arg TORCH_CUDA_ARCH=cu121 \ + --build-arg CUDA_VERSION=12.4.0 \ + --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/deepwok/mase-docker-gpu:$TAG . || exit 1 diff --git a/Dockerfile-gpu b/Dockerfile-gpu index a5e6ce9..2caebe4 100644 --- a/Dockerfile-gpu +++ b/Dockerfile-gpu @@ -1,10 +1,10 @@ # GPU Dockerfile for MASE # # Build arguments: -# CUDA_VERSION - CUDA version matching your host driver (default: 12.1.1) +# CUDA_VERSION - CUDA version matching your host driver (default: 12.4.0) # Run `nvidia-smi` on the host to check the max supported CUDA version. # Common options: 11.8.0 / 12.1.1 / 12.4.0 -# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag matching CUDA_VERSION (default: cu121) +# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag matching CUDA_VERSION (default: cu124) # Common options: cu118 / cu121 / cu124 # # Example build command: @@ -15,8 +15,8 @@ # --tag aaronyirenzhao/mase-gpu:latest \ # --push . -ARG CUDA_VERSION=12.1.1 -FROM nvidia/cuda:${CUDA_VERSION}-cudnn8-devel-ubuntu22.04 +ARG CUDA_VERSION=12.4.0 +FROM nvidia/cuda:${CUDA_VERSION}-cudnn9-devel-ubuntu22.04 USER root # Install system packages @@ -28,7 +28,7 @@ RUN pip3 install --upgrade pip # Install CUDA-enabled PyTorch. # The wheel index must match CUDA_VERSION (e.g. cu121 for CUDA 12.1). # If your host driver is older, override TORCH_CUDA_ARCH at build time. -ARG TORCH_CUDA_ARCH=cu121 +ARG TORCH_CUDA_ARCH=cu124 RUN pip3 install --no-cache-dir \ torch==2.6 torchvision==0.21.0 torchaudio==2.6.0 \ --index-url https://download.pytorch.org/whl/${TORCH_CUDA_ARCH} diff --git a/README.md b/README.md index 67eb83e..e5e3b90 100644 --- a/README.md +++ b/README.md @@ -60,8 +60,8 @@ DOCKER_USER= TAG=$(date -u +%Y%m%d%H%M%S) && \ docker buildx build --no-cache \ --platform linux/amd64 \ - --build-arg CUDA_VERSION=12.1.1 \ - --build-arg TORCH_CUDA_ARCH=cu121 \ + --build-arg CUDA_VERSION=12.4.0 \ + --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/$DOCKER_USER/mase-gpu:$TAG \ --tag docker.io/$DOCKER_USER/mase-gpu:latest \ From f817142cde8f55b720a818b18d5c88071c24ef3f Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:08:35 +0100 Subject: [PATCH 4/7] Fix GPU base image: use cudnn8 instead of cudnn9 for CUDA 12.4 Made-with: Cursor --- Dockerfile-gpu | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/Dockerfile-gpu b/Dockerfile-gpu index 2caebe4..16faf09 100644 --- a/Dockerfile-gpu +++ b/Dockerfile-gpu @@ -16,7 +16,7 @@ # --push . ARG CUDA_VERSION=12.4.0 -FROM nvidia/cuda:${CUDA_VERSION}-cudnn9-devel-ubuntu22.04 +FROM nvidia/cuda:${CUDA_VERSION}-cudnn8-devel-ubuntu22.04 USER root # Install system packages From c299d1e730945e920632a3a22d3f75bfe09dcf3d Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:11:54 +0100 Subject: [PATCH 5/7] Simplify GPU Dockerfile: use ubuntu:22.04 base, rely on PyTorch wheel for CUDA Made-with: Cursor --- .github/workflows/docker-image-gpu.yml | 1 - Dockerfile-gpu | 24 ++++++++++++++---------- README.md | 1 - 3 files changed, 14 insertions(+), 12 deletions(-) diff --git a/.github/workflows/docker-image-gpu.yml b/.github/workflows/docker-image-gpu.yml index 664711b..4fe1c6f 100644 --- a/.github/workflows/docker-image-gpu.yml +++ b/.github/workflows/docker-image-gpu.yml @@ -19,7 +19,6 @@ jobs: TAG=$(date -u +%Y%m%d%H%M%S) docker build --no-cache \ - --build-arg CUDA_VERSION=12.4.0 \ --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/deepwok/mase-docker-gpu:$TAG . || exit 1 diff --git a/Dockerfile-gpu b/Dockerfile-gpu index 16faf09..a606b0e 100644 --- a/Dockerfile-gpu +++ b/Dockerfile-gpu @@ -1,22 +1,26 @@ # GPU Dockerfile for MASE # +# Uses a plain Ubuntu base image. CUDA support is provided by the PyTorch +# wheel which bundles the necessary CUDA runtime libraries. +# +# Requirements: +# - NVIDIA GPU with drivers installed on the host +# - Run with: docker run --gpus all ... +# # Build arguments: -# CUDA_VERSION - CUDA version matching your host driver (default: 12.4.0) -# Run `nvidia-smi` on the host to check the max supported CUDA version. -# Common options: 11.8.0 / 12.1.1 / 12.4.0 -# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag matching CUDA_VERSION (default: cu124) +# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag (default: cu124) +# Run `nvidia-smi` on the host to check your max supported +# CUDA version, then pick the matching tag. # Common options: cu118 / cu121 / cu124 # # Example build command: # docker buildx build --platform linux/amd64 \ -# --build-arg CUDA_VERSION=12.1.1 \ -# --build-arg TORCH_CUDA_ARCH=cu121 \ +# --build-arg TORCH_CUDA_ARCH=cu124 \ # -f Dockerfile-gpu \ # --tag aaronyirenzhao/mase-gpu:latest \ # --push . -ARG CUDA_VERSION=12.4.0 -FROM nvidia/cuda:${CUDA_VERSION}-cudnn8-devel-ubuntu22.04 +FROM ubuntu:22.04 USER root # Install system packages @@ -26,8 +30,8 @@ RUN bash install-pkgs.sh RUN pip3 install --upgrade pip # Install CUDA-enabled PyTorch. -# The wheel index must match CUDA_VERSION (e.g. cu121 for CUDA 12.1). -# If your host driver is older, override TORCH_CUDA_ARCH at build time. +# The wheel bundles the CUDA runtime - no NVIDIA base image needed. +# Override TORCH_CUDA_ARCH at build time to match your host driver. ARG TORCH_CUDA_ARCH=cu124 RUN pip3 install --no-cache-dir \ torch==2.6 torchvision==0.21.0 torchaudio==2.6.0 \ diff --git a/README.md b/README.md index e5e3b90..499e476 100644 --- a/README.md +++ b/README.md @@ -60,7 +60,6 @@ DOCKER_USER= TAG=$(date -u +%Y%m%d%H%M%S) && \ docker buildx build --no-cache \ --platform linux/amd64 \ - --build-arg CUDA_VERSION=12.4.0 \ --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/$DOCKER_USER/mase-gpu:$TAG \ From 2ebbd622f9dfde8682006fa0ce371e1815f7a0fd Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:20:08 +0100 Subject: [PATCH 6/7] GPU: auto-install CUDA torch via default PyPI wheel, remove TORCH_CUDA_ARCH arg Made-with: Cursor --- .github/workflows/docker-image-gpu.yml | 1 - Dockerfile-gpu | 17 +++-------------- README.md | 1 - 3 files changed, 3 insertions(+), 16 deletions(-) diff --git a/.github/workflows/docker-image-gpu.yml b/.github/workflows/docker-image-gpu.yml index 4fe1c6f..05428f4 100644 --- a/.github/workflows/docker-image-gpu.yml +++ b/.github/workflows/docker-image-gpu.yml @@ -19,7 +19,6 @@ jobs: TAG=$(date -u +%Y%m%d%H%M%S) docker build --no-cache \ - --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/deepwok/mase-docker-gpu:$TAG . || exit 1 diff --git a/Dockerfile-gpu b/Dockerfile-gpu index a606b0e..8a5bf27 100644 --- a/Dockerfile-gpu +++ b/Dockerfile-gpu @@ -7,15 +7,8 @@ # - NVIDIA GPU with drivers installed on the host # - Run with: docker run --gpus all ... # -# Build arguments: -# TORCH_CUDA_ARCH - PyTorch CUDA wheel tag (default: cu124) -# Run `nvidia-smi` on the host to check your max supported -# CUDA version, then pick the matching tag. -# Common options: cu118 / cu121 / cu124 -# # Example build command: # docker buildx build --platform linux/amd64 \ -# --build-arg TORCH_CUDA_ARCH=cu124 \ # -f Dockerfile-gpu \ # --tag aaronyirenzhao/mase-gpu:latest \ # --push . @@ -29,13 +22,9 @@ RUN bash install-pkgs.sh RUN pip3 install --upgrade pip -# Install CUDA-enabled PyTorch. -# The wheel bundles the CUDA runtime - no NVIDIA base image needed. -# Override TORCH_CUDA_ARCH at build time to match your host driver. -ARG TORCH_CUDA_ARCH=cu124 -RUN pip3 install --no-cache-dir \ - torch==2.6 torchvision==0.21.0 torchaudio==2.6.0 \ - --index-url https://download.pytorch.org/whl/${TORCH_CUDA_ARCH} +# Install PyTorch with CUDA support. +# pip installs the CUDA-enabled wheel by default on Linux (includes bundled CUDA runtime). +RUN pip3 install --no-cache-dir torch torchvision torchaudio # Install MASE and its dependencies ADD install-pips.sh install-pips.sh diff --git a/README.md b/README.md index 499e476..f1d51f4 100644 --- a/README.md +++ b/README.md @@ -60,7 +60,6 @@ DOCKER_USER= TAG=$(date -u +%Y%m%d%H%M%S) && \ docker buildx build --no-cache \ --platform linux/amd64 \ - --build-arg TORCH_CUDA_ARCH=cu124 \ -f Dockerfile-gpu \ --tag docker.io/$DOCKER_USER/mase-gpu:$TAG \ --tag docker.io/$DOCKER_USER/mase-gpu:latest \ From 6e5cce8c3f5197ed81b1770ad4549d5b692a3241 Mon Sep 17 00:00:00 2001 From: Lou Date: Tue, 21 Apr 2026 20:24:38 +0100 Subject: [PATCH 7/7] Re-enable CPU Docker workflow Made-with: Cursor --- .github/workflows/docker-image-cpu.yml | 1 + 1 file changed, 1 insertion(+) diff --git a/.github/workflows/docker-image-cpu.yml b/.github/workflows/docker-image-cpu.yml index 19bd882..3366313 100644 --- a/.github/workflows/docker-image-cpu.yml +++ b/.github/workflows/docker-image-cpu.yml @@ -33,3 +33,4 @@ jobs: else echo "Skipped pushing: DOCKER_HUB_PASSWORD secret not configured." fi +