From beac654c4132a146fc56a7bf9cc7307f7a36a575 Mon Sep 17 00:00:00 2001 From: Yi Yan Date: Tue, 26 Aug 2025 15:07:04 +0800 Subject: [PATCH] fix: update base image uri --- src/emd/cfn/shared/ecs_cluster.yaml | 2 +- src/pipeline/backend/huggingface/embedding/Dockerfile | 2 +- src/pipeline/backend/huggingface/llm/Dockerfile | 2 +- src/pipeline/backend/huggingface/llm/Dockerfile_baichuan | 2 +- src/pipeline/backend/huggingface/rerank/Dockerfile | 2 +- src/pipeline/backend/llama_cpp/Dockerfile | 2 +- src/pipeline/backend/lmdeploy/Dockerfile | 2 +- src/pipeline/backend/lmdeploy/Dockerfile_internvl2 | 2 +- src/pipeline/backend/ollama/Dockerfile | 2 +- src/pipeline/backend/tgi/Dockerfile | 2 +- src/pipeline/backend/vllm/Dockerfile | 2 +- src/pipeline/backend/vllm/Dockerfile_baichuan_m1 | 2 +- src/pipeline/backend/vllm/Dockerfile_dots_ocr | 2 +- src/pipeline/backend/vllm/Dockerfile_gemma3 | 2 +- src/pipeline/backend/vllm/Dockerfile_higgs_audio | 2 +- src/pipeline/backend/vllm/Dockerfile_qwen25_vl | 2 +- 16 files changed, 16 insertions(+), 16 deletions(-) diff --git a/src/emd/cfn/shared/ecs_cluster.yaml b/src/emd/cfn/shared/ecs_cluster.yaml index 70653d3f..85fa8aff 100644 --- a/src/emd/cfn/shared/ecs_cluster.yaml +++ b/src/emd/cfn/shared/ecs_cluster.yaml @@ -10,7 +10,7 @@ Parameters: APIRouterImageURI: Type: String Description: The URI of OpenAI compatbile API router image. If provided, the router will be deployed. - Default: "public.ecr.aws/aws-gcr-solutions/dmaa/api-router:latest" + Default: "quay.io/dmaa/api-router:latest" UseSpot: Type: String Description: Use Fargate Spot capacity? diff --git a/src/pipeline/backend/huggingface/embedding/Dockerfile b/src/pipeline/backend/huggingface/embedding/Dockerfile index 6066c887..53e824ad 100644 --- a/src/pipeline/backend/huggingface/embedding/Dockerfile +++ b/src/pipeline/backend/huggingface/embedding/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base +FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base # Create a new stage based on vllm-base FROM huggingface-base AS sagemaker-serving diff --git a/src/pipeline/backend/huggingface/llm/Dockerfile b/src/pipeline/backend/huggingface/llm/Dockerfile index 3064f1a7..f764fb4f 100644 --- a/src/pipeline/backend/huggingface/llm/Dockerfile +++ b/src/pipeline/backend/huggingface/llm/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base +FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base # Create a new stage based on vllm-base FROM huggingface-base AS sagemaker-serving diff --git a/src/pipeline/backend/huggingface/llm/Dockerfile_baichuan b/src/pipeline/backend/huggingface/llm/Dockerfile_baichuan index 743b8c22..95f5eab9 100644 --- a/src/pipeline/backend/huggingface/llm/Dockerfile_baichuan +++ b/src/pipeline/backend/huggingface/llm/Dockerfile_baichuan @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base +FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base # Create a new stage based on vllm-base FROM huggingface-base AS sagemaker-serving diff --git a/src/pipeline/backend/huggingface/rerank/Dockerfile b/src/pipeline/backend/huggingface/rerank/Dockerfile index 6066c887..53e824ad 100644 --- a/src/pipeline/backend/huggingface/rerank/Dockerfile +++ b/src/pipeline/backend/huggingface/rerank/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base +FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base # Create a new stage based on vllm-base FROM huggingface-base AS sagemaker-serving diff --git a/src/pipeline/backend/llama_cpp/Dockerfile b/src/pipeline/backend/llama_cpp/Dockerfile index 461995f5..6fa022d7 100644 --- a/src/pipeline/backend/llama_cpp/Dockerfile +++ b/src/pipeline/backend/llama_cpp/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/llama-cpp:{{VERSION}} +FROM quay.io/dmaa/llama-cpp:{{VERSION}} # Ensure the serve script has executable permissions # RUN chmod +x /usr/bin/serve diff --git a/src/pipeline/backend/lmdeploy/Dockerfile b/src/pipeline/backend/lmdeploy/Dockerfile index bd353640..3e2c22a8 100644 --- a/src/pipeline/backend/lmdeploy/Dockerfile +++ b/src/pipeline/backend/lmdeploy/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base +FROM quay.io/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base # Create a new stage based on vllm-base FROM lmdeploy-base AS sagemaker-serving diff --git a/src/pipeline/backend/lmdeploy/Dockerfile_internvl2 b/src/pipeline/backend/lmdeploy/Dockerfile_internvl2 index b9b9148b..b35839c7 100644 --- a/src/pipeline/backend/lmdeploy/Dockerfile_internvl2 +++ b/src/pipeline/backend/lmdeploy/Dockerfile_internvl2 @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base +FROM quay.io/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base # Create a new stage based on vllm-base FROM lmdeploy-base AS sagemaker-serving diff --git a/src/pipeline/backend/ollama/Dockerfile b/src/pipeline/backend/ollama/Dockerfile index 794d19a1..eafc3014 100644 --- a/src/pipeline/backend/ollama/Dockerfile +++ b/src/pipeline/backend/ollama/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/ollama:{{VERSION}} AS ollama-base +FROM quay.io/dmaa/ollama:{{VERSION}} AS ollama-base # Ensure the serve script has executable permissions # RUN chmod +x /usr/bin/serve diff --git a/src/pipeline/backend/tgi/Dockerfile b/src/pipeline/backend/tgi/Dockerfile index 4466e1e8..7ea25316 100644 --- a/src/pipeline/backend/tgi/Dockerfile +++ b/src/pipeline/backend/tgi/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/text-generation-inference:{{VERSION}} AS tgi-base +FROM quay.io/dmaa/huggingface/text-generation-inference:{{VERSION}} AS tgi-base # Create a new stage based on vllm-base FROM tgi-base AS sagemaker-serving diff --git a/src/pipeline/backend/vllm/Dockerfile b/src/pipeline/backend/vllm/Dockerfile index c0e3e63c..fcc47ed4 100644 --- a/src/pipeline/backend/vllm/Dockerfile +++ b/src/pipeline/backend/vllm/Dockerfile @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base +FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base # Create a new stage based on vllm-base FROM vllm-base AS sagemaker-serving diff --git a/src/pipeline/backend/vllm/Dockerfile_baichuan_m1 b/src/pipeline/backend/vllm/Dockerfile_baichuan_m1 index 10535bc6..452e2d88 100644 --- a/src/pipeline/backend/vllm/Dockerfile_baichuan_m1 +++ b/src/pipeline/backend/vllm/Dockerfile_baichuan_m1 @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base +FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base # Create a new stage based on vllm-base FROM vllm-base AS sagemaker-serving diff --git a/src/pipeline/backend/vllm/Dockerfile_dots_ocr b/src/pipeline/backend/vllm/Dockerfile_dots_ocr index 9197b986..10530886 100644 --- a/src/pipeline/backend/vllm/Dockerfile_dots_ocr +++ b/src/pipeline/backend/vllm/Dockerfile_dots_ocr @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base +FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base WORKDIR /opt/ml/code diff --git a/src/pipeline/backend/vllm/Dockerfile_gemma3 b/src/pipeline/backend/vllm/Dockerfile_gemma3 index 6782c2b1..2eb8de66 100644 --- a/src/pipeline/backend/vllm/Dockerfile_gemma3 +++ b/src/pipeline/backend/vllm/Dockerfile_gemma3 @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base +FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base FROM vllm-base AS sagemaker-serving diff --git a/src/pipeline/backend/vllm/Dockerfile_higgs_audio b/src/pipeline/backend/vllm/Dockerfile_higgs_audio index 2572c5d5..ab36af75 100644 --- a/src/pipeline/backend/vllm/Dockerfile_higgs_audio +++ b/src/pipeline/backend/vllm/Dockerfile_higgs_audio @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa/higgs-audio-vllm:latest AS base +FROM quay.io/dmaa/higgs-audio-vllm:latest AS base WORKDIR /opt/ml/code diff --git a/src/pipeline/backend/vllm/Dockerfile_qwen25_vl b/src/pipeline/backend/vllm/Dockerfile_qwen25_vl index 980f075b..2f025577 100644 --- a/src/pipeline/backend/vllm/Dockerfile_qwen25_vl +++ b/src/pipeline/backend/vllm/Dockerfile_qwen25_vl @@ -1,4 +1,4 @@ -FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base +FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base # Create a new stage based on vllm-base FROM vllm-base AS sagemaker-serving