Skip to content
This repository was archived by the owner on Sep 20, 2025. It is now read-only.
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion src/emd/cfn/shared/ecs_cluster.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -10,7 +10,7 @@ Parameters:
APIRouterImageURI:
Type: String
Description: The URI of OpenAI compatbile API router image. If provided, the router will be deployed.
Default: "public.ecr.aws/aws-gcr-solutions/dmaa/api-router:latest"
Default: "quay.io/dmaa/api-router:latest"
UseSpot:
Type: String
Description: Use Fargate Spot capacity?
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/huggingface/embedding/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base
FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base

# Create a new stage based on vllm-base
FROM huggingface-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/huggingface/llm/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base
FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base

# Create a new stage based on vllm-base
FROM huggingface-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/huggingface/llm/Dockerfile_baichuan
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base
FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base

# Create a new stage based on vllm-base
FROM huggingface-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/huggingface/rerank/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base
FROM quay.io/dmaa/huggingface/transformers-pytorch-gpu:{{VERSION}} AS huggingface-base

# Create a new stage based on vllm-base
FROM huggingface-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/llama_cpp/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/llama-cpp:{{VERSION}}
FROM quay.io/dmaa/llama-cpp:{{VERSION}}

# Ensure the serve script has executable permissions
# RUN chmod +x /usr/bin/serve
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/lmdeploy/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base
FROM quay.io/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base

# Create a new stage based on vllm-base
FROM lmdeploy-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/lmdeploy/Dockerfile_internvl2
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base
FROM quay.io/dmaa/openmmlab/lmdeploy:{{VERSION}} AS lmdeploy-base

# Create a new stage based on vllm-base
FROM lmdeploy-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/ollama/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/ollama:{{VERSION}} AS ollama-base
FROM quay.io/dmaa/ollama:{{VERSION}} AS ollama-base

# Ensure the serve script has executable permissions
# RUN chmod +x /usr/bin/serve
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/tgi/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/huggingface/text-generation-inference:{{VERSION}} AS tgi-base
FROM quay.io/dmaa/huggingface/text-generation-inference:{{VERSION}} AS tgi-base

# Create a new stage based on vllm-base
FROM tgi-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base
FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base

# Create a new stage based on vllm-base
FROM vllm-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile_baichuan_m1
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base
FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base

# Create a new stage based on vllm-base
FROM vllm-base AS sagemaker-serving
Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile_dots_ocr
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base
FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base

WORKDIR /opt/ml/code

Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile_gemma3
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base
FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base

FROM vllm-base AS sagemaker-serving

Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile_higgs_audio
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa/higgs-audio-vllm:latest AS base
FROM quay.io/dmaa/higgs-audio-vllm:latest AS base

WORKDIR /opt/ml/code

Expand Down
2 changes: 1 addition & 1 deletion src/pipeline/backend/vllm/Dockerfile_qwen25_vl
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
FROM public.ecr.aws/aws-gcr-solutions/dmaa-vllm/vllm-openai:{{VERSION}} AS vllm-base
FROM quay.io/dmaa/vllm-openai:{{VERSION}} AS vllm-base

# Create a new stage based on vllm-base
FROM vllm-base AS sagemaker-serving
Expand Down