Skip to content

Commit 16239ad

Browse files
authored
fix: Atchernych/5673959 docs (#4482)
Signed-off-by: Anna Tchernych <[email protected]>
1 parent 96afba9 commit 16239ad

File tree

2 files changed

+3
-2
lines changed

2 files changed

+3
-2
lines changed

recipes/README.md

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -176,12 +176,13 @@ For Llama-3-70B with vLLM (Aggregated), an example of integration with the Infer
176176

177177
Follow to Follow [Deploy Inference Gateway Section 2](../deploy/inference-gateway/README.md#2-deploy-inference-gateway) to install GAIE. Then apply manifests.
178178
Update the containers.epp.image in the deployment file, i.e. llama-3-70b/vllm/agg/gaie/k8s-manifests/epp/deployment.yaml
179+
This should be the same image you have used for your deployment.
179180

180181
```bash
181182
export DEPLOY_PATH=llama-3-70b/vllm/agg/
182183
#DEPLOY_PATH=<model>/<framework>/<mode>/
183184
kubectl apply -R -f "$DEPLOY_PATH/gaie/k8s-manifests" -n "$NAMESPACE"
184-
185+
```
185186

186187
### DeepSeek-R1 on GB200 (Multi-node)
187188

recipes/llama-3-70b/vllm/agg/gaie/k8s-manifests/epp/deployment.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -38,7 +38,7 @@ spec:
3838

3939
containers:
4040
- name: epp
41-
image: nvcr.io/nvstaging/ai-dynamo/dynamo-frontend:0.7.0rc2-amd64
41+
image: nvcr.io/nvidia/ai-dynamo/vllm-runtime:my-tag
4242
imagePullPolicy: IfNotPresent
4343
resources:
4444
requests:

0 commit comments

Comments
 (0)