Skip to content

Commit 0523af7

Browse files
committed
[ci] Remove --runtime=nvidia from launch_container.sh
1 parent e8bc912 commit 0523af7

File tree

2 files changed

+1
-9
lines changed

2 files changed

+1
-9
lines changed

tests/integration/launch_container.sh

Lines changed: 0 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -84,10 +84,7 @@ if [[ "$(support_nvme)" == *"true"* ]]; then
8484
fi
8585

8686
is_llm=false
87-
if [[ "$platform" == *"-gpu"* ]]; then # if the platform has cuda capabilities
88-
runtime="nvidia"
8987
elif [[ "$platform" == *"lmi"* || "$platform" == *"trtllm"* || "$platform" == *"tensorrt-llm"* ]]; then # Runs multi-gpu
90-
runtime="nvidia"
9188
is_llm=true
9289
if [[ "$(is_p4d_or_p5)" == *"true"* || $is_multi_node ]]; then
9390
shm="20gb"
@@ -157,7 +154,6 @@ if $is_multi_node; then
157154
-e DJL_LEADER_ADDR=${leader_hostname} \
158155
-e DJL_WORKER_ADDR_FORMAT="${LWS_NAME}-${GROUP_INDEX}-%d.${LWS_NAME}.${NAMESPACE}" \
159156
${env_file} \
160-
${runtime:+--runtime="${runtime}"} \
161157
${shm:+--shm-size="${shm}"} \
162158
${host_device:+ ${host_device}} \
163159
"${docker_image}" "service ssh start; djl-serving"
@@ -179,7 +175,6 @@ if $is_multi_node; then
179175
-e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
180176
-e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
181177
${env_file} \
182-
${runtime:+--runtime="${runtime}"} \
183178
${shm:+--shm-size="${shm}"} \
184179
${host_device:+ ${host_device}} \
185180
"${docker_image}" "service ssh start; /usr/bin/python3 /opt/djl/partition/run_multi_node_setup.py 2>&1 | tee /opt/djl/logs/lmi-worker.log; tail -f"
@@ -207,7 +202,6 @@ elif $is_sm_neo_context; then
207202
-e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
208203
-e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
209204
${env_file} \
210-
${runtime:+--runtime="${runtime}"} \
211205
${shm:+--shm-size="${shm}"} \
212206
${host_device:+ ${host_device}} \
213207
"${docker_image}"
@@ -228,7 +222,6 @@ elif $is_partition; then
228222
-e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID \
229223
-e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
230224
-e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
231-
${runtime:+--runtime="${runtime}"} \
232225
${shm:+--shm-size="${shm}"} \
233226
${host_device:+ ${host_device}} \
234227
"${docker_image}" \
@@ -247,7 +240,6 @@ elif [[ "$docker_image" == *"text-generation-inference"* ]]; then
247240
-e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID \
248241
-e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
249242
-e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
250-
${runtime:+--runtime="${runtime}"} \
251243
${shm:+--shm-size="${shm}"} \
252244
"${docker_image}" \
253245
${args})
@@ -272,7 +264,6 @@ else
272264
-e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
273265
-e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
274266
$uid_mapping \
275-
${runtime:+--runtime="${runtime}"} \
276267
${shm:+--shm-size="${shm}"} \
277268
${host_device:+ ${host_device}} \
278269
"${docker_image}" \

tests/integration/llm/prepare.py

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -458,6 +458,7 @@
458458
"option.trust_remote_code": True,
459459
"option.max_model_len": 16384,
460460
"option.gpu_memory_utilization": "0.9",
461+
"option.enable_expert_parallel": True,
461462
},
462463
"llama-4-scout-17b-16e-instruct": {
463464
"option.model_id": "s3://djl-llm/Llama-4-Scout-17B-16E-Instruct/",

0 commit comments

Comments
 (0)