@@ -84,10 +84,7 @@ if [[ "$(support_nvme)" == *"true"* ]]; then
8484fi
8585
8686is_llm=false
87- if [[ " $platform " == * " -gpu" * ]]; then # if the platform has cuda capabilities
88- runtime=" nvidia"
8987elif [[ " $platform " == * " lmi" * || " $platform " == * " trtllm" * || " $platform " == * " tensorrt-llm" * ]]; then # Runs multi-gpu
90- runtime=" nvidia"
9188 is_llm=true
9289 if [[ " $( is_p4d_or_p5) " == * " true" * || $is_multi_node ]]; then
9390 shm=" 20gb"
@@ -157,7 +154,6 @@ if $is_multi_node; then
157154 -e DJL_LEADER_ADDR=${leader_hostname} \
158155 -e DJL_WORKER_ADDR_FORMAT=" ${LWS_NAME} -${GROUP_INDEX} -%d.${LWS_NAME} .${NAMESPACE} " \
159156 ${env_file} \
160- ${runtime: +--runtime=" ${runtime} " } \
161157 ${shm: +--shm-size=" ${shm} " } \
162158 ${host_device: + ${host_device} } \
163159 " ${docker_image} " " service ssh start; djl-serving"
@@ -179,7 +175,6 @@ if $is_multi_node; then
179175 -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
180176 -e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
181177 ${env_file} \
182- ${runtime: +--runtime=" ${runtime} " } \
183178 ${shm: +--shm-size=" ${shm} " } \
184179 ${host_device: + ${host_device} } \
185180 " ${docker_image} " " service ssh start; /usr/bin/python3 /opt/djl/partition/run_multi_node_setup.py 2>&1 | tee /opt/djl/logs/lmi-worker.log; tail -f"
@@ -207,7 +202,6 @@ elif $is_sm_neo_context; then
207202 -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
208203 -e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
209204 ${env_file} \
210- ${runtime: +--runtime=" ${runtime} " } \
211205 ${shm: +--shm-size=" ${shm} " } \
212206 ${host_device: + ${host_device} } \
213207 " ${docker_image} "
@@ -228,7 +222,6 @@ elif $is_partition; then
228222 -e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID \
229223 -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
230224 -e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
231- ${runtime: +--runtime=" ${runtime} " } \
232225 ${shm: +--shm-size=" ${shm} " } \
233226 ${host_device: + ${host_device} } \
234227 " ${docker_image} " \
@@ -247,7 +240,6 @@ elif [[ "$docker_image" == *"text-generation-inference"* ]]; then
247240 -e AWS_ACCESS_KEY_ID=$AWS_ACCESS_KEY_ID \
248241 -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
249242 -e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
250- ${runtime: +--runtime=" ${runtime} " } \
251243 ${shm: +--shm-size=" ${shm} " } \
252244 " ${docker_image} " \
253245 ${args} )
272264 -e AWS_SECRET_ACCESS_KEY=$AWS_SECRET_ACCESS_KEY \
273265 -e AWS_SESSION_TOKEN=$AWS_SESSION_TOKEN \
274266 $uid_mapping \
275- ${runtime: +--runtime=" ${runtime} " } \
276267 ${shm: +--shm-size=" ${shm} " } \
277268 ${host_device: + ${host_device} } \
278269 " ${docker_image} " \
0 commit comments