From 3c55c053eeabce7dceb979c1be463caa78a69015 Mon Sep 17 00:00:00 2001 From: Jim Burtoft <39492751+jimburtoft@users.noreply.github.com> Date: Fri, 18 Jul 2025 22:39:45 -0400 Subject: [PATCH] Change batch size sort in Neuron config Fixing https://github.com/huggingface/text-generation-inference/issues/3299 --- backends/neuron/server/text_generation_server/tgi_env.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/backends/neuron/server/text_generation_server/tgi_env.py b/backends/neuron/server/text_generation_server/tgi_env.py index ee97f180eee..3d68c92d16f 100644 --- a/backends/neuron/server/text_generation_server/tgi_env.py +++ b/backends/neuron/server/text_generation_server/tgi_env.py @@ -112,7 +112,7 @@ def neuron_config_to_env(neuron_config): def sort_neuron_configs(dictionary): - return -dictionary["tp_degree"], -dictionary["batch_size"] + return -dictionary["tp_degree"], dictionary["batch_size"] def lookup_compatible_cached_model(