Skip to content
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
9 changes: 6 additions & 3 deletions src/lighteval/models/transformers/transformers_model.py
Original file line number Diff line number Diff line change
Expand Up @@ -466,13 +466,16 @@ def _init_max_length(self) -> int:
return self.config.max_length

# Try to get the sequence length from the model config.
text_model_config = self.transformers_config.get_text_config()

seqlen_config_attrs = ("n_positions", "max_position_embeddings", "n_ctx")
for attr in seqlen_config_attrs:
if hasattr(self.transformers_config, attr):
return getattr(self.transformers_config, attr)
if hasattr(text_model_config, attr):
return getattr(text_model_config, attr)

logger.warning(
"No max_length attribute found in the model config. Using the default max sequence length setting {2048}. It is recomended to set max_length through the model args"
"No max_length attribute found in the model config. Using the default max sequence length setting `2048`. "
"It is recommended to set max_length trough the model args: max_length=..."
)

return 2048
Expand Down
Loading