Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 4 additions & 0 deletions tests/slow/test_grpo_slow.py
Original file line number Diff line number Diff line change
Expand Up @@ -19,8 +19,10 @@
import numpy as np
import pytest
import torch
import transformers
from accelerate.utils.memory import release_memory
from datasets import Dataset, Features, Image, Value, load_dataset
from packaging.version import Version
from parameterized import parameterized
from transformers import (
AutoModelForCausalLM,
Expand Down Expand Up @@ -171,6 +173,8 @@ def test_training_with_liger_grpo_loss_and_peft(self, model_name):
@parameterized.expand(MODELS_TO_TEST)
def test_training_with_transformers_paged(self, model_name):
"""Test that training works with transformers paged implementation (requires GPU)."""
if Version(transformers.__version__) < Version("4.56.2"):
pytest.xfail("Upstream bug in transformers (GH#40692). Fix merged; awaiting release >= 4.56.2")
training_args = GRPOConfig(
output_dir=self.tmp_dir,
learning_rate=0.1, # increase the learning rate to speed up the test
Expand Down
5 changes: 4 additions & 1 deletion tests/test_online_dpo_trainer.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,9 +12,10 @@
# See the License for the specific language governing permissions and
# limitations under the License.


import pytest
import transformers
from datasets import Dataset, features, load_dataset
from packaging.version import Version
from parameterized import parameterized
from transformers import AutoModelForCausalLM, AutoModelForSequenceClassification, AutoTokenizer
from transformers.testing_utils import require_peft, require_torch_accelerator, require_vision
Expand Down Expand Up @@ -421,6 +422,8 @@ def test_generation_config_setup(self):
@require_torch_accelerator
@parameterized.expand([("standard_prompt_only",), ("conversational_prompt_only",)])
def test_training_with_transformers_paged(self, config_name):
if Version(transformers.__version__) < Version("4.56.2"):
pytest.xfail("Upstream bug in transformers (GH#40692). Fix merged; awaiting release >= 4.56.2")
training_args = OnlineDPOConfig(
output_dir=self.tmp_dir,
per_device_train_batch_size=2,
Expand Down
Loading