[breaking] support transformers 4.48 (#6628)

Former-commit-id: 15357cdad9
This commit is contained in:
hoshi-hiyouga
2025-01-31 01:36:33 +08:00
committed by GitHub
parent 245de012ca
commit f6779b0e0c
17 changed files with 53 additions and 105 deletions

View File

@@ -14,8 +14,10 @@
import os
import pytest
from transformers.utils import is_flash_attn_2_available, is_torch_sdpa_available
from llamafactory.extras.packages import is_transformers_version_greater_than
from llamafactory.train.test_utils import load_infer_model
@@ -27,6 +29,7 @@ INFER_ARGS = {
}
@pytest.mark.xfail(is_transformers_version_greater_than("4.48"), reason="Attention refactor.")
def test_attention():
attention_available = ["disabled"]
if is_torch_sdpa_available():