mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-14 07:42:49 +08:00
[ci] Change macOS version (#9229)
This commit is contained in:
parent
2e2f92701f
commit
af8437095a
9
.github/workflows/tests.yml
vendored
9
.github/workflows/tests.yml
vendored
@ -30,7 +30,7 @@ jobs:
|
||||
os:
|
||||
- "ubuntu-latest"
|
||||
- "windows-latest"
|
||||
- "macos-13"
|
||||
- "macos-latest"
|
||||
transformers:
|
||||
- null
|
||||
include: # test backward compatibility
|
||||
@ -75,10 +75,11 @@ jobs:
|
||||
run: |
|
||||
python -m pip install "transformers==${{ matrix.transformers }}"
|
||||
|
||||
- name: Install transformers to avoid mac os ci errors
|
||||
if: ${{ matrix.os == 'macos-13' }}
|
||||
- name: Update accelerate to avoid mac os ci errors (before accelerate 1.11.0)
|
||||
if: ${{ matrix.os == 'macos-latest' }}
|
||||
run: |
|
||||
python -m pip install "transformers<=4.51.3"
|
||||
python -m pip uninstall -y accelerate
|
||||
python -m pip install "git+https://github.com/huggingface/accelerate.git"
|
||||
|
||||
- name: Cache files
|
||||
id: hf-hub-cache
|
||||
|
@ -1,7 +1,8 @@
|
||||
# core deps
|
||||
transformers>=4.49.0,<=4.56.2,!=4.52.0
|
||||
transformers>=4.49.0,<=4.56.2,!=4.52.0; python_version < '3.10'
|
||||
transformers>=4.49.0,<=4.57.0,!=4.52.0; python_version >= '3.10'
|
||||
datasets>=2.16.0,<=4.0.0
|
||||
accelerate>=1.3.0,<=1.10.1
|
||||
accelerate>=1.3.0,<=1.11.0
|
||||
peft>=0.14.0,<=0.17.1
|
||||
trl>=0.8.6,<=0.9.6
|
||||
# gui
|
||||
|
@ -94,9 +94,9 @@ def check_version(requirement: str, mandatory: bool = False) -> None:
|
||||
|
||||
def check_dependencies() -> None:
|
||||
r"""Check the version of the required packages."""
|
||||
check_version("transformers>=4.49.0,<=4.56.2")
|
||||
check_version("transformers>=4.49.0,<=4.57.0")
|
||||
check_version("datasets>=2.16.0,<=4.0.0")
|
||||
check_version("accelerate>=1.3.0,<=1.10.1")
|
||||
check_version("accelerate>=1.3.0,<=1.11.0")
|
||||
check_version("peft>=0.14.0,<=0.17.1")
|
||||
check_version("trl>=0.8.6,<=0.9.6")
|
||||
|
||||
|
@ -376,7 +376,7 @@ def test_qwen2_vl_plugin():
|
||||
@pytest.mark.skipif(not is_transformers_version_greater_than("4.57.0"), reason="Requires transformers>=4.57.0")
|
||||
def test_qwen3_vl_plugin():
|
||||
frame_seqlen = 1
|
||||
tokenizer_module = _load_tokenizer_module(model_name_or_path="Qwen/Qwen3-VL-235B-A22B-Instruct")
|
||||
tokenizer_module = _load_tokenizer_module(model_name_or_path="Qwen/Qwen3-VL-30B-A3B-Instruct")
|
||||
qwen3_vl_plugin = get_mm_plugin(name="qwen3_vl", video_token="<|video_pad|>")
|
||||
check_inputs = {"plugin": qwen3_vl_plugin, **tokenizer_module}
|
||||
check_inputs["expected_mm_messages"] = [
|
||||
|
@ -1,2 +1,2 @@
|
||||
# change if test fails or cache is outdated
|
||||
0.9.4.102
|
||||
0.9.4.103
|
||||
|
Loading…
x
Reference in New Issue
Block a user