mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-04 20:52:59 +08:00
Merge pull request #2730 from cx2333-gt/main
fix flash_attn in train_web Former-commit-id: 2185855bdb7d4cb55f3af796e35fb1b0e8dce5e3
This commit is contained in:
commit
90e66c8d94
@ -113,7 +113,7 @@ class Runner:
|
||||
quantization_bit=int(get("top.quantization_bit")) if get("top.quantization_bit") in ["8", "4"] else None,
|
||||
template=get("top.template"),
|
||||
rope_scaling=get("top.rope_scaling") if get("top.rope_scaling") in ["linear", "dynamic"] else None,
|
||||
flash_attn=(get("top.booster") == "flash_attn"),
|
||||
flash_attn=(get("top.booster") == "flashattn"),
|
||||
use_unsloth=(get("top.booster") == "unsloth"),
|
||||
dataset_dir=get("train.dataset_dir"),
|
||||
dataset=",".join(get("train.dataset")),
|
||||
@ -200,7 +200,7 @@ class Runner:
|
||||
quantization_bit=int(get("top.quantization_bit")) if get("top.quantization_bit") in ["8", "4"] else None,
|
||||
template=get("top.template"),
|
||||
rope_scaling=get("top.rope_scaling") if get("top.rope_scaling") in ["linear", "dynamic"] else None,
|
||||
flash_attn=(get("top.booster") == "flash_attn"),
|
||||
flash_attn=(get("top.booster") == "flashattn"),
|
||||
use_unsloth=(get("top.booster") == "unsloth"),
|
||||
dataset_dir=get("eval.dataset_dir"),
|
||||
dataset=",".join(get("eval.dataset")),
|
||||
|
Loading…
x
Reference in New Issue
Block a user