fix inference in llamaboard

Former-commit-id: f36057ea0300ab089ded568fa170682e9e19c4ee
This commit is contained in:
hoshi-hiyouga 2024-04-24 13:53:39 +08:00 committed by GitHub
parent fff1fb1232
commit 4a854dfe27

View File

@ -222,7 +222,7 @@ class Runner:
quantization_bit=int(get("top.quantization_bit")) if get("top.quantization_bit") in ["8", "4"] else None,
template=get("top.template"),
rope_scaling=get("top.rope_scaling") if get("top.rope_scaling") in ["linear", "dynamic"] else None,
flash_attn=(get("top.booster") == "flashattn"),
flash_attn="fa2" if get("top.booster") == "flashattn2" else "auto",
use_unsloth=(get("top.booster") == "unsloth"),
dataset_dir=get("eval.dataset_dir"),
dataset=",".join(get("eval.dataset")),