mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 00:28:10 +08:00
fix inference in llamaboard
Former-commit-id: 5e631915157083b61e2d5a183e0c91f2d11f416e
This commit is contained in:
parent
a5eabbe933
commit
7cfcd69c64
@ -222,7 +222,7 @@ class Runner:
|
|||||||
quantization_bit=int(get("top.quantization_bit")) if get("top.quantization_bit") in ["8", "4"] else None,
|
quantization_bit=int(get("top.quantization_bit")) if get("top.quantization_bit") in ["8", "4"] else None,
|
||||||
template=get("top.template"),
|
template=get("top.template"),
|
||||||
rope_scaling=get("top.rope_scaling") if get("top.rope_scaling") in ["linear", "dynamic"] else None,
|
rope_scaling=get("top.rope_scaling") if get("top.rope_scaling") in ["linear", "dynamic"] else None,
|
||||||
flash_attn=(get("top.booster") == "flashattn"),
|
flash_attn="fa2" if get("top.booster") == "flashattn2" else "auto",
|
||||||
use_unsloth=(get("top.booster") == "unsloth"),
|
use_unsloth=(get("top.booster") == "unsloth"),
|
||||||
dataset_dir=get("eval.dataset_dir"),
|
dataset_dir=get("eval.dataset_dir"),
|
||||||
dataset=",".join(get("eval.dataset")),
|
dataset=",".join(get("eval.dataset")),
|
||||||
|
Loading…
x
Reference in New Issue
Block a user