mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 00:28:10 +08:00
fix Llama lora merge crash
Former-commit-id: a8bd8e9149ff79a2707fec9c6d006761cfdd0dee
This commit is contained in:
parent
edeed55664
commit
2eba1c6851
@ -71,7 +71,7 @@ def export_model(args: Optional[Dict[str, Any]] = None):
|
|||||||
(config.top_p is not None and config.top_p != 1.0) or
|
(config.top_p is not None and config.top_p != 1.0) or
|
||||||
(config.typical_p is not None and config.typical_p != 1.0)
|
(config.typical_p is not None and config.typical_p != 1.0)
|
||||||
):
|
):
|
||||||
config.do_sample = False
|
config.do_sample = True
|
||||||
|
|
||||||
model.save_pretrained(
|
model.save_pretrained(
|
||||||
save_directory=model_args.export_dir,
|
save_directory=model_args.export_dir,
|
||||||
|
Loading…
x
Reference in New Issue
Block a user