From 4e68828e46f7c955ed30324c482fa5e9c3eeef26 Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Mon, 10 Mar 2025 16:46:08 +0800 Subject: [PATCH] [config] fix export max len (#7230) Former-commit-id: 728c2f68190129e88ac375b453c60ed0d6bebbd8 --- src/llamafactory/hparams/parser.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/src/llamafactory/hparams/parser.py b/src/llamafactory/hparams/parser.py index 0e22868c..dc7f6c1b 100644 --- a/src/llamafactory/hparams/parser.py +++ b/src/llamafactory/hparams/parser.py @@ -418,7 +418,8 @@ def get_infer_args(args: Optional[Union[Dict[str, Any], List[str]]] = None) -> _ if model_args.export_dir is not None and model_args.export_device == "cpu": model_args.device_map = {"": torch.device("cpu")} - model_args.model_max_length = data_args.cutoff_len + if data_args.cutoff_len != DataArguments().cutoff_len: # override cutoff_len if it is not default + model_args.model_max_length = data_args.cutoff_len else: model_args.device_map = "auto"