From f6014742fa8b2b94adcf88fc2e347bb00f598818 Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Thu, 5 Sep 2024 18:08:09 +0800 Subject: [PATCH] fix #5366 Former-commit-id: e9bda48c6d7bde135df6456513708a997ada916c --- src/llamafactory/train/tuner.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llamafactory/train/tuner.py b/src/llamafactory/train/tuner.py index e0d7a7c9..9928e7b9 100644 --- a/src/llamafactory/train/tuner.py +++ b/src/llamafactory/train/tuner.py @@ -72,7 +72,7 @@ def export_model(args: Optional[Dict[str, Any]] = None) -> None: tokenizer_module = load_tokenizer(model_args) tokenizer = tokenizer_module["tokenizer"] processor = tokenizer_module["processor"] - get_template_and_fix_tokenizer(tokenizer, data_args.template) + get_template_and_fix_tokenizer(tokenizer, data_args) model = load_model(tokenizer, model_args, finetuning_args) # must after fixing tokenizer to resize vocab if getattr(model, "quantization_method", None) is not None and model_args.adapter_name_or_path is not None: