From 77d16ada1e60dbc0effe50bed805ae62c8f2ff9f Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Wed, 10 Apr 2024 00:57:51 +0800 Subject: [PATCH] Update adapter.py Former-commit-id: 98bc97d8d218182c026e9f57bbcbf40ab1e0bc87 --- src/llmtuner/model/adapter.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llmtuner/model/adapter.py b/src/llmtuner/model/adapter.py index bf206907..bf6f9381 100644 --- a/src/llmtuner/model/adapter.py +++ b/src/llmtuner/model/adapter.py @@ -145,7 +145,7 @@ def init_adapter( if model_args.use_unsloth: from unsloth import FastLanguageModel # type: ignore - unsloth_peft_kwargs = {"model": model, "max_seq_length": model_args.model_max_length} + unsloth_peft_kwargs = {"model": model, "max_seq_length": model_args.model_max_length} model = FastLanguageModel.get_peft_model(**peft_kwargs, **unsloth_peft_kwargs) else: lora_config = LoraConfig(