mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-12-15 03:10:35 +08:00
Update adapter.py
This commit is contained in:
@@ -145,7 +145,7 @@ def init_adapter(
|
|||||||
if model_args.use_unsloth:
|
if model_args.use_unsloth:
|
||||||
from unsloth import FastLanguageModel # type: ignore
|
from unsloth import FastLanguageModel # type: ignore
|
||||||
|
|
||||||
unsloth_peft_kwargs = {"model": model, "max_seq_length": model_args.model_max_length}
|
unsloth_peft_kwargs = {"model": model, "max_seq_length": model_args.model_max_length}
|
||||||
model = FastLanguageModel.get_peft_model(**peft_kwargs, **unsloth_peft_kwargs)
|
model = FastLanguageModel.get_peft_model(**peft_kwargs, **unsloth_peft_kwargs)
|
||||||
else:
|
else:
|
||||||
lora_config = LoraConfig(
|
lora_config = LoraConfig(
|
||||||
|
|||||||
Reference in New Issue
Block a user