use peft 0.7.0, fix #1561 #1764

This commit is contained in:
hiyouga
2023-12-11 17:13:40 +08:00
parent 28d5de7e78
commit 9ce1b0e2f2
3 changed files with 5 additions and 2 deletions

View File

@@ -102,6 +102,9 @@ def init_adapter(
)
model = get_peft_model(model, lora_config)
for param in filter(lambda p: p.requires_grad, model.parameters()):
param.data = param.data.to(torch.float32)
if model_args.checkpoint_dir is not None:
logger.info("Loaded fine-tuned model from checkpoint(s): {}".format(",".join(model_args.checkpoint_dir)))