From bad35d173013e111019e968c5789e1d4417625c5 Mon Sep 17 00:00:00 2001 From: hoshi-hiyouga Date: Sat, 8 Jun 2024 00:45:02 +0800 Subject: [PATCH] fix #4139 Former-commit-id: cfd62283a9772fc854b852d2a1b71699f79a0048 --- src/llamafactory/model/adapter.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/llamafactory/model/adapter.py b/src/llamafactory/model/adapter.py index d17873f7..bd14a52f 100644 --- a/src/llamafactory/model/adapter.py +++ b/src/llamafactory/model/adapter.py @@ -239,7 +239,7 @@ def init_adapter( ) model = get_peft_model(model, lora_config) - if cast_trainable_params_to_fp32: + if is_trainable and cast_trainable_params_to_fp32: for param in filter(lambda p: p.requires_grad, model.parameters()): param.data = param.data.to(torch.float32)