From 06860e8f0f86b678f86f68c3dda10313630dbdb6 Mon Sep 17 00:00:00 2001 From: hiyouga Date: Fri, 15 Mar 2024 15:06:30 +0800 Subject: [PATCH] fix export Former-commit-id: 6bc2c23b6d26b52f54ac37fa6149e6eb3cc18ee6 --- src/llmtuner/train/tuner.py | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/src/llmtuner/train/tuner.py b/src/llmtuner/train/tuner.py index cde608ea..2abffdfc 100644 --- a/src/llmtuner/train/tuner.py +++ b/src/llmtuner/train/tuner.py @@ -60,8 +60,9 @@ def export_model(args: Optional[Dict[str, Any]] = None): if getattr(model, "quantization_method", None) is None: # cannot convert dtype of a quantized model output_dtype = getattr(model.config, "torch_dtype", torch.float16) - model = model.to(output_dtype) setattr(model.config, "torch_dtype", output_dtype) + for param in model.parameters(): + param.data = param.data.to(output_dtype) model.save_pretrained( save_directory=model_args.export_dir,