From 3d483e09142ec920290634e1619f0961f7bf379a Mon Sep 17 00:00:00 2001 From: hiyouga Date: Sun, 17 Mar 2024 22:32:03 +0800 Subject: [PATCH] fix packages Former-commit-id: 8e04794b2da067a4123b9d7091a54c5647f44244 --- src/llmtuner/train/sft/metric.py | 5 +++++ src/llmtuner/train/tuner.py | 2 +- src/llmtuner/train/utils.py | 2 +- 3 files changed, 7 insertions(+), 2 deletions(-) diff --git a/src/llmtuner/train/sft/metric.py b/src/llmtuner/train/sft/metric.py index d1af4c17..35f89f56 100644 --- a/src/llmtuner/train/sft/metric.py +++ b/src/llmtuner/train/sft/metric.py @@ -2,6 +2,7 @@ from dataclasses import dataclass from typing import TYPE_CHECKING, Dict, Sequence, Tuple, Union import numpy as np +from transformers.utils.versions import require_version from ...extras.constants import IGNORE_INDEX from ...extras.packages import is_jieba_available, is_nltk_available, is_rouge_available @@ -32,6 +33,10 @@ class ComputeMetrics: r""" Uses the model predictions to compute metrics. """ + require_version("jieba", "To fix: pip install jieba") + require_version("nltk", "To fix: pip install nltk") + require_version("rouge_chinese", "To fix: pip install rouge-chinese") + preds, labels = eval_preds score_dict = {"rouge-1": [], "rouge-2": [], "rouge-l": [], "bleu-4": []} diff --git a/src/llmtuner/train/tuner.py b/src/llmtuner/train/tuner.py index 2abffdfc..1b8e3cb7 100644 --- a/src/llmtuner/train/tuner.py +++ b/src/llmtuner/train/tuner.py @@ -62,7 +62,7 @@ def export_model(args: Optional[Dict[str, Any]] = None): output_dtype = getattr(model.config, "torch_dtype", torch.float16) setattr(model.config, "torch_dtype", output_dtype) for param in model.parameters(): - param.data = param.data.to(output_dtype) + param.data = param.data.to(output_dtype) model.save_pretrained( save_directory=model_args.export_dir, diff --git a/src/llmtuner/train/utils.py b/src/llmtuner/train/utils.py index 09572ff7..e8c88331 100644 --- a/src/llmtuner/train/utils.py +++ b/src/llmtuner/train/utils.py @@ -160,7 +160,7 @@ def _create_galore_optimizer( training_args: "Seq2SeqTrainingArguments", finetuning_args: "FinetuningArguments", ) -> "torch.optim.Optimizer": - require_version("galore_torch", "To fix: pip install git+https://github.com/hiyouga/GaLore.git") + require_version("galore_torch", "To fix: pip install galore-torch") if len(finetuning_args.galore_target) == 1 and finetuning_args.galore_target[0] == "all": galore_targets = find_all_linear_modules(model)